Preferences

spaceman_2020 parent
If you compare to the amount of effort required in Photoshop to achieve the same results, still a vast improvement

qingcharles
I work in Photoshop all day, and I 100% agree. Also, I just retried a task that wouldn't work last night on nano-banana and it worked first time on the released model, so I'm wondering if there were some changes to the released version?
spaceman_2020 OP
We had an exhibition some time back where I used AI to generate the posters for our product. This is a side project and not something we do seriously, but the results were outstanding - better than what the majority of much bigger exhibitors had.

It took me a LOT of time to get things right, but if I was to get an actual studio to make those images, it would have cost me a thousands of dollars

Bombthecat
Yeah, played around with it, it created an amazing poster for starfinder ttrpg ( something like DND) with specifies who looked really! Good. Usually stuff likes this fails hard, since there isn't much training data of unique fantasy creatures.

But flash 2.5? Worked! It did it, crazy stuff

Bombthecat
How many times did you tried? I uploaded a black and white photo and let it colourize, something like 20 percent were still black and white.
echelon
Vibe coding might not be real, but vibe graphics design certainly is.

https://imgur.com/a/internet-DWzJ26B

Anyone can make images and video now.

cwmoore
Are those oil derricks, or wind turbines? Who cares! Graphic design is easy now!
viraptor
They're Australian farm windmills https://media.istockphoto.com/id/959193466/photo/australian-...

(But yeah, some got a generator attached...)

lebimas
What tools did you use to make those videos from the PG image?
echelon
I used a bunch of models in conjunction:

- Midjourney (background)

- Qwen Image (restyle PG)

- Gemini 2.5 Flash (editing in PG)

- Gemini 2.5 Flash (adding YC logo)

- Kling Pro (animation)

I didn't spend too much time correcting mistakes.

I used a desktop model aggregation and canvas tool that I wrote [1] to iterate and structure the work. I'll be open sourcing it soon.

[1] https://getartcraft.com

kstenerud
The app looks interesting, but I think it needs some documentation. I think I generated something? Maybe? I saw a spinny thing for awhile, but then nothing.

I couldn't get the 3d thing to do much. I had assets in the scene but I couldn't for the life of me figure out how to use the move, rotate or scale tools. And the people just had their arms pointing outward. Are you supposed to pose them somehow? Maybe I'm supposed to ask the AI to pose them?

Inpainting I couldn't figure out either... It's for drawing things into an existing image (I think?) but it doesn't seem to do anything other than show a spinny thing for awhile...

I didn't test the video tool because I don't have a midjourney account.

unixhero
What is PG?
sethaurus
In this context, it's Paul Graham, the head Y Combinator guy whose cartoon likeness appears in the generated video: https://news.ycombinator.com/user?id=pg
Paul Graham, Y Combinator founder.
spaceman_2020 OP
Midjourney with style references is just about the easiest way right now for an absolute noob to get good aesthetics
bacchusracine
This post may or may not violate our community standards so we aren't going to display it.
throwaway638637
What is up with that T rex's arms?
benreesman
I think much like coding, the top of the game is all the old stuff and a bunch of new stuff that is impossible to master without some real math or at least outlier mathematical intuition.

The old top of the game is available to more people (though mid level people trying to level up now face a headwind in a further decoupling of easily read signals and true taste, making the old way of developing good taste harder).

This stuff makes people who were already "master rate" who are also nontrivially sophisticated machine learning hobbyists minimum and drives their peak and frontier out, drives break even collaboration overhead down.

It's always been possible to DIY code or graphic design, it's always been possible to tell the efforts of dabblers and pros apart, and unlike many commodities? There is rarely a "good enough". In software this is because compute is finite and getting more out of it pays huge, uneven returns, in graphic design its because extreme quality work is both aesthetically pleasing as well as a mark of quality (imperfect but a statement someone will commit resources).

And it's just hard to see it being different in any field. Lawyers? Opposing counsel has the best AI, your lawyer better have it too. Doctors? No amount of health is "enough" (in general).

I really think HN in particular but to some extent all CNBC-adjacent news (CEO OnlyFans stuff of all categories) completely misses the forest (the gap between intermediate and advanced just skyrocketed) for the trees (space-filling commodity knowledge work just plummeted in price).

But "commodity knowledge work" was always kind of an oxymoron, David Graeber called such work "bullshit jobs". You kinda need it to run a massive deficit in an over-the-hill neoliberal society, it's part of the " shift from production to consumption" shell game. But it's a very recent, very brief thing that's already looking more than wobbly. Outside of that? Apprentices, journeymen, masters is the model that built the world.

AI enables a new even more extreme form of mastery, blurs the line between journeyman and dabbler, and makes taking on apprentices a much longer-term investment (one of many reasons the PRC seems poised to enjoy a brief hegemony before demographics do in the Middle Kingdom for good, in China, all the GPUs run Opus, none run GPT-5 or LLaMA Behemoth).

The thing I really don't get is why CEOs are so excited about this and I really begin to suspect they haven't as a group thought it through (Zuckerberg maybe has, he's offering Tulloch a billion): the kind of CEO that manages a big pile of "bullshit jobs"?

AI can do most of their job today. Claude Opus 4.1? It sounds like if a mid-range CEO was exhaustively researched and gaff immune. Ditto career machine politicians. AI non practitioner prognosticators. That crowd.

But the top graphic communications people and CUDA kernel authors? Now they have to master ComfyUI or whatever and the color theory to get anything from it that stands out.

This is not a democratizing thing. And I cannot see it accruing to the Zuckerberg side of the labor/capital divvy up without a truly durable police state. Zuck offering my old chums nation state salaries is an extreme and likely transitory thing, but we know exactly how software professional economics work when it buckets as "sorcery" and "don't bother": that's 1950 to whenever we mark the start of the nepohacker Altman Era, call it 2015. In that world good hackers can do whatever they want, whenever they want, and the money guys grit their teeth. The non-sorcery bucket has paper mache hack-magnet hackathon projects in it at a fraction of the old price. So disruption, wow.

Whether that's good or bad is a value judgement I'll save for another blog post (thank you for attending my TED Talk).

captnFwiffo
Sure, now the client wants 130 edits without losing coherency with the original. What does a vibe designer do? Just keep re-prompting and re-generating until it works? Sounds hard to me.
Filligree
They use Kontext, Qwen-Edit or Gemini.
petralithic
Why would you compare it to Photoshop? If you compare it to other tools in the same category, of image generation, you will find models like Flux and Qwen do much better.

This item has no comments currently.