upvote
One-shot prompting: agreed.

Using a node based workflow with comfyUI, also being able to draw, also being able to train on your own images in a lora, and effectively using control nets and masks: different story...

I see, in the near future, a workflow by artists, where they themselves draw a sketch, with composition information, then use that as a base for 'rendering' the image drawn, with clean up with masking and hand drawing. lowering the time to output images.

Commercial artists will be competing, on many aspects that have nothing to do with the quality of their art itself. One of those factors is speed, and quantity. Other non-artistic aspects artists compete with are marketing, sales and attention.

Just like the artisan weavers back in the day were competing with inferior quality automatic loom machines. Focusing on quality over all others misses what it means to be in a society and meeting the needs of society.

Sometimes good enough is better than the best if it's more accessible/cheaper.

I see no such tooling a-la comfyUI available for text generation... everyone seems to be reliant on one-shot-ting results in that space.

reply
I've tried at least 4 other tools/SAASs and I'm just not seeing it. I've tried training models in other tools with input images, sketches, and long prompts built from other LLMs and the output is usually really bad if you want something even remotely novel.

Aside for the terrible name, what does comfyUI add? This[1] all screams AI slop to me.

[1]https://www.comfy.org/gallery

reply
It's a node based UI. So you can use multiple models in succession, for parts of the image or include a sketch like the person you're responding to said. You can also add stages to manipulate your prompt.

Basically it's way beyond just "typing a prompt and pressing enter" you control every step of the way

reply
right, but how is it better than Lovart AI, Freepik, Recraft, or any of the others?
reply
Your question is a bit like asking how a word processor is better than a typewriter... they both produce typed text, but otherwise not comparable.
reply
I'm looking at their blog[1] and yeah it looks like they're doing literally the exact same thing the other tools I named are doing but with a UI inspired by things like shader pipeline tools in game engines. It isn't clear how it's doing all of the things the grandparent is claiming.

[1]https://blog.comfy.org/p/nano-banana-via-comfyui-api-nodes

reply
Interesting, have you used both? A typewriter types when the key is pressed, a word processor sends an interrupt though the keyboard into the interrupt device through a bus and from there its 57 different steps until it shows up on the screen.

They’re about as similar as oil and water.

reply