It's a node based UI. So you can use multiple models in succession, for parts of the image or include a sketch like the person you're responding to said. You can also add stages to manipulate your prompt.
Basically it's way beyond just "typing a prompt and pressing enter" you control every step of the way
I'm looking at their blog[1] and yeah it looks like they're doing literally the exact same thing the other tools I named are doing but with a UI inspired by things like shader pipeline tools in game engines. It isn't clear how it's doing all of the things the grandparent is claiming.
Interesting, have you used both? A typewriter types when the key is pressed, a word processor sends an interrupt though the keyboard into the interrupt device through a bus and from there its 57 different steps until it shows up on the screen.