Hacker News new | past | comments | ask | show | jobs | submit login

It looks like you have selected for it to use a segmentation mask, and not to use text.



I have no idea what that means, but the fact that it both told me to enter a text prompt and actually let me do it while not being in whatever magical mode it should have been in in order to actually use the text prompt is another point that can be added to my above rant.

Alright, I've uttered the incantation for it to do the thing. I still don't get it. [0] https://imgur.com/4zbaiH0

I also tried another example prompt, which bared a striking similarity to the previous result. I don't know if it's persisting the result (It shouldn't - I didn't click the re-use image button), but the strange life-raft looking artifact is very persistent. [1] https://imgur.com/KTCM4xH


Yeah, the text-to-image seems to be highly dependent on whether the generator knows how to generate the specific objects the text model thinks should be in the image. I got much more consistent results using the semantic segmentation drawing as input:

https://imgur.com/QC13zml

(and for what it's worth, you're totally right that the UI is just an absolute disaster)


The picture you drew and had it turn into rocks is actually really cool!

I think I would have been more generous to the project had I known it could do that. Maybe I X'd out of the frustrating tutorial too early? :)


Literally the first step after accepting the TOS:

https://i.imgur.com/oH4P1Xc.png


Yes but the selection is lost if you press enter out of habit. I had the same frustration until I realized what happened which turned that into anger :)

Maybe I'm just too old for tech demos.


Nah, the UI is not very intuitive, so there's plenty of blame to go around :-)




Consider applying for YC's Spring batch! Applications are open till Feb 11.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: