Hacker Newsnew | past | comments | ask | show | jobs | submitlogin

Stability AI with Stable Diffusion is already at the finish line in this race, by being $0, open source and not being exclusively a cloud-based AI model and can be used offline.

Anything else that is 'open source' AI and allows on-device AI systems eventually brings the cost to $0.



I agree. I am barely excited for DALL-E 3 because I know it's going to be run by OpenAI who have repeatedly made me dislike them more and more over the last year plus. My thoughts are: "Cool. Another closed system like MidJourney. Chat integration would be cool but it's still going to likely be crazy expensive per image versus infinite possibilities with Stable Diffusion."

Especially with DALL-E. Honestly I'd be more excited if MidJourney released something new. DALL-E was the first but, in my experience, the lower-quality option. It felt like a toy, MidJourney felt like a top-tier product akin to Photoshop Express on mobile, still limited but amazing results every time, and Stable Diffusion feels like photoshop allowing endless possibilities locally without restrictions except it's FREE!


They all have their place. OpenAI literally started every major AI revolution including image gen with DALL-E. Let them be the peleton while SD and others follow closely and overtake eventually.


I like painting them as the peleton! You're not wrong it's just not super exciting for me


What's a peleton? Genuine question from a non native English speaker.


I think they mean peloton. It's not English, I believe it is French, meaning a group of bicyclers. In this context it refers to the peloton leading the race.+-


there has a to be a way to link the API from automatic1111 and a "gpt" or "bert" model, to allow similar flexibility, right? The only issue i see is training the llm on the rules of image composition correlated to what CLIP/Deepbooru sees. maybe there will be a leak, or someone can convince one of the "AI Art example/display" sites to give them a dump of images with all metadata. enough of that and this sort of thing seems like a "gimme".

I just started training LoRA, and the lack of definitive information is a burden; but following the commonalities in all of the guides i was able to get a LoRA to model a person i know, extremely accurately, about 1/8th of the time. in my experience, getting a "5 star" image 12% of the time is outstanding. And when i say 1/8th i mean i ran off 100 images, and hand ranked the ones that actually used the LoRA correctly, and 4 and 5 star ranks were 21/100 - i just double checked my numbers!


That's a different race entirely; most people don't even have the hardware, let alone the knowledge to run SD.


unless it's way worse


SDXL and older models are quite good. It's not yet very user friendly but will get there.




Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: