×
all 10 comments

[–]Airbus480[S] 16 points17 points  (5 children)

Source: https://twitter.com/EMostaque/status/1554011833320837120

It seems that NovelAI devs got access to StableDiffusion, and with prompt engineering(?) they were able to generate this. I am already amazed, this imo beats DALLE-2 on anime. I already signed up for the StableDiffusion beta, can't wait.

[–]gwern 19 points20 points  (2 children)

this imo beats DALLE-2 on anime.

Note that this is unfinetuned on Danbooru20xx or anything like that, it's just the StableDiffusion-small default baseline quality obtained from web scrape corpuses like LAION-400M.

The fact that you can get such a great sample out of it, despite still being in training (I think), unfinetuned, and way smaller than DALL-E 2, proves what I've been saying all along about DALL-E 2 anime: something went terribly wrong in the DALL-E 2 data curation or training process, because the anime results are far worse than they ought to be. A tiny unfinetuned (but uncensored) model should not be able to kick its ass like this!

[–]MayRue 0 points1 point  (1 child)

i'm here because i spent 20 credits trying to get decent anime style art out of DALLE-2. And i thought perhaps there might be a way that i missed to make it better.

But its all of us getting kinda trash results?

[–]Incognit0ErgoSum 3 points4 points  (0 children)

its_happening.gif

[–]Sashinii 8 points9 points  (2 children)

I recently gained access to Dall-E, tried making anime and manhwa art, and all of it is terrible.

This new AI looks promising; I can't wait to be able to use AI to create art that actually looks good like this image of Kurisu, which is really impressive, so much so that Stable Diffusion is most likely currently the best image synthesis for anime.

[–]gwern 6 points7 points  (0 children)

Stable Diffusion is most likely currently the best image synthesis for anime.

It might not be better than the rival diffusion model Waifu Labs has been training.

[–]zadesawa 1 point2 points  (0 children)

Manga is more semantic than just pictures. I think that’s why NN image generators do poorly because they only understands raster data currently