Super conditioning is very powerful 😎
See the effects of the conditioning scale 👉 wandb.ai/dalle-mini/dalle-mi…
For each prompt, we generate images using a different conditioning scale (1 row per scale).
May 9, 2022 · 9:29 PM UTC
We could notice a bigger difference in early training. Now it's more obvious on difficult prompts.
A higher conditioning scale is supposed to match better the prompt, but with a lower diversity.
In some cases, we can be lucky and have a great sample with no conditioning scale, but the image will more often be missing details from the prompt.
During training, you drop the prompt 10-20% of the time and at generation, you do:
logits = logits_uncond + cond_scale * (logits_cond - logits_uncond)
So for cond_scale > 1, you shift logits distribution closer to the condition and away from unconditional distribution
You can find more details about super conditioning in this great thread from @RiversHaveWings