DALL-E 2 is very good at generating images to match text descriptions but I felt
like using DALL-E 2 to mess up some brands. Here's breakfast cereals!
"A box of lucky charms cereal on a grocery store shelf""A box of Froot Loops on
a
I was testing DALL-E 2 to see if it would be subject to some common incorrect
assumptions about the sizes of things. For example if you asked people what size
a kiwi bird is, they tend to assume it's a smallish bird, maybe around the size
of a
Recently I've been experimenting with DALL-E 2, one of the models that uses CLIP
to generate images from my text descriptions. It was trained on internet text
and images, so there's a lot it can do, and a lot of ways it can remix the stuff
One thing I've noticed with image-generating algorithms is that the more of
something they have to put in an image, the worse it is.
I first noticed this with the kitten-generating variant of StyleGAN, which often
does okay on one cat:
alternative for shocked_pikachu.pngbut is terrible
DALL-E (and other text-to-image generators) will often add text to their images
even when you don't ask for any. Ask for a picture of a Halifax Pier
[https://www.aiweirdness.com/the-terror-of-the-sea/] and it could end up covered
in messy writing, variously legible versions of "Halifax"