Categories: Uncategorized

The Kitten Effect

One thing I’ve noticed with image-generating algorithms is that the more of something they have to put in an image, the worse it is.

I first noticed this with the kitten-generating variant of StyleGAN, which often does okay on one cat:

alternative for shocked_pikachu.png

but is terrible at a crowd of kittens.

A few years later, Dall-E 2 is a LOT more coherent despite having a larger job to do. But it’s still susceptible to the kitten effect.

One kitten:

“A cute kitten in a basket”

Two kittens:

“Two cute kittens in a basket”

Ten kittens:

“Ten cute kittens in a basket”. Note: the kittens are all happy and doing great because when you are a virtual kitten it doesn’t matter if you have extra earholes in the middle of your face.

A vast herd of cute kittens thundering across the plain:

“A vast herd of cute kittens thundering across the plain” These virtual kittens are also all healthy and happy. The one with extra eyes in its stomach is great at pouncing.

Similarly, there’s a huge difference in quality between a single dog and a herd of them.

A dog running across a field, matte painting
Vast herd of millions of dogs thundering across the plain, matte painting. They may look a bit odd, but they’re all very good dogs.

And it’s not just animals, of course. Here’s the kitten effect played out in lucky charms marshmallows:

A single lucky charms marshmallow piece on a plate:

a single lucky charms marshmallow piece on a plate, labeled with the name of the shape:

A labeled list of lucky charms marshmallows:

It was not obvious to me that this should be so. If it can make one kitten, why can’t it make ten of equal quality?

My theory is this is may actually be not a numbers thing but a size thing. When I asked Dalle-2 to generate a kitten that takes up less of an image, the kitten gets way, way worse.

“A kitten sitting at the far end of a large room”.

It’s not out of pixels to make the detail with – note the sharpness of the wood grain. But at this scale it has run out of something nonetheless.

It does make DALL-E 2’s eye test charts particularly mean.

Bonus post! In which I get DALL-E 2 to look increasingly closely at a giraffe.

AI Generated Robotic Content

Recent Posts

Workflow upscale/magnify video from Sora with Wan , based on cseti007

📦 : https://github.com/lovisdotio/workflow-magnify-upscale-video-comfyui-lovis I did this ComfyUI workflow for Sora 2 upscaling 🚀 ( or…

19 hours ago

The Complete Guide to Pydantic for Python Developers

Python's flexibility with data types is convenient when coding, but it can lead to runtime…

19 hours ago

Rooms from Motion: Un-posed Indoor 3D Object Detection as Localization and Mapping

We revisit scene-level 3D object detection as the output of an object-centric framework capable of…

19 hours ago

Inside the AIPCon 8 Demos Transforming Manufacturing, Insurance, and Construction

Editor’s Note: This is the second in a two-part series highlighting demo sessions from AIPCon…

19 hours ago

Responsible AI design in healthcare and life sciences

Generative AI has emerged as a transformative technology in healthcare, driving digital transformation in essential…

19 hours ago

5 ad agencies used Gemini 2.5 Pro and gen media models to create an “impossible ad”

The conversation around generative AI in the enterprise is getting creative.  Since launching our popular…

19 hours ago