Back to AI Timeline
2021Model

DALL-E: Text to Image

OpenAI unveiled DALL-E, a model capable of generating images from text descriptions by combining language understanding with image generation. Users could describe scenes that had never existed and receive plausible visual representations. DALL-E demonstrated that AI could bridge the gap between language and visual creativity in ways previously thought to be uniquely human.

In January 2021, OpenAI unveiled DALL-E, a neural network that could generate images from text descriptions. Named as a portmanteau of Salvador Dali and the Pixar character WALL-E, the system could create plausible images of scenes that had never existed -- "an armchair in the shape of an avocado" or "a snail made of a harp" -- demonstrating a remarkable ability to combine concepts in visually coherent ways.

How DALL-E Worked

The original DALL-E was based on a modified version of GPT-3 with 12 billion parameters. It treated image generation as a sequence prediction problem: text tokens were followed by image tokens, and the model learned to predict the image tokens that should follow a given text description. The images were represented as sequences of discrete visual tokens using a technique called dVAE (discrete variational autoencoder). This approach allowed DALL-E to leverage the same Transformer architecture that had proven so successful for language.

The Demonstrations

OpenAI's announcement showcased DALL-E's ability to handle a wide range of prompts. It could combine unrelated concepts ("a baby daikon radish in a tutu walking a dog"), render text in images, apply transformations ("the same cat from different angles"), and generate images in various styles ("a painting of a fox in the style of Starry Night"). The diversity and quality of the outputs amazed both researchers and the public.

DALL-E 2

In April 2022, OpenAI released DALL-E 2, which used a completely different approach based on diffusion models and CLIP (Contrastive Language-Image Pre-training). DALL-E 2 produced dramatically higher-resolution and more photorealistic images. It could also edit existing images, extending them beyond their borders (outpainting) or modifying specific regions based on text instructions (inpainting). The quality improvement over the original was striking.

The Creative Implications

DALL-E raised profound questions about creativity and authorship. Could a machine be creative? Who owned the copyright to AI-generated images? Were artists being replaced or empowered? These debates intensified as the technology improved and became more widely available. Some artists embraced AI as a new creative tool, while others saw it as a threat to their livelihoods and an appropriation of their styles.

Safety and Restrictions

OpenAI implemented various restrictions on DALL-E's use. The system was designed to decline requests for violent, sexual, or politically sensitive content. It avoided generating images of real public figures to prevent deepfakes. Access was initially limited to approved users, with a gradual rollout to the public. These restrictions reflected growing awareness of the potential for misuse of generative AI tools.

Competition and Democratization

DALL-E's announcement triggered intense competition. Midjourney, Stable Diffusion, and Google's Imagen followed within a year or two. The open-source release of Stable Diffusion in 2022 particularly accelerated the field, enabling anyone to generate images locally without API access. Text-to-image generation went from a research curiosity to a mainstream capability in less than two years.

Legacy

DALL-E demonstrated that AI could bridge the gap between language and visual creativity, opening up entirely new possibilities for creative expression, design, and communication. It was a pivotal moment in the emergence of generative AI as a transformative technology, directly inspiring the wave of creative AI tools that followed.

Key Figures

Aditya RameshMikhail PavlovGabriel GohScott Gray

Lasting Impact

DALL-E proved that AI could generate creative visual content from text descriptions, launching the text-to-image revolution. It opened entirely new possibilities for creative expression and sparked critical debates about AI creativity, copyright, and the future of visual arts.

Related Events

2014Research
GANs Introduced by Ian Goodfellow

Ian Goodfellow and colleagues introduced Generative Adversarial Networks, a framework where two neural networks compete against each other to generate realistic data. A generator creates fake samples while a discriminator tries to distinguish them from real ones, driving both to improve. GANs would go on to revolutionize image generation, style transfer, and synthetic data creation.

2022Model
Stable Diffusion Goes Open Source

Stability AI released Stable Diffusion as an open-source image generation model, democratizing access to high-quality AI art creation. Unlike proprietary alternatives, anyone could download, run, and modify the model on consumer hardware. The release sparked an explosion of creative applications, fine-tuned models, and community-driven innovation.

2023Model
Midjourney V5

Midjourney released version 5 of its AI image generation tool, producing photorealistic images that were often indistinguishable from photographs. The leap in quality raised new questions about AI-generated media and authenticity. Midjourney V5 became a go-to tool for artists, designers, and creative professionals worldwide.

2024Model
Sora Announced (OpenAI Video)

OpenAI announced Sora, a text-to-video generation model capable of producing realistic, minute-long videos from text prompts. Sora demonstrated an understanding of physics, object permanence, and complex scene composition that surpassed all prior video generation models. The announcement intensified discussions about the future of media, film production, and synthetic content.