OpenAI’s DALL-E 2 explained

Vishal Rajput
AIGuys
Published in
7 min readMay 27, 2022

--

We all know that DALL-E 2 is one of the craziest models we have seen this year. DALL-E 2 is a text-to-image generation model developed by Open-AI researchers. It’s the second installment in the series of text-to-image generation models. DALL-E 2 is an extremely capable model with the ability to draw a large variety of images with never before seen text prompts. DALL-E 2 can create images that don’t exist on the internet, it’s more like an actual painter who can paint anything with his/her imagination. So, without any further ado, let’s dive straight into the details of how DALL-E 2 works.

Some results from DALL-E 2 model (Image Src- https://cdn.openai.com/papers/dall-e-2.pdf)

Before we go into the workings of DALL-E 2, let’s have a look at some awesome results from the model.

You can check all the results here: DALL-E 2

Different variations of the famous painting “The Kiss” created by DALL-E 2 (Image Src- https://openai.com/dall-e-2/)
Image created by DALL-E 2 from a text description (Image Src- https://openai.com/dall-e-2/)

DALL-E 2 can understand images and text in great detail thus it can not only create images out of the text but also create other versions of famous paintings, understand different art styles, and also understand the lighting conditions and create shadows accordingly. It can also be…

--

--