Introduction to Generative AI Models

Baljinder Kaur Smagh
3 min readSep 28, 2021

--

Image link

Generative AI is one of the most significant advances in the artificial intelligence field due to its ability to create new things. Generative AI models are mostly unsupervised algorithms. In contract discriminative techniques that learn to classify the data, generative AI techniques are mostly involved in creating new data from the training data.

This Framework enables computers to learn the underlying pattern associated with the input and then generate similar content.

There are three popular types of generative models:

  1. Generative Adversarial Networks(GANs)
  2. Autoregressive Convolutional neural networks (AR-CNNs)
  3. Transformer-based Models

Generative Adversarial Networks:

GANs employ two neural networks that compete with one another, a generator and a discriminator. The generator, also known as the generative network, is a neural network that is in charge of generating new data or content that is similar to the source data. The discriminator, also known as the discriminative network, is a neural network that distinguishes between the source data and the generated data. The Competition between these two networks to improve their methods until they manage to generate Indistinguishable data from the source content.

image credit:Thalles Silva

Use cases:

  • Drug Discovery
  • 3D Object Creation

Models:

  • DCGAN(Deep Convulutional GAN)
  • ProGAN(Progressively Growing GAN)
  • BigGAN

Autoregressive Convolutional neural networks:

Autoregressive means regression on self. Here Autoregression term is used to imply predicting future outcomes of a sequence from the previously observed outcomes of that sequence.AR-CNNs are used to examine systems that evolve over time and assume that the probability of certain data is only dependent on what has happened previously. They rely on previous time-series data to generate accurate new data. The main autoregressive architectures are RNNs and casual convolutional networks.

WaveNet Model.Image Credit

Use Cases:

  • Weather Forecasting
  • Stock Prediction
  • Music Generation

Models:

  • PixelRNN
  • PixelCNN
  • WaveNet

Transformer-based Models

Transformer-based models are most commonly used to analyze data with a sequential structure (such as the sequence of words in a sentence). In the modern period, transformer-based techniques have become a common tool for modeling natural language. The ability of the transformer models to attend to different positions of the input sequence in order to compute a representation of that sequence is core to their architecture.

Image Credit

Use Cases:

  • Text Generation
  • Text Summarization
  • Text Translation

Models:

  • BERT(Bidirectional Encoder Representation from Transformers)
  • RoBERTA(Robustly Optimized BERT)

--

--