Published in


OpenAI’s unCLIP Text-to-Image System Leverages Contrastive and Diffusion Models to Achieve SOTA Performance

Contrastive vision-language models such as OpenAI’s CLIP (Contrastive Language–Image Pre-training, 2021) have garnered much attention in the computer vision research community thanks to their impressive capabilities in zero-shot learning and learning robust representations of images that capture both…




We produce professional, authoritative, and thought-provoking content relating to artificial intelligence, machine intelligence, emerging technologies and industrial insights.

Recommended from Medium

Longformer: A Transformer for Long Form Documents

Keras Model For Tensorflow Serving (TF serving with Keras)

An IISc Lecture: Deep Learning Research- Representation Learning

Building Production Machine Learning Systems on Google Cloud Platform (Part 4)

Demystifying Batch Normalization

Breaking the Game: Pendragon Four Rise of Merlin

Time Series Forecast in Python

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store


AI Technology & Industry Review — | Newsletter: | Share My Research | Twitter: @Synced_Global

More from Medium

Google Builds Language Models with Socratic Dialogue to Improve Zero-Shot Multimodal Reasoning…

Chinese Prominent AI Lab Plagiarizes Big Model Paper; Microsoft Research Asia Halts Internship…

DeepMind’s Clever Idea to Master Asymmetric Games

Explaining the code of the popular text-to-image algorithm (VQGAN+CLIP in PyTorch)