DataSeries
Published in

DataSeries

Five Key Facts Wu Dao 2.0: The Largest Transformer Model Ever Built

The record-setting model combines some clever research and engineering methods.

Image Source: https://www.forbes.com/sites/alexzhavoronkov/2021/07/19/wu-dao-20bigger-stronger-faster-ai-from-china/?sh=4a5264ed6fb2

I recently started an AI-focused educational newsletter, that already has over 100,000 subscribers. TheSequence is a no-BS (meaning no hype, no news etc) ML-oriented newsletter that takes 5 minutes to read. The goal is…

--

--

--

Imagine the future of data

Recommended from Medium

Image Classifier using Deep Learning

SIRENs — Implicit Neural Representations with Periodic Activation Functions

Visualization of Hough Transformation to Detect Lines

yellow line is just a single point in the Hough Space

Semantic Segmentation in Seismic Images

Detecting anomalies in data streams using half space trees

Pytorch: Step by Step implementation 3D Convolution Neural Network

A Practical Introduction to Random Forest Classifiers from scikit-learn

Distance Metrics: Concepts and Uses in Machine Learning Models

Get the Medium app

A button that says 'Download on the App Store', and if clicked it will lead you to the iOS App store
A button that says 'Get it on, Google Play', and if clicked it will lead you to the Google Play store
Jesus Rodriguez

Jesus Rodriguez

CEO of IntoTheBlock, Chief Scientist at Invector Labs, I write The Sequence Newsletter, Guest lecturer at Columbia University, Angel Investor, Author, Speaker.

More from Medium

DeepMind and OpenAI Ideas to Incorporate Human Feedback in Reinforcement Learning Agents

Microsoft’s DeepSpeed-MoE Makes Massive MoE Model Inference up to 4.5x Faster and 9x Cheaper

Retrieval-Enhanced Transformers: DeepMind paper summary

How DeepMind Uses Transformer Models to Help Restore Ancient Inscriptions