An Unassuming Genius: the Man behind Google’s AutoML
Get tickets now at AI Frontiers Conference to hear from this speaker.
Slim, quiet, and wearing a pair of thick glasses, Quoc Le does not strike you as someone who is leading a revolution in the AI field.
In 2011, Le co-founded Google Brain, together with his Ph.D. advisor Andrew Ng, Google Fellow Jeff Dean and Google Researcher Greg Corrado. The goal was exploring deep learning in the context of Google’s gigantic data. Before that, Le has done some pioneering work at Stanford on unsupervised deep learning.
In 2012, Le published a paper (presented at ICML) that started the keen interest on deep learning: He has developed a deep neural network model that can recognize cats based on 10 million digital images from Youtube, as well as over 3,000 objects in the ImageNet dataset. The giant system consists of 16,000 machines and one billion synapses, 100 times larger than anyone who had ever tried. Le’s paper was followed by the AlexNet paper later that year, and the whole field of deep learning started to generation traction.
While the unsupervised learning approach was later proved to be impractical for commercial use — at least during that period — Le said in a 2015 Wired interview that “it would be wonderful if we could have an algorithm that can discover that — that can learn in the same way — because more practically, we have much more unlabeled data than labeled data.”
Sequence to Sequence Learning
After formally joining Google as a research scientist upon his graduation in 2013, Le soon achieved stunning breakthroughs in machine translation, one of the most active research areas in the machine learning community.
To achieve this result, he has to move beyond the deep learning methods that worked well with images and speech, which can be analyzed with fixed-size input. With natural languages, the length of the sentence varies and the order of words matters.
In 2014, Le proposed sequence to sequence (seq2seq) learning with Ilya Sutskever and Oriol Vinyals. It is a general-purpose encoder-decoder framework that trains models to convert sequences from one domain to another, such as sentences between different languages.
The seq2seq learning requires fewer engineering design choices and allows Google translation system to efficiently and accurately work on very large datasets. It is mainly used for machine translation systems and is proved to be applicable in a broader range of tasks, including text summarization, conversational AI, and question-answering.
Le further invented doc2vec, an unsupervised algorithm learns fixed-length feature representations from variable-length pieces of texts, such as sentences, paragraphs, and documents. Doc2vec is an extension of word2vec, which was presented in 2013 by Google fellow Tomas Mikolov. The idea is that each word can be represented by a vector, which can be automated learn from a collection text. Le added paragraph vector so the model can generate a representation of the document, regardless of its length.
Le’s research efforts paid off. In 2016, Google announced Neural Machine Translation system, which uses AI to learns over time to create better and more natural translations.
In 2015, Le made the MIT Technology Review’s “Innovators Under 35” list in the Visionaries category, for aiming “make software smart enough to assist people in their everyday lives feels pretty good.”
AutoML: Neural Network Learns to Improve Itself
Training a deep neural network requires a large amount of labeled data and back-and-forth experiments: You choose an architecture, build hidden layers, and adjust weights based on outputs. For people with limited machine learning expertise, the training process is somehow painstaking and time-consuming.
In 2016, Le and another Google fellow Barret Zoph proposed neural architecture search. They use a recurrent network to generate the model descriptions of neural networks and train this RNN with reinforcement learning to maximize the expected accuracy of the generated architectures on a validation set.
The new approach can help researchers design a novel network architecture that matches the best human-invented architecture in test set accuracy on the CIFA-10 dataset. One year later, Le and Zoph took their study to the next level by proposing NASNet-A, a transferable architecture for large-scale image datasets.
Le’s research lays the groundwork for AutoML, a suite of Google products designed for developers who have limited machine learning expertise and resources. At an early stage though, AutoML is now solving real-life problems: a data scientist used AutoML to build a model that identifies the restaurant based on noodle images, with almost 95 % accuracy; a Japanese developer built a model that can classify images with its brand name, with AutoML.
Google launched AutoML Vision earlier this year. Last month at Google’s Cloud Next conference, the company released tools of translation and natural language.
For the last six years, Le has stayed at the forefront of deep learning development. The 36-year-old Google research scientist now is poised to take deep learning to the next level. Would machine replace data scientists eventually? This is a question that awaits answers.
At the upcoming AI Frontiers Conference on Nov 9 in San Jose, California, Le will take about “Using machine learning to automate machine learning”, with special focus on Neural Architecture Search and AutoAugment (automated data augmentation).
Quoc Le will speak at AI Frontiers Conference on Nov 9, 2018 in San Jose, California.
AI Frontiers Conference brings together AI thought leaders to showcase cutting-edge research and products. This year, our speakers include: Ilya Sutskever (Founder of OpenAI), Jay Yagnik (VP of Google AI), Kai-Fu Lee(CEO of Sinovation), Mario Munich (SVP of iRobot), Quoc Le (Google Brain), Pieter Abbeel (Professor of UC Berkeley) and more.
Buy tickets at aifrontiers.com. For question and media inquiry, please contact: firstname.lastname@example.org