NVIDIA’s GameGAN Uses AI to Recreate Pac-Man and Other Game Environments

Synced
SyncedReview
Published in
4 min readMay 26, 2020

Before AI agents are deployed in the real world they must undergo extensive testing in challenging simulated environments. Writing the code to build good simulators however is usually highly time-consuming and requires skilled graphics experts. A more scalable way going forward is to learn to simulate by simply observing the dynamics of the real world. While works such as Intel Labs and the University of Texas’ Learning by Cheating approaches the challenge by learning behaviours, these require much supervision.

Aiming at training a game simulator that can model both the deterministic and stochastic nature of environments, researchers from NVIDIA, University of Toronto, Vector Institute and MIT have proposed a simulator that learns by simply watching an agent interact with its environment.

Focusing on games as a proxy of real environments — and particularly on the seminal Pac-Man, which turns 40 this year — the researchers propose GameGAN, a generative model that learns to visually imitate video game environments by ingesting screenplay and keyboard actions during training. GameGAN consists of three models: a dynamics engine to maintain an internal state variable that is recurrently updated, an external memory module to remember what the model has generated, and a rendering engine to decode the output image at each time instance.

The core modules are neural networks that are trained end-to-end. During training, GameGAN takes in user commands — screenplay and keyboard actions — then conditions on these to predict the next frame. GameGAN can thus learn from rollouts of image and action pairs without requiring access to the underlying game logic or engine.

Researchers tested the GameGAN system on a modified version of Pac-Man and the VizDoom environment, conducting both quantitative and qualitative evaluations with four models: Action-LSTM, World Model, GameGAN-M (GameGAN without the memory module and with a simple rendering engine), and the full GameGAN model.

In the experiments, the full GameGAN was shown to produce higher quality results, while supporting multiple practical applications such as transferring a given game from one operating system to another without re-writing any code. In the future, the researchers hope to extend the model to capture more complex real-world environments.

The paper Learning to Simulate Dynamic Environments with GameGAN was accepted to CVPR 2020 and is on arXiv. There is also a project page on GitHub.

Author: Yuqing Li | Editor: Michael Sarazen

We know you don’t want to miss any story. Subscribe to our popular Synced Global AI Weekly to get weekly AI updates.

Share Your Research With Synced Review

Share My Research is Synced’s new column that welcomes scholars to share their own research breakthroughs with over 1.5M global AI enthusiasts. Beyond technological advances, Share My Research also calls for interesting stories behind the research and exciting research ideas. Share your research with us by clicking here.

Need a comprehensive review of the past, present and future of modern AI research development? Trends of AI Technology Development Report is out!

2018 Fortune Global 500 Public Company AI Adaptivity Report is out!
Purchase a Kindle-formatted report on Amazon.
Apply for Insight Partner Program to get a complimentary full PDF report.

--

--

Synced
SyncedReview

AI Technology & Industry Review — syncedreview.com | Newsletter: http://bit.ly/2IYL6Y2 | Share My Research http://bit.ly/2TrUPMI | Twitter: @Synced_Global