AI : less than two minutes of videogame footage to recreate game engine

0
801

Game studios and enthusiasts may soon have a new tool at their disposal to speed up game development and experiment with different styles of play.

Georgia Institute of Technology researchers have developed a new approach using an artificial intelligence to learn a complete game engine, the basic software of a game that governs everything from character movement to rendering graphics.

To get their AI agent to create an accurate predictive model that could account for all the physics of a 2D platform-style game, the team trained the AI on a single “speedrunner” video, where a player heads straight for the goal.

This made “the training problem for the AI as difficult as possible.”

Their current work uses Super Mario Bros. and they’ve started replicating the experiments with Mega Man and Sonic the Hedgehog as well.

The same team first used AI and Mario Bros. gameplay video to create unique game level designs.

The researchers found that their game engine predicted video frames significantly more similar to those in the original game when compared to the same test on a neural network. This gave them an accurate, general model of a game using only the video footage.

Artificial intelligence uses less than two minutes of videogame footage to recreate game engine
A section of gameplay video (top) shows the original Mega Man in the “Bomberman” stage. Output from the cloned game engine (bottom) attempts to replicate this gameplay section. Despite some failures on predicting animation frames …more

“Our AI creates the predictive model without ever accessing the game’s code, and makes significantly more accurate future event predictions than those of convolutional neural networks,” says Matthew Guzdial, lead researcher and Ph.D. student in computer science. “A single video won’t produce a perfect clone of the game engine, but by training the AI on just a few additional videos you get something that’s pretty close.”

They next tested how well the cloned engine would perform in actual gameplay.

They employed a second AI to play the game level and ensure the game’s protagonist wouldn’t fall through solid floors or go undamaged if hit by an enemy.

The results: the AI playing with the cloned engine proved indistinguishable compared to an AI playing the original game engine.

“The technique relies on a relatively simple search algorithm that searches through possible sets of rules that can best predict a set of frame transitions,” says Mark Riedl, associate professor of Interactive Computing and co-investigator on the project.

“To our knowledge this represents the first AI technique to learn a game engine and simulate a game world with gameplay footage.

Artificial intelligence uses less than two minutes of videogame footage to recreate game engine
A section of gameplay video (top) is produced by the original Super Mario Bros. engine, and the cloned engine (bottom) demonstrates the ability to accurately predict animation states. Credit: Georgia Institute of Technology
Artificial intelligence uses less than two minutes of videogame footage to recreate game engine
A section of gameplay video (top) is produced by the original Super Mario Bros. engine, and the cloned engine (bottom) demonstrates the ability to accurately predict animation states. Credit: Georgia Institute of Technology

The current cloning technique works well with games where much of the action happens on-screen.

Guzdial says Clash of Clans or other games with action taking place off-screen might prove difficult for their system.

“Intelligent agents need to be able to make predictions about their environment if they are to deliver on the promise of advancing different technology applications,” he says.

“Our model can be used for a variety of tasks in training or education scenarios, and we think it will scale to many types of games as we move forward.”

The research was presented at the International Joint Conference on Artificial Intelligence, Aug. 19-25, in Melbourne, Australia. The paper, “Game Engine Learning from Video,” was authored by Matthew Guzdial, Boyang Li, and Mark Riedl.

LEAVE A REPLY

Please enter your comment!
Please enter your name here

Questo sito usa Akismet per ridurre lo spam. Scopri come i tuoi dati vengono elaborati.