• Hey, guest user. Hope you're enjoying NeoGAF! Have you considered registering for an account? Come join us and add your take to the daily discourse.

DeepMind and Blizzard open StarCraft II as an AI research environment

Wok

Member
SC2LE is a set of tools that should accelerate AI research in the real-time strategy game StarCraft II.

The SC2LE release includes:
  • A Machine Learning API developed by Blizzard that gives researchers and developers hooks into the game. This includes the release of tools for Linux for the first time.
  • A dataset of anonymised game replays, which will increase from 65k to more than half a million in the coming weeks.
  • An open source version of DeepMind's toolset, PySC2, to allow researchers to easily use Blizzard's feature-layer API with their agents.
  • A series of simple RL mini-games to allow researchers to test the performance of agents on specific tasks.
  • A joint paper that outlines the environment, and reports initial baseline results on the mini-games, supervised learning from replays, and the full 1v1 ladder game against the built-in AI.

The release also contains a series of ‘mini-games' - an established technique for breaking down the game into manageable chunks that can be used to test agents on specific tasks, such as moving the camera, collecting mineral shards or selecting units. We hope that researchers can test their techniques on these as well as propose new mini-games for other researchers to compete and evaluate on.

Our initial investigations show that our agents perform well on these mini-games. But when it comes to the full game, even strong baseline agents, such as A3C, cannot win a single game against even the easiest built-in AI. For instance, the following video shows an early-stage training agent (left) which fails to keep its workers mining, a task that humans find trivial. After training (right), the agents perform more meaningful actions, but if they are to be competitive, we will need further breakthroughs in deep RL and related areas.

One technique that we know allows our agents to learn stronger policies is imitation learning. This kind of training will soon be far easier thanks to Blizzard, which has committed to ongoing releases of hundreds of thousands of anonymised replays gathered from the StarCraft II ladder. These will not only allow researchers to train supervised agents to play the game, but also opens up other interesting areas of research such as sequence prediction and long-term memory.

References:
http://us.battle.net/sc2/en/blog/20944009
https://deepmind.com/blog/deepmind-and-blizzard-open-starcraft-ii-ai-research-environment/

igmxT8B.gif

Actions available to both humans and agents depend on the units selected

zqzDrgw.gif

Simple RL mini-games will allow researchers to test the performance of agents on specific tasks
 
Top Bottom