Turn off the Ad Banner  

To print: Select File and then Print from your browser's menu.

    -----------------------------------------------
This story was printed from CdrInfo.com,
located at http://www.cdrinfo.com.
-----------------------------------------------

Appeared on: Monday, June 25, 2018
AI Algorithms Crushed Human Players in Dota 2 Video Game

Researchers at OpenAI have created a bot which beats the world's top professionals at 1v1 matches of Dota 2 under standard tournament rules.

OpenAI, a nonprofit based in San Francisco California backed by billionaire Elon Musk, developed software that can master complex games combining fast, real-time action, longer-term strategy, imperfect information and team play.

OPenAI's team of five neural networks, OpenAI Five, has started to defeat amateur human teams at Dota 2. The company aims to beat a team of top professionals at The International in August subject only to a limited set of heroes.

Dota 2 is one of the most popular and complex esports games in the world, with creative and motivated professionals who train year-round to earn part of Dota's annual $40M prize pool (the largest of any esports game).

Dota 2 is played between two teams of five players, with each player controlling a character called a "hero". The tournament version pits two competing five-player teams. Each team is assigned a base on opposing ends of a map that can only be learned through exploration. Each player controls a separate character with unique powers and weapons. Each team must battle to reach the opposing team's territory and destroy a structure called an Ancient.

The Dota rules are very complex - the game has been actively developed for over a decade, with game logic implemented in hundreds of thousands of lines of code. This logic takes milliseconds per tick to execute, versus nanoseconds for Chess or Go engines. The game also gets an update about once every two weeks, constantly changing the environment semantics.

A Dota-playing AI must master the following:

OpenAI's software learned solely through trial-and-error while playing against itself. The software plays 180 years worth of games against itself every day, learning via self-play. It trains using a scaled-up version of Proximal Policy Optimization running on 256 GPUs and 128,000 CPU cores - a larger-scale version of the system we built to play the much-simpler solo variant of the game last year. Using a separate LSTM for each hero and no human data, it learns recognizable strategies. This indicates that reinforcement learning can yield long-term planning with large but achievable scale - without fundamental advances.

The ability to learn these kinds of video games at human or super-human levels is important for the advancement of AI because they more closely approximate the uncertainties and complexity of the real world than games.

OpenAI said its software in mid-June beat a semi-professional team that is ranked among the top 1 percent of Dota 2 players and an amateur team ranked in the top 10 percent -- both times winning two games to one in a best-of-three series. Earlier in the month, OpenAI's bot crushed three amateur teams.

The sort of reinforcement learning OpenAI used could be promising for solving real-world situations, particularly those that could be couched as games. But have in mind that the amount of data and computing power required to use the technique effectively has limited its applications.



Home | News | All News | Reviews | Articles | Guides | Download | Expert Area | Forum | Site Info
Site best viewed at 1024x768+ - CDRINFO.COM 1998-2024 - All rights reserved -
Privacy policy - Contact Us .