1. TechSpot is dedicated to computer enthusiasts and power users. Ask a question and give support. Join the community here.
    TechSpot is dedicated to computer enthusiasts and power users.
    Ask a question and give support.
    Join the community here, it only takes a minute.
    Dismiss Notice

OpenAI's Dota 2 agents display their dominance over mere mortals one last time

By Cal Jeffrey · 9 replies
Apr 23, 2019
Post New Reply
  1. In total, the AI team, called OpenAI Five, played 7,257 matches and only lost 42 for a win ratio of 99.4 percent. The agents won 4,075 games outright. Human players quit, admitting defeat, 3,140 times. OpenAI Five’s dominance against amateur players is not surprising.

    Last summer the bots were pitted against a team of professional Dota players and defeated them 2-1 in a best of three series. Even though the AI did lose one match, it was constrained to a choice of only 18 of the 100 heroes and had its reaction time increased to 200ms from its usual 80ms.

    More recently, OpenAI’s agents were tested not only against professionals but with The International 2018 champion, team OG. In the best-of-three match, OpenAI Five swept OG in the first two games making a third round unnecessary.

    The Verge notes that at the time of the pairing between OpenAI Five and OG, the agents had already had the equivalent of 45,000 years worth of training.

    “OpenAI Five is powered by deep reinforce learning, which means we didn’t code it how to play. We coded it how to learn,” said Greg Brockman, OpenAI’s CTO. “In its 10 months of existence, it’s already played 45,000 years of Dota 2 gameplay. That’s a lot — it hasn’t grown bored yet.”

    The OpenAI bots can also play in cooperative mode with humans. Logically, the amateur players enjoyed playing with the agents, more than playing against them. The Verge notes that the cooperative version of the AI received far more attention with more than 35,000 games played.

    For now, OpenAI Five is retired. The team will review the data accumulated from the matches and “will use Dota 2 as a research platform for further development of powerful AI systems.”

    Permalink to story.

  2. ShagnWagn

    ShagnWagn TS Guru Posts: 733   +566

    So the whole idea behind "AI" (which was created from human thought - not 'artificial') is not how fast it reacts, but intelligence. Letting it react as fast as a computer is not a rating of intelligence. This is like pitting a human against a robot in strength. This was very obvious when they first brought openai against people. Real human intelligence won over sheer power.

    "had its reaction time increased to 200ms from its usual 80ms."

    Is 200ms typical of real players? If so, then it would really be a fair contest of human-programmed intelligence to real-time human intelligence.

    If "AI" really is intelligent, why are phones so stupid? :) When will we have AI phones instead of "smart" ID10T phones.
  3. syrious01

    syrious01 TS Rookie Posts: 24   +18

    I guess if you do some extremely dirty math you can see where AI has the learning advantage.
    Average game = 1hr
    Hours in a year = 8760hrs (not factoring in sleeping, pissing, and eating)
    Potential AI games in 45,000 years = 394,200,000

    Even if you played dota every minute of every date until you were 35, you'd still only have 306,600 hours of experience.
  4. ShagnWagn

    ShagnWagn TS Guru Posts: 733   +566

    I guess I did not explain the overall point very well.

    Yours: 394,200,000 games and it still is getting beaten? Ouch, this explains one of my points.
  5. J spot

    J spot TS Maniac Posts: 221   +141

    I don't think you truly understand the concept of AI, yet.
  6. gamerk2

    gamerk2 TS Maniac Posts: 259   +161

    The problem with current AI is that no implementation is really learning. The current preferred method is reinforcement learning, which does an OK job of making a competent AI that can do one specific thing, but also has significant limitations (bad behaviors that are learned early will never be unlearned) and can be exploited once understood.

    For example, the SC2 AI lost a game because the human player figured out the AI would disengage it's entire army to defend it's base, even if just one unit was attacking it. This was a learned behavior, and the human punished it by preventing the AI from attacking until the human could win through sheer numbers. A "smart" AI would have figured this out, but the AI was bound by it's past experiences and didn't have any "thought" to see, in realtime, what was happening.

    Older AIs simply used databases they would access to determine what the proper results should be. Most chess engines fall into this territory, as does Siri/Alexia and the most other commercial implementations. They do OK at what they do, but a very limited in function as a result.

    An AI with actual independent thought is well beyond what we are capable of at the moment. Putting aside the processing and storage requirements, no one is even in agreement on the best way to do that. That is honestly just as much a sociological question as an engineering one.
  7. shm0wie

    shm0wie TS Member Posts: 24   +20

    Comparing AI to the software in phones is like comparing a model airplane to an F18, and asking why the model airplane won't fly.
  8. J spot

    J spot TS Maniac Posts: 221   +141

    There are two top AI chess engines, Alpha Zero and Leela Zero, and they are much more impressive than traditional brute force chess engines. Their "thinking" is phenomenal. And as far as a brute force chess engines, like Stockfish, no human that ever lived has a chance against it.
  9. gamerk2

    gamerk2 TS Maniac Posts: 259   +161

    All those engines do is essentially play themselves over and over, acquiring a massive database of past knowledge that is used to make the correct move in any situation. As I noted, this is the way AIs are currently moving toward, and even then mistakes are made from time to time (and yes, I watched the entire Alpha Zero/Stockfish series; I noted some moves from both engines that appear sub-optimal).
  10. J spot

    J spot TS Maniac Posts: 221   +141

    I can't take the comment that you saw some moves from both engines that appeared sub-optimal. There's just no way for you as a human, who's not even a professional that dedicates their whole life to chess and chess theory, and plays eight hour games of pure thinking and evaluating, is in any position to know whether any move that Alpha Zero made was sub optimal or not.

Add your comment to this article

You need to be a member to leave a comment. Join thousands of tech enthusiasts and participate.
TechSpot Account You may also...