A.I. is Now 10 Times Better Than a Pro Poker Player


Scientists at the University of Alberta are cracking away at the complexities of artificial intelligence with their new "DeepStack" system, which can not only play a round of poker with you, but walk away with all of your money. This new technology builds upon the legacy of systems like IBM's Deep Blue, which was the first program to beat a world champion, Gary Kasparov, at chess in 1996. As Michael Bowling, co-author of the research and leader of the Computer Poker Research Group at Alberta, puts it: poker is the next big step for designing AI. In a game of Heads Up No Limit poker, DeepStack was able to win against professional poker players at a rate of 49 big blinds per 100. "We are winning by 49 per 100, that's like saying whatever the players were doing was not that much more effective than if they just folded every hand," Bowling tells Inverse.

Time to Fold, Humans: Poker-Playing AI Beats Pros at Texas Hold'em


It is no mystery why poker is such a popular pastime: the dynamic card game produces drama in spades as players are locked in a complicated tango of acting and reacting that becomes increasingly tense with each escalating bet. The same elements that make poker so entertaining have also created a complex problem for artificial intelligence (AI). A study published today in Science describes an AI system called DeepStack that recently defeated professional human players in heads-up, no-limit Texas hold'em poker, an achievement that represents a leap forward in the types of problems AI systems can solve. DeepStack, developed by researchers at the University of Alberta, relies on the use of artificial neural networks that researchers trained ahead of time to develop poker intuition. During play, DeepStack uses its poker smarts to break down a complicated game into smaller, more manageable pieces that it can then work through on the fly.

AI takes on video games in quest for common sense


Next week, scientists working on artificial intelligence (AI) and games will be watching the latest human-machine matchup. But instead of a single pensive player squaring off against a computer, a team of five top video game players will be furiously casting magic spells and lobbing (virtual) fireballs at a team of five AIs called OpenAI Five. They'll be playing the real-time strategy game Dota 2 at The International in Vancouver, Canada, an annual e-sports tournament that draws professional gamers who compete for millions of dollars. In 1997, IBM's Deep Blue AI bested chess champion Garry Kasparov. In 2016, DeepMind's AlphaGo AI beat Lee Sedol, a world master, at the traditional Chinese board game Go.

Checkers computer becomes invincible

AITopics Original Links

An invincible checkers-playing program named Chinook has solved a game whose origins date back several millennia, scientists reported Thursday on the journal Science's Web site. By playing out every possible move -- about 500 billion billion in all -- the computer proved it can never be beaten. Even if its opponent also played flawlessly, the outcome would be a draw. Chinook, created by computer scientists from the University of Alberta in 1989, wrapped up its work less than three months ago. In doing so, its programmers say the newly crowned checkers king has solved the most challenging game yet cracked by a machine -- even outdoing the chess-playing wizardry of IBM's Deep Blue.

Regret Minimization in Multiplayer Extensive Games

AAAI Conferences

The counterfactual regret minimization (CFR) algorithm is state-of-the-art for computing strategies in large games and other sequential decision-making problems. Little is known, however, about CFR in games with more than 2 players. This extended abstract outlines research towards a better understanding of CFR in multiplayer games and new procedures for computing even stronger multiplayer strategies. We summarize work already completed that investigates techniques for creating "expert" strategies for playing smaller sub-games, and work that proves CFR avoids classes of undesirable strategies. In addition, we provide an outline of our future research direction. Our goals are to apply regret minimization to the problem of playing multiple games simultaneously, and augment CFR to achieve effective on-line opponent modelling of multiple opponents. The objective of this research is to build a world-class computer poker player for multiplayer Limit Texas Hold'em.