Scientists at the University of Alberta are cracking away at the complexities of artificial intelligence with their new "DeepStack" system, which can not only play a round of poker with you, but walk away with all of your money. This new technology builds upon the legacy of systems like IBM's Deep Blue, which was the first program to beat a world champion, Gary Kasparov, at chess in 1996. As Michael Bowling, co-author of the research and leader of the Computer Poker Research Group at Alberta, puts it: poker is the next big step for designing AI. In a game of Heads Up No Limit poker, DeepStack was able to win against professional poker players at a rate of 49 big blinds per 100. "We are winning by 49 per 100, that's like saying whatever the players were doing was not that much more effective than if they just folded every hand," Bowling tells Inverse.
It is no mystery why poker is such a popular pastime: the dynamic card game produces drama in spades as players are locked in a complicated tango of acting and reacting that becomes increasingly tense with each escalating bet. The same elements that make poker so entertaining have also created a complex problem for artificial intelligence (AI). A study published today in Science describes an AI system called DeepStack that recently defeated professional human players in heads-up, no-limit Texas hold'em poker, an achievement that represents a leap forward in the types of problems AI systems can solve. DeepStack, developed by researchers at the University of Alberta, relies on the use of artificial neural networks that researchers trained ahead of time to develop poker intuition. During play, DeepStack uses its poker smarts to break down a complicated game into smaller, more manageable pieces that it can then work through on the fly.
Mirowski cites Turing as author of the paragraph containing this remark. The paragraph appeared in , in a chapter with Turing listed as one of three contributors. Which parts of the chapter are the work of which contributor, particularly the introductory material containing this quote, is not made explicit.
An invincible checkers-playing program named Chinook has solved a game whose origins date back several millennia, scientists reported Thursday on the journal Science's Web site. By playing out every possible move -- about 500 billion billion in all -- the computer proved it can never be beaten. Even if its opponent also played flawlessly, the outcome would be a draw. Chinook, created by computer scientists from the University of Alberta in 1989, wrapped up its work less than three months ago. In doing so, its programmers say the newly crowned checkers king has solved the most challenging game yet cracked by a machine -- even outdoing the chess-playing wizardry of IBM's Deep Blue.
Playing against a top Go player, Google DeepMind's AlphaGo artificial-intelligence program has puzzled commentators with moves that are often described as "beautiful," but do not fit into the usual human style of play. Artificial-intelligence experts think these moves reflect a key AI strength of AlphaGo, its ability to learn from its experience. Such moves cannot be produced by just incorporating human knowledge, said Doina Precup, associate professor in the School of Computer Science at McGill University in Quebec, in an email interview. "AlphaGo represents not only a machine that thinks, but one that can learn and strategize," agreed Howard Yu, professor of strategic management and innovation at IMD business school. AlphaGo won three games consecutively against Lee Se-dol last week in Seoul, securing the tournament and US$1 million in prize money that Google plans to give to charities.