With artificial intelligence having beaten humans in board games, what's next?

March 16, 2016 by Tracey Lien, Los Angeles Times

When a person's intelligence is tested, there are exams. IQ tests, general knowledge quizzes, SATs.

When artificial intelligence is tested, there are games. Checkers, chess, Go.

But what happens when computer programs beat humans at all of those games? This is the question AI experts must ask after a Google-developed program called AlphaGo defeated a world champion Go player in four out of five matches in a series that concluded Tuesday.

Long a yardstick for advances in AI, the era of testing has come to an end, said Murray Campbell, an IBM research scientist who was part of the team that developed Deep Blue, the first to beat a world chess champion.

"Games are fun and they're easy to measure," said Campbell. "It's clear who won and who lost, and you always have the human benchmark," he said. "Can you do better than a human?"

For checkers, chess, and now Go, it seems the answer is now a resounding yes. Computer algorithms beat world champion-level human players in checkers and chess in the 1990s.

Go - an ancient board game developed in China that is more complex than chess - was seen as one of the last board game hurdles.

Board games, Campbell said, were perfect tests because they have clear rules and nothing is hidden from players. The real world is much messier and full of unknowns. What's next, it seems, is for AI to get messy.

With AI having conquered what experts call "complete information" games - the kind in which players can see what their opponents are doing - Tuomas Sandholm, a professor at Carnegie Mellon University who studies artificial intelligence, said the next step is "incomplete information games" like poker.

"The game of two-player-limit Texas hold 'em poker has almost been solved," said Sandholm, who described "solving" a game as finding the optimal way of playing it. "In the larger game of two-player no-limit Texas hold 'em poker, we're right at the cusp of it. We currently have the world's best computer program, but we are still not better than the very best dozen or so humans."

Games are typically chosen for the specific challenges that researchers want their AI to be able to overcome. With board games such as chess and Go, computer programs are put through the ringer to see if they can learn from past matches and determine the best next move. With Texas hold 'em, it's about interpreting actions as signals, and figuring out the next best move without knowing what the opponent has in their hand.

All these things have real-world applications, Sandholm said. In complete information games, AI can help people search through large databases and do calculations and modeling. In incomplete information games, it can be used in situations where there are lots of unknown factors, such as negotiations, cybersecurity and auctions, and even in planning medical treatment.

Some robotics experts believe AI will one day get messier than that, taking algorithms out of controlled game environments and into the open world.

"That's next," said Thomas Johnson, founder of MotionFigures, a startup that is bringing robotics and AI to toys. "The challenge will be putting AI algorithms into practice in open environments where the rules are not all given to it upfront, and adaptation is required to be successful."

Johnson points to DARPA challenges - competitions run by the U.S. Department of Defense - as an example of robots and AI being put to the test in the real world. Past DARPA challenges forced researchers to build robots that can walk up and down hills; many fell over or staggered like toddlers learning to walk.

"We take for granted things like balance and vision," he said. "But for a robot, to walk up and down hills requires so many complicated decisions to be made in real time, and it's really difficult to do."

There's still more that can be done in controlled environments, though.

Algorithms such as AlphaGo only know how to play Go. Oren Etzioni, executive director of the Allen Institute for Artificial Intelligence, believes the next step could be for AI to learn to play (and beat world champions) at any game. Or, as his institute is doing, putting AI through standardizing testing. As in, the SATs. Or eighth grade science tests.

"The scientific part of it isn't complicated," Etzioni said.

After all, it's not hard to get a computer program to remember and regurgitate facts. What is hard is getting computers to apply their knowledge to everyday situations.

"The question in the test doesn't require the computer to give a definition of gravity or recite an equation, but to describe a situation," he said. "For example, 'There's a ball rolling down a hill.' This is the paradox: The hard part for the machine is easy for the human. The machine is struggling to figure out what does it mean when it says 'the ball is rolling down the hill'?"

There are no shortages of benchmarks and tests that come after Go. And that doesn't even get into benchmarks for different types of , such as emotionally intelligent AI, speech recognizing robots, or computers designed to understand language.

"It's an exciting field to be in," said Johnson. "I think it's incredible and it makes me stop and think what the next 10, 20 years is gonna bring in AI."

Explore further: Google's Go triumph is a milestone for artificial intelligence research

Related Stories

AI beats human Go grandmaster... again (Update)

March 10, 2016

A Google-developed supercomputer bested a South Korean Go grandmaster again Thursday, taking a commanding 2-0 lead in a five-game series that has become a stunning global debut for a new style of "intuitive" artificial intelligence ...

Recommended for you

Cryptocurrency rivals snap at Bitcoin's heels

January 14, 2018

Bitcoin may be the most famous cryptocurrency but, despite a dizzying rise, it's not the most lucrative one and far from alone in a universe that counts 1,400 rivals, and counting.

Top takeaways from Consumers Electronics Show

January 13, 2018

The 2018 Consumer Electronics Show, which concluded Friday in Las Vegas, drew some 4,000 exhibitors from dozens of countries and more than 170,000 attendees, showcased some of the latest from the technology world.

Finnish firm detects new Intel security flaw

January 12, 2018

A new security flaw has been found in Intel hardware which could enable hackers to access corporate laptops remotely, Finnish cybersecurity specialist F-Secure said on Friday.

1 comment

Adjust slider to filter visible comments by rank

Display comments: newest first

not rated yet Mar 16, 2016
Wow. Now here's a journalist who is really confused. Since intelligence is required to walk up a hill, neither squirrels nor beetles are capable of it, right? Not to mention sight, sound, touch, smell,... Confounding autonomic sensory perception with intelligence isn't very intelligent. Nor is it very 'smart' to confuse IQ tests (which test something we unfortunately (and both ignorantly and misleadingly) call "Intelligence Quotient" with intelligence metrics. IQ correlates well with is grades in school. School grades do NOT strongly correlate with life success (although those who do the worst in testing typically also have low life-time accomplishment ratings). Critical thinking, effective thinking, etc. are not well measured by IQ tests. Going after the wrong target (or suggesting that A.I. researchers are pursuing them) is not intelligent.

Please sign in to add a comment. Registration is free, and takes less than a minute. Read more

Click here to reset your password.
Sign in to get notified via email when new comments are made.