*** Computer beats human champ in strategy challenge | THE DAILY TRIBUNE | KINGDOM OF BAHRAIN

Computer beats human champ in strategy challenge

In a milestone for artificial intelligence, a computer has beaten a human champion at a strategy game that requires "intuition" rather than brute processing power to prevail, its makers said Wednesday.

Dubbed AlphaGo, the system honed its own skills through a process of trial and error, playing millions of games against itself until it was battle-ready, and surprised even its creators with its prowess.

"AlphaGo won five-nil, and it was stronger than perhaps we were expecting," said Demis Hassabis, the chief executive of Google DeepMind, a British artificial intelligence (AI) company.

A computer defeating a professional human player at the 3,000-year-old Chinese board game known as Go, was thought to be about a decade off.

The clean-sweep victory over three-time European Go champion Fan Hui "signifies a major step forward in one of the great challenges in the development of artificial intelligence -- that of game-playing," the British Go Association said in a statement.

The two-player game is described as perhaps the most complex ever designed, with more configurations possible than there are atoms in the Universe, Hassabis says.

Players take turns placing stones on a board, trying to surround and capture the opponent's stones, with the aim of controlling more than 50 percent of the board.

There are hundreds of places where a player can place the first stone, black or white, with hundreds of ways in which the opponent can respond to each of these moves and hundreds of possible responses to each of those in turn.

"But as simple as the rules are, Go is a game of profound complexity. There are 1,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000,000 possible positions," Hassabis explained in a blog. 

Such a search base is "too enormous and too vast for brute force approaches to have any chance," added his colleague David Silver, who co-authored the paper in the science journal Nature.

So the team sought to create an AI system with a "more human-like" approach to a game Hassabis said "is played primarily through intuition and feel."

AlphaGo uses two sets of "deep neural networks" containing millions of neuron-like connections to reduce the search base to something more manageable.

The first, "policy network" narrows the search at each turn to only those moves most likely to lead to a win.

The second, "value network", estimates a winner from each move made, "rather than searching all the way to the end of the game," said Silver. 

"AlphaGo looks ahead by playing out the remainder of the game in its imagination many times over," he explained. 

"The search process itself is not based on brute force, it's based on something more akin to imagination."

AlphaGo was programmed with 30 million moves from games played by human experts, and then left to do some self-coaching.

It played "thousands and thousands of games between its neural networks, gradually improving them using a trial-and-error process known as reinforcement learning," said Silver.

The result: The value networks are able to "very accurately" estimate the eventual winner from any Go position, "a problem that was so hard it was believed to be impossible."

Most Read