Newest version of self-taught, ‘superhuman’ AI now even smarter: makers

PARIS – The updated version of the computer that stunned humanity by beating the best mortal players at a strategy board game requiring intuition has become even smarter, its makers said Wednesday.

Even more startling, the new version of AlphaGo is entirely self-taught — a major step toward the rise of machines that achieve superhuman abilities “with no human input,” they reported in the science journal Nature.

Dubbed AlphaGo Zero, the artificial intelligence system learned by itself, within days, to master the ancient Chinese board game known as go — said to be the most complex two-person game ever invented.

It came up with its own, novel moves to eclipse all the go acumen that humans have acquired over thousands of years.

After just three days of self-training, it was put to the ultimate test against its forerunner AlphaGo, which had dethroned the top human champs.

AlphaGo Zero won by 100 games to zero.

“AlphaGo Zero not only rediscovered the common patterns and openings that humans tend to play … it ultimately discarded them in preference for its own variants, which humans don’t even know about or play at the moment,” said AlphaGo lead researcher David Silver.

The 3,000-year-old game, played with black and white stones on a board, has more possible configurations of moves than there are atoms in the universe.

AlphaGo made world headlines with its shock 4-1 victory in March 2016 over 18-time go champion Lee Sedol, one of the game’s all-time masters.

Lee’s defeat showed that AI was progressing faster than widely thought, experts said at the time. Some have called for rules to make sure that artificial intelligence always remains completely under human control.

In May this year, an updated AlphaGo Master program beat world No. 1 Ke Jie in three matches out of three.

Unlike its predecessors, which trained on data from thousands of human games before practicing by playing against itself, AlphaGo Zero did not learn from humans or by playing against them, according to researchers at DeepMind, the British artificial intelligence company developing the system.

“All previous versions of AlphaGo … were told: ‘Well, in this position the human expert played this particular move, and in this other position the human expert played here,’ ” Silver said in a video explaining the advance.

AlphaGo Zero skipped this step.

Instead, it was programmed to respond to reward — a positive point for a win versus a negative point for a loss.

Starting with just the rules of go and no instructions, the system learned the game, devised strategies and improved as it competed against itself, starting with “completely random play.” This is a trial-and-error process known as “reinforcement learning.”

Unlike its predecessors, AlphaGo Zero “is no longer constrained by the limits of human knowledge,” Silver and DeepMind CEO Demis Hassabis wrote in a blog.

AlphaGo Zero used a single machine — a human brain-mimicking “neural network” — compared to the multiple-machine “brain” that beat Lee.

It had four data processing units, compared to AlphaGo’s 48, and played 4.9 million training games over three days, compared to 30 million over several months.

“People tend to assume that machine learning is all about big data and massive amounts of computation, but actually what we saw with AlphaGo Zero is that algorithms matter much more,” said Silver.

The findings suggested that AI based on reinforcement learning performed better than those that rely on human expertise, Satinder Singh of the University of Michigan wrote in a commentary also carried by Nature.

“However, this is not the beginning of any end because AlphaGo Zero, like all other successful AI so far, is extremely limited in what it knows and in what it can do compared with humans and even other animals,” he said.

AlphaGo Zero’s ability to learn on its own “might appear creepily autonomous,” added Anders Sandberg of the Future of Humanity Institute at Oxford University.

But there is an important difference, he said, “between the general-purpose smarts humans have and the specialized smarts” of computer software.

“What DeepMind has demonstrated over the past years is that one can make software that can be turned into experts in different domains … but it does not become generally intelligent.”

It was also worth noting that AlphaGo was not programming itself, said Sandberg.

“The clever insights making Zero better was due to humans, not any piece of software suggesting that this approach would be good. I would start to get worried when that happens.”