DeepMind’s AI is an Atari gaming pro now

Google-backed startup DeepMind Technologies has built an artificial intelligence agent that can learn to successfully play forty nine classic Atari games by itself, with minimal input.

Cofounder and WIRED2014 speaker Demis Hassabis called the stir, detailed in a paper published in Nature, “the very first significant rung on the ladder to proving general learning systems can work”. “It’s the very first time that anyone has built a single general learning system that can learn directly from practice,” he told journalists ahead of the announcement. “The ultimate aim is to build general purpose wise machines — that’s many decades away.

ADVERTISEMENT

But this is going from pixels to deeds, and it can work on a challenging task even humans find difficult. It’s a baby step, but an significant one.”

Google acquired the London-based startup for a reported sum of £300 million in January 2014, following rumours that Facebook was also interested. Later that year, speaking at WIRED2014, former child chess prodigy Hassabis spoke about how DeepMind’s AI — or agent, as it is referred to internally — had developed a flawless

Breakout strategy (engineering a tunnel so the ball hits the top of the screen) after being left to play the Atari game overnight. “It’s now better at playing the game than any human. It has ideally modelled this sophisticated stream,” Hassabis said at the time.

In the Nature paper published today (25 February), however, Hassabis and his coauthors expose how deep Q-network (DQN) combined a very human type of learning known as reinforcement learning, with deep learning — the method Google employed back in

ADVERTISEMENT

2012 to train its AI to recognise pictures of cats in YouTube movies. Hassabis noted this is the very first time an open system has combined the two approaches.

ADVERTISEMENT

Hassabis, who also has a PhD in cognitive neuroscience from University College London, believes focusing on the biological functions of learning could be the key to cracking AI. “We learn through things like memory replay through the hippocampus, so there are crossovers inbetween neuroscience and this,” he told journalists.

DQN was only given pixel and score information, but was otherwise left to its own devices to create strategies and play forty nine Atari games. This is compared to much-publicised AI systems such as IBM’s Watson or Deep Blue, which rely on pre-programmed information to hone their abilities. “With Deep Blue there were chess grandmasters on the development team distilling their chess skill into the programme and it executed it without learning anything,” said Hassabis. “Ours learns from the ground up. We give it a perceptual practice and it learns from that directly. It learns and adapts from unexpected things, and programme designers don’t have to know the solution themselves.” “The interesting and cool thing about AI tech is that it can actually train you, as the creator, something fresh. I can’t think of many other technologies that can do that.”

ADVERTISEMENT

As a result of this treatment DQN — which trained on each game for two weeks — achieved more than seventy five percent of the human score on more than half the games, and achieved better results than AIs using just reinforcement learning. It was even able to come up with loopholes in the games that the team did not know about. A supercomputer was not used to process the computations, but that would make progress even quicker the team suggests. “It is worth noting that the games in which DQN excels are enormously varied in their nature, from side-scrolling shooters (Sea Raid) to boxing games (Boxing) and three-dimensional car-racing games (Enduro),” the team writes in the Nature paper. This is of paramount importance because DeepMind is of the belief that its AI is on the road to creating a general AI that can be applied to any decision-making situation. Any information could be used as the input, to help with these general applications — the team just chose to only provide pixel and score feedback in this example.

Before being acquired by Google, the company was hopeful its technology could one day be applied to climate science or disease modelling. But for now the team will be moving from Atari games to games of the 90s — including 3D and racing games “where the challenge is much greater”. The long-term purpose is to then apply what is learned to Google’s own products, including Search, Translate, and presumably its driverless car tech.

Related movie:

Related Posts

Zipcar Introduces Car Sharing Program in Collingswood, NJ CAMBRIDGE, Mass. , Sept. 29, two thousand eleven /PRNewswire/ — Zipcar, Inc. (Nasdaq: ZIP), the world’s leading car sharing network, and the Borough of Collingswood , today announced Zipcar’s expansion into Collingswood, Fresh Jersey , located just ten minutes outside of Philadelphia .

WSB Atlanta SIGN IN Sign in using your wsbtv profile Sign in using you account with: Sign Up / Sign In Welcome Back Sign Up / Sign In Welcome back. Please sign in You’re Almost Done! Please confirm the information below before signing in. REGISTER By submitting your registration information, you agree to our Terms […]

Woman stabbed in Huntington – WBOY – Clarksburg, Morgantown: News, Sports, Weather UPDATE: According to police, the woman received a petite cut on her arm from a box cutter. They tell us they’re receiving conflicting stories. No charges will be filed and no arrests will be made. ORIGINAL: According to Cabell County Dispatchers, a woman […]

Leave a Reply

Your email address will not be published. Required fields are marked *