Post Thumb

DeepMind's new algorithm adds 'memory' to AI

Share it

The system could learn to defeat the games, and score higher than humans, but not remember how it had done so.

“Previously, we had a system that could learn to play any game, but it could only learn to play one game,” James Kirkpatrick, a research scientist at DeepMind and the lead author of its new research paper, tells WIRED. “Here we are demonstrating a system that can learn to play several games one after the other”.

The work, published in the Proceedings of the National Academy of Sciences journal, explains how DeepMind’s AI can learn in sequences using supervised learning and reinforcement learning tests.

“That way there is room to learn the new task but the changes we’ve applied do not override what we’ve learned before”. “Previously, DQN had to learn how to play each game individually,” the paper says. Essentially, the deep neural network using the EWC algorithm was able to learn to play one game and then transfer what it had learnt to play a brand new game.

While it is able to learn from its previous experiences and retain the most useful information, it isn’t able to perform as well as a neural network that completes just one game.

read more...

Article originally posted at www.wired.co.uk

Post Author: Carla Parsons

2 thoughts on “DeepMind's new algorithm adds 'memory' to AI

    Daniel Kirstein

    (October 29, 2017 - 7:44 am)

    Hope they find the plug, not they decide we good energy source
    to them

    Keith Gaddis

    (October 29, 2017 - 7:44 am)

    Very cool

Leave a Reply

Your email address will not be published. Required fields are marked *