{"id":186391,"date":"2015-02-25T13:40:58","date_gmt":"2015-02-25T18:40:58","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/ai-masters-49-atari-2600-games-without-instructions.php"},"modified":"2015-02-25T13:40:58","modified_gmt":"2015-02-25T18:40:58","slug":"ai-masters-49-atari-2600-games-without-instructions","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/ai-masters-49-atari-2600-games-without-instructions.php","title":{"rendered":"AI masters 49 Atari 2600 games without instructions"},"content":{"rendered":"<p><p>    The venerable Atari 2600.  <\/p>\n<p>    Artificial intelligence, machines and software with the ability    to think for themselves, can be used for a variety of    applications ranging from military technology to everyday    serviceslike automated telephone systems. However, none    of the systems that currently exist exhibit learning abilities    that would match the human intelligence. Recently, scientists    have wondered whether an artificial agent could be given a tiny    bit of human-like intelligence by modeling the algorithm on    aspects of the primate neural system.  <\/p>\n<p>    Using a bio-inspired system architecture, scientists have    created a single algorithm that is actually able to develop    problem-solving skills when presented with challenges that can    stump some humans. And then they immediately put it to use    learning a set of classic video games.  <\/p>\n<p>    Scientists developed the novel agent (they called it the Deep    Q-network), one that combined reinforcement learning with    what's termed a \"deep convolutional network,\" a layered system    of artificial neural networks. Deep-Q is able to understand    spatial relationships between different objects in an image,    such as distance from one another, in such a sophisticated way    that it can actually re-envision the scene from a    differentviewpoint. This type of system was inspired by    early work done on the visual cortex.  <\/p>\n<p>    Scientists considered tasks in which Deep-Q was able to    interact with the environment through a sequence of    observations, actions, and rewards, with an ultimate goal of    interacting in a way to maximize reward. Reinforcement learning    systems sound like a simple approach to developing artificial    intelligenceafter all, we have all seen that small children    are able to learn from their mistakes. Yet when it comes to    designing artificial intelligence, it is much trickier to    ensure all the components necessary for this type of learning    are actually included. As a result, artificial reinforcement    learning systems are usually quite unstable.  <\/p>\n<p>    Here, these scientists addressed previous instability issues in    creatingDeep-Q. One important mechanism that they    specifically added to Deep-Q was experience replay. This    element allows the system to store visual information about    experiences and transitions much like our memory works. For    example, if a small child leaves home to go to a playground, he    will still remember what home looks like at the playground. If    he is running and he trips over a tree root, he will remember    that bad outcome and try to avoid tree roots in the future.  <\/p>\n<p>    Using these abilities, Deep-Q is able to    performreinforcement learning, using rewards to    continuously establishvisual relationships between    objects and actions within the convolution network. Over time,    it identifiesvisual aspects of the environment that would    promote good outcomes.  <\/p>\n<p>    This bio-inspired approach is based on evidence that rewards    during perceptual learning may influence the way images and    sequences of events or resulting outcomes are processed within    the primate visual cortex. Additionally, evidence suggests that    in the mammalian brain, the hippocampus may actually support    the physical realization of the processes involved in the    experience replay algorithm.  <\/p>\n<p>    It takes a few hundred tries, but the neural networks    eventually figure out the rules, then later discover    strategies.  <\/p>\n<p>    Scientists tested Deep Qs problem-solving abilities on the    Atari 2600 gaming platform. Deep-Q learned not only the rules    for a variety of games (49 games in total) in a range of    different environments, but the behaviors required to maximize    scores. It did so with minimal prior knowledge, receiving only    visual images (in pixel form) and the game score as inputs. In    these experiments, the authors used the same algorithm, network    architecture, and hyperparameters on each gamethe exact same    limitations a human player would have, given we can't swap    brains out. Notably, these game genres varied from boxing to    car-racing, representing a tremendous range of inputs and    challenges.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Continued here: <\/p>\n<p><a target=\"_blank\" href=\"http:\/\/arstechnica.com\/science\/2015\/02\/ai-masters-49-atari-2600-games-without-instructions\" title=\"AI masters 49 Atari 2600 games without instructions\">AI masters 49 Atari 2600 games without instructions<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> The venerable Atari 2600.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/ai-masters-49-atari-2600-games-without-instructions.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-186391","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/186391"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=186391"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/186391\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=186391"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=186391"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=186391"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}