{"id":177791,"date":"2017-02-15T21:20:27","date_gmt":"2017-02-16T02:20:27","guid":{"rendered":"http:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/googles-ai-learns-betrayal-and-aggressive-actions-pay-off-big-think\/"},"modified":"2017-02-15T21:20:27","modified_gmt":"2017-02-16T02:20:27","slug":"googles-ai-learns-betrayal-and-aggressive-actions-pay-off-big-think","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/ai\/googles-ai-learns-betrayal-and-aggressive-actions-pay-off-big-think\/","title":{"rendered":"Google&#8217;s AI Learns Betrayal and &quot;Aggressive&quot; Actions Pay Off &#8211; Big Think"},"content":{"rendered":"<p><p>    As the development of artificial intelligence continues at    breakneck speed, questions about whether we understand what we    are getting ourselves into persist. One fear is that    increasingly intelligent robots will take all our jobs. Another    fear is that we will create a world where a superintelligence    will one day decide that it has no need for humans. This fear    is well-explored in popular culture, through books and films    like the Terminator series.  <\/p>\n<p>    Another possibility is maybe the one that makes the most sense    - since humans are the ones creating them, the machines and    machine intelligences are likely to behave just like humans.    For better or worse. DeepMind, Googles cutting-edge AI    company, has shown just that.  <\/p>\n<p>    The accomplishments of the DeepMind program so far include    learning from its memory, mimicking human voices, writing music, and beating the best    Go player in the world.  <\/p>\n<p>    Recently, the DeepMind team ran a series of tests to    investigate how the AI would respond when faced with certain    social dilemmas. In particular, they wanted to find out whether    the AI is more likely to cooperate or compete.  <\/p>\n<p>    One of the tests involved 40 million instances    of playing the computer game Gathering, during which    DeepMind showed how far its willing to go to get what it    wants. The game was chosen because it encapsulates aspects of    the classic Prisoners Dilemma from game theory.  <\/p>\n<p>    Pitting AI-controlled characters (called agents) against each    other, DeepMind had them compete to gather the most virtual    apples. Once the amount of available apples got low, the AI    agents started to display \"highly aggressive\" tactics,    employing laser beams to knock each other out. They would also    steal the opponents apples.  <\/p>\n<p>    Heres how one of those games played out:  <\/p>\n<p>    The DeepMind AI agents are in blue and red. The apples are    green, while the laser beams are yellow.  <\/p>\n<p>    The DeepMind team described their test in a blog postthis way:  <\/p>\n<p>      We let the agents      play this game many thousands of times and let them learn how      to behave rationally using deep multi-agent      reinforcement learning. Rather naturally, when there are      enough apples in the environment, the agents learn to      peacefully coexist and collect as many apples as they can.      However, as the number of apples is reduced, the agents learn      that it may be better for them to tag the other agent to give      themselves time on their own to collect the scarce      apples.    <\/p>\n<p>    Interestingly, what appears to have happened is that the AI    systems began to develop some forms of human behavior.  <\/p>\n<p>      This model... shows that some aspects of      human-like behaviour emerge as a product of      the environment and learning. Less aggressive policies emerge      from learning in relatively abundant environments with less      possibility for costly action.The greed motivation      reflects the temptation to take out a rival and collect all      the apples oneself, said Joel Z. Leibo from the DeepMind      team to Wired.    <\/p>\n<p>    Besides the fruit gathering, the AI was also tested via a    Wolfpack hunting game. In it, two AI characters in the    form of wolves chased a third AI agent - the prey. Here the    researchers wanted to see if the AI characters would choose to    cooperate to get the prey because they were rewarded for    appearing near the prey together when it was being    captured.  <\/p>\n<p>      \"The idea is that the prey is dangerous - a      lone wolf can overcome it, but is at risk of losing the      carcass to scavengers. However, when the two wolves capture      the prey together, they can better protect the carcass from      scavengers, and hence receive a higher reward, wrote the researchers in      their paper.    <\/p>\n<p>    Indeed, the incentivized cooperation strategy won out in this    instance, with the AI choosing to work together.  <\/p>\n<p>    This is how that test panned out:  <\/p>\n<p>    The wolves are red, chasing the blue dot (prey), while    avoiding grey obstacles.  <\/p>\n<p>    If you are thinking Skynet is here, perhaps the silver lining    is that the second test shows how AIs self-interest can    include cooperation rather than the all-out competitiveness of    the first test. Unless, of course, its cooperation to hunt down    humans.  <\/p>\n<p>    Here's a chart showing the results of the game tests that shows    a clear increase in aggression during \"Gathering\":  <\/p>\n<\/p>\n<p>    Movies aside, the researchers are working to figure    out how AI can eventually control complex multi-agent systems such as the economy,    traffic systems, or the ecological health of our planet  all    of which depend on our continued cooperation.  <\/p>\n<p>    One nearby AI implementation    where this could be relevant - self-driving cars which will    have to choose safest routes, while keeping the objectives of    all the parties involved under consideration.  <\/p>\n<p>    The warning from the tests is that if the objectives are not    balanced out in the programming, the AI might act selfishly,    probably not for everyones benefit.  <\/p>\n<p>    Whats next for the DeepMind team? Joel Leibo    wants the AI to go deeper into the motivations behind    decision-making:  <\/p>\n<p>      Going forward it would be interesting to      equip agents with the ability to reason about other agents      beliefs and goals, said Leibo to      Bloomberg.    <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Read more:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow\" href=\"http:\/\/bigthink.com\/paul-ratner\/googles-deepmind-ai-displays-highly-aggressive-human-like-behavior-in-tests\" title=\"Google's AI Learns Betrayal and &quot;Aggressive&quot; Actions Pay Off - Big Think\">Google's AI Learns Betrayal and &quot;Aggressive&quot; Actions Pay Off - Big Think<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> As the development of artificial intelligence continues at breakneck speed, questions about whether we understand what we are getting ourselves into persist. One fear is that increasingly intelligent robots will take all our jobs <a href=\"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/ai\/googles-ai-learns-betrayal-and-aggressive-actions-pay-off-big-think\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":8,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[187743],"tags":[],"class_list":["post-177791","post","type-post","status-publish","format-standard","hentry","category-ai"],"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/177791"}],"collection":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/users\/8"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/comments?post=177791"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/177791\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/media?parent=177791"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/categories?post=177791"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/tags?post=177791"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}