{"id":202243,"date":"2015-10-18T20:40:55","date_gmt":"2015-10-19T00:40:55","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/are-we-smart-enough-to-control-artificial-intelligence-mit.php"},"modified":"2015-10-18T20:40:55","modified_gmt":"2015-10-19T00:40:55","slug":"are-we-smart-enough-to-control-artificial-intelligence-mit","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/are-we-smart-enough-to-control-artificial-intelligence-mit.php","title":{"rendered":"Are We Smart Enough to Control Artificial Intelligence? | MIT &#8230;"},"content":{"rendered":"<p><p>    Years ago I had coffee with a friend who ran a startup. He had    just turned 40. His father was ill, his back was sore, and he    found himself overwhelmed by life. Dont laugh at me, he    said, but I was counting on the singularity.  <\/p>\n<p>    My friend worked in technology; hed seen the changes that    faster microprocessors and networks had wrought. It wasnt that    much of a step for him to believe that before he was beset by    middle age, the intelligence of machines would exceed that of    humansa moment that futurists call the singularity. A    benevolent superintelligence might analyze the human genetic    code at great speed and unlock the secret to eternal youth. At    the very least, it might know how to fix your back.  <\/p>\n<p>    But what if it wasnt so benevolent? Nick Bostrom, a    philosopher who directs the Future of Humanity Institute at the    University of Oxford, describes the following scenario in his    book Superintelligence, which has prompted a great    deal of debate about the future of artificial intelligence.    Imagine a machine that we might call a paper-clip    maximizerthat is, a machine programmed to make as many paper    clips as possible. Now imagine that this machine somehow became    incredibly intelligent. Given its goals, it might then decide    to create new, more efficient paper-clip-manufacturing    machinesuntil, King Midas style, it had converted essentially    everything to paper clips.  <\/p>\n<p>    No worries, you might say: you could just program it to make    exactly a million paper clips and halt. But what if it makes    the paper clips and then decides to check its work? Has it    counted correctly? It needs to become smarter to be sure. The    superintelligent machine manufactures some as-yet-uninvented    raw-computing material (call it computronium) and uses that    to check each doubt. But each new doubt yields further digital    doubts, and so on, until the entire earth is converted to    computronium. Except for the million paper clips.  <\/p>\n<p>    Bostrom does not believe that the paper-clip maximizer will    come to be, exactly; its a thought experiment, one designed to    show how even careful system design can fail to restrain    extreme machine intelligence. But he does believe that    superintelligence could emerge, and while it could be great, he    thinks it could also decide it doesnt need humans around. Or    do any number of other things that destroy the world. The title    of chapter 8 is: Is the default outcome doom?  <\/p>\n<p>    If this sounds absurd to you, youre not alone. Critics such as    the robotics pioneer Rodney Brooks say that people who fear a    runaway AI misunderstand what computers are doing when we say    theyre thinking or getting smart. From this perspective, the    putative superintelligence Bostrom describes is far in the    future and perhaps impossible.  <\/p>\n<p>    Yet a lot of smart, thoughtful people agree with Bostrom and    are worried now. Why?  <\/p>\n<p>    Volition  <\/p>\n<p>    The question Can a machine think? has shadowed computer    science from its beginnings. Alan Turing proposed in 1950 that    a machine could be taught like a child; John McCarthy, inventor    of the programming language LISP, coined the term artificial    intelligence in 1955. As AI researchers in the 1960s and 1970s    began to use computers to recognize images, translate between    languages, and understand instructions in normal language and    not just code, the idea that computers would eventually develop    the ability to speak and thinkand thus to do evilbubbled into    mainstream culture. Even beyond the oft-referenced HAL from    2001: A Space Odyssey, the 1970 movie Colossus:    The Forbin Project featured a large blinking mainframe    computer that brings the world to the brink of nuclear    destruction; a similar theme was explored 13 years later in    WarGames. The androids of 1973s Westworld    went crazy and started killing.  <\/p>\n<p>      Extreme AI predictions are comparable to seeing more      efficient internal combustion engines and jumping to the      conclusion that the warp drives are just around the corner,      Rodney Brooks writes.    <\/p>\n<p>    When AI research fell far short of its lofty goals, funding    dried up to a trickle, beginning long AI winters. Even so,    the torch of the intelligent machine was carried forth in the    1980s and 90s by sci-fi authors like Vernor Vinge, who    popularized the concept of the singularity; researchers like    the roboticist Hans Moravec, an expert in computer vision; and    the engineer\/entrepreneur Ray Kurzweil, author of the 1999 book    The Age of Spiritual Machines. Whereas Turing had    posited a humanlike intelligence, Vinge, Moravec, and Kurzweil    were thinking bigger: when a computer became capable of    independently devising ways to achieve goals, it would very    likely be capable of introspectionand thus able to modify its    software and make itself more intelligent. In short order, such    a computer would be able to design its own hardware.  <\/p>\n<p>    As Kurzweil described it, this would begin a beautiful new era.    Such machines would have the insight and patience (measured in    picoseconds) to solve the outstanding problems of    nanotechnology and spaceflight; they would improve the human    condition and let us upload our consciousness into an immortal    digital form. Intelligence would spread throughout the cosmos.  <\/p>\n<p>    You can also find the exact opposite of such sunny optimism.    Stephen Hawking has warned that because people would be unable    to compete with an advanced AI, it could spell the end of the human race. Upon    reading Superintelligence, the entrepreneur Elon Musk    tweeted: Hope were not just the biological boot loader for    digital superintelligence. Unfortunately, that is increasingly    probable. Musk then followed with a $10 million grant to the    Future of Life Institute. Not to be confused with Bostroms    center, this is an organization that says it is working to    mitigate existential risks facing humanity, the ones that    could arise from the development of human-level artificial    intelligence.  <\/p>\n<p>    No one is suggesting that anything like superintelligence    exists now. In fact, we still have nothing approaching a    general-purpose artificial intelligence or even a clear path to    how it could be achieved. Recent advances in AI, from automated    assistants such as Apples Siri to Googles driverless cars,    also reveal the technologys severe limitations; both can be    thrown off by situations that they havent encountered before.    Artificial neural networks can learn for themselves to recognize cats in photos. But they must be    shown hundreds of thousands of examples and still end up much    less accurate at spotting cats than a child.  <\/p>\n<p>    This is where skeptics such as Brooks, a founder of    iRobot and Rethink Robotics, come in. Even if its    impressiverelative to what earlier computers could managefor    a computer to recognize a picture of a cat, the machine has no    volition, no sense of what cat-ness is or what else is    happening in the picture, and none of the countless other    insights that humans have. In this view, AI could possibly lead    to intelligent machines, but it would take much more work than    people like Bostrom imagine. And even if it could happen,    intelligence will not necessarily lead to sentience.    Extrapolating from the state of AI today to suggest that    superintelligence is looming is comparable to seeing more    efficient internal combustion engines appearing and jumping to    the conclusion that warp drives are just around the corner,    Brooks wrote recently on Edge.org. Malevolent AI is nothing to worry    about, he says, for a few hundred years at least.  <\/p>\n<p>    Insurance policy  <\/p>\n<p>    Even if the odds of a superintelligence arising are very long,    perhaps its irresponsible to take the chance. One person who    shares Bostroms concerns is Stuart J. Russell, a professor of    computer science at the University of California, Berkeley.    Russell is the author, with Peter Norvig (a peer of Kurzweils    at Google), of Artificial Intelligence: A Modern    Approach, which has been the standard AI textbook for two    decades.  <\/p>\n<p>    There are a lot of supposedly smart public intellectuals who    just havent a clue, Russell told me. He pointed out that AI    has advanced tremendously in the last decade, and that while    the public might understand progress in terms of Moores Law    (faster computers are doing more), in fact recent AI work has    been fundamental, with techniques like deep learning laying the    groundwork for computers that can automatically increase their    understanding of the world around them.  <\/p>\n<p>      Bostroms book proposes ways to align computers with human      needs. Were basically telling a god how wed like to be      treated.    <\/p>\n<p>    Because Google, Facebook, and other companies are actively    looking to create an intelligent, learning machine, he    reasons, I would say that one of the things we ought not to do    is to press full steam ahead on building superintelligence    without giving thought to the potential risks. It just seems a    bit daft. Russell made an analogy: Its like fusion research.    If you ask a fusion researcher what they do, they say they work    on containment. If you want unlimited energy youd better    contain the fusion reaction. Similarly, he says, if you want    unlimited intelligence, youd better figure out how to align    computers with human needs.  <\/p>\n<p>    Bostroms book is a research proposal for doing so. A    superintelligence would be godlike, but would it be animated by    wrath or by love? Its up to us (that is, the engineers). Like    any parent, we must give our child a set of values. And not    just any values, but those that are in the best interest of    humanity. Were basically telling a god how wed like to be    treated. How to proceed?  <\/p>\n<p>    Bostrom draws heavily on an idea from a thinker named Eliezer    Yudkowsky, who talks about coherent extrapolated volitionthe    consensus-derived best self of all people. AI would, we hope,    wish to give us rich, happy, fulfilling lives: fix our sore    backs and show us how to get to Mars. And since humans will    never fully agree on anything, well sometimes need it to    decide for usto make the best decisions for humanity as a    whole. How, then, do we program those values into our    (potential) superintelligences? What sort of mathematics can    define them? These are the problems, Bostrom believes, that    researchers should be solving now. Bostrom says it is the    essential task of our age.  <\/p>\n<p>    For the civilian, theres no reason to lose sleep over scary    robots. We have no technology that is remotely close to    superintelligence. Then again, many of the largest corporations    in the world are deeply invested in making their computers more    intelligent; a true AI would give any one of these companies an    unbelievable advantage. They also should be attuned to its    potential downsides and figuring out how to avoid them.  <\/p>\n<p>    This somewhat more nuanced suggestionwithout any claims of a    looming AI-mageddonis the basis of an    open letter on the website of the Future of Life Institute,    the group that got Musks donation. Rather than warning of    existential disaster, the letter calls for more research into    reaping the benefits of AI while avoiding potential pitfalls.    This letter is signed not just by AI outsiders such as Hawking,    Musk, and Bostrom but also by prominent computer scientists    (including Demis Hassabis, a top AI researcher). You    can see where theyre coming from. After all, if they develop    an artificial intelligence that doesnt share the best human    values, it will mean they werent smart enough to control their    own creations.  <\/p>\n<p>    Paul Ford, a freelance writer in New York, wrote about    Bitcoin in March\/April 2014.  <\/p>\n<p>    Gain the insight you need on artificial intelligence at EmTech    MIT.  <\/p>\n<p>    Register today  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Read the rest here:<\/p>\n<p><a target=\"_blank\" href=\"http:\/\/www.technologyreview.com\/review\/534871\/our-fear-of-artificial-intelligence\/\" title=\"Are We Smart Enough to Control Artificial Intelligence? | MIT ...\">Are We Smart Enough to Control Artificial Intelligence? | MIT ...<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Years ago I had coffee with a friend who ran a startup. He had just turned 40. His father was ill, his back was sore, and he found himself overwhelmed by life.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/are-we-smart-enough-to-control-artificial-intelligence-mit.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-202243","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/202243"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=202243"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/202243\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=202243"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=202243"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=202243"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}