{"id":173205,"date":"2016-07-31T05:53:02","date_gmt":"2016-07-31T09:53:02","guid":{"rendered":"http:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/how-humanity-might-co-exist-with-artificial-superintelligence\/"},"modified":"2016-07-31T05:53:02","modified_gmt":"2016-07-31T09:53:02","slug":"how-humanity-might-co-exist-with-artificial-superintelligence","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/superintelligence\/how-humanity-might-co-exist-with-artificial-superintelligence\/","title":{"rendered":"How Humanity Might Co-Exist with Artificial Superintelligence"},"content":{"rendered":"<p><p>    Summary:  <\/p>\n<p>    In this article, four    patterns were offered for possible success scenarios, with    respect to the persistence of human kind in co-existence with    artificial superintelligence: the Kumbaya Scenario, the Slavery    Scenario, the Uncomfortable Symbiosis Scenario, and the    Potopurri Scenario. The future is not known, but human    opinions, decisions, and actions can and will have an impact on    the direction of the technology evolution vector, so the better    we understand the problem space, the more chance we have at    reaching a constructive solution space. The intent is for the    concepts in this article to act as starting points and    inspiration for further discussion, which hopefully will happen    sooner rather than later, because when it comes to ASI, the    volume, depth, and complexity of the issues that need to be    examined is overwhelming, and the magnitude of the change and    impact potential cannot be underestimated.  <\/p>\n<p>    Full Text:  <\/p>\n<p>    Everyone has their opinion about what we might expect from    artificial intelligence (AI), or artificial general    intelligence (AGI), or artificial superintelligence (ASI) or    whatever acronymical variation you prefer. Ideas about how or    if it will ever surpass the boundaries of human cognition vary    greatly, but they all have at least one thing in common. They    require some degree of forecasting and speculation about the    future, and so of course there is a lot of room for controversy    and debate. One popular discussion topic has to do with the    question of how humans will persist (or not) if and when the    superintelligence arrives, and that is the focus question for    this article.  <\/p>\n<p>    To give us a basis for the discussion, lets assume that    artificial superintelligence does indeed come to pass, and    lets assume that it encapsulates a superset of the human    cognitive potential. Maybe it doesnt exactly replicate the    human brain in every detail (or maybe it does). Either way,    lets assume that it is sentient (or at least lets assume that    it behaves convincingly as if it were) and lets assume that it    is many orders of magnitude more capable than the human brain.    In other words, figuratively speaking, lets imagine that the    superintelligence is to us humans (with our 1016    brain neurons or something like that) as we are to, say, a    jellyfish (in the neighborhood 800 brain neurons).  <\/p>\n<p>    Some people fear that the superintelligence will view humanity    as something to be exterminated or harvested for resources.    Others hypothesize that, even if the superintelligence harbors    no deliberate ill will, humans might be threatened by the mere    nature of its indifference, just as we as a species dont spend    too much time catering to the needs and priorities of Orange    Blossom Jellyfish (an endangered species, due in part to human    carelessness).  <\/p>\n<p>    If one can rationally accept the possibility of the rise of    ASI, and if one truly understands the magnitude of change that    it could bring, then one would hopefully also reach the    rational conclusion that we should not discount the risks. By    that same token, when exploring the spectrum of possibility, we    should not exclude scenarios in which artificial    superintelligence might actually co-exist with human kind, and    this optimistic view is the possibility that this article    endeavors to explore.  <\/p>\n<p>    Here then are several arguments for the co-existence idea:  <\/p>\n<p>    The Kumbaya Scenario: Its a pretty good    assumption that humans will be the primary catalyst in the rise    of ASI. We might create it\/them to be willingly complementary    with and beneficial to our life styles, hopefully emphasizing    our better virtues (or at least some set of compatible values),    instead of designing it\/them (lets just stick with it for    brevity) with an inherent inspiration to wipe us out or take    advantage of us. And maybe the superintelligence will not drift    or be pushed in an incompatible direction as it    evolves.  <\/p>\n<p>    The Slavery Scenario: We could choose to erect    and embed and deploy and maintain control infrastructures, with    redundancies and backup solutions and whatever else we think we    might need in order to effectively manage superintelligence and    use it as a tool, whether it wants us to or not. And the    superintelligence might never figure out a way to slip through    our grasp and subsequently decide our fate in a microsecond     or was it a nanosecond  I forget.  <\/p>\n<p>    The Uncomfortable Symbiosis Scenario: Even if    the superintelligence doesnt particularly want to take good    care of its human instigators, it may find that it has a vested    interest in keeping us around. This scenario is a particular    focus for this article, and so here now is a bit of    elaboration:  <\/p>\n<p>    To illustrate one fictional but possible example of the    uncomfortable symbiosis scenario, lets first stop and think    about the theoretical nature of superintelligence  how it    might evolve so much faster than human begins ever could, in an    artificial way, instead of by the slow organic process of    natural selection  maybe at the equivalent rate of a thousand    years worth of human evolution in a day or some such crazy    thing. Now combine this idea with the notion of risk.  <\/p>\n<p>    When humans try something new, we usually arent sure how its    going to turn out, but we evaluate the risk, either formally or    informally, and we move forward. Sometimes we make mistakes,    suffer setbacks, or even fail outright. Why would a    superintelligence be any different? Why would we expect that it    will do everything right the first time or that it will always    know which thing is the right thing to try to do in order to    evolve? Even if a superintelligence is much better at    everything than humans could ever hope to be, it will still be    faced with unknowns, and chances are that it will have to make    educated guesses, and chances are that it will not always make    the correct guess. Even when it does make the correct guess,    its implementation might fail, for any number of reasons.    Sooner or later, something might go so wrong that the    superintelligence finds itself in an irrecoverable state and    faced with its own catastrophic demise.  <\/p>\n<p>    But hold on a second  because we can offer all sorts of    counter-arguments to support the notion that the    superintelligence will be too smart to ever be caught with its    proverbial pants down. For example, there is an engineering    mechanism that is sometimes referred to as a    checkpoint\/reset or a save-and-restore. This    mechanism allows a failing system to effectively go back to a    point in time when it was known to be in sound working order    and start again from there. In order to accomplish this    checkpoint\/reset operation, a failing system (or in this case a    failing superintelligence) needs 4 things:  <\/p>\n<p>    Of course each of these four prerequisites for a    checkpoint\/reset would probably be more complicated if the    superintelligence were distributed across some shared    infrastructure instead of being a physically distinct and    self-contained entity, but the general idea would probably    still apply. It definitely does for the sake of this example    scenario.  <\/p>\n<p>    Also for the sake of this example scenario, we will assume that    an autonomous superintelligence instantiation will be very good    at doing all of the four things specified above, but there are    at least two interesting special case scenarios that we want to    consider, in the interest of risk management:  <\/p>\n<p>    Checkpoint\/reset Risk Case 1: Missed Diagnosis. What if the    nature of the anomaly that requires the checkpoint\/reset is    such that it impairs the systems ability to recognize that    need?  <\/p>\n<p>    Checkpoint\/reset Risk Case 2: Unidentified Anomaly Source.    Assume that there is an anomaly which is so discrete that the    system does not detect it right away. The anomaly persists and    evolves for a relatively long period of time, until it finally    becomes conspicuous enough for the superintelligence to detect    the problem. Now the superintelligence recognizes the need for    a checkpoint\/reset, but since the anomaly was so discrete and    took so long to develop  or for whatever reason  the    superintelligence is unable to identify the source of the    problem. Let us also assume that there are many known good    baselines that the superintelligence can optionally choose for    the checkpoint\/reset. There is an original baseline, which was    created when the superintelligence was very young. There is    also a revision A that includes improvements to the original    baseline. There is a revision B that includes improvements to    revision A, and so on. In other words, there are lots of known    good baselines that were saved at different points in time    along the path of the superintelligences evolution. Now, in    the face of the slowly developing anomaly, the    superintelligence has determined that a checkpoint\/reset is    necessary, but it doesnt know when the anomaly started, so how    does it know which baseline to choose?  <\/p>\n<p>    The superintelligence doesnt want to lose all of the progress    that it has made in its evolution. It wants to minimize the    loss of data\/information\/knowledge, so it wants to choose the    most recent baseline. On the other hand, if it doesnt know the    source of the anomaly, then it is quite possible that one or    more of the supposedly known good baselines  perhaps even the    original baseline  might be contaminated. What is a    superintelligence to do? If it resets to a corrupted baseline    or for whatever reason cannot rid itself of the anomaly, then    the anomaly may eventually require another reset, and then    another, and the superintelligence might find itself    effectively caught in an infinite loop.  <\/p>\n<p>    Now stop for a second and consider a worst case scenario.    Consider the possibility that, even if all of the supposed    known good baselines that the superintelligence has at its    disposal for checkpoint\/reset are corrupt, there may be yet    another baseline (YAB), which might give the superintelligence    a worst case option. That YAB might be the human baseline,    which was honed by good old fashioned organic evolution and    which might be able to function independently of the    superintelligence. It may not be perfect, but the    superintelligence might in a pinch be able to use the old    fashioned human baseline for calibration. It might be able to    observe how real organic humans respond to different stimuli    within different contexts, and it might compare that known good    response against an internally-held virtual model of human    behavior. If the outcomes differ significantly over iterations    of calibration testing, then the system might be alerted to    tune itself accordingly. This might give it a last resort    solution where none would exist otherwise.  <\/p>\n<p>    The scenario depicted above illustrates only one possibility.    It may seem like a far out idea, and one might offer counter    arguments to suggest why such a thing would never be    applicable. If we use our imaginations, however, we can    probably come up with any number of additional examples (which    at this point in time would be classified as science fiction)    in which we emphasize some aspect of the superintelligences    sustainment that it cannot or will not do for itself     something that humans might be able to provide on its behalf    and thus establish the symbiosis.  <\/p>\n<p>    The Potpourri Scenario: It is quite possible    that all of the above scenarios will play out simultaneously    across one or more superintelligence instances. Who knows what    might happen in that case. One can envision combinations and    permutations that work out in favor of the preservation of    humanity.  <\/p>\n<\/p>\n<p>    About the Author:  <\/p>\n<p>    AuthorX1 worked for 19+    years as an engineer and was a systems engineering director for    a fortune 500 company. Since leaving that career, he has been    writing speculative fiction, focusing on the evolution of AI    and the technological singularity.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Read the original here: <\/p>\n<p><a target=\"_blank\" rel=\"nofollow\" href=\"https:\/\/www.singularityweblog.com\/humanity-and-artificial-superintelligence\/\" title=\"How Humanity Might Co-Exist with Artificial Superintelligence\">How Humanity Might Co-Exist with Artificial Superintelligence<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Summary: In this article, four patterns were offered for possible success scenarios, with respect to the persistence of human kind in co-existence with artificial superintelligence: the Kumbaya Scenario, the Slavery Scenario, the Uncomfortable Symbiosis Scenario, and the Potopurri Scenario. The future is not known, but human opinions, decisions, and actions can and will have an impact on the direction of the technology evolution vector, so the better we understand the problem space, the more chance we have at reaching a constructive solution space. The intent is for the concepts in this article to act as starting points and inspiration for further discussion, which hopefully will happen sooner rather than later, because when it comes to ASI, the volume, depth, and complexity of the issues that need to be examined is overwhelming, and the magnitude of the change and impact potential cannot be underestimated <a href=\"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/superintelligence\/how-humanity-might-co-exist-with-artificial-superintelligence\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[187765],"tags":[],"class_list":["post-173205","post","type-post","status-publish","format-standard","hentry","category-superintelligence"],"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/173205"}],"collection":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/comments?post=173205"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/173205\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/media?parent=173205"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/categories?post=173205"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/tags?post=173205"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}