A superintelligence is a hypothetical agent that possesses intelligence far surpassing that of the brightest and most gifted human minds. "Superintelligence" may also refer to a property of problem-solving systems (e.g., superintelligent language translators or engineering assistants) whether or not these high-level intellectual competencies are embodied in agents that act in the world. A superintelligence may or may not be created by an intelligence explosion and associated with a technological singularity.
University of Oxford philosopher Nick Bostrom defines superintelligence as "any intellect that greatly exceeds the cognitive performance of humans in virtually all domains of interest". The program Fritz falls short of superintelligence even though it is much better than humans at chess because Fritz cannot outperform humans in other tasks. Following Hutter and Legg, Bostrom treats superintelligence as general dominance at goal-oriented behavior, leaving open whether an artificial or human superintelligence would possess capacities such as intentionality (cf. the Chinese room argument) or first-person consciousness (cf. the hard problem of consciousness).
Technological researchers disagree about how likely present-day human intelligence is to be surpassed. Some argue that advances in artificial intelligence (AI) will probably result in general reasoning systems that lack human cognitive limitations. Others believe that humans will evolve or directly modify their biology so as to achieve radically greater intelligence. A number of futures studies scenarios combine elements from both of these possibilities, suggesting that humans are likely to interface with computers, or upload their minds to computers, in a way that enables substantial intelligence amplification.
Some researchers believe that superintelligence will likely follow shortly after the development of artificial general intelligence. The first generally intelligent machines are likely to immediately hold an enormous advantage in at least some forms of mental capability, including the capacity of perfect recall, a vastly superior knowledge base, and the ability to multitask in ways not possible to biological entities. This may give them the opportunity toeither as a single being or as a new speciesbecome much more powerful than humans, and to displace them.
A number of scientists and forecasters argue for prioritizing early research into the possible benefits and risks of human and machine cognitive enhancement, because of the potential social impact of such technologies.
Philosopher David Chalmers argues that artificial general intelligence is a very likely path to superhuman intelligence. Chalmers breaks this claim down into an argument that AI can achieve equivalence to human intelligence, that it can be extended to surpass human intelligence, and that it can be further amplified to completely dominate humans across arbitrary tasks.
Concerning human-level equivalence, Chalmers argues that the human brain is a mechanical system, and therefore ought to be emulatable by synthetic materials. He also notes that human intelligence was able to biologically evolve, making it more likely that human engineers will be able to recapitulate this invention. Evolutionary algorithms in particular should be able to produce human-level AI. Concerning intelligence extension and amplification, Chalmers argues that new AI technologies can generally be improved on, and that this is particularly likely when the invention can assist in designing new technologies.
If research into strong AI produced sufficiently intelligent software, it would be able to reprogram and improve itself a feature called "recursive self-improvement". It would then be even better at improving itself, and could continue doing so in a rapidly increasing cycle, leading to a superintelligence. This scenario is known as an intelligence explosion. Such an intelligence would not have the limitations of human intellect, and may be able to invent or discover almost anything.
Computer components already greatly surpass human performance in speed. Bostrom writes, "Biological neurons operate at a peak speed of about 200 Hz, a full seven orders of magnitude slower than a modern microprocessor (~2 GHz)." Moreover, neurons transmit spike signals across axons at no greater than 120 m/s, "whereas existing electronic processing cores can communicate optically at the speed of light". Thus, the simplest example of a superintelligence may be an emulated human mind that's run on much faster hardware than the brain. A human-like reasoner that could think millions of times faster than current humans would have a dominant advantage in most reasoning tasks, particularly ones that require haste or long strings of actions.
Another advantage of computers is modularity, that is, their size or computational capacity can be increased. A non-human (or modified human) brain could become much larger than a present-day human brain, like many supercomputers. Bostrom also raises the possibility of collective superintelligence: a large enough number of separate reasoning systems, if they communicated and coordinated well enough, could act in aggregate with far greater capabilities than any sub-agent.
There may also be ways to qualitatively improve on human reasoning and decision-making. Humans appear to differ from chimpanzees in the ways we think more than we differ in brain size or speed. Humans outperform non-human animals in large part because of new or enhanced reasoning capacities, such as long-term planning and language use. (See evolution of human intelligence and primate cognition.) If there are other possible improvements to reasoning that would have a similarly large impact, this makes it likelier that an agent can be built that outperforms humans in the same fashion humans outperform chimpanzees.
All of the above advantages hold for artificial superintelligence, but it is not clear how many hold for biological superintelligence. Physiological constraints limit the speed and size of biological brains in many ways that are inapplicable to machine intelligence. As such, writers on superintelligence have devoted much more attention to superintelligent AI scenarios.
Carl Sagan suggested that the advent of Caesarean sections and in vitro fertilization may permit humans to evolve larger heads, resulting in improvements via natural selection in the heritable component of human intelligence. By contrast, Gerald Crabtree has argued that decreased selection pressure is resulting in a slow, centuries-long reduction in human intelligence, and that this process instead is likely to continue into the future. There is no scientific consensus concerning either possibility, and in both cases the biological change would be slow, especially relative to rates of cultural change.
Selective breeding, nootropics, NSI-189, MAO-I's, epigenetic modulation, and genetic engineering could improve human intelligence more rapidly. Bostrom writes that if we come to understand the genetic component of intelligence, pre-implantation genetic diagnosis could be used to select for embryos with as much as 4 points of IQ gain (if one embryo is selected out of two), or with larger gains (e.g., up to 24.3 IQ points gained if one embryo is selected out of 1000). If this process is iterated over many generations, the gains could be an order of magnitude greater. Bostrom suggests that deriving new gametes from embryonic stem cells could be used to iterate the selection process very rapidly. A well-organized society of high-intelligence humans of this sort could potentially achieve collective superintelligence.
Alternatively, collective intelligence might be constructible by better organizing humans at present levels of individual intelligence. A number of writers have suggested that human civilization, or some aspect of it (e.g., the Internet, or the economy), is coming to function like a global brain with capacities far exceeding its component agents. If this systems-based superintelligence relies heavily on artificial components, however, it may qualify as an AI rather than as a biology-based superorganism.
A final method of intelligence amplification would be to directly enhance individual humans, as opposed to enhancing their social or reproductive dynamics. This could be achieved using nootropics, somatic gene therapy, or braincomputer interfaces. However, Bostrom expresses skepticism about the scalability of the first two approaches, and argues that designing a superintelligent cyborg interface is an AI-complete problem.
Most surveyed AI researchers expect machines to eventually be able to rival humans in intelligence, though there is little consensus on when this will likely happen. At the 2006 AI@50 conference, 18% of attendees reported expecting machines to be able "to simulate learning and every other aspect of human intelligence" by 2056; 41% of attendees expected this to happen sometime after 2056; and 41% expected machines to never reach that milestone.
In a survey of the 100 most cited authors in AI (as of May 2013, according to Microsoft academic search), the median year by which respondents expected machines "that can carry out most human professions at least as well as a typical human" (assuming no global catastrophe occurs) with 10% confidence is 2024 (mean 2034, st. dev. 33 years), with 50% confidence is 2050 (mean 2072, st. dev. 110 years), and with 90% confidence is 2070 (mean 2168, st. dev. 342 years). These estimates exclude the 1.2% of respondents who said no year would ever reach 10% confidence, the 4.1% who said 'never' for 50% confidence, and the 16.5% who said 'never' for 90% confidence. Respondents assigned a median 50% probability to the possibility that machine superintelligence will be invented within 30 years of the invention of approximately human-level machine intelligence.
Bostrom expressed concern about what values a superintelligence should be designed to have. He compared several proposals:
Responding to Bostrom, Santos-Lang raised concern that developers may attempt to start with a single kind of superintelligence.
Learning computers that rapidly become superintelligent may take unforeseen actions or robots might out-compete humanity (one potential technological singularity scenario). Researchers have argued that, by way of an "intelligence explosion" sometime over the next century, a self-improving AI could become so powerful as to be unstoppable by humans.
Concerning human extinction scenarios, Bostrom (2002) identifies superintelligence as a possible cause:
When we create the first superintelligent entity, we might make a mistake and give it goals that lead it to annihilate humankind, assuming its enormous intellectual advantage gives it the power to do so. For example, we could mistakenly elevate a subgoal to the status of a supergoal. We tell it to solve a mathematical problem, and it complies by turning all the matter in the solar system into a giant calculating device, in the process killing the person who asked the question.
In theory, since a superintelligent AI would be able to bring about almost any possible outcome and to thwart any attempt to prevent the implementation of its goals, many uncontrolled, unintended consequences could arise. It could kill off all other agents, persuade them to change their behavior, or block their attempts at interference.
Eliezer Yudkowsky explains: "The AI does not hate you, nor does it love you, but you are made out of atoms which it can use for something else."
This presents the AI control problem: how to build a superintelligent agent that will aid its creators, while avoiding inadvertently building a superintelligence that will harm its creators. The danger of not designing control right "the first time", is that a misprogrammed superintelligence might rationally decide to "take over the world" and refuse to permit its programmers to modify it once it has been activated. Potential design strategies include "capability control" (preventing an AI from being able to pursue harmful plans), and "motivational control" (building an AI that wants to be helpful).
Bill Hibbard advocates for public education about superintelligence and public control over the development of superintelligence.
- Familiarity goes beyond the field for Parkland footballs Santos vs. Freedom - lehighvalleylive.com - November 13th, 2019
- Internet freedom declined in the US and worldwide this year: report | TheHill - The Hill - November 13th, 2019
- Embrace The Freedom Of Honoring Your Word Impeccably - Inc. - November 13th, 2019
- Hayek, Republican Freedom, and the Universal Basic Income - Niskanen Center - November 13th, 2019
- Lula's Free, and He's Promising to Fight - The Nation - November 13th, 2019
- Five things to know about Freedom Never Dies, by the Sojourners - Vancouver Sun - November 13th, 2019
- The wages of freedom - The Boston Globe - November 13th, 2019
- In our opinion: Because of those who sacrifice for freedom, the importance of Veterans Day will never cease - Deseret News - November 13th, 2019
- America's Security Is Still Tied to the Fate of Freedom - The American Interest - November 13th, 2019
- CSG and Freedom expand partnership to provide business and IT services - MobileSyrup - November 13th, 2019
- Mormon quest for peace and freedom in Mexico shattered by violence and adversity - CNN - November 13th, 2019
- Geingob congratulates Angola on 44 years of freedom - New Era Live - November 13th, 2019
- The bar has always been high for Freedom footballs Jenkins; thats where he likes it - lehighvalleylive.com - November 13th, 2019
- Does Freedom Have a Future in China? - The Wall Street Journal - November 13th, 2019
- Democracy doesnt matter to the defenders of economic freedom - The Guardian - November 13th, 2019
- Religious-Freedom Voters Will Vote Trump - National Review - November 13th, 2019
- Donald Trump plans to make foreign aid conditional on religious freedom - The Guardian - November 13th, 2019
- Reauthorizing the USA Freedom Act of 2015 FBI - Federal Bureau of Investigation - November 13th, 2019
- The Chinese Government Cannot Be Allowed to Undermine Academic Freedom - The Nation - November 13th, 2019
- 'A story about freedom': artist set to re-enact largest slave revolt in US history - The Guardian - November 13th, 2019
- Just How Far Should the Freedom of Information Act Go? - WVTF - October 16th, 2019
- How we ditched debt: Little splurges on the path to freedom - USA TODAY - October 16th, 2019
- USMNT loss, D.C. United, and Beckhams forest stadium: Freedom Kicks! - Black And Red United - October 16th, 2019
- The OSCE Produces Guidance On Freedom Of Religion Or Belief And Security - Forbes - October 16th, 2019
- Finding the freedom to 'f*ck like an animal' - PGH City Paper - October 16th, 2019
- The first Steering Committee meeting of the Action Freedom of Expression and Freedom of the Media in North Macedonia (JUFREX 2) held in Skopje -... - October 16th, 2019
- Freedom at a price to be paid by others - Financial Times - October 16th, 2019
- Cyrus Wilson ready for another chance to be granted his freedom - NewsChannel5.com - October 16th, 2019
- Russia Turns the Screws on Internet Freedom - VOA News - October 16th, 2019
- Tom Steyers Tie: Theyll Never Take His Freedom! - The New York Times - October 16th, 2019
- Charting Americas path to freedom on a road trip through the Deep South - Irish Examiner - October 16th, 2019
- As the World Marches for Freedom, Where is Trump? - The American Interest - October 16th, 2019
- Rembrandt Painted The Best Portrait Of Freedom Ever, And Here It Is - The Federalist - October 16th, 2019
- Crawling to freedom? Boris Johnson compares Brexit to tunnel escape from The Shawshank Redemption - RT - October 16th, 2019
- What Ed Meeses Presidential Medal of Freedom Says About the G.O.P. and Impeachment - The New Yorker - October 16th, 2019
- Sneaking to freedom from East Berlin to West in a modified BMW Isetta - CNET - October 16th, 2019
- The Temporary Freedom of Leaving Prison to Attend a Funeral - The Marshall Project - October 16th, 2019
- Land development body should be covered by Freedom of Information ombudsman - The Irish Times - October 16th, 2019
- The Jolt: More than the freedom to do as youre told - Atlanta Journal Constitution - October 16th, 2019
- Deck by deck changes coming to Freedom of the Seas - Royal Caribbean Blog - October 16th, 2019
- Watch SpaceX Test Its Crew Dragon’s Escape System - September 14th, 2019
- Airbus Planes Will Track How Often Passengers Go to the Bathroom - September 14th, 2019
- Watch a Robot Fish Fly by Shooting Water out of Its Rear - September 14th, 2019
- Residents Are Trying to Flee Russian Town Where Snow Turned Black - September 14th, 2019
- This Startup Will Send DNA From Your Spit to the Moon for $99 - September 14th, 2019
- FBI Investigation Targets Trump Booster Peter Thiel’s VC Firm - September 14th, 2019
- YouTube Influencers Are Mentally Collapsing From Stress - September 14th, 2019
- MIT Team “Accidentally” Invents Blackest Material in Existence - September 14th, 2019
- A Driver Tricked Uber’s Algorithm, Sexually Assaulted a Passenger - September 14th, 2019
- Toyota Wants to Slather Solar Panels All Over Its Prius Hybrid - September 14th, 2019
- A Runaway Star Is Escaping a Black Hole at 1.2 Million MPH - September 14th, 2019
- Putin Critic Uses Drone to Save Hard Drives Before Police Raid - September 14th, 2019
- French Gov Official Warns Facebook: Libra Is Not Welcome Here - September 14th, 2019
- Watch a Tesla Model X Blast Through Deep Flood Waters - September 14th, 2019
- Milky Way’s Giant Black Hole Lets out Two Giant, Radioactive Burps - September 14th, 2019
- Instagram Keeps Accidentally Flagging Fish Photos as Offensive - September 14th, 2019
- MIT Community Horrified by Famed Researcher’s Epstein Outburst - September 14th, 2019
- Unlike MIT, Harvard Cut Off Epstein Donations After Conviction - September 14th, 2019
- Horrifying Study: Corpses Thrash Around For a Year After Death - September 14th, 2019
- atheism | Definition, Philosophy, & Comparison to ... - May 28th, 2019
- Atheism - Simple English Wikipedia, the free encyclopedia - May 28th, 2019
- atheism r/atheism - reddit: the front page of the internet - May 28th, 2019
- Atheism | CARM.org - May 28th, 2019
- Ripple Price Forecast: XRP vs SWIFT, SEC Updates, and More - May 25th, 2019
- Cryptocurrency News: Looking Past the Bithumb Crypto Hack - May 25th, 2019
- Cryptocurrency News: This Week on Bitfinex, Tether, Coinbase, & More - May 25th, 2019
- Cryptocurrency News: Bitcoin ETFs, Andreessen Horowitz, and Contradictions in Crypto - May 25th, 2019
- Cryptocurrency News: What You Need to Know This Week - May 25th, 2019
- Cryptocurrency News: XRP Validators, Malta, and Practical Tokens - May 25th, 2019
- Cryptocurrency News: New Exchanges Could Boost Crypto Liquidity - May 25th, 2019
- Cryptocurrency News: Bitcoin ETF Rejection, AMD Microchip Sales, and Hedge Funds - May 25th, 2019
- Cryptocurrency News: Vitalik Buterin Doesn’t Care About Bitcoin ETFs - May 25th, 2019
- Bitcoin Rise: Is the Recent Bitcoin Price Surge a Sign of Things to Come or Another Misdirection? - May 25th, 2019
- Ripple Price Forecast: XRP vs SWIFT, SEC Updates, and More - April 29th, 2019
- Cryptocurrency News: Looking Past the Bithumb Crypto Hack - April 29th, 2019
- Cryptocurrency News: This Week on Bitfinex, Tether, Coinbase, & More - April 29th, 2019
- Cryptocurrency News: XRP Validators, Malta, and Practical Tokens - April 29th, 2019
- Cryptocurrency News: Bitcoin ETFs, Andreessen Horowitz, and Contradictions in Crypto - April 29th, 2019
- Cryptocurrency News: Bitcoin ETF Rejection, AMD Microchip Sales, and Hedge Funds - April 29th, 2019
- Cryptocurrency News: What You Need to Know This Week - April 29th, 2019