{"id":218685,"date":"2017-06-11T16:23:08","date_gmt":"2017-06-11T20:23:08","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/why-are-ai-predictions-so-terrible-venturebeat.php"},"modified":"2022-06-26T21:46:01","modified_gmt":"2022-06-27T01:46:01","slug":"why-are-ai-predictions-so-terrible-venturebeat","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/why-are-ai-predictions-so-terrible-venturebeat.php","title":{"rendered":"Why are AI predictions so terrible? &#8211; VentureBeat"},"content":{"rendered":"<p><p>    In 1997, IBMs Deep Blue beat world chess    champion Gary Kasparov, the first time an AI technology was    able to outperform a world expert in a highly complicated    endeavor. It was even more impressive when you    considerthey were using 1997 computational power. In    1997, my computer could barely connect to the internet; long    waits of agonizing beeps and buzzes made it clear the computer    was struggling under the weight of the task.  <\/p>\n<p>    Even in the wake of Deep Blues literally game-changing    victory, most experts remained unconvinced. Piet Hut, an    astrophysicist at the Institute for Advanced Study in New    Jersey, told the NY Times in 1997 that it would still be    another hundred years before a computer beats a human at Go.  <\/p>\n<p>    Admittedly, the ancient game of Go is infinitely more    complicated than chess. Even in 2014, the common consensus was    that an AI victory in Go was still decades away. The world    reigning champion, Lee Sedol, gloated in an article for    Wired, There is chess in the western world, but Go is    incomparably more subtle and intellectual.  <\/p>\n<p>    Then AlphaGo, Googles AI platform, defeated him a mere    two years later. Hows that for subtlety?  <\/p>\n<p>    In recent years, it is becoming increasingly well known that AI    is able to outperform humans in much more than board games.    This has led to a growing anxiety among the working public that    their very livelihood may soon be automated.  <\/p>\n<p>    Countless publications have been quick to seize on this fear to    drive pageviews. It seems like every day there is a new article    claiming to know definitively which jobs will survive the AI    revolution and which will not. Some even go so far to express    their percentage predictions down to the decimal point    giving the whole activity a sense of gravitas. However,    if you compare their conclusions, the most striking aspect is    how wildly inconsistent the results are.  <\/p>\n<p>    One of the latest entries into the mire is a Facebook quiz    aptly named Will Robots take My    Job?. Naturally, I looked up writers and I received back    a comforting 3.8%. After all, if a doctor told me I had a 3.8%    chance of succumbing to a disease, I would hardly be in a hurry    to get my affairs in order.  <\/p>\n<p>    There is just one thing keeping me from patting myself on the    back: AI writers already exist and are being widely used by    major publications. In this way, their prediction would be like    a doctor declaring there was only a 3.8% chance of my disease    getting worseat my funeral.  <\/p>\n<p>    All this begs the question: why are these predictions about AI    so bad?  <\/p>\n<p>    Digging into the sources from Will Robots take My Job gives    us our first clue. The predictions are based on a research    paper. This is at the root of most bad AI predictions.    Academics tend to view the world very differently from Silicon    Valley entrepreneurs. Where in academia just getting a project    approved may take years, tech entrepreneurs operate on the idea    of what can we get built and shipped by Friday?    Therefore, asking academics for predictions on the    proliferation of industry is like asking your local DMV about    how quickly Uber may be able to gain market share in China.    They may be experts in the vertical, but they are still worlds    away from the move fast and break stuff mentality that    pervades the tech community.  <\/p>\n<p>    As a result, their predictions are as good as random guesses,    colored by their understanding of a world that moves at a    glacial pace.  <\/p>\n<p>    Another contributing factor to bad AI predictions is human    bias. When the question is between who will win, man or    machine,we cant help but to root for the home team. It    has been said, that it is very hard to make someone believe    something when their job is dependent on them not understanding    it. Meaning the banter around the water-cooler at oil companies    rarely turns to concerns about climate change. AI poses a    threat to the very notion of human based jobs, so the stakes    are much higher. When you ask people who work for a university    the likelihood of AI automating all jobs, it is all but    impossible for them to be objective.  <\/p>\n<p>    Hence the conservative estimations to admit that any job    that can be taught to a person can obviously also be taught to    an AI would fill the researcher with existential dread. Better    to sidestep the whole issue and say that it wont happen for    another 50 years, hoping theyll be dead by then and it will be    the next guys problem.  <\/p>\n<p>    Which brings us to our final contributing factor, that humans    are really bad at understanding exponential growth. The    research paper that Will Robots Take My Job was from 2013.    The last four years in AI might well have been 40 years based    on how much has changed. In fact, their bad predictions make    more sense through this lens. There is an obvious bias for    assuming jobs that require decision making as more safe than    those that are straight routine. However, the proliferation of    neural net resources are showing that AI is actually very good    at decision making, when the task is well defined.  <\/p>\n<p>    The problem is our somewhat primitive reasoning tends to view    the world in linear reasoning. Take this example often used on    logic tests. If the number of lily pads on a lake double every    day, and the lake will be full at 30 days, how many days will    it take for the lake to be half full? A depressingly high    number of peoples knee jerk response would be 15. The real    answer is 29. In fact, if you were viewing the pond the lily    pads wouldnt appear to be growing at all until about the 26th    day. If you were to ask the average person on day 25 how many    days until the pond was full they might rightfully conclude    decades.  <\/p>\n<p>    The reality is AI tools are growing exponentially. Even in    their current iteration, they have the power to automate at    least part of all human jobs. The uncomforting truth that all    these AI predictions seek to distract us from is that no job is    safe from automation. Collectively we are like Lee Sedol in    2014, smug in our sense of superiority. The coming    proliferation of AI is perhaps best summed up in the sentiments    of Nelson Mandela: It always seems impossible until is it    done.  <\/p>\n<p>    Aiden Livingston is the founder of Casting.AI, the first chatbot talent    agent.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>See more here:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/venturebeat.com\/2017\/06\/10\/why-are-ai-predictions-so-terrible\/\" title=\"Why are AI predictions so terrible? - VentureBeat\">Why are AI predictions so terrible? - VentureBeat<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> In 1997, IBMs Deep Blue beat world chess champion Gary Kasparov, the first time an AI technology was able to outperform a world expert in a highly complicated endeavor. It was even more impressive when you considerthey were using 1997 computational power.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/why-are-ai-predictions-so-terrible-venturebeat.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-218685","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":"Danzig","_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/218685"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=218685"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/218685\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=218685"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=218685"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=218685"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}