{"id":99463,"date":"2014-01-09T01:41:30","date_gmt":"2014-01-09T06:41:30","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/computer-science-the-learning-machines.php"},"modified":"2014-01-09T01:41:30","modified_gmt":"2014-01-09T06:41:30","slug":"computer-science-the-learning-machines","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/computer-science-the-learning-machines.php","title":{"rendered":"Computer science: The learning machines"},"content":{"rendered":"<p><p>        BRUCE ROLFF\/SHUTTERSTOCK      <\/p>\n<p>    Three years ago, researchers at the secretive Google X lab in    Mountain View, California, extracted some 10 million still    images from YouTube videos and fed them into Google Brain  a    network of 1,000 computers programmed to soak up the world much    as a human toddler does. After three days looking for recurring    patterns, Google Brain decided, all on its own, that there were    certain repeating categories it could identify: human faces,    human bodies and  cats1.  <\/p>\n<p>    Google Brain's discovery that the Internet is full of cat    videos provoked a flurry of jokes from journalists. But it was    also a landmark in the resurgence of deep learning: a    three-decade-old technique in which massive amounts of data and    processing power help computers to crack messy problems that    humans solve almost intuitively, from recognizing faces to    understanding language.  <\/p>\n<p>    Deep learning itself is a revival of an even older idea for    computing: neural networks. These systems, loosely inspired by    the densely interconnected neurons of the brain, mimic human    learning by changing the strength of simulated neural    connections on the basis of experience. Google Brain, with    about 1 million simulated neurons and 1 billion simulated    connections, was ten times larger than any deep neural network    before it. Project founder Andrew Ng, now director of the    Artificial Intelligence Laboratory at Stanford University in    California, has gone on to make deep-learning systems ten times    larger again.  <\/p>\n<p>    Such advances make for exciting times in artificial    intelligence (AI)  the often-frustrating attempt to get    computers to think like humans. In the past few years,    companies such as Google, Apple and IBM have been aggressively    snapping up start-up companies and researchers with    deep-learning expertise. For everyday consumers, the results    include software better able to sort through photos, understand    spoken commands and translate text from foreign languages. For    scientists and industry, deep-learning computers can search for    potential drug candidates, map real neural networks in the    brain or predict the functions of proteins.  <\/p>\n<p>    AI has gone from failure to failure, with bits of progress.    This could be another leapfrog, says Yann LeCun, director of    the Center for Data Science at New York University and a    deep-learning pioneer.  <\/p>\n<p>    Over the next few years we'll see a feeding frenzy. Lots of    people will jump on the deep-learning bandwagon, agrees    Jitendra Malik, who studies computer image recognition at the    University of California, Berkeley. But in the long term, deep    learning may not win the day; some researchers are pursuing    other techniques that show promise. I'm agnostic, says Malik.    Over time people will decide what works best in different    domains.  <\/p>\n<p>    Back in the 1950s, when computers were new, the first    generation of AI researchers eagerly predicted that fully    fledged AI was right around the corner. But that optimism faded    as researchers began to grasp the vast complexity of real-world    knowledge  particularly when it came to perceptual problems    such as what makes a face a human face, rather than a mask or a    monkey face. Hundreds of researchers and graduate students    spent decades hand-coding rules about all the different    features that computers needed to identify objects. Coming up    with features is difficult, time consuming and requires expert    knowledge, says Ng. You have to ask if there's a better    way.  <\/p>\n<p>        IMAGES: ANDREW NG      <\/p>\n<p>    In the 1980s, one better way seemed to be deep learning in    neural networks. These systems promised to learn their own    rules from scratch, and offered the pleasing symmetry of using    brain-inspired mechanics to achieve brain-like function. The    strategy called for simulated neurons to be organized into    several layers. Give such a system a picture and the first    layer of learning will simply notice all the dark and light    pixels. The next layer might realize that some of these pixels    form edges; the next might distinguish between horizontal and    vertical lines. Eventually, a layer might recognize eyes, and    might realize that two eyes are usually present in a human face    (see 'Facial recognition').  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Continued here: <\/p>\n<p><a target=\"_blank\" href=\"http:\/\/www.nature.com\/doifinder\/10.1038\/505146a\" title=\"Computer science: The learning machines\">Computer science: The learning machines<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> BRUCE ROLFF\/SHUTTERSTOCK Three years ago, researchers at the secretive Google X lab in Mountain View, California, extracted some 10 million still images from YouTube videos and fed them into Google Brain a network of 1,000 computers programmed to soak up the world much as a human toddler does.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/computer-science-the-learning-machines.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-99463","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/99463"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=99463"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/99463\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=99463"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=99463"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=99463"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}