{"id":233865,"date":"2017-08-10T13:33:49","date_gmt":"2017-08-10T17:33:49","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/deepmind-ai-teaches-itself-about-the-world-by-watching-videos-new-scientist.php"},"modified":"2022-12-29T02:04:36","modified_gmt":"2022-12-29T07:04:36","slug":"deepmind-ai-teaches-itself-about-the-world-by-watching-videos-new-scientist","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/deepmind-ai-teaches-itself-about-the-world-by-watching-videos-new-scientist.php","title":{"rendered":"DeepMind AI teaches itself about the world by watching videos &#8211; New Scientist"},"content":{"rendered":"<p><p>Lions roar: video clips beat labels for AIs seeking knowledge    <\/p>\n<p>      OE KLAMAR\/AFP\/Gettyvide    <\/p>\n<p>    By Matt Reynolds  <\/p>\n<p>    To an untrained AI, the world is a blur of confusing data    streams. Most humans have no problem making sense of the sights    and sounds around them, but algorithms tend only to acquire    this skill if those sights and sounds are explicitly labelled    for them.  <\/p>\n<p>    Now DeepMind has developed    an AI that teaches itself to recognise a range of visual and    audio concepts just by watching tiny snippets of video. This AI    can grasp the concept of lawn mowing or tickling, for example,    but it hasnt been taught the words to describe what its    hearing or seeing.  <\/p>\n<p>    We want to build machines that continuously learn about their    environment in an autonomous manner, says Pulkit    Agrawal at the University of California, Berkeley. Agrawal,    who wasnt involved with the work, says this project takes us    closer to the goal of creating AI that can teach itself by    watching and listening to the world around it.  <\/p>\n<p>    Most     computer vision algorithms need to be fed lots of labelled    images so it can tell different objects apart. Show an    algorithm thousands of cat photos labelled cat and soon    enough itll learn to recognise cats even in images it hasnt    seen before.  <\/p>\n<p>    But this way of teaching algorithms  called supervised    learning  is cheating, says Relja Arandjelovi who led the project    at DeepMind. Instead of relying on human-labelled datasets, his    algorithm learns to recognise images and sounds by matching up    what it sees with what it hears.  <\/p>\n<p>    Humans are particularly good at this kind of learning , says    Paolo    Favaro at the University of Bern in Switzerland. We dont    have somebody following us around and telling us what    everything is, he says.  <\/p>\n<p>    Arandjelovi created his    algorithm by starting with two networks  one that specialised    in recognising images and another that did a similar job with    audio. He showed the image recognition network stills taken    from short videos while the audio recognition network was    trained on 1-second audio clips taken from the same point in    each video.  <\/p>\n<p>    A third network compared still images with audio clips to learn    which sounds corresponded with which sights in the videos. In    all, the system was trained on 60 million still-audio pairs    taken from 400,000 videos.  <\/p>\n<p>    The algorithm learned to recognise audio and visual concepts,    including crowds, tap dancing and water, without ever seeing a    specific label for a single concept. When shown a photo of    someone clapping, for example, most of the time it knew which    sound was associated with that image.  <\/p>\n<p>    This kind of co-learning approach could be extended to include    senses other than sight and hearing, says Agarwal. Learning    visual and touch features simultaneously can, for example,    enable the agent to search for objects in the dark and learn    about material properties such as friction, he says.  <\/p>\n<p>    DeepMind will present the study at the International Conference on    Computer Vision which takes place in Venice, Italy, in late    October.  <\/p>\n<p>    While the AI in the DeepMind project doesnt interact with the    real world, Agarwal says that perfecting self-supervised    learning will eventually let us create AI that can operate in    the real world and learn from what it sees and hears.  <\/p>\n<p>    But until we reach that point, self-supervised learning might    be a good way of training image and audio recognition    algorithms without input from vast amounts of human-labelled    data. The DeepMind algorithm can correctly categorise an audio    clip nearly 80 per cent of the time, making it better at    audio-recognition than many algorithms trained on labelled    data.  <\/p>\n<p>    Such promising results suggest that similar algorithms might be    able to learn something by crunching through huge unlabelled    datasets like YouTubes millions of online videos. Most of the    data in the world is unlabelled and therefore it makes sense to    develop systems that can learn from unlabelled data, Agrawal    says.  <\/p>\n<p>    Journal reference: arxiv.org  <\/p>\n<p>    Read more:     Curious AI learns by exploring game worlds and making    mistakes  <\/p>\n<p>    More on these topics:  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>More here:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"https:\/\/www.newscientist.com\/article\/2143498\/\" title=\"DeepMind AI teaches itself about the world by watching videos - New Scientist\">DeepMind AI teaches itself about the world by watching videos - New Scientist<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Lions roar: video clips beat labels for AIs seeking knowledge OE KLAMAR\/AFP\/Gettyvide By Matt Reynolds To an untrained AI, the world is a blur of confusing data streams.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/deepmind-ai-teaches-itself-about-the-world-by-watching-videos-new-scientist.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-233865","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":"Danzig","_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/233865"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=233865"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/233865\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=233865"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=233865"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=233865"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}