{"id":217844,"date":"2017-06-08T23:31:45","date_gmt":"2017-06-09T03:31:45","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/deepmind-shows-ai-has-trouble-seeing-homer-simpsons-actions-ieee-spectrum.php"},"modified":"2022-11-13T02:07:31","modified_gmt":"2022-11-13T07:07:31","slug":"deepmind-shows-ai-has-trouble-seeing-homer-simpsons-actions-ieee-spectrum","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/deepmind-shows-ai-has-trouble-seeing-homer-simpsons-actions-ieee-spectrum.php","title":{"rendered":"DeepMind Shows AI Has Trouble Seeing Homer Simpson&#8217;s Actions &#8211; IEEE Spectrum"},"content":{"rendered":"<p><p>    The best artificial intelligence still has trouble visually    recognizing many of Homer Simpsons favorite behaviors such as    drinking beer, eating chips, eating doughnuts, yawning, and the    occasional face-plant. Those findings from DeepMind, the    pioneering London-based AI lab, also suggest the motive behind    why DeepMind has created a huge new dataset of YouTube clips to    help train AI on identifying human actions in videos that go    well beyond Mmm, doughnuts or Doh!  <\/p>\n<p>    The most popular AI used by Google, Facebook, Amazon, and other    companies beyond Silicon Valley is based on deep learning    algorithms that can learn to identify patterns in huge amounts    of data. Over time, such algorithms can become much better at a    wide variety of tasks such as     translating between English and Chinese for Google    Translateor automatically     recognizing the faces of friends in Facebook    photos.But even the most finely tuned deep learning    relies on having lots of quality data to learn    from.To help improve AIscapability to    recognizehuman actions in motion,DeepMind has    unveiled itsKinetics dataset consisting of 300,000 video    clips and 400 human action classes.  <\/p>\n<p>    AI systems are now very good at recognizing objects in images,    but still have trouble making sense of videos, says    aDeepMind spokesperson.One of the main reasons for    this is that the research community has so far lacked a large,    high-quality video dataset.  <\/p>\n<p>    DeepMind enlisted the help of online workers through Amazons    Mechanical Turk service to help correctly identify and label    the actions inthousands of YouTube clips. Each of the 400    human action classes in the Kinetics dataset has at least 400    video clips, with each clip lasting around 10 seconds and taken    from separate YouTube videos. More details can be found in a    DeepMind    paper on the arXiv preprint server.  <\/p>\n<p>    The new Kinetics dataset seems likely to represent a new    benchmark for training datasets intended to improve AI computer    vision for video. It has far more video clips and action    classes than the HMDB-51 and UCF-101 datasets that previously    formed the benchmarks for the research community. DeepMind also    made a point of ensuring it had a diverse datasetone that did    not include multiple clips from the same YouTube videos.  <\/p>\n<p>    Tech giants such as Googlea sister company to DeepMind under    the umbrella Alphabet grouparguably have the best access to    large amounts of video data that could prove helpful in    training AI. Alphabets ownership of YouTube, the incredibly    popular, online, video-streaming service, does not hurt either.    But other companies and independent research groups must rely    on publicly available datasets to train their deep learning    algorithms.  <\/p>\n<p>    Early training and testing with the Kinetics dataset showed    some intriguing results. For example, deep learning algorithms    showed accuracies of 80percent or greater in classifying    actions such as playing tennis, crawling baby, presenting    weather forecast, cutting watermelon, and bowling. But the    classification accuracy dropped to around 20 percent or less    for the Homer Simpson actions, including slapping and    headbutting, and an assortment of other actions such as making    a cake, tossing coin and fixing hair.  <\/p>\n<p>    AI faces special challenges with classifying actions such as    eating because it may not be able to accurately identify the    specific food being consumedespecially if the hot dog or    burger is already partially consumed or appears very small    within the overall video. Dancing classes and actions focused    on a specific part of the body can also prove tricky. Some    actions also occur fairly quickly and are only visible for a    small number of frames within a video clip, according to a    DeepMind spokesperson.  <\/p>\n<p>    DeepMind also wanted to see if the new Kinetics dataset has    enough gender balance to allow for accurate AI training. Past    cases have shown how imbalanced training datasets can lead to    deep learning algorithms performing worse at recognizing the    faces of certain ethnic groups. Researchers have also shown how    such algorithms can pick up     gender and racial biases from language.  <\/p>\n<p>    A preliminary study showed that the new Kinetics dataset seems    to fairly balanced. DeepMind researchers found that no single    gender dominated within 340 out of the 400 action classesor    else it was not possible to determine gender in those actions.    Those action classes that did end up gender imbalanced included    YouTube clips of actionssuch as shaving beard or    dunking basketball (mostly male) and filling eyebrows or    cheerleading (mostly female).  <\/p>\n<p>    But even action classes that had gender imbalance did not show    much evidence of classifier bias. This means that even the    Kinetics action classes featuring mostly male participantssuch    as playing poker or hammer throwdid not seem to bias AI to    the point where the deep learning algorithms had trouble    recognizing female participants performing the same actions.  <\/p>\n<p>    DeepMind hopes that outside researchers can help suggest new    human action classes for the Kinetics dataset. Any improvements    may enable AI trained on Kinetics to better recognize both the    most elegant of actions and the clumsier moments in videos that    lead people to say doh! In turn, that could lead to new    generations of computer software and robots with the capacity    to recognize what all those crazy humans are doing on YouTube    or in other video clips.  <\/p>\n<p>    Video understanding represents a significant challenge for the    research community, and we are in the very early stages with    this, according to the DeepMind spokesperson. Any real-world    applications are still a really long way off, but you can see    potential in areas such as medicine, for example, aiding the    diagnosis of heart problems in echocardiograms.  <\/p>\n<p>      IEEE Spectrums general technology blog, featuring      news, analysis, and opinions about engineering, consumer      electronics, and technology and society, from the editorial      staff and freelance contributors.    <\/p>\n<p>      Sign up for the Tech Alert newsletter and receive      ground-breaking technology and science news from IEEE      Spectrum every Thursday.    <\/p>\n<\/p>\n<p>    A deep learning approach could make self-driving cars better at    adapting to new situations 26Apr2016  <\/p>\n<\/p>\n<p>    A tech startup aims to spread the wealth of deep learning AI to    many industries 3Mar2016  <\/p>\n<\/p>\n<p>    Google engineers balanced speed and accuracy to deploy deep    learning in Chinese-to-English translations 3Oct2016  <\/p>\n<\/p>\n<p>    If machine learning systems can be taught using simulated data    from Grand Theft Auto V instead of data annotated by humans, we    could get to reliable vehicle autonomy much faster 8Jun  <\/p>\n<\/p>\n<p>    Adversarial grasping helps robots learn better ways of picking    up and holding onto objects 5Jun  <\/p>\n<\/p>\n<p>    Reverse engineering 1 cubic millimeter of brain tissue could    lead to better artificial neural networks 30May  <\/p>\n<\/p>\n<p>    The FDA needs computer experts with industry experience to help    oversee AI-driven health apps and wearables software    29May  <\/p>\n<\/p>\n<p>    The prototype chip learns a style of music, then composes its    own tunes 23May  <\/p>\n<\/p>\n<p>    Crashing into objects has taught this drone to fly    autonomously, by learning what not to do 10May  <\/p>\n<\/p>\n<p>    Silicon Valley startup Verdigris cloud-based analysis can tell    whether youre using a Chromebook or a Mac, or whether a motor    is running fine or starting to fail 3May  <\/p>\n<\/p>\n<p>    An artificial intelligence program correctly identifies 355    more patients who developed cardiovascular disease 1May  <\/p>\n<\/p>\n<p>    MITs WiGait wall sensor can unobtrusively monitor people for    many health conditions based on their walking patterns    1May  <\/p>\n<\/p>\n<p>    Facebook's Yael Maguire talks about millimeter wave networks,    Aquila, and flying tethered antennas at the F8 developer    conference 19Apr  <\/p>\n<\/p>\n<p>    Machine learning uses data from smartphones and wearables to    identify signs of relationship conflicts 18Apr  <\/p>\n<\/p>\n<p>    Machine-learning algorithms that readily pick up cultural    biases may pose ethical problems 13Apr  <\/p>\n<\/p>\n<p>    AI and robots have to work in a way that is beneficial to    people beyond reaching functional goals and addressing    technical problems 29Mar  <\/p>\n<\/p>\n<p>    Understanding when they don't understand will help make robots    more useful 15Mar  <\/p>\n<\/p>\n<p>    Palo Alto startup twoXAR partners with Santen Pharmaceutical to    identify new glaucoma drugs; efforts on rare skin disease,    liver cancer, atherosclerosis, and diabetic nephropathy also    under way 13Mar  <\/p>\n<\/p>\n<p>    And they have a new piece of hardwarethe Jetson TX2that they    hope everyone will use for this edge processing 8Mar  <\/p>\n<\/p>\n<p>    A deep-learning AI has beaten human poker pros with the    hardware equivalent of a gaming laptop 2Mar  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>View post:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow noopener\" href=\"http:\/\/spectrum.ieee.org\/tech-talk\/robotics\/artificial-intelligence\/deepmind-shows-ai-has-trouble-seeing-homer-simpson-actions\" title=\"DeepMind Shows AI Has Trouble Seeing Homer Simpson's Actions - IEEE Spectrum\">DeepMind Shows AI Has Trouble Seeing Homer Simpson's Actions - IEEE Spectrum<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> The best artificial intelligence still has trouble visually recognizing many of Homer Simpsons favorite behaviors such as drinking beer, eating chips, eating doughnuts, yawning, and the occasional face-plant. Those findings from DeepMind, the pioneering London-based AI lab, also suggest the motive behind why DeepMind has created a huge new dataset of YouTube clips to help train AI on identifying human actions in videos that go well beyond Mmm, doughnuts or Doh! The most popular AI used by Google, Facebook, Amazon, and other companies beyond Silicon Valley is based on deep learning algorithms that can learn to identify patterns in huge amounts of data <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/artificial-intelligence\/deepmind-shows-ai-has-trouble-seeing-homer-simpsons-actions-ieee-spectrum.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[13],"tags":[],"class_list":["post-217844","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence"],"modified_by":"Danzig","_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/217844"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=217844"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/217844\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=217844"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=217844"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=217844"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}