{"id":191981,"date":"2017-05-09T15:31:36","date_gmt":"2017-05-09T19:31:36","guid":{"rendered":"http:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/how-facebooks-ai-ambitions-will-boost-nvidia-motley-fool\/"},"modified":"2017-05-09T15:31:36","modified_gmt":"2017-05-09T19:31:36","slug":"how-facebooks-ai-ambitions-will-boost-nvidia-motley-fool","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/ai\/how-facebooks-ai-ambitions-will-boost-nvidia-motley-fool\/","title":{"rendered":"How Facebook&#8217;s AI Ambitions Will Boost NVIDIA &#8211; Motley Fool"},"content":{"rendered":"<p><p>    Facebook    (NASDAQ:FB)    has been doubling down on artificial intelligence    (AI) to process the large amount of content users post to its    platform, trying to make sense of all of that data to make    communication easier. For instance, AI helps the social media    specialist classify live videos in real time, while also    helping in speech and text translations.  <\/p>\n<p>    One of the ways to take advantage of Facebook's increasing AI    adoption is through NVIDIA (NASDAQ:NVDA),    as its graphics processing units (GPUs) are playing a    mission-critical role in the fast processing of huge data sets.  <\/p>\n<p>    Back in March, Facebook announced that it is using NVIDIA's    GPUs to power its next-generation GPU server -- Big Basin -- so    it can train bigger machine learning models for faster    processing of photos, text, and videos. NVIDIA supplied eight    of its Tesla P100 GPU accelerators for the server, along with    its high-speed NVLink technology that enables ultra-fast    communication between the GPUs by removing any    connection-related bottlenecks.  <\/p>\n<p>    NVIDIA's Tesla GPUs and the NVLink interconnect technology are    allowing Facebook to train 30% larger AI models, thanks to a    33% jump in the bandwidth memory as compared to the previous    generation -- Big Sur processor. As it turns out, Big Basin can    perform 100% faster than Big Sur in certain scenarios,    processing more complex models in a shorter time frame.  <\/p>\n<p>      Image source: NVIDIA.    <\/p>\n<p>    What's more, NVIDIA and Facebook have now taken their AI    relationship further with the Caffe2, a scalable deep learning    AI framework that gives developers more power in training and    iterating AI models. Caffe2 connects eight of Facebook's Big    Basin servers, giving users the capability of using 64 NVIDIA    Tesla GPU accelerators and allowing them to train AI models    seven times faster with the help of a supercomputer.  <\/p>\n<p>    Facebook will need more high-performance servers going forward    thanks to booming mobile data traffic and a huge user base. The    social media specialist has 1.23 billion daily active users who    post 300 million photos a day and 510,000 comments each second.    What's more, the company is betting big on video, and its    \"Live\" service has seen a 400% surge in streaming since launch.  <\/p>\n<p>    Facebook's growth is not going to stop anytime soon as its    emerging markets user base is growing at a terrific pace.    Research firm eMarketer forecasts that countries such as India,    Indonesia, Mexico, and the Philippines will become its    fastest-growing markets until 2020, leading to a spurt in    content posted onto the platform, especially due to growing    smartphone penetration.  <\/p>\n<p>    Facebook, therefore, will need more capable servers to tackle    the growing data volume and complexity. This is good news for    NVIDIA's professional visualization business, which houses the    Tesla GPU unit. The Tesla GPUs are aimed at accelerating    high-performance computing and hyperscale data center workloads    -- allowing them to crunch huge amounts of data at a fast pace    -- so Facebook is going to need more of them as its workload    grows.  <\/p>\n<p>    As the likes of Facebook and others start using AI to train    their analytics models, NVIDIA will find a bigger market to    sell its GPU accelerators. Markets and Markets forecasts that    the AI chipset market will grow at over 60% a year until 2022,    hitting a size of $16 billion. GPU accelerators could make up a    big part of this market thanks to the crucial role they play in    the AI space.  <\/p>\n<p>    This should supercharge NVIDIA's professional visualization    business, which is already reaping the benefits of growing data center    workloads. In fact, the Tesla GPUs are being used by cloud    service providers such as Amazon Web Services,    Google Cloud, and Microsoft    Azure, and Facebook will further boost the segment's growth    thanks to its growing AI bets.  <\/p>\n<p>    Suzanne Frey, an executive    at Alphabet, is a member of The Motley Fools board of    directors. Teresa Kersten is an employee of LinkedIn and is a    member of The Motley Fools board of directors. LinkedIn is    owned by Microsoft. Harsh    Chauhan has no position in any stocks mentioned. The Motley    Fool owns shares of and recommends Alphabet (A shares), Amazon,    Facebook, and Nvidia. The Motley Fool has a disclosure    policy.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Follow this link:<\/p>\n<p><a target=\"_blank\" rel=\"nofollow\" href=\"https:\/\/www.fool.com\/investing\/2017\/05\/08\/how-facebooks-ai-ambitions-will-boost-nvidia.aspx\" title=\"How Facebook's AI Ambitions Will Boost NVIDIA - Motley Fool\">How Facebook's AI Ambitions Will Boost NVIDIA - Motley Fool<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> Facebook (NASDAQ:FB) has been doubling down on artificial intelligence (AI) to process the large amount of content users post to its platform, trying to make sense of all of that data to make communication easier. For instance, AI helps the social media specialist classify live videos in real time, while also helping in speech and text translations <a href=\"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/ai\/how-facebooks-ai-ambitions-will-boost-nvidia-motley-fool\/\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":4,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[187743],"tags":[],"class_list":["post-191981","post","type-post","status-publish","format-standard","hentry","category-ai"],"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/191981"}],"collection":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/users\/4"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/comments?post=191981"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/posts\/191981\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/media?parent=191981"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/categories?post=191981"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/prometheism-transhumanism-posthumanism\/wp-json\/wp\/v2\/tags?post=191981"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}