{"id":214498,"date":"2017-03-09T10:05:20","date_gmt":"2017-03-09T15:05:20","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/microsoft-facebook-build-dualing-open-standard-gpu-servers-for-cloud-top500-news.php"},"modified":"2017-03-09T10:05:20","modified_gmt":"2017-03-09T15:05:20","slug":"microsoft-facebook-build-dualing-open-standard-gpu-servers-for-cloud-top500-news","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/super-computer\/microsoft-facebook-build-dualing-open-standard-gpu-servers-for-cloud-top500-news.php","title":{"rendered":"Microsoft, Facebook Build Dualing Open Standard GPU Servers for Cloud &#8211; TOP500 News"},"content":{"rendered":"<p><p>    It was only a matter of time until someone came up with an Open    Compute Project (OCP) design for a GPU-only accelerator box for    the datacenter. That time has come.  <\/p>\n<p>    In this case though, it was two someones: Microsoft and    Facebook. This week at the Open Compute Summit in Santa Clara,    California, both hyperscalers announced different OCP designs    for putting eight of NVIDIAs Tesla P100 GPUs into a single    chassis. Both fill the role of a GPU expansion box that can be    paired with CPU-based servers in need of compute acceleration.    The idea is to disaggregate the GPUs and CPUs in cloud    environments so that users may flexibly mix these processors in    different ratios, depending upon the demands of the particular    workload.  <\/p>\n<p>    The principle application target is machine learning, one of    the P100s major areas of expertise. An eight-GPU configuration    of these devices will yield over 80 teraflops at single    precision and over 160 teraflops at half precision.  <\/p>\n<\/p>\n<p>    Source: Microsoft  <\/p>\n<\/p>\n<p>    Microsofts OCP contribution is known as     HGX-1. Its principle innovation is that it can dynamically    serve up as many GPUs to a CPU-based host as it may need     well, up to eight, at least. It does this via four PCIe    switches, an internal NVLink mesh network, plus a fabric    manager to route the data through the appropriate connections.    Up to four of the HGX-1 expansion boxes can be glued together    for a total of 32 GPUs. Ingrasys, a Foxconn subsidiary will be    the initial manufacturer of the HGX-1 chassis.  <\/p>\n<p>    The Facebook version, which is called     Big Basin, looks quite similar. Again, P100 devices are    glued together vial an internal mesh, which they describe as    similar to the design of the DGX-1, NVIDIAs in-house server    designed for AI research. A CPU server can be connected to the    Big Basin chassis via one or more PCIe cable. Quanta Cloud    Technology will initially manufacture the Big Basin servers.  <\/p>\n<\/p>\n<p>    Source: Facebook  <\/p>\n<\/p>\n<p>    Facebook said they were able to achieve a 100 percent    performance improvement on ResNet50, an image classification    model, using Big Basin, compared to its older Big Sur server,    which uses the Maxwell-generation Tesla M40 GPUs. Besides image    classification, Facebook will use the new boxes for other sorts    deep learning training, such as text translation, speech    recognition, and video classification, to name a few.  <\/p>\n<p>    In Microsofts case, the HGX-1 appears to be the first of    multiple OCP designs that will fall under its Project Olympus    initiative, which the company     unveiled last October. Essentially, Project Olympus is a    related set of OCP hardware building blocks for cloud hardware.    Although HGX-1 is suitable for many compute-intensive    workloads, Microsoft is promoting it for artificial    intelligence work, calling it the Project Olympus hyperscale    GPU accelerator chassis for AI, according to a     blog posted by Azure Hardware Infrastructure GM Kushagra    Vaid.  <\/p>\n<p>    Vaid also set the stage for what will probably become other    Project Olympus OCP designs, hinting at future platforms that    will include the upcoming Intel Skylake Xeon and AMD Naples    processors. He also left open the possibility that Intel FPGAs    or Nervana accelerators could work their way into some of these    designs.  <\/p>\n<p>    In addition, Vail brought up the possibility of a ARM-based OCP    server via the companys engagement with chipmaker Cavium. The    software maker has already announced its using Qualcomms new    ARM chip, the Centriq 2400, in Azure instances. Clearly,    Microsoft is keeping its cloud options open.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>See more here:<\/p>\n<p><a target=\"_blank\" href=\"https:\/\/www.top500.org\/news\/microsoft-facebook-build-dualing-open-standard-gpu-servers-for-cloud\/\" title=\"Microsoft, Facebook Build Dualing Open Standard GPU Servers for Cloud - TOP500 News\">Microsoft, Facebook Build Dualing Open Standard GPU Servers for Cloud - TOP500 News<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> It was only a matter of time until someone came up with an Open Compute Project (OCP) design for a GPU-only accelerator box for the datacenter. That time has come. In this case though, it was two someones: Microsoft and Facebook.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/super-computer\/microsoft-facebook-build-dualing-open-standard-gpu-servers-for-cloud-top500-news.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[41],"tags":[],"class_list":["post-214498","post","type-post","status-publish","format-standard","hentry","category-super-computer"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/214498"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=214498"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/214498\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=214498"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=214498"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=214498"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}