{"id":229243,"date":"2017-07-21T02:59:04","date_gmt":"2017-07-21T06:59:04","guid":{"rendered":"http:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/uncategorized\/with-epyc-amd-offers-serious-competition-to-intel-in-hpc-top500-news.php"},"modified":"2017-07-21T02:59:04","modified_gmt":"2017-07-21T06:59:04","slug":"with-epyc-amd-offers-serious-competition-to-intel-in-hpc-top500-news","status":"publish","type":"post","link":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/super-computer\/with-epyc-amd-offers-serious-competition-to-intel-in-hpc-top500-news.php","title":{"rendered":"With EPYC, AMD Offers Serious Competition to Intel in HPC &#8211; TOP500 News"},"content":{"rendered":"<p><p>    While a number of commentators have written off AMDs prospects    of competing against Intel in HPC, testing of the latest server    silicon from each chipmaker has revealed that the EPYC chip    offers some surprising performance advantages against Intels    newest \"Skylake\" Xeon destined for the datacenter.  <\/p>\n<\/p>\n<p>    Since Intel integrated the 512-bit Advanced Vector    Extensions (AVX-512) feature into its new Xeon Skylake scalable    processor (SP) platform, it can theoretically double    floating-point performance (and integer performance) compared    its previous Broadwell generation Xeon line. The latter chips    supported vector widths of only 256 bits. With EPYC, AMD    decided to forego any extra-wide vector support, implementing    its floating-point unit with a modest 128-bit capability. That    leaves it with a distinct disadvantage on vector-friendly    codes.  <\/p>\n<p>    However, the majority of HPC codes dont take advantage of    AVX-512 today, since prior to Skylake the only platform that    supported it was Intels Knights Landing Xeon Phi, a    processor specifically designed for vector-intensive software.    Many HPC applications could certainly be enhanced to use the    extra-wide vectors, although for others, like sparse matrix    codes, it may not be worth the trouble. In any case, adding    AVX-512 support to the code base will be done one application    at a time.  <\/p>\n<p>    Without the performance boost from extra-wide vector    instructions, the theoretical floating-point advantage of the    new Xeon over the AMD EPYC processor disappears. At least that    is what can be concluded from     testing done by the gang over at Anandtech. They recently    ran a series of floating-point-intensive tests, among other,    pitting the EPYC 7601 (32 cores, 2.2 GHz) against the    comparable Xeon Platinum 8176 (28 cores, 2.1 GHz). Both are    considered high bin server chips from their respective product    lines.  <\/p>\n<p>    The testing comprised benchmarks based on three real-world    codes: C-ray, a ray-tracing code that runs out of L1 cache;    POV-Ray, a ray-tracing code that runs out of L2 cache; and    NAMD, a molecular dynamics code that requires consistent use of    main memory. The tests were performed on dual-socket servers    running Ubuntu Linux.  <\/p>\n<p>    Somewhat surprising, the EPYC processor outran the Xeon in all    three floating-point benchmarks. For C-ray, the 7601 delivered    about 50 percent more renders than the 8176 in a given amount    of time, while for POV-Ray, the 7601 scored a more modest 16    percent performance advantage. For NAMD, Anandtech used two    implementations, a non-AVX version and an AVX-version that uses    Intels compiler vectorization smarts (but not specifically for    AVX-512). In both cases, the EPYC processor prevailed  by 41    percent, with the older implementation, and by 22 percent with    AVX turned on. Anandtechs conclusion was that even though the    Zen FP might not have the highest peak FLOPS in theory, there    is lots of FP code out there that runs best on EPYC.  <\/p>\n<p>    Its worth noting that in Anandtech also performed a Big Data    benchmark, in which the Xeon edged the EPYC by a little less    than 5 percent. In this case, the test was a collection of five    Spark-based codes, which measured mostly integer performance    and memory accesses. In general, the EPYC processors should do    better on data-demanding codes due to its superior memory    bandwidth, but it was not clear how memory intensive these    particular codes were. It would be interesting to see how these    two architectures match up on in-memory database benchmarks.  <\/p>\n<p>    Execution speed aside, AMD silicon looks even more attractive    when you consider price-performance. The Xeon 8176 lists for    $8,719, while the EPYC 7601 is priced at $4,000. With the Xeon    line, you could move up to a faster clock (2.5 GHz) with the    top-of-the-line 8180 for around $10,000, or move down to the    Xeon 8160 (same clock, 24 cores) for $4,700. But either way,    AMD looks to be undercutting Intel on price for comparably    performaning server silicon.  <\/p>\n<p>    Of course, if an application can take full advantage of    AVX-512, the performance advantage would shift to Intel.    (Perhaps not a price-performance advantage though.) One other    thing to consider is for AVX-512-friendly codes, the Xeon Phi    itself offers the best performance and price, not to mention    energy efficiency. The only caveat here is threads on the Xeon    Phi execute about 1 GHz slower than on their Xeon counterparts,    so if single-threaded performance is critical to some portion    of your code or codes, youre going to take a pretty    significant performance hit.  <\/p>\n<p>    In     a discussion posted on Facebook earlier this week, Forrest    Norrod, SVP and GM of Enterprise, Embedded & Semi-Custom    Products, said he was pleased with how their new server chip is    positioned against its rival. He made particular mention of the    favorable floating-point performance, noting the results on    EPYC have been tremendous, head-to-head, against the    competitor.  <\/p>\n<p>    He went on to explain that while the EPYC design team    considered implementing a wide vector capability, they felt it    was too expensive in terms of die area and power requirements    to load down the CPU with such a capability. Instead they opted    for a more general-purpose floating-point unit, plumbed with    dedicated FP pipes to improve performance.  <\/p>\n<p>    Also part of the Facebook discussion was AMD Engineering Fellow    Kevin Lepak, who explained that another facet of the decision    to keep the EPYC floating-point unit more generalized was due    to AMDs GPU computing product line, which essentially fulfills    the role of a dedicated vector processor. The company felt it    didnt make much sense to overlap this capability in their CPU    platform as long as they were offering both. As noted earlier,    Intel made the exact opposite decision, vis--vis their Xeon    and Xeon Phi lines.  <\/p>\n<p>    Norrod and Lepak also delved into the rationale for    implementing EPYC as a multi-chip module (MCM) processor,    rather than as a monolithic chip, as Intel has done with its    SkylakeXeons. A 32-core EPYC processor, for    example, is comprised of four eight-core dies glued together    with the Infinity Fabric. Intel has been critical of AMD for    its MCM approach, claiming it hinders performance at various    choke points. AMD counters that its a more effective way    to get its extra-large feature set  eight memory channels, 128    PCIe lanes, built-in encryption, and so on  into the    processor, while also serving to lower costs via increased    manufacturing yields.  <\/p>\n<p>    None of these technical arguments amount to much for customers,    who will be focused on performance, price-performance and    performance-per-watt across their own applications. If AMD can    deliver superior numbers on even two ofthese criteria,    Intel will likely lose its 90 percent-plus market share in    HPCfor the first time in nearly ten years. And that would    be a true EPYC event.  <\/p>\n<p><!-- Auto Generated --><\/p>\n<p>Excerpt from: <\/p>\n<p><a target=\"_blank\" href=\"https:\/\/www.top500.org\/news\/with-epyc-amd-can-offer-serious-competition-to-intel-in-hpc\/\" title=\"With EPYC, AMD Offers Serious Competition to Intel in HPC - TOP500 News\">With EPYC, AMD Offers Serious Competition to Intel in HPC - TOP500 News<\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p> While a number of commentators have written off AMDs prospects of competing against Intel in HPC, testing of the latest server silicon from each chipmaker has revealed that the EPYC chip offers some surprising performance advantages against Intels newest \"Skylake\" Xeon destined for the datacenter. Since Intel integrated the 512-bit Advanced Vector Extensions (AVX-512) feature into its new Xeon Skylake scalable processor (SP) platform, it can theoretically double floating-point performance (and integer performance) compared its previous Broadwell generation Xeon line. The latter chips supported vector widths of only 256 bits.  <a href=\"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/super-computer\/with-epyc-amd-offers-serious-competition-to-intel-in-hpc-top500-news.php\">Continue reading <span class=\"meta-nav\">&rarr;<\/span><\/a><\/p>\n","protected":false},"author":1,"featured_media":0,"comment_status":"closed","ping_status":"closed","sticky":false,"template":"","format":"standard","meta":{"limit_modified_date":"","last_modified_date":"","_lmt_disableupdate":"","_lmt_disable":"","footnotes":""},"categories":[41],"tags":[],"class_list":["post-229243","post","type-post","status-publish","format-standard","hentry","category-super-computer"],"modified_by":null,"_links":{"self":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/229243"}],"collection":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/comments?post=229243"}],"version-history":[{"count":0,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/posts\/229243\/revisions"}],"wp:attachment":[{"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/media?parent=229243"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/categories?post=229243"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.euvolution.com\/futurist-transhuman-news-blog\/wp-json\/wp\/v2\/tags?post=229243"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}