Breaking News

Razer Unveils the Ultra-Lightweight DeathAdder V4 Pro Sony launches a high-resolution shotgun microphone with superior sound quality and compact design. Arctic announces New Liquid Freezer III Pro 280 and Pro 420 Silicon Power Launches Hypera microSDXC Express Card Samsung announces Watch8, Z Fold7 and Z Flip7

logo

  • Share Us
    • Facebook
    • Twitter
  • Home
  • Home
  • News
  • Reviews
  • Essays
  • Forum
  • Legacy
  • About
    • Submit News

    • Contact Us
    • Privacy

    • Promotion
    • Advertise

    • RSS Feed
    • Site Map

Search form

Nvidia Says Intel Is Misleading With Outdated Deep Learning Benchmarks

Nvidia Says Intel Is Misleading With Outdated Deep Learning Benchmarks

Enterprise & IT Aug 18,2016 0

Nvidia, which has been promoting the use of its GPUs for deep learning applications, claims that Intel published some incorrect "facts" about Xeon Phi processors and and benchmarks related to about deep learning. Deep learning has the potential to revolutionize computing, improve the efficiency and intelligence of our business systems, and deliver advancements that will help humanity in profound ways.

Nvidia has been pitching high-performance GPUs with its Tesla products for deep learning, and Intel has just announced at IDF a specialized Xeon Phi chip called Knights Mill optimized for deep learning. Knights Mill is scheduled for release in 2017.

Intel has not shied away from comparing their tech to GPUs and touting why they believe Xeon Phi to be superior.

For example, Knights Mill is capable of acting as a host processor. So expect to see Intel promoting the benefits of not needing separate host processors & co-processors, and how Knights Mill can be attached directly to system RAM. This, along with the performance differences between the GPU architectures and Knights Mill, will be a recurring fight between the two companies both now and next year.

Nvidia used a blog post to outlibe Intel's "Deep Learning benchmark mistakes".

Intel recently published some benchmarks to make three claims about deep learning performance with Knights Landing Xeon Phi processors:

  • Xeon Phi is 2.3x faster in training than GPUs
  • Xeon Phi offers 38% better scaling that GPUs across nodes
  • Xeon Phi delivers strong scaling to 128 nodes while GPUs do not

Accordign to Nvidia, Intel used Caffe AlexNet data that is 18 months old, comparing a system with four Maxwell GPUs to four Xeon Phi servers. "With the more recent implementation of Caffe AlexNet, Intel would have discovered that the same system with four Maxwell GPUs delivers 30% faster training time than four Xeon Phi servers," Nvidia said. "In fact, a system with four Pascal-based NVIDIA TITAN X GPUs trains 90% faster and a single NVIDIA DGX-1 is over 5x faster than four Xeon Phi servers," the company continued.

Intel is comparing Caffe GoogleNet training performance on 32 Xeon Phi servers to 32 servers from Oak Ridge National Laboratory’s Titan supercomputer. Titan uses four-year-old GPUs (Tesla K20X) and an interconnect technology inherited from the prior Jaguar supercomputer. Xeon Phi results were based on recent interconnect technology.

According to Nvidia, using more recent Maxwell GPUs and interconnect, Baidu has shown that their speech training workload scales almost linearly up to 128 GPUs.

"Scalability relies on the interconnect and architectural optimizations in the code as much as the underlying processor. GPUs are delivering great scaling for customers like Baidu," Nvidia claims.

Intel says that 128 Xeon Phi servers deliver 50x faster performance compared with a single Xeon Phi server, while no such scaling data exists for GPUs. But as Nvidia noted above, Baidu already published results showing near-linear scaling up to 128 GPUs.

According to Nvidia:

"For strong-scaling, we believe strong nodes are better than weak nodes. A single strong server with numerous powerful GPUs delivers superior performance than lots of weak nodes, each with one or two sockets of less-capable processors, like Xeon Phi. For example, a single DGX-1 system offers better strong-scaling performance than at least 21 Xeon Phi servers (DGX-1 is 5.3x faster than 4 Xeon Phi servers)"

It is obvious that Nvidia and Intel have chosen different approaches in order to address the demands of deep learning apps.

Regarding the new Intel Knights Mill chip, scheduled for release in 2017, it will be 10nm chip.

Among the features/design tweaks for the new processor, Intel is adding what they are calling "variable precision" support. What that fully entails isn’t clear, but the use of lower precision modes has been a major factor in the development and subsequent high performance of machine learning-focused processors, so it’s likely that this means that Intel is adding FP16 and possibly other lower-precision modes, something the current Knights Landing lacks.

Also on the feature list is improved scale-out performance. It’s not clear right now if this is some kind of fabric/interconnect change. But the ultimate goal is to make clusters of Xeon Phi processors perform better, which is an important factor in bringing down the training time of very large and complex datasets. Meanwhile there are also unspecified memory changes for Knights Mill, with Intel touting the chip’s "flexible, high capacity memory."

Tags: NvidiaIntel
Previous Post
AT&T Changes Cellular Data Plans, Eliminates Overage Fees
Next Post
AMD TrueAudio Next Promises To Bring Realistic Audio to VR

Related Posts

  • An Intel-HP Collaboration Delivers Next-Gen AI PCs

  • New Intel Xeon 6 CPUs to Maximize GPU-Accelerated AI Performance

  • Intel Unveils New GPUs for AI and Workstations at Computex 2025

  • G.SKILL Releases DDR5 Memory Support List for Intel 200S Boost

  • Intel and its partners release BIOS update for Intel 15th Gen to increase performance

  • PNY Announces Support for the New NVIDIA RTX PRO Blackwell Graphics Card Family

  • KIOXIA flash memory and SSD solutions empower AI applications at NVIDIA GTC 2025

  • INNO3D GEFORCE RTX 50 SERIES IS HERE!

Latest News

Razer Unveils the Ultra-Lightweight DeathAdder V4 Pro
PC components

Razer Unveils the Ultra-Lightweight DeathAdder V4 Pro

Sony launches a high-resolution shotgun microphone with superior sound quality and compact design.
Cameras

Sony launches a high-resolution shotgun microphone with superior sound quality and compact design.

Arctic announces New Liquid Freezer III Pro 280 and Pro 420
Cooling Systems

Arctic announces New Liquid Freezer III Pro 280 and Pro 420

Silicon Power Launches Hypera microSDXC Express Card
Cameras

Silicon Power Launches Hypera microSDXC Express Card

Samsung announces Watch8, Z Fold7 and Z Flip7
Smartphones

Samsung announces Watch8, Z Fold7 and Z Flip7

Popular Reviews

be quiet! Light Loop 360mm

be quiet! Light Loop 360mm

be quiet! Dark Mount Keyboard

be quiet! Dark Mount Keyboard

be quiet! Light Mount Keyboard

be quiet! Light Mount Keyboard

Noctua NH-D15 G2

Noctua NH-D15 G2

Soundpeats Pop Clip

Soundpeats Pop Clip

be quiet! Light Base 600 LX

be quiet! Light Base 600 LX

Crucial T705 2TB NVME White

Crucial T705 2TB NVME White

be quiet! Pure Base 501

be quiet! Pure Base 501

Main menu

  • Home
  • News
  • Reviews
  • Essays
  • Forum
  • Legacy
  • About
    • Submit News

    • Contact Us
    • Privacy

    • Promotion
    • Advertise

    • RSS Feed
    • Site Map
  • About
  • Privacy
  • Contact Us
  • Promotional Opportunities @ CdrInfo.com
  • Advertise on out site
  • Submit your News to our site
  • RSS Feed