Breaking News

Samsung Galaxy S25 Edge Features New Corning Gorilla Glass Ceramic 2 for Enhanced Durability Razer announces Clio Chair Accessory for Audio Immersion Razer Unveils Ergonomic Gaming Mouse and Keyboard for Gaming on the Go Noctua releases NH-D15 G2 specific offset LGA1851 mounting bars for improved cooling performance ADATA Launches T7 and T5 Enterprise SSD Series

logo

  • Share Us
    • Facebook
    • Twitter
  • Home
  • Home
  • News
  • Reviews
  • Essays
  • Forum
  • Legacy
  • About
    • Submit News

    • Contact Us
    • Privacy

    • Promotion
    • Advertise

    • RSS Feed
    • Site Map

Search form

Intel Says CPU Outperforms NVIDIA GPU on ResNet-50 Deep Learning Inference

Intel Says CPU Outperforms NVIDIA GPU on ResNet-50 Deep Learning Inference

Enterprise & IT May 21,2019 0

Intel compared the inference performance of two of their most expensive CPUs to NVIDIA GPUs, and claimed a victory, although Nvidia disagrees.

Intel said it had achieved leadership performance of 7878 images per second on ResNet-50 with its latest generation of Intel Xeon Scalable processors, outperforming 7844 images per second on NVIDIA Tesla V100, the best GPU performance as published by NVIDIA on its website including T4.

Specifically, Intel used a two-socket Intel Xeon 9282a processor, a high core-count multi-chip packaged server multiprocessor, using Intel Optimization for Caffe. Intel achieved 7878 images per second by simultaneously running 28 software instances each one across four cores with batch size 11. The performance on NVIDIA Tesla V100 is 7844 images per second and NVIDIA Tesla T4 is 4944 images per second per NVIDIA's published numbers.

However, to achieve the performance of a single mainstream NVIDIA V100 GPU, Intel combined two power-hungry, highest-end CPUs with an estimated price of $50,000-$100,000.

"Intel’s performance comparison highlighted the clear advantage of NVIDIA T4 GPUs, which are built for inference. When compared to a single highest-end CPU, they’re not only faster but also 7x more energy-efficient and an order of magnitude more cost-efficient," Nvidia said.

Intel Xeon Scalable processors are widely available in clouds and data centers. However, Intel says that having the CPU with high deep learning capabilities gives AI customers flexibility to manage their compute infrastructure uniformly and cost effectively.

Deep learning is used in image/video processing, natural language processing, personalized recommender systems, and reinforcement learning. The types of workloads and algorithms are rapidly expanding. Intel claims that a general purpose CPU is very adaptable to this dynamically changing environment.

Obviously, Intel’s latest Cascade Lake CPUs include new instructions that improve inference, making them the best CPUs for inference. However, Nvidia says that these Xeons are hardly competitive with NVIDIA deep learning-optimized Tensor Core GPUs.

Inference (also known as prediction), in simple terms, is the “pattern recognition” that a neural network does after being trained. It’s where AI models provide intelligent capabilities in applications, like detecting fraud in financial transactions, conversing in natural language to search the internet, and predictive analytics to fix manufacturing breakdowns before they even happen.

While most AI inference today happens on CPUs, NVIDIA Tensor Core GPUs are also being adopted across the full range of AI models. Tensor Core has transformed NVIDIA GPUs to highly efficient AI processors. Tensor Cores do multi-precision calculations at high rates to provide optimal precision for diverse AI models and have automatic support in popular AI frameworks.

A measure of the complexity of AI models is the number of parameters they have. Parameters in an AI model are the variables that store information the model has learned. While ResNet-50 has 25 million parameters, BERT has 340 million, a 13x increase.

So Nvidia striked back to Intel, saying that on an advanced model like BERT, a single NVIDIA T4 GPU is 56x faster than a dual-socket CPU server (Dual Intel Xeon Gold 6240) and 240x more power-efficient.

Another key usage of AI is in recommendation systems, which are used to provide relevant content recommendations on video sharing sites, news feeds on social sites and product recommendations on e-commerce sites.

Neural collaborative filtering, or NCF, is a recommender system that uses the prior interactions of users with items to provide recommendations. According to Nvidia, when running inference on the NCF model that is a part of the MLPerf 0.5 training benchmark, NVIDIA T4 brings 12x more performance and 24x higher energy efficiency than CPUs (Single Intel Xeon Gold 6140)

Tags: deep learningIntel XeonNvidia Tesla
Previous Post
Apple Expands Keyboard Repairs to Newer MacBooks
Next Post
HONOR Unveils New Flagship HONOR 20 Series

Related Posts

  • Intel Launches New Xeon W-3400 and Xeon W-2400 Workstation Processors

  • SK hynix Obtains Industry’s First Validation for 1anm DDR5 DRAM on the 4th Gen Intel Xeon Scalable Processor

  • Intel Announces New Xeon W-3300 Processors

  • Researchers Use Analog AI hardware to Support Deep Learning Inference Without Great Accuracy

  • Intel Launches 10th Gen Intel Core vPro Processors For the Modern Workforce

  • Intel Announces Portfolio for 5G Network Infrastructure, New 2nd Generation Xeon Scalable Platforms

  • Toshiba Establishes a Data Corporation

  • CES: Pqlabs Say New QuantaFlow AI Interface Accelerator is 10X Faster than NVIDIA V100

Latest News

Samsung Galaxy S25 Edge Features New Corning Gorilla Glass Ceramic 2 for Enhanced Durability
Smartphones

Samsung Galaxy S25 Edge Features New Corning Gorilla Glass Ceramic 2 for Enhanced Durability

Razer announces Clio Chair Accessory for Audio Immersion
Consumer Electronics

Razer announces Clio Chair Accessory for Audio Immersion

Razer Unveils Ergonomic Gaming Mouse and Keyboard for Gaming on the Go
PC components

Razer Unveils Ergonomic Gaming Mouse and Keyboard for Gaming on the Go

Noctua releases NH-D15 G2 specific offset LGA1851 mounting bars for improved cooling performance
Cooling Systems

Noctua releases NH-D15 G2 specific offset LGA1851 mounting bars for improved cooling performance

ADATA Launches T7 and T5 Enterprise SSD Series
Enterprise & IT

ADATA Launches T7 and T5 Enterprise SSD Series

Popular Reviews

be quiet! Light Loop 360mm

be quiet! Light Loop 360mm

be quiet! Dark Rock 5

be quiet! Dark Rock 5

G.skill Trident Z5 Neo RGB DDR5-6000 64GB CL30

G.skill Trident Z5 Neo RGB DDR5-6000 64GB CL30

Arctic Liquid Freezer III 420 - 360

Arctic Liquid Freezer III 420 - 360

be quiet! Dark Mount Keyboard

be quiet! Dark Mount Keyboard

Crucial Pro OC 32GB DDR5-6000 CL36 White

Crucial Pro OC 32GB DDR5-6000 CL36 White

Crucial T705 2TB NVME White

Crucial T705 2TB NVME White

be quiet! Light Base 600 LX

be quiet! Light Base 600 LX

Main menu

  • Home
  • News
  • Reviews
  • Essays
  • Forum
  • Legacy
  • About
    • Submit News

    • Contact Us
    • Privacy

    • Promotion
    • Advertise

    • RSS Feed
    • Site Map
  • About
  • Privacy
  • Contact Us
  • Promotional Opportunities @ CdrInfo.com
  • Advertise on out site
  • Submit your News to our site
  • RSS Feed