Friday, January 19, 2018
Search
  
Submit your own News for
inclusion in our Site.
Click here...
Breaking News
Italy Probes Apple and Samsung Over Smartphone Software Updates
SoftBank Becomes Uber's Largest Shareholder
Congress Votes to Extend NSA Spying
IBM Reports Revenue After Six Years
Western Digital Expands Mid-Range Enterprise Air-based Hard Drives Offerings
Amazon Eyes the Advertising Business: report
European Regulators Approves Qualcomm's Acquisition of NXP, Subject to Conditions
Intel Says That Patches for Spectre, Meltdown Affect Newer Chips
Active Discussions
Which of these DVD media are the best, most durable?
How to back up a PS2 DL game
Copy a protected DVD?
roxio issues with xp pro
Help make DVDInfoPro better with dvdinfomantis!!!
menu making
Optiarc AD-7260S review
cdrw trouble
 Home > News > General Computing > AMD, In...
Last 7 Days News : SU MO TU WE TH FR SA All News

Tuesday, October 10, 2017
AMD, Intel, ARM, IBM and Others Support the Open Neural Network Exchange Format for AI


AMD, ARM, Huawei, IBM and Intel have announced their support for the Open Neural Network Exchange (ONNX) format, which was co-developed by Microsoft and Facebook in order to reduce friction for developing and deploying AI.

Introduced last month, the Open Neural Network Exchange (ONNX) format is a standard for representing deep learning models that enables models to be transferred between frameworks (PyTorch, Caffe2, and Cognitive Toolkit). ONNX is the first step toward an open ecosystem where AI developers can easily move between tools and choose the combination that is best for them.

Standardization is good for both the compute industry and for developers because it enables a level of interoperability between various products and frameworks, while streamlining the path from development to production.

By joining the project, Intel plans to further expand the choices developers have on top of frameworks powered by the Intel Nervana Graph library and deployment through the company's Deep Learning Deployment Toolkit.

Intel plans to enable users to convert ONNX models to and from Intel Nervana Graph models, giving users an even broader selection of choice in their deep learning toolkits.

Arm is already engaged to accelerate Caffe2 for its Arm Cortex-A CPUs as well as for Arm Mali GPU-based devices which currently use the Facebook application.



Previous
Next
QLC NAND Flash to Succeed TLC NAND Next Year        All News        Intel Delivers 17-Qubit Superconducting Chip with Advanced Packaging
Apple Enters into Deal Content With Spielberg, NBCUniversal: report     General Computing News      Intel Delivers 17-Qubit Superconducting Chip with Advanced Packaging

Get RSS feed Easy Print E-Mail this Message

Related News
Google Offers AI Image-Detection Tool to Cloud Customers
Microsoft AI Can Read a Document and Answer Questions About it as Well as a Person
AI-Powered Smartphones and Their Uses
LG Electronics Launches The ThinQ Brand
Google Opening Artificial Intelligence Research Center in China
IBM Says New POWER9-based AC922 Power Systems Offer 4x Deep-learning Framework Performance Over x86
IBM Scientists Demonstrate 10x Faster Machine Learning Using GPUs
Intel Advances Artificial Intelligence With Nervana Neural Network Processor
Intel's New Loihi Self-Learning Chip Promises to Accelerate Artificial Intelligence
Facebook Opens AI Research in Montreal
IFA: Huawei Says it Has an Edge Over Apple With New Kirin 970 Mobile AI Chipset
Microsoft Unveils Project Brainwave for Real-time AI

Most Popular News
 
Home | News | All News | Reviews | Articles | Guides | Download | Expert Area | Forum | Site Info
Site best viewed at 1024x768+ - CDRINFO.COM 1998-2018 - All rights reserved -
Privacy policy - Contact Us .