Electronics Maker
  • Home
  • Electronics News
  • Articles
  • Magazine
  • Events
  • Interview
  • Electronics Projects
No Result
View All Result
  • Home
  • Electronics News
  • Articles
  • Magazine
  • Events
  • Interview
  • Electronics Projects
No Result
View All Result
Electronics Maker
No Result
View All Result
Home Electronics News

NVIDIA Launches Magnum IO Software Suite to Help Eliminate Data Bottlenecks for Data Scientists and AI, HPC Researchers

Electronics Maker by Electronics Maker
November 19, 2019
in Electronics News
0
0
SHARES
121
VIEWS
Share on FacebookShare on Twitter

New Offering Moves Vast Amounts of Data up to 20x Faster Than Previously Possible

jkDENVER—SC19—Nov. 18, 2019—NVIDIA today introduced NVIDIA Magnum IO, a suite of software to help data scientists and AI and high performance computing researchers process massive amounts of data in minutes, rather than hours.

Optimized to eliminate storage and input/output bottlenecks, Magnum IO delivers up to 20x faster data processing for multi-server, multi-GPU computing nodes when working with massive datasets to carry out complex financial analysis, climate modeling and other HPC workloads.

NVIDIA has developed Magnum IO in close collaboration with industry leaders in networking and storage, including DataDirect Networks, Excelero, IBM, Mellanox and WekaIO.

“Processing large amounts of collected or simulated data is at the heart of data-driven sciences like AI,” said Jensen Huang, founder and CEO of NVIDIA. “As the scale and velocity of data grow exponentially, processing it has become one of data centers’ great challenges and costs.

“Extreme compute needs extreme I/O. Magnum IO delivers this by bringing NVIDIA GPU acceleration, which has revolutionized computing, to I/O and storage. Now, AI researchers and data scientists can stop waiting on data and focus on doing their life’s work,” he said.

At the heart of Magnum IO is GPUDirect, which provides a path for data to bypass CPUs and travel on “open highways” offered by GPUs, storage and networking devices. Compatible with a wide range of communications interconnects and APIs — including NVIDIA NVLink™ and NCCL, as well as OpenMPI and UCX — GPUDirect is composed of peer-to-peer and RDMA elements.

Its newest element is GPUDirect Storage, which enables researchers to bypass CPUs when accessing storage and quickly access data files for simulation, analysis or visualization.

NVIDIA Magnum IO software is available now, with the exception of GPUDirect Storage, which is currently available to select early-access customers. Broader release of GPUDirect Storage is planned for the first half of 2020.

Ecosystem Support

“Modern HPC and AI research relies upon an incredible amount of data, often more than a petabyte in scale, which requires a new level of technology leadership to best handle the challenge. DDN, by taking advantage of NVIDIA’s Magnum IO suite of software along with our parallel EXA5-enabled storage architecture, is paving the way to a new direct data path which makes petabyte-scale data stores directly accessible to the GPU at high bandwidth, an approach that was not previously possible.”

— Sven Oehme, chief research officer, DDN

“The amount of data that leading HPC and AI researchers now need to access continues to grow by leaps and bounds, making I/O a complex challenge for many to manage. IBM Spectrum Scale is designed to address the needs of any organization looking to accelerate AI and run data-intensive workloads. The use of IBM Spectrum Scale and NVIDIA GPU acceleration can help customers alleviate I/O bottlenecks and get the insights needed from their data faster.”

— Sam Werner, vice president of storage offering management, IBM

“Leading HPC and AI researchers choose Mellanox to provide them with the most advanced technology to move and process immense amounts of data as efficiently and quickly as possible. We have been working together with NVIDIA to ensure Magnum IO works seamlessly with Mellanox’s state-of-the-art InfiniBand and Ethernet interconnect solutions and to enable our mutual customers to overcome data bottlenecks altogether and advance their science, research and product development activities.”

— Dror Goldenberg, senior vice president of software architecture, Mellanox Technologies

Tags: AIEthernet interconnect solutions
Electronics Maker

Electronics Maker

Subscribe Newsletter

Subscribe to our mailing list to receives daily updates direct to your inbox!

*we hate spam as much as you do

Electronics Maker

It is a professionally managed electronics print media with highly skilled staff having experience of this respective field.

Subscribe Newsletter

Subscribe to our mailing list to receives daily updates direct to your inbox!

*we hate spam as much as you do

Quick Links

  • »   Electronics News
  • »   Articles
  • »   Magazine
  • »   Events
  • »   Interview
  • »   About Us
  • »   Contact Us

Contact Us

EM Media LLP
  210, II nd Floor, Sager Plaza – 1, Road No-44,, Plot No-16, Pitampura, New Delhi - 110034
  01145629941
  info@electronicsmaker.com
  www.electronicsmaker.com

  • Advertise with Us
  • Careers
  • Terms Of Use
  • Privacy Policy
  • Disclaimer
  • Cookie Policy

© 2020 Electronics Maker. All rights reserved.

No Result
View All Result
  • Home
  • Electronics News
  • Articles
  • Magazine
  • Events
  • Interview
  • Electronics Projects

© 2020 Electronics Maker. All rights reserved.