ScaleFlux Computational Storage With Support for Nvidia Magnum IO GPUDirect Storage
CSD2000 allows users to leverage computational storage for AI/ML and analytics acceleration with GPUs.
This is a Press Release edited by StorageNewsletter.com on March 25, 2021 at 2:31 pmScaleFlux, Inc. announced its CSD 2000 solution will support Nividia Magnum IO GPUDirect Storage, enabling users to take advantage of ‘Computational Storage’ for AI and ML and analytics acceleration.
As Nvidia Magnum IO – the IO subsystem for the GPU-accelerated data center – gains increasing popularity, users can combine it with the company’s CSD 2000 to take their data center to the next level with Computational Storage: a new class of storage drive that brings compute right to the data.
Training with larger and more diverse data can both reduce bias and improve statistical variance overall improving the accuracy of an AI/ML model. While effective in boosting accuracy, massive increases in data set sizes introduce new challenges around storing, preparing, and delivering the data to the GPUs. Developers often fetch compressed data from remote storage, however, this leads to a decompression burden in order to continue with the preparation process to generate the appropriate training data.
With this release, the company addresses this problem by delivering the use of Computational Storage to advance AI/ML and data analytics with GPUs. CSD 2000 deploys transparent compression/decompression, which differs from alternative solutions in that it requires no code changes to the application, does not incur latency or performance penalties, reduces data movement, and scales throughput with storage capacity. It offloads and improves parallelism in the processor-intensive compression/decompression functions, freeing up GPU cycles to perform analytics and model training instead of bogging down in data preparation. It also expands the capacity per flash bit by 3-5x, without the added latency or reduced throughput of other compression options.
“As our data sets grow larger and more complex, we are constantly looking for ways to improve AI initiatives with the latest technology,” said Jeff Hookailo, CEO, Middle Canyon. “We are already seeing positive results in our testing from the compression/decompression function in the ScaleFlux CSD, combined with the direct transfer of data between the CSD and the Nvidia GPUs using Magnum IO GPUDirect Storage. We are excited to see the ways this combination will enhance how we train and work with AI moving forward.“
“Modern AI and data science workloads are powered by vast amounts of data, which makes it critical to enable fast communications between GPU computing and data center storage systems,” said Kushal Datta, senior product manager, Nvidia Corp. “ScaleFlux’s addition of Nvidia Magnum IO GPUDirect Storage in their flagship CSD 2000 solution helps support AI deployments by boosting system bandwidth while decreasing latency in the data center.“
Key features of CSD 2000 with Nvidia Magnum IO GPUDirect Storage include:
-
Combines PCIe SSD performance levels with a flash mapping architecture and built-in compression/decompression engines
-
Achieves ‘penalty-free compression’, so users can scale compression/decompression throughput as they add storage capacity and deliver compression/decompression without hurting latency – a benefit that is impossible to achieve with host-based software compression
-
Enables users to take advantage of data compressibility to reduce the cost of storing each byte of user data, shaving up to 70% off the costs of ordinary enterprise SSD storage.
“Accelerated computing with Nvidia GPUs is increasingly critical to data-driven businesses. However, data preparation and decompression consumes precious time. This is where CSD 2000 adds value to users. It handles the decompression process and eliminates up to 87% of the data loading time so the GPU can get to work faster on the training activity,” said Hao Zhong, co-founder and CEO. “We at ScaleFlux have been collaborating with the Nvidia team for the past year, and are thrilled to support Nvidia Magnum IO GPUDirect Storage with the innovative capabilities of Computational Storage.“
The firm’s Computational Storage Drive CSD 2000 Series brings performance, scalability, and TCO savings to mainstream flash deployments. Theses drives combine up to 8TB of the latest 3D NAND flash technology with hardware-accelerated compute engines, achieving data R/W speeds and low latency.