Cloudian Delivers Breakthrough AI Performance with New PyTorch Connector Leveraging Nvidia GPUDirect Storage Technology
74% performance boost and 43% reduction in CPU utilization accelerates machine learning workflows
This is a Press Release edited by StorageNewsletter.com on July 18, 2025 at 2:01 pmCloudian, Inc., provider in enterprise-grade object storage solutions, announced the availability of its new PyTorch connector with Remote Direct Memory Access (RDMA) support, delivering performance improvements for AI and ML workloads.
Built on Nvidia GPUDirect Storage technology and optimized for Nvidia Spectrum-X networking infrastructure, the breakthrough solution demonstrates a 74% increase in data processing performance while reducing processor utilization by 43%, representing a significant advancement in AI workflow acceleration.
Testing conducted using TorchBench, an open source PyTorch performance measurement tool, showed remarkable improvements in image processing capabilities. The new RDMA-enabled connector, built on Nvidia GPUDirect Storage technology, processed 52,000 images per second compared to 30,000 images per second using the default S3 connector-a substantial performance gain that directly translates to faster model training and reduced computational costs for AI practitioners.
“This represents a fundamental breakthrough in how AI workloads access and process data through advanced Nvidia networking acceleration,” said Neil Stobart, CTO, Cloudian. “By leveraging Nvidia GPUDirect Storage technology to eliminate traditional network bottlenecks, we’re enabling data scientists and AI engineers to supercharge their workflows while reducing infrastructure costs through direct GPU-to-storage communication.”
With RDMA, the enhanced connector is able to bypass traditional CPU-intensive network protocols, enabling direct memory-to-memory data transfers between Cloudian storage systems and GPU-accelerated AI frameworks running on Nvidia network infrastructure including Nvidia Spectrum-X Ethernet switches and Nvidia ConnectX SuperNICs. This architectural advancement proves particularly significant for PyTorch users leveraging Nvidia accelerated computing, who represent a substantial portion of the ML community including researchers at major technology companies, academic institutions, and AI-focused start-ups.
Benchmark testing was conducted using Cloudian HyperStore 8.2.2 software running on six Supermicro servers equipped with Nvidia networking platforms in an all-flash media configuration, representing enterprise-grade storage infrastructure commonly deployed for GPU-accelerated AI workloads.
The PyTorch ecosystem serves millions of developers worldwide, from individual researchers to large-scale enterprise AI operations utilizing Nvidia accelerated computing infrastructure. Organizations implementing computer vision, natural language processing, and deep learning applications on Nvidia platforms stand to benefit from the reduced training times and lower computational overhead delivered by the Nvidia GPUDirect Storage connector.
The integration with Nvidia GPUDirect Storage technology ensures optimal data path efficiency for AI workloads, eliminating unnecessary data copies and reducing latency in GPU-centric ML pipelines. This direct storage-to-GPU communication pathway maximizes the performance potential of Nvidia’s advanced networking and computing infrastructure.
The Cloudian PyTorch connector is available for evaluation, enabling organizations to assess the performance benefits within their Nvidia-accelerated AI environments.