SAN MATEO, CA – Nov 18, 2024 – Cloudian, the global leader in secure S3-compatible AI data lake platforms, today announced its integration with NVIDIA Magnum IO GPUDirectÒ Storage technology, delivering breakthrough performance of over 200GB/s from a HyperStore system. Cloudian HyperStore® with GPUDirect access simplifies the management of AI training and inference datasets – at petabyte and exabyte scales – while reducing costs by eliminating the need for complex data migrations and legacy file storage layers.
Key Benefits of NVIDIA GPUDirect Storage for Object Storage in AI Training and Inference Workflows:
- Limitless Scalability: Expands effortlessly to exabyte scale without disruption, supporting growing AI datasets without adding management complexity.
- Reduced Costs and No Data Migrations: Removes legacy file layers and enables a single, unified data lake without the need for constant data movement between tiers.
- High Performance: Delivers over 200GB/s from a single system with performance sustained over a 30-minute period without the use of data caching.
- Maximized CPU for AI Workloads: Slashes CPU overhead by 45% during data transfers, freeing computational resources for AI processing.
- No Kernel Modifications: Eliminates the security exposure of vendor-specific kernel modifications, reducing security vulnerabilities
- Integrated Metadata: Rich metadata facilitates rapid search without the need for external databases.S3
“Cloudian is proud to be at the forefront of transforming how enterprises and AI hyperscalers harness data to realize the power of AI,” said Michael Tso, CEO of Cloudian. “For too long, AI users have been saddled with the unnecessary complexity and performance bottlenecks of legacy storage solutions. With GPUDirect Storage integration, we are enabling AI workflows to directly leverage a simply scalable storage architecture so organizations can unleash the full potential of their data.”
“At Supermicro, we’re committed to delivering the most advanced and efficient solutions for AI and deep learning,” said Michael McNerney, SVP of Marketing and Network Security at Supermicro. “Cloudian’s integration of NVIDIA GPUDirect Storage with the HyperStore line of object storage appliances based on Supermicro systems – including the Hyper 2U and 1U servers, the high-density SuperStorage 90-bay storage servers, and the Simply Double 2U 24-bay storage servers – represents a significant innovation in the use of object storage for AI workloads. This will enable our mutual customers to deploy more powerful and cost-effective AI infrastructure at scale.”
“Fast, consistent, and scalable performance in object storage systems is crucial for AI workflows,” said Rob Davis, vice president of storage technology, NVIDIA. “It enables real-time processing and decision-making, which are essential for applications like fraud detection and personalized recommendations.”
Legacy file-based storage systems in AI workflows often require frequent data movement between long-term and high-speed storage, adding management complexity. With Cloudian’s solution, AI training and inference happen directly on the data in-place, accelerating workflows and eliminating frequent migration. Cloudian HyperStore’s limitless scalability enables AI data lakes to grow to exabyte levels, while its centralized management ensures simple, unified control across multi-data center and multi-tenant environments.
NVIDIA GPUDirect Storage with NVIDIA ConnectX and NVIDIA BlueField networking technologies optimize data transfer speeds by enabling direct communication between NVIDIA GPUs and multiple Cloudian storage nodes, bypassing the CPU. This direct parallel data transfer delivers consistent and scalable performance over 200GB/s from a HyperStore system—as measured on the industry-standard GOSBench benchmark over a sustained period without the use of data caching. As throughput can be easily and economically scaled, organizations can achieve better GPU utilization and lower GPU communications latency.
Managing the enormous datasets needed for AI workflows can be both costly and resource intensive. Cloudian’s software-defined platform helps address these challenges by eliminating the need for a separate file storage layer. With AI workflows occurring directly within the object-based data lake, organizations can streamline data management while significantly reducing operational and capital expenses, as well as overall complexity.
GPUDirect for Object Storage requires no vendor-driven kernel-level modifications. Unlike file solutions, this approach reduces potential vulnerabilities typically associated with kernel changes. By eliminating the need for such alterations, it simplifies system administration, decreases attack surfaces, and lowers the risk of security breaches.
Metadata plays a crucial role in AI workflows by enabling rapid data discovery, retrieval, and access control. Cloudian accelerates AI data searches with integrated metadata support that allows for easy tagging, classification, and indexing of large datasets. Unlike legacy file-based systems, which depend on rigid directory structures and separate databases for metadata management, Cloudian natively handles metadata within the object storage platform, simplifying workflows and speeding up AI training and inference processes.
Data privacy and security are top priorities for enterprises adopting AI, as noted by Forrester analysts. Cloudian addresses these concerns with the industry’s most comprehensive range of security features. These include advanced access controls, encryption protocols, integrated key management, and S3 Object Lock for ransomware protection, helping ensure that sensitive AI data remains safe and secure throughout its lifecycle.
Cloudian’s integration with NVIDIA’s GPUDirect Storage technology enables direct data transfers between storage systems and GPU memory, bypassing the CPU. This direct path reduces CPU utilization by 45% during data transfers, allowing the CPU to focus on other tasks and improving overall system efficiency.
Cloudian HyperStore with NVIDIA Magnum IO GPUDirect Storage technology is available now.