Cloudian Announces Support For NVIDIA GPUDirect Acceleration For Object Storage
AUCKLAND – 19 November 2024 – Cloudian, the global leader in secure S3-compatible AI data lake platforms, today announced its integration with NVIDIA Magnum IO™ GPUDirectÒ Storage technology, delivering breakthrough performance of over 200GB/s from a HyperStore system. Cloudian HyperStore® with GPUDirect access simplifies the management of AI training and inference datasets – at petabyte and exabyte scales – while reducing costs by eliminating the need for complex data migrations and legacy file storage layers.
Key Benefits of NVIDIA GPUDirect Storage for Object Storage in AI Training and Inference Workflows:
- Limitless Scalability: Expands effortlessly to exabyte scale without disruption, supporting growing AI datasets without adding management complexity.
- Reduced Costs and No Data Migrations: Removes legacy file layers and enables a single, unified data lake without the need for constant data movement between tiers.
- High Performance: Delivers over 200GB/s from a single system with performance sustained over a 30-minute period without the use of data caching.
- Maximised CPU for AI Workloads: Slashes CPU overhead by 45 per cent during data transfers, freeing computational resources for AI processing.
- No Kernel Modifications: Eliminates the security exposure of vendor-specific kernel modifications, reducing security vulnerabilities
- Integrated Metadata: Rich metadata facilitates rapid search without the need for external databases.
“Cloudian is proud to be at the forefront of transforming how enterprises and AI hyperscalers harness data to realise the power of AI,” said Michael Tso, CEO of Cloudian. “For too long, AI users have been saddled with the unnecessary complexity and performance bottlenecks of legacy storage solutions. With GPUDirect Storage integration, we are enabling AI workflows to directly leverage a simply scalable storage architecture so organisations can unleash the full potential of their data.”
“At Supermicro, we're committed to delivering the most advanced and efficient solutions for AI and deep learning,” said Michael McNerney, SVP of Marketing and Network Security at Supermicro. “Cloudian's integration of NVIDIA GPUDirect Storage with the HyperStore line of object storage appliances based on Supermicro systems – including the Hyper 2U and 1U servers, the high-density SuperStorage 90-bay storage servers, and the Simply Double 2U 24-bay storage servers – represents a significant innovation in the use of object storage for AI workloads. This will enable our mutual customers to deploy more powerful and cost-effective AI infrastructure at scale.”
“Fast, consistent, and scalable performance in object storage systems is crucial for AI workflows,” said Rob Davis, vice president of storage technology, NVIDIA. “It enables real-time processing and decision-making, which are essential for applications like fraud detection and personalised recommendations.”
Simplifies Data
Management, Exabyte Scale Eliminates Data
Migration
Legacy file-based storage systems in
AI workflows often require frequent data movement between
long-term and high-speed storage, adding management
complexity. With Cloudian’s solution, AI training and
inference happen directly on the data in-place, accelerating
workflows and eliminating frequent migration. Cloudian
HyperStore’s limitless scalability enables AI data lakes
to grow to exabyte levels, while its centralised management
ensures simple, unified control across multi-data centre and
multi-tenant environments.
Fast Throughput for
Higher GPU Utilisation
NVIDIA
GPUDirect Storage with
NVIDIA ConnectX and
NVIDIA BlueField networking technologies optimise data
transfer speeds by enabling direct communication between
NVIDIA GPUs and multiple Cloudian storage nodes, bypassing
the CPU. This direct parallel data transfer delivers
consistent and scalable performance over 200GB/s from a
HyperStore system—as measured on the industry-standard
GOSBench benchmark over a sustained period without the use
of data caching. As throughput can be easily and
economically scaled, organisations can achieve better GPU
utilisation and lower GPU communications
latency.
Reduces Storage
Costs
Managing the enormous datasets needed for
AI workflows can be both costly and resource intensive.
Cloudian’s software-defined platform helps address these
challenges by eliminating the need for a separate file
storage layer. With AI workflows occurring directly within
the object-based data lake, organisations can streamline
data management while significantly reducing operational and
capital expenses, as well as overall complexity.
No Kernel Level Modifications
GPUDirect for Object Storage requires no vendor-driven kernel-level modifications. Unlike file solutions, this approach reduces potential vulnerabilities typically associated with kernel changes. By eliminating the need for such alterations, it simplifies system administration, decreases attack surfaces, and lowers the risk of security breaches.
Integrated Metadata for Simplicity
and Accelerated Search
Metadata plays a crucial
role in AI workflows by enabling rapid data discovery,
retrieval, and access control. Cloudian accelerates AI data
searches with integrated metadata support that allows for
easy tagging, classification, and indexing of large
datasets. Unlike legacy file-based systems, which depend on
rigid directory structures and separate databases for
metadata management, Cloudian natively handles metadata
within the object storage platform, simplifying workflows
and speeding up AI training and inference
processes.
Enhanced Data
Security
Data privacy and security are top
priorities for enterprises adopting AI, as noted by
Forrester analysts. Cloudian addresses these concerns with
the industry’s most comprehensive range of security
features. These include advanced access controls, encryption
protocols, integrated key management, and S3 Object Lock for
ransomware protection, helping ensure that sensitive AI data
remains safe and secure throughout its
lifecycle.
Reduced CPU Consumption
Cloudian's integration with NVIDIA's GPUDirect Storage technology enables direct data transfers between storage systems and GPU memory, bypassing the CPU. This direct path reduces CPU utilisation by 45 per cent during data transfers, allowing the CPU to focus on other tasks and improving overall system efficiency.
Cloudian HyperStore with NVIDIA Magnum IO GPUDirect Storage technology is available now.
“At Softsource vBridge, we've seen firsthand how data management challenges can hinder AI adoption,” said David Small, Group Technology Officer of Softsource vBridge. “Cloudian's GPUDirect for Object Storage will simplify the entire AI data lifecycle, which could be the key to democratising AI across various business sectors, allowing companies of all sizes to harness the power of their data. We're particularly excited about how this could accelerate AI projects for our mid-market clients who have previously found enterprise AI solutions out of reach.”