Home / Posts / Dell-NVIDIA 6TB/s Storage

Dell & NVIDIA's Exascale Storage: Powering the Next Generation of AI Factories at 6TB/s

March 20, 2026 Dillip Chowdary

The race for Artificial General Intelligence (AGI) has long been throttled not just by compute cycles, but by the physical limitations of data movement. Today, Dell Technologies and NVIDIA have fundamentally redefined the ceiling of AI infrastructure with the announcement of their new Exascale Storage Fabric (ESF). Capable of sustained throughput of 6TB/s (Terabytes per second), this system is specifically engineered to feed the voracious appetites of Blackwell-class AI Factories.

The Architecture of 6TB/s: Overcoming the PCIe Bottleneck

Achieving 6TB/s isn't merely a matter of adding more drives; it requires a radical rethinking of the storage controller and the interconnect. The ESF utilizes a Massively Parallel Architecture where storage is no longer a peripheral but an integrated part of the compute fabric. By leveraging NVIDIA Magnum IO GPUDirect Storage (GDS), data bypasses the CPU entirely, moving directly from the Dell PowerScale nodes into the H100/B200 GPU memory.

At the core of this performance leap is the implementation of PCIe Gen7 and 2.4Tbps Infiniand NDR-800 interconnects. Dell’s custom-designed Aura Controllers manage metadata across billions of small files with sub-microsecond latency, a critical requirement for checkpointing large language models (LLMs) with over 10 trillion parameters. The internal bus architecture has been redesigned to support non-blocking switching, ensuring that any GPU in the cluster can access any storage node with uniform latency.

Performance Metric

The system delivers a staggering 120 million IOPS at 4KB random reads, ensuring that even the most complex multi-modal training jobs never experience IO wait states. This represents a 5x improvement over the previous generation of high-end storage arrays.

AI Factories: From Data Centers to Synthesis Labs

The concept of the "AI Factory" moves beyond traditional cloud computing. These are purpose-built facilities where data is the raw material and intelligence is the finished product. Dell’s PowerScale F910 nodes, integrated into the NVIDIA Spectrum-4 Ethernet fabric, allow for seamless scaling from a single rack to an entire campus. The F910 introduces a multi-tenant isolation feature that allows multiple training jobs to share the same physical storage without performance interference, using hardware-level Quality of Service (QoS).

The software layer is equally impressive. Dell OneFS 10.5 introduces "Smart-Pathing," an AI-driven routing algorithm that predicts data access patterns and pre-caches datasets on 32TB SLC NVMe drives. This reduces the time-to-first-token in inference workloads by as much as 45%. OneFS now includes a native tensor-aware filesystem that understands the structure of weighted matrices, allowing for optimized block placement that aligns with the way GPUs ingest data during backpropagation.

The Sustainability Challenge: Cooling 6TB/s

Moving 6TB of data every second generates immense thermal energy, both in the silicon controllers and the NAND gates. Dell has addressed this with Direct Liquid Cooling (DLC) throughout the storage rack. By circulating a dielectric fluid directly over the ASICs and flash modules, the ESF achieves a PUE (Power Usage Effectiveness) of 1.04. This is critical as the total power draw of a fully populated exascale storage rack can exceed 150kW.

The sustainability strategy also extends to the data itself. Dell’s AI-Driven De-duplication uses a lightweight neural network to identify redundant data patterns across multi-petabyte datasets without the performance penalty of traditional hashing algorithms. This can reduce the physical storage footprint by up to 40%, significantly lowering the total cost of ownership (TCO) for large-scale AI projects.

The Future: From Training to Real-Time Synthesis

As we look toward 2027, the collaboration between Dell and NVIDIA signals a future where storage is no longer a bottleneck but a catalyst for discoveries. The ESF is already being used in early-access programs for computational biology, where researchers are training models on billions of protein folding sequences. The ability to stream this data at 6TB/s allows for real-time model adaptation, a feat that was previously impossible.

The "Storage-as-a-Compute" (SaaC) trend is also emerging. Dell is experimenting with Near-Data Processing (NDP), where simple data preprocessing tasks like normalization and augmentation are performed directly on the storage controller's ARM cores, further offloading the GPUs. This converged architecture is the blueprint for the next decade of High-Performance Computing (HPC).

Create Stunning AI Videos

Bring your technical blogs to life with AI Video Generator. Convert text to high-fidelity video in minutes.

Try AI Video Generator →