Nvidia's RTX IO promise: What happened to direct SSD-to-GPU data streaming?

09/26/2025

Initially, Nvidia's RTX IO was presented as a groundbreaking feature designed to streamline data flow directly from solid-state drives (SSDs) to graphics processing unit (GPU) memory, effectively bypassing the central processing unit (CPU). This innovation promised accelerated load times and enhanced performance for PC gaming. Concurrently, a new PCIe 5.0 switch, primarily for data centers, has emerged, leveraging Nvidia's GPUDirect technology to achieve a similar direct data transfer between storage and GPU. The striking resemblance between these two technologies, coupled with the apparent omission of direct SSD-to-GPU streaming from RTX IO's current public descriptions, has sparked considerable curiosity regarding the evolution and future of this once-hyped gaming feature.

The HighPoint Rocket 7638D, a PCIe 5.0 switch designed for data centers, has recently garnered attention. This switch facilitates high-speed data transfer of up to 64 GB/s from storage devices directly to AI GPUs, utilizing Nvidia's GPUDirect technology. GPUDirect, a feature available since 2019, establishes a direct conduit for data between storage and a GPU's video random-access memory (VRAM), circumventing the CPU and system memory, which traditionally act as intermediaries. This mechanism is crucial for demanding applications in artificial intelligence and scientific computing where rapid data access is paramount.

In contrast, Nvidia's RTX IO, first introduced with the GeForce RTX 30-series GPUs, was initially advertised with a similar capability: streaming data directly from SSDs to GPU memory. This was a significant selling point, promising to revolutionize game loading and asset streaming. However, a review of Nvidia's current documentation for RTX IO reveals a notable absence of this specific feature. Instead, the focus has shifted to other performance-enhancing aspects, such as GPU decompression and the offloading of CPU cycles. This change in emphasis has led many to question why a feature once highlighted has seemingly been de-emphasized or removed.

A public demonstration of RTX IO using Valve's Portal showcased its ability to rapidly load assets into VRAM, but even then, the discussion primarily revolved around GPU decompression and CPU cycle savings. The much-touted direct SSD-to-GPU data pathway was conspicuously absent from the discourse. This raises several questions: Why does Nvidia's data center ecosystem benefit from GPUDirect's direct data streaming, while the gaming-focused RTX IO has seemingly lost this capability? Was the feature technically challenging to implement for consumer-grade hardware, or did it face lukewarm reception from game developers?

A plausible explanation for this divergence might lie with Microsoft's DirectStorage API. DirectStorage offers a more optimized method for data transfer from storage to GPU memory. Although it does not entirely bypass the CPU and system RAM like GPUDirect, it processes data calls in a highly parallel fashion, significantly reducing the CPU's workload. Recent revisions to DirectStorage have integrated GPU decompression, allowing game assets to remain compressed until they are loaded into VRAM and then decompressed by the GPU's shader cores. Notably, one of the compression formats supported by DirectStorage, GDeflate, was developed by Nvidia specifically for RTX IO. This suggests a potential strategic shift or an integration of RTX IO's core concepts within Microsoft's broader DirectStorage framework.

The precise reasons behind the apparent sidelining of the direct SSD-to-GPU streaming aspect of RTX IO remain speculative. It could be due to implementation difficulties, a lack of widespread adoption by game developers, or simply a strategic pivot towards integrating its functionalities within more universally applicable APIs like DirectStorage. Despite the advanced capabilities of modern gaming hardware and software, the promise of instantaneous game loading and seamless asset streaming, as initially envisioned with RTX IO, has yet to fully materialize for the mainstream PC gaming audience. Meanwhile, console gamers and data centers continue to enjoy the benefits of highly optimized storage-to-GPU data pathways, leaving PC enthusiasts hopeful for a similar future.