NVIDIA, Kioxia Partner on 100M IOPS SSDs for AI Servers by 2027

Artificial intelligence thrives on data, and lots of it. Whether it’s the large language models we chat with or the systems that help us find information, these AI applications need to access data at incredible speeds. Sometimes, the fastest GPUs aren’t enough, because the data simply can’t get to them quickly enough. This bottleneck, especially when data is spread out or sitting in slower memory, can really slow things down.

A recent report from Tom’s Hardware on Friday, April 12, highlights a major push to solve this problem. Kioxia, a Japanese storage company, is joining forces with NVIDIA, a powerhouse in AI hardware. Their mission? To develop a solid-state drive (SSD) that can hit a mind-blowing 100 million input/output operations per second (IOPS) by 2027. This isn’t for your average home computer; it’s specifically for AI servers that demand intense, random data reads.

To grasp how huge this is, consider that today’s top-performing data center SSDs typically achieve around 2 to 3 million IOPS for random reads of small data blocks (like 4 KB or 512 B). The target of 100 million IOPS represents a leap of about 30 to 33 times what we see now. It’s a massive shift in how AI systems will access information. Large AI models or real-time inference applications, such as language models, vector search, embeddings, and data retrieval, rely heavily on these small, random reads with very low delays.

Chasing the Speed Record

Image: Courtesy of NVIDIA

So, how do Kioxia and NVIDIA plan to achieve such a massive speed boost? Their strategy involves several key technologies. First, they’re looking at something called XL-Flash memory. This is a special type of SLC NAND memory designed with many internal sections, or ‘planes,’ which allows for extreme parallel processing. Imagine dozens of workers doing tasks at the same time, instead of just a few.

Next, they’ll use super-fast connections. This includes the cutting-edge PCIe 7.0 interface. What’s even more important is a direct ‘peer-to-peer’ link between the SSD and the GPU. This setup bypasses the main computer processor (CPU), cutting down on delays caused by data having to travel through extra components. It’s like having a dedicated high-speed lane for data, directly from storage to the AI brain.

They might also use server designs with multiple controllers or SSD modules working together. This would help spread out the workload. However, making all these parts work perfectly at this scale – from the controller chips to the software and the NAND memory itself – presents its own set of tough engineering puzzles. Achieving 100 million IOPS linearly will be a significant technological challenge.

The Road Ahead: Hurdles to Overcome

While the vision is exciting, building such advanced storage comes with significant hurdles. The first big worry is power and heat. SSDs operating at these extreme speeds will generate a lot of heat. Keeping them cool will require very efficient cooling systems and smart power designs. Nobody wants a server that melts down.

Then there’s the cost. Producing memory with so many internal layers, powerful controllers, and the latest interfaces won’t be cheap. If the price tag is too high, only a very small group of specialized users might be able to afford it, limiting its widespread adoption.

Another challenge is the gap between lab results and real-world performance. In perfect test environments, hitting 100 million IOPS might be possible. But in actual use, with varying workloads, many competing data requests, or unexpected glitches, the actual performance could be much lower. It’s easy to make a car go fast on a test track; it’s harder to do it in everyday traffic.

Finally, compatibility and integration pose a big question. Current servers aren’t built for a direct SSD-GPU connection or the new PCIe 7.0 standard. To use these new SSDs, companies will need to invest in redesigning their entire server infrastructure. This means big investments across the whole tech supply chain, from chip makers to server builders.

Recent Articles

Related News

Leave A Reply

Please enter your comment!
Please enter your name here