Here’s why 100TB+ SSDs will play a big role in ultra-large language models in the near future


  • Kioxia unveils new project called AiSAQ, which wants to replace RAM with SSDs for AI computing
  • Larger (read: 100TB+) SSDs could improve RAG at a lower cost than just using memory
  • No timeline has been given, but expect Kioxia’s rivals to offer similar technology

Large language models often generate plausible but factually incorrect output – in other words, they make things up. These “hallucinations” can damage the reliability of information-critical tasks such as medical diagnosis, legal analysis, financial reporting and scientific research.

Retrieval-Augmented Generation (RAG) addresses this problem by integrating external data sources, giving LLMs access to real-time information during generation, reducing errors and, by grounding output in current data, improving contextual accuracy. Implementing RAG efficiently requires significant memory and storage resources, and this is especially true for large-scale vector data and indexes. Traditionally, this data has been stored in DRAM, which, while fast, is both expensive and limited in capacity.

Leave a Comment

Your email address will not be published. Required fields are marked *