Kioxia unveils new project called AiSAQ, which wants to replace RAM with SSDs for AI computing
Larger (read: 100TB+) SSDs could improve RAG at a lower cost than just using memory
No timeline has been given, but expect Kioxia’s rivals to offer similar technology
Large language models often generate plausible but factually incorrect output – in other words, they make things up. These “hallucinations” can damage the reliability of information-critical tasks such as medical diagnosis, legal analysis, financial reporting and scientific research.
Retrieval-Augmented Generation (RAG) addresses this problem by integrating external data sources, giving LLMs access to real-time information during generation, reducing errors and, by grounding output in current data, improving contextual accuracy. Implementing RAG efficiently requires significant memory and storage resources, and this is especially true for large-scale vector data and indexes. Traditionally, this data has been stored in DRAM, which, while fast, is both expensive and limited in capacity.