Here’s why 100TB+ SSDs will play a huge role in ultra large language models in the near future

0
1

  • Kioxia reveals new project called AiSAQ which wants to substitute RAM with SSDs for AI data processing
  • Bigger (read: 100TB+) SSDs could improve RAG at a lower cost than using memory only
  • No timeline has been given, but expect Kioxia’s rivals to offer similar tech

Large language models often generate plausible but factually incorrect outputs – in other words, they make stuff up. These “hallucination”s can damage reliability in information-critical tasks such as medical diagnosis, legal analysis, financial reporting, and scientific research.

Retrieval-Augmented Generation (RAG) mitigates this issue by integrating external data sources, allowing LLMs to access real-time information during generation, reducing errors, and, by grounding outputs in current data, improving contextual accuracy. Implementing RAG effectively requires substantial memory and storage resources, and this is particularly true for large-scale vector data and indices. Traditionally, this data has been stored in DRAM, which, while fast, is both expensive and limited in capacity.

LEAVE A REPLY

Please enter your comment!
Please enter your name here