|
This video is part of the appearance, “Solidigm Presents at AI Infrastructure Field Day 2“. It was recorded as part of AI Infrastructure Field Day 2 at 13:30 - 15:30 on April 23, 2025.
Watch on YouTube
Watch on Vimeo
Solidigm focused on the evolving role of storage in AI, specifically highlighting its significance in the AI data pipeline through 2025. The presentation emphasizes that the AI workflow involves a series of distinct tasks, each with unique demands on hardware, and that it’s often distributed among different organizations. The presentation identifies six primary stages in this pipeline: ingestion, data preparation, model training, fine-tuning, inference, and archiving. Solidigm’s perspective underscores the increasing data intensity of these steps, especially during data ingestion, inference, and archiving.
Solidigm explores how storage choices impact these stages, distinguishing between direct-attached storage within GPU servers and network-attached storage. Direct-attached storage is optimized for performance-intensive tasks, while network-attached storage is used for larger datasets and offers capacity and cost-effectiveness. Ace Stryker highlights the rising importance of network storage, driven by advances in network bandwidth and the growing size and complexity of AI models and larger datasets for RAG. The key takeaway is that great storage facilitates larger models, longer interactions, and improved outputs on the inference side.
Finally, the presentation showcases a collaboration with Metrum AI, presenting a real-world demo that addresses the challenges of data-intensive inference by offloading model weights and RAG data to SSDs. This allows running 70 billion parameter models on less powerful hardware, which would have been impossible without storage-based offloading, saving on costs and hardware requirements. The demo emphasizes the potential of SSDs to enhance performance in AI applications by reducing GPU memory usage. The collaboration offers insights into the benefits of leveraging storage in AI, as the partnership also showed that offloading to SSD could provide the same or better performance as DRAM in this application of retrieval augmented generation.
Personnel: Ace Stryker, Steen Graham