Supermicro, Inc. is launching full stack optimized storage solution for AI and ML data pipelines from data collection to high performance data delivery. This new solution maximizes AI time-to-value by keeping GPU data pipelines fully saturated. For AI training, massive amounts of raw data at petascale capacities can be collected, transformed, and loaded into an organization's AI workflow pipeline.

This multi-tiered Supermicro solution has been proven to deliver multi-petabyte data for AIOps and MLOps in production environments. The entire multi-rack scale solution from Supermicro is designed to reduce implementation risks, enable organizations to train models faster, and quickly use the resulting data for AI inference. Petabytes of unstructured data used in large-scale AI training processing must be available to the GPU servers with low latencies and high bandwidth to keep the GPUs productive.

Supermicro's extensive portfolio of Intel and AMD based storage servers is a crucial element of the AI pipeline. These include the Supermicro Petascale All-Flash storage servers, which have a capacity of 983.04 TB per server of NVMe Gen 5 flash capacity and deliver up to 230 GB/s of read bandwidth and 30 million IOPS. This solution also includes the Supermicro SuperServer 90 drive bay storage servers for the capacity object tier.

This complete and tested solution is available worldwide for customers in ML, GenAI, and other computationally complex workloads. The new storage solution consists of: All-Flash tier - Supermicro Petascale Storage Servers; Application tier ? Supermicro 8U GPU Servers: AS -8125GS-TNHR and SYS-821GE-TNHR; Object tier - Supermicro 90 drive bay 4U SuperStorage Server running Quantum ActiveScale object storage; Software: WEKA Data Platform and Quantum ActiveScale object storage; and Switches: Supermicro InfiniBand and Ethernet Switches.