The modern enterprise defines itself by its data. This requires a data infrastructure for AI/ML as well as a data infrastructure that is the foundation for a Modern Datalake capable of supporting business intelligence, data analytics, and data science. This is true if they are behind, getting started or using AI for advanced insights. For the foreseeable future, this will be the way that enterprises are perceived. There are multiple dimensions or stages to the larger problem of how AI goes to market in the enterprise. Those include data ingestion, transformation, training, inferencing, production, and archiving, with data shared across each stage. As these workloads scale the complexity of the underlying AI data infrastructure increases. This creates the need for high performance infrastructure while minimizing total cost of ownership (TCO).
MinIO has created a comprehensive blueprint for data infrastructure to support exascale AI and other large scale data lake workloads. It is called the MinIO DataPod. The unit of measurement it uses is 100 PiB. Why? Because the reality is that this is common today in the enterprise. Here are some quick examples:
A North American automobile manufacturer with nearly an exabyte of car video
A German automobile manufacturer with more than 50 PB of car telemetry
A biotech firm with more than 50 PB of biological, chemical, & patient-centric data
A cybersecurity company with more than 500 PB of log files
A media streaming company with more than 200 PB of video
A defense contractor with more than 80 PB of geospatial, log and telemetry data from aircraft
Even if they are not at 100 PB today, they will be within a few quarters. The average firm is growing at 42% a year, data-centric firms are growing at twice that rate, if not more.
The MinIO Datapod reference architecture can be stacked in different ways to achieve almost any scale - indeed we have customers that have built off of this blueprint - all the way past an exabyte and with multiple hardware vendors. The MinIO DataPod offers an end-to-end architecture that enables infrastructure administrators to deploy cost-efficient solutions for a variety of AI and ML workloads.Here is the rationale for our architecture.
AI workloads, especially generative AI, inherently require GPUs for compute. They are spectacular devices with incredible throughput, memory bandwidth and parallel processing capabilities. Keeping up with GPUs that are getting faster and faster requires high-speed storage. This is especially true when training data cannot fit into memory and training loops have to make more calls to storage. Furthermore, enterprises require more than performance, they also need security, replication, and resiliency.
The enterprise storage requirement demands that the architecture fully disaggregate storage from compute. This allows for storage to scale independently of the compute and given that storage growth is generally one or more orders of magnitude more than compute growth, this approach ensures the best economics through superior capacity utilization.
Networking infrastructure has standardized on 100 Gigabits per second (Gbps) bandwidth links for AI workload deployments. Modern day NVMe drives provide 7GBps throughput on average making the network bandwidth between the storage servers and the GPU compute servers the bottleneck for AI pipeline execution performance.
Solving this problem with complex networking solutions like Infiniband (IB) has real limitations. We recommend that enterprises leverage existing, industry-standard Ethernet-based solutions (e.g., HTTP over TCP) that work out of the box to deliver data at high throughput for GPUs for the following reasons:
It is not a coincidence that AI data infrastructure in public clouds are all built on top of object stores. Nor is it a coincidence that every major foundational model was trained on an object store. This is a function of the fact that POSIX is too chatty to work at the data scale required by AI - despite what the chorus of legacy filers will claim.
The same architecture that delivers AI in the public cloud should be applied to the private cloud and obviously the hybrid cloud. Object stores excel at handling various data formats and large volumes of unstructured data and can effortlessly scale to accommodate growing data without compromising performance. Their flat namespace and metadata capabilities enable efficient data management and processing that is crucial for AI tasks requiring fast access to large datasets.
As high-speed GPUs evolve and network bandwidth standardizes at 200/400/800 Gbps and beyond, modern object stores will be the only solution that meets the performance SLAs and scale of AI workloads.
We know that GPUs are the star of the show and that they are hardware. But even Nvidia will tell you the secret sauce is CUDA. Move outside the chip, however, and the infrastructure world is increasingly software-defined. Nowhere is this more true than storage.Software-defined storage solutions are essential for scalability, flexibility, and cloud integration, surpassing traditional appliance-based models for the following reasons:
Cloud Compatibility: Software-defined storage aligns with cloud operations, unlike appliances that cannot run across multiple clouds.
Containerization: Appliances cannot be containerized, losing cloud-native advantages and preventing Kubernetes orchestration.
Hardware Flexibility: Software-defined storage supports a wide range of hardware, from edge to core, accommodating diverse IT environments.
Adaptive Performance: Software-defined storage offers unmatched flexibility, efficiently managing different capacities and performance needs across various chipsets.
At exabyte scale, simplicity and a cloud-based operating model are crucial. Object storage, as a software-defined solution, should work seamlessly on commodity off-the-shelf (COTS) hardware and any compute platform, be it bare metal, virtual machines, or containers.
Custom-built hardware appliances for object storage often compensate for poorly designed software with costly hardware and complex solutions, resulting in a high total cost of ownership (TCO).
Enterprise customers using MinIO for AI initiatives build exabyte scale data infrastructure as repeatable units of 100PiB. This helps infrastructure administrators ease the process of deployment, maintenance and scaling as the AI data grows exponentially over a period of time. Below is the bill of materials (BOM) for building a 100PiB scale data infrastructure.
Component |
Quantity |
---|---|
Total number of Racks |
30 |
Total number of Storage Servers |
330 |
Total number of storage servers per rack |
11 |
Total number of TOR switches |
60 |
Total number of Spine switches |
10 |
Erasure Code Stripe Size |
10 |
Erasure Code Parity |
4 |
Component |
Description |
Quantity |
---|---|---|
Rack Enclosure |
42U/45U slot Rack |
1 |
Storage Server |
2U form factor |
11 |
Top Of the Rack Switches |
Layer 2 switch |
2 |
Management Switch |
Combined Layer 2 and Layer 3 |
1 |
Network Cables |
AOC cables |
30-40 |
Power |
Dual power supply with RPDU |
17kW - 20kW |
Component |
Specification |
---|---|
Server |
2U, single socket |
CPU |
64 core, 128 * PCIe 4.0 lanes |
Memory |
256 GB |
Network |
Dual port, 200gbe NIC |
Drive Bays |
24 hot-swap 2.5" U.2 NVMe |
Drives |
30TB * 24 NVMe |
Power |
1600W Redundant Power Supplies |
Total Raw Capacity |
720 TB |
Dell: PowerEdge R7615 Rack Server
Supermicro: A+ Server 2114S-WN24RT
Component |
Specification |
|
---|---|---|
Top of the Rack (TOR) Switch |
32 * 100GbE QSFP 28 Ports |
|
Spine Switch |
64 * 100GbE QSFP 28 Ports |
|
Cable |
100G QSFP 28 AOC |
|
Power |
500 Watts per switch |
|
MinIO has validated this architecture with multiple customers and would expect others to see the following average price per terabyte per month. This is an average street price and the actual price may vary depending on the configuration and the hardware vendor relationship.
Scale |
Storage Hardware price**(Per TB/month)** |
MinIO Software Price**(Per TB/month)** |
---|---|---|
100PiB |
$1.50 |
$3.54 |
Vendor specific turnkey hardware appliances for AI will result in high TCO and is not scalable from an unit economics standpoint for large data AI initiatives at exabyte scale.
Data Infrastructure setup at exabyte scale while meeting the TCO objectives for all AI/ML workloads can be complex and hard to get right. MinIO’s DataPOD infrastructure blueprint makes it simple and straightforward for Infrastructure administrators to set up the required commodity off the shelf hardware with highly scalable, performant cost effective S3 compatible MinIO enterprise object store resulting in improved overall time-to-market and faster time to value from AI initiatives across organizations within the enterprise landscape.