⚙️ Hardware Platform Compatibility
AI/HPC accelerators with minimum required CUDA, ROCm, OFED, and Open MPI versions. Next-generation entries are based on public roadmap disclosures and may change.
NVIDIA Platforms
CUDA · Tensor Cores · NVLinkBlackwell
Blackwell Architecture (2024–2025)
sm_100 (B100/B200) / sm_100a (GB200)
📦 Products / SKUs
- NVIDIA B100 SXM
- NVIDIA B200 SXM
- NVIDIA B200 NVL
- NVIDIA GB200 Grace Blackwell NVL72
- NVIDIA GB200 NVL2
🖥️ Software Requirements
| CUDA | 12.4+ |
| Driver (Linux) | 550.54.14 |
| Linux Kernel | 5.15+ |
| OFED | 23.10-0.5.5.0+ |
| Open MPI | 4.1.6+ |
🔗 Interconnect
- InfiniBand NDR400 (400 Gb/s)
- PCIe Gen 6
- NVLink-C2C 900 GB/s (Grace-Blackwell only)
🔌 NVLink: NVLink 5.0 / NVLink-C2C (GB200)
📋 Notes
GB200 Grace Blackwell pairs an ARM Neoverse V2 CPU with a Blackwell GPU via NVLink-C2C. Requires CUDA 12.4+ and driver >= 550.54.14. DOCA 2.6+ required for full NIC/DPU feature support with ConnectX-7.
Hopper
Hopper Architecture (2022–2023)
sm_90 (H100) / sm_90a (GH200)
📦 Products / SKUs
- NVIDIA H100 SXM5
- NVIDIA H100 PCIe
- NVIDIA H200 SXM5
- NVIDIA H200 NVL
- NVIDIA GH200 Grace Hopper Superchip
🖥️ Software Requirements
| CUDA | 11.8+ |
| Driver (Linux) | 520.61.05 |
| Linux Kernel | 5.4+ |
| OFED | 5.8-3.0.7.0+ |
| Open MPI | 4.1.4+ |
🔗 Interconnect
- InfiniBand NDR200 / NDR400 (200–400 Gb/s)
- PCIe Gen 5
- NVLink-C2C 900 GB/s (GH200 only)
🔌 NVLink: NVLink 4.0 / NVLink-C2C (GH200)
📋 Notes
H100/H200 require CUDA 11.8+ for basic support; 12.0+ recommended. GH200 Grace Hopper requires ARM-native toolchain (aarch64). Transformer Engine (FP8) available from CUDA 11.8+ with cuDNN 8.7+.
Ada Lovelace (Enterprise/HPC)
Ada Lovelace Architecture (2022–2023)
sm_89
📦 Products / SKUs
- NVIDIA L40S
- NVIDIA L40
- NVIDIA RTX 6000 Ada
🖥️ Software Requirements
| CUDA | 11.8+ |
| Driver (Linux) | 520.61.05 |
| Linux Kernel | 5.4+ |
| OFED | 5.7-1.0.2.0+ |
| Open MPI | 4.1.4+ |
🔗 Interconnect
- PCIe Gen 4
- InfiniBand HDR/NDR (host NIC)
🔌 NVLink: N/A (PCIe only for HPC SKUs)
📋 Notes
L40S targets AI inference and training; no HBM memory. NVLink not available on PCIe form factor.
Ampere
Ampere Architecture (2020–2021)
sm_80 (A100/A800) / sm_86 (A30)
📦 Products / SKUs
- NVIDIA A100 SXM4 80GB
- NVIDIA A100 SXM4 40GB
- NVIDIA A100 PCIe 80GB
- NVIDIA A800 SXM4 80GB
- NVIDIA A30
🖥️ Software Requirements
| CUDA | 11.0+ |
| Driver (Linux) | 450.36.06 |
| Linux Kernel | 4.18+ |
| OFED | 5.2-2.2.3.0+ |
| Open MPI | 4.0.7+ |
🔗 Interconnect
- InfiniBand HDR (200 Gb/s)
- PCIe Gen 4
🔌 NVLink: NVLink 3.0 (600 GB/s)
📋 Notes
A100 is widely deployed in HPC clusters. TF32 and BF16 introduced. A800 is the export-controlled variant of A100 (NVLink bandwidth reduced). CUDA 11.0+ required; 11.4+ recommended for full bf16 support.
Volta
Volta Architecture (2017–2018)
sm_70
📦 Products / SKUs
- NVIDIA V100 SXM2 32GB
- NVIDIA V100 SXM2 16GB
- NVIDIA V100 PCIe 32GB
- NVIDIA V100S PCIe
🖥️ Software Requirements
| CUDA | 9.0+ |
| Driver (Linux) | 384.81 |
| Linux Kernel | 3.10+ |
| OFED | 4.7-1.0.0.1+ |
| Open MPI | 3.1.6+ |
🔗 Interconnect
- InfiniBand EDR / HDR100 (100 Gb/s)
- PCIe Gen 3
🔌 NVLink: NVLink 2.0 (300 GB/s)
📋 Notes
Tensor Cores first introduced in Volta. Still widely deployed in production HPC clusters. CUDA 9.0+ required; CUDA 11.x highly recommended. V100 PCIe lacks NVLink; only SXM2 form factor supports NVLink 2.0.
AMD Platforms
ROCm · CDNA · Infinity FabricMI300 Series (CDNA3)
CDNA3 Architecture (2023–2024)
gfx940 (MI300A) / gfx941 / gfx942 (MI300X)
📦 Products / SKUs
- AMD Instinct MI300X
- AMD Instinct MI300A (APU)
- AMD Instinct MI308X
🖥️ Software Requirements
| ROCm | 6.0+ |
| Driver (Linux) | 6.3.0 (amdgpu-dkms) |
| Linux Kernel | 5.15+ |
| OFED | 23.10-0.5.5.0+ |
| Open MPI | 4.1.6+ |
🔗 Interconnect
- InfiniBand NDR200 / NDR400
- PCIe Gen 5
- Infinity Fabric (MI300A CPU-GPU)
📋 Notes
MI300X has the largest HBM capacity (192 GB) of any GPU as of 2024. MI300A is an APU combining 24 Zen 4 CPU cores with CDNA3 GPU dies on one package via Infinity Fabric. ROCm 6.0+ required for full MI300 support; ROCm 6.1+ recommended for MI300X production workloads.
MI250 Series (CDNA2)
CDNA2 Architecture (2021–2022)
gfx90a
📦 Products / SKUs
- AMD Instinct MI250X
- AMD Instinct MI250
- AMD Instinct MI210
🖥️ Software Requirements
| ROCm | 5.0+ |
| Driver (Linux) | 5.13.0 (amdgpu-dkms) |
| Linux Kernel | 5.15+ |
| OFED | 5.4-3.5.8.0+ |
| Open MPI | 4.1.3+ |
🔗 Interconnect
- InfiniBand HDR (200 Gb/s)
- PCIe Gen 4
- AMD Infinity Fabric (inter-GCD)
📋 Notes
MI250X consists of two GCDs (Graphics Compute Dies) per OAM module. The Frontier supercomputer (1st exascale system) uses MI250X. ROCm 5.0+ required; 5.4+ recommended for stable production use. GPU-aware MPI requires Open MPI 4.1.3+ with ROCm-aware UCX.
MI100 (CDNA1)
CDNA1 Architecture (2020)
gfx908
📦 Products / SKUs
- AMD Instinct MI100
🖥️ Software Requirements
| ROCm | 4.0+ |
| Driver (Linux) | 5.10.0 (amdgpu-dkms) |
| Linux Kernel | 5.4+ |
| OFED | 5.2-2.2.3.0+ |
| Open MPI | 4.0.7+ |
🔗 Interconnect
- InfiniBand HDR (200 Gb/s)
- PCIe Gen 4
📋 Notes
First CDNA-architecture GPU; matrix cores introduced. CDNA1 uses a separate die architecture from RDNA consumer GPUs. ROCm 4.0+ required; ROCm 5.x provides improved HIP compatibility.
🔌 API Access
Hardware platform data is also available via the REST API (requires API key):
GET /api/v1/hardware
GET /api/v1/hardware/NVIDIA
GET /api/v1/hardware/AMD
GET /api/v1/hardware/NVIDIA/Blackwell
Add Accept: application/yaml to receive YAML output.