AMD Instinct MI300X
by AMD · paid · Last verified 2026-03-17
AMD's flagship AI accelerator based on CDNA3 architecture with a chiplet design integrating 192GB HBM3 memory — the highest capacity of any GPU accelerator. Its massive memory capacity makes it uniquely suited for serving very large models without model parallelism.
https://www.amd.com/en/products/accelerators/instinct/mi300/mi300x.html ↗B
B—Above Average
Adoption: BQuality: A+Freshness: ACitations: B+Engagement: F
Specifications
- License
- Proprietary
- Pricing
- paid
- Capabilities
- ai-training, inference, fp8-compute, high-capacity-memory
- Integrations
- rocm, pytorch, tensorflow, vllm, triton
- Use Cases
- large-model-inference, llm-training, hpc
- API Available
- No
- Tags
- gpu, data-center, training, inference, amd, cdna3
- Added
- 2026-03-17
- Completeness
- 100%
Index Score
60Adoption
60
Quality
90
Freshness
85
Citations
72
Engagement
0
Put AI to work for your business
Deploy this hardware alongside autonomous AaaS agents that handle tasks end-to-end — no babysitting required.