
Intel Gaudi 3 AI Accelerator
Intel's third-gen AI accelerator optimized for LLM training and inference. Cost-effective alternative with Ethernet-based scaling. Contact us for evaluation units and pricing.
Price Available Upon Request
Contact us for custom pricing and availability
Ships directly from distributor
Price Available Upon Request
This cutting-edge product requires custom configuration and pricing. Contact our team for availability and quotes.
Or call us at (850) 407-7265
What it's for
Intel Gaudi 3 AI Accelerator delivers exceptional performance for training and inference of large language models and generative AI. Built on TSMC's 5nm process, Gaudi 3 combines 64 tensor processor cores, 8 matrix multiplication engines, and 128GB of HBM2e memory with industry-leading 3.7TB/s bandwidth. With integrated 24x 200GbE networking and competitive pricing, Gaudi 3 offers compelling value for AI infrastructure. Break free from proprietary interconnects. Gaudi 3 scales on standard Ethernet at half the cost. Call (555) 123-4567 for demo units and TCO analysis.
Key Features
- ✓64 tensor processor cores (TPCs) with 8 matrix multiplication engines (MMEs)
- ✓128GB HBM2e memory with 3.7TB/s bandwidth
- ✓1.8 petaFLOPS FP8 and BF16 compute performance
- ✓Integrated 24x 200 GbE networking for scale-out
- ✓96MB on-die SRAM cache for low-latency operations
- ✓14 integrated media engines (H.265, H.264, JPEG, VP9)
- ✓TSMC 5nm process node technology
- ✓Available in PCIe (600W) and OAM (900W) form factors
Use Cases
- →Large language model training (LLaMA, GPT-style models)
- →Generative AI inference at scale
- →Computer vision and video processing
- →Recommendation systems and embeddings
- →Multi-modal AI workloads
- →Cost-optimized AI infrastructure deployment
Technical Specifications
| Architecture | Gaudi 3 (TSMC 5nm) |
| GPU Memory | 128 GB HBM2e |
| Memory Bandwidth | 3.7 TB/s |
| Tensor Processor Cores | 64 TPCs |
| Matrix Multiply Engines | 8 MMEs |
| FP8 Performance | 1.8 petaFLOPS |
| BF16 Performance | 1.8 petaFLOPS |
| On-die SRAM | 96 MB |
| Network Interfaces | 24x 200 GbE |
| Media Engines | 14 (H.265/H.264/JPEG/VP9) |
| Max TDP (PCIe) | 600W |
| Max TDP (OAM) | 900W |
| Thermal Solution | Air cooling (PCIe) / Liquid (OAM) |
| Form Factor | PCIe Gen5 AIC |
| PCIe Interface | PCIe Gen5 x16 |
Related Products

NVIDIA H100 80GB PCIe Gen5
NVIDIA H100 Tensor Core GPU in PCIe form factor with 80GB HBM3 memory. Ideal for deploying AI inference and training in standard servers without NVLink clustering requirements. Contact our sales team for volume pricing and immediate availability.
$29,999

NVIDIA H100 80GB SXM5
NVIDIA H100 Tensor Core GPU in SXM5 form factor with NVLink for multi-GPU scaling. Designed for HGX server platforms and large-scale AI training clusters. Enterprise volume discounts available - contact sales for custom configurations.
$32,999

NVIDIA H200 141GB HBM3e SXM5
Industry-leading Hopper architecture GPU with 141GB HBM3e memory and 4.8TB/s bandwidth. Perfect for large language models, generative AI, and high-performance computing workloads. In stock now - contact us for immediate delivery and competitive pricing.
$39,999

NVIDIA B200 192GB Blackwell
Revolutionary Blackwell architecture with 192GB HBM3e and FP4 precision for next-gen AI. Pre-order now for 2025 delivery - reserve your allocation with our sales team.
$0