NVIDIA B100
The NVIDIA B100, powered by cutting-edge Blackwell architecture and boasting real-time inference to accelerate trillion-parameter large language models, is the world’s most powerful GPU for AI and HPC.
NVIDIA B100
Purpose-built B100 clusters, designed and managed by CUDO
Deployed across 16 ISO-certified data centres
From 8 to 1,000+ GPUs in a single deployment
NVIDIA Quantum-X800 InfiniBand or Spectrum-X Ethernet networking
Expert rack-level design, installation, and benchmarking before handoff
24/7 monitoring, management, and engineering support
Compatible with Slurm, Kubernetes, and NVIDIA Base Command
Available at the most cost-effective pricing
Launch your AI products faster with on-demand GPUs and a global network of data center partners
Bare metal
Powered by renewable energy
No noisy neighbors
SpectrumX local networking
300Gbps external connectivity
NVMe SSD storage
Enterprise
Powerful GPU clusters
Scalable data center colocation
Large quantities of GPUs and hardware
Optimize to your requirements
Expert installation
Scale as your demand grows
Specifications
Browse specifications for the NVIDIA B200 GPU
Starting from
Contact us for pricing
Architecture
NVIDIA Blackwell
GPU
8x NVIDIA Blackwell GPUs
GPU memory
Up to 1.5 TB total (192 GB per GPU), 64 TB/s HBM3e aggregate bandwidth (Note: 8 TB/s per GPU)
FP4 tensor core performance
112 petaFLOPS
FP8 tensor core performance
56 petaFLOPS
NVIDIA NVSwitch
2x
NVIDIA NVLink bandwidth
14.4 TB/s aggregate bandwidth
System power usage
~10kW max
CPU
2x Intel Xeon Platinum 8570 Processors or AMD EPYC equivalents
System memory
2 TB, configurable to 4 TB
Networking
Up to 8x OSFP ports serving NVIDIA ConnectX-7 or ConnectX-8 SuperNICs (Up to 400 Gb/s or 800 Gb/s depending on partner configuration). 2x dual-port NVIDIA BlueField-3 DPUs.
Management network
1GbE onboard network interface card (NIC) with RJ45 1GbE RJ45 host baseboard management controller (BMC)
Storage
OS: 2x 1.9 TB NVMe M.2, internal storage: 8x 3.84 TB NVMe U.2
Software
NVIDIA AI Enterprise, NVIDIA Base Command, NVIDIA DGX OS / Ubuntu
Rack units (RU)
8 - 10
Operating temperature
10-35°C / 50-90°F
Ideal uses cases for the NVIDIA B100 GPU
Explore uses cases for the NVIDIA B100 including Frontier model training, Production inference at scale, Scientific computing and HPC, Sovereign and regulated AI.
AI inference
AI developers can utilize the NVIDIA B100 to accelerate AI inference workloads, such as image and speech recognition, at lightning speed. The B100 GPU’s powerful Tensor Cores enable it to quickly process large amounts of data, making it perfect for real-time inference applications.
Deep learning
The NVIDIA B100 will empower data scientists and researchers to achieve groundbreaking milestones in deep learning. Its massive memory and processing power guarantee significantly reduced training and deployment times for complex, large-scale models and enables model training on significantly larger datasets.
High-performance computing
From complex scientific simulations to weather forecasting and intricate financial modeling, the B100 will empower diverse organizations to accelerate high-performance computing tasks. Its unmatched memory bandwidth and processing capabilities ensure smooth operation for workloads of any scale, allowing you to achieve unmatched results faster than ever.
Blog
Resources
- Emmanuel Ohiri
Resources
- Emmanuel Ohiri
Resources
- Emmanuel Ohiri
Resources
- Emmanuel Ohiri
Resources
- Emmanuel Ohiri
Resources
- Emmanuel Ohiri
Browse alternative GPU solutions for your workloads
Access a wide range of performant NVIDIA and AMD GPUs to accelerate your AI, ML & HPC workloads
NVIDIA H100 PCIe
Price on request
Scale with high performance H100 GPUs on our reserved cloud.