NVIDIA B100

The NVIDIA B100, powered by cutting-edge Blackwell architecture and boasting real-time inference to accelerate trillion-parameter large language models, is the world’s most powerful GPU for AI and HPC.

 

NVIDIA B100

Purpose-built B100 clusters, designed and managed by CUDO

Deployed across 16 ISO-certified data centres

From 8 to 1,000+ GPUs in a single deployment

NVIDIA Quantum-X800 InfiniBand or Spectrum-X Ethernet networking

Expert rack-level design, installation, and benchmarking before handoff

24/7 monitoring, management, and engineering support

Compatible with Slurm, Kubernetes, and NVIDIA Base Command

Available at the most cost-effective pricing

Launch your AI products faster with on-demand GPUs and a global network of data center partners

Bare metal

Complete control over a physical machine for more control.

Powered by renewable energy

No noisy neighbors

SpectrumX local networking

300Gbps external connectivity

NVMe SSD storage

Enterprise

We offer a range of solutions for enterprise customers.

Powerful GPU clusters

Scalable data center colocation

Large quantities of GPUs and hardware

Optimize to your requirements

Expert installation

Scale as your demand grows

Specifications

Browse specifications for the NVIDIA B200 GPU

Starting from

Contact us for pricing

Architecture

NVIDIA Blackwell

GPU

8x NVIDIA Blackwell GPUs

GPU memory

Up to 1.5 TB total (192 GB per GPU), 64 TB/s HBM3e aggregate bandwidth (Note: 8 TB/s per GPU)

FP4 tensor core performance

112 petaFLOPS

FP8 tensor core performance

56 petaFLOPS

NVIDIA NVSwitch

2x

NVIDIA NVLink bandwidth

14.4 TB/s aggregate bandwidth

System power usage

~10kW max

CPU

2x Intel Xeon Platinum 8570 Processors or AMD EPYC equivalents

System memory

2 TB, configurable to 4 TB

Networking

Up to 8x OSFP ports serving NVIDIA ConnectX-7 or ConnectX-8 SuperNICs (Up to 400 Gb/s or 800 Gb/s depending on partner configuration). 2x dual-port NVIDIA BlueField-3 DPUs.

Management network

1GbE onboard network interface card (NIC) with RJ45 1GbE RJ45 host baseboard management controller (BMC)

Storage

OS: 2x 1.9 TB NVMe M.2, internal storage: 8x 3.84 TB NVMe U.2

Software

NVIDIA AI Enterprise, NVIDIA Base Command, NVIDIA DGX OS / Ubuntu

Rack units (RU)

8 - 10

Operating temperature

10-35°C / 50-90°F

Ideal uses cases for the NVIDIA B100 GPU

Explore uses cases for the NVIDIA B100 including Frontier model training, Production inference at scale, Scientific computing and HPC, Sovereign and regulated AI.

AI inference

AI developers can utilize the NVIDIA B100 to accelerate AI inference workloads, such as image and speech recognition, at lightning speed. The B100 GPU’s powerful Tensor Cores enable it to quickly process large amounts of data, making it perfect for real-time inference applications.

Deep learning

The NVIDIA B100 will empower data scientists and researchers to achieve groundbreaking milestones in deep learning. Its massive memory and processing power guarantee significantly reduced training and deployment times for complex, large-scale models and enables model training on significantly larger datasets.

High-performance computing

From complex scientific simulations to weather forecasting and intricate financial modeling, the B100 will empower diverse organizations to accelerate high-performance computing tasks. Its unmatched memory bandwidth and processing capabilities ensure smooth operation for workloads of any scale, allowing you to achieve unmatched results faster than ever.

Blog

Browse alternative GPU solutions for your workloads

Access a wide range of performant NVIDIA and AMD GPUs to accelerate your AI, ML & HPC workloads

Discuss your infrastructure requirements

Scroll to Top