CUDO Compute Boosts GPU Fleet: More NVIDIA A40s, A6000s and V100s Now Available to Power AI and HPC Globally

In line with our commitment to meet the surging demand for GPUs for AI and HPC acceleration, we're excited to announce that we've added a fleet of on-demand GPUs to service customer needs globally.

3 min read

Pete Hill

Pete Hill

CUDO Compute Boosts GPU Fleet: More NVIDIA A40s, A6000s and V100s Now Available to Power AI and HPC Globally cover photo

In line with our commitment to meet the surging demand for GPUs for AI and HPC acceleration, we're excited to announce that we've added a fleet of on-demand GPUs to service customer needs globally. These additions ensure that our customers have immediate access to the most powerful, efficient, and versatile computing resources available anytime and affordably.

New GPU additions

NVIDIA A40 and A6000 GPUs are now available on-demand on our platform. Both GPUs are built on NVIDIA’s Ampere architecture, which is a significant improvement over its predecessors in terms of energy efficiency, computational power, and AI training and inference.

NVIDIA A40

The NVIDIA RTX A40 delivers 37.4 teraflops of FP32 (single-precision floating-point) performance, with 10,752 CUDA cores that facilitate parallel computing. Additionally, it includes 336 Tensor Cores designed for deep learning performance. It is also equipped with 48 GB of GDDR6 memory, offering a bandwidth of 696 GB/s. This vast memory pool and high bandwidth efficiently handle large datasets and complex AI models.

SpecificationNVIDIA RTX A40
FP32 Performance37.4 TFLOPS
CUDA Cores10,752
Tensor Cores336
Memory Type48 GB GDDR6
Memory Bandwidth696 GB/s

The RTX A40 is currently available on demand from $0.79 /hour. Configure Now.

NVIDIA A6000

The NVIDIA RTX A6000, on the other hand, is a versatile GPU that can be used for AI, HPC, and graphics-intensive applications, like video editing (VFX) and 3D rendering. Its Ampere architecture, 10,752 CUDA processing cores, and 48GB of GDDR6 memory provide a broad bandwidth of 768.0 GB/s, enabling it to handle large datasets and intensive AI workloads.

SpecificationNVIDIA RTX A6000
CUDA Cores10,752
Memory Type48 GB GDDR6
Memory Bandwidth768.0 GB/s
Ray Tracing Cores84 (Next-Generation)

The RTX A6000 is also equipped with 84 next-generation Ray Tracing (RT) cores, which makes it excellent for VFX and complex 3D models.

You can get started using the RTX A6000 now from $0.79 /hour. Start now.

NVIDIA V100

The NVIDIA Tesla V100 is designed with a strong emphasis on AI and deep learning applications. It delivers up to 14 TFLOPS of FP32 performance with 5,120 CUDA cores. Furthermore, it features 640 Tensor Cores, specifically engineered for accelerated deep learning performance. The V100 is also equipped with 16 GB or 32 GB of HBM2 memory, providing a bandwidth of up to 900 GB/s. This extensive memory capacity and superior bandwidth are adept at efficiently managing large datasets and complex AI models.

SpecificationNVIDIA Tesla V100
FP32 PerformanceUp to 14 TFLOPS
CUDA Cores5,120
Tensor Cores640
Memory32 GB HBM2
Memory BandwidthUp to 900 GB/s

These additions underscore our commitment to delivering cutting-edge technology that meets the needs of new and existing customers.

Get access to the NVIDIA V100 now from $0.49 /hour. Start now.

More A5000 GPUs coming soon

We are thrilled to announce that we’ll be adding more NVIDIA A5000 GPUs to our extensive roster. The A5000 is designed to tackle a wide array of professional graphics and compute tasks, ensuring versatility and reliability. These forthcoming additions will further boost our users' capabilities. Stay tuned!

US Locations Update

We're delighted to share that we're expanding our network with new strategic locations while bolstering current hubs to elevate our service, availability, and performance.

New Los Angeles Data Center Now Live!

Our new data center, us-losangeles-1, is now fully operational and strategically situated to bolster the capabilities of our US-based customers.

Equipped with 16x Tesla V100 GPUs and 10 dedicated CPU compute hosts, us-losangeles-1 guarantees robust computing resources primed for the most demanding tasks.

The LA data center brings several advantages to our users in the United States, including reduced latency and increased overall capacity to handle peak demand scenarios.

GPUs added to New York, Santa Clara, and Irvne Locations

Lastly, we have added more GPUs to our New York and Santa Clara locations which are now live. We will also be adding more GPUs to our Irvine location this month. This expansion ensures that our customers can access high-performance computing resources whenever and wherever they need them, all while maintaining the reliability and efficiency for which we are known.

Get in touch with us today!

Learn more about CUDO Compute: Website, LinkedIn, Twitter, YouTube, Get in touch.

Subscribe to our Newsletter

Subscribe to the CUDO Compute Newsletter to get the latest product news, updates and insights.