NVIDIA HGX H200

Accelerate Generative AI Training and Inference Workloads

Gain Access to the NVIDIA HGX H200

The NVIDIA H200 Tensor Core GPU supercharges generative AI and HPC with game-changing performance and memory capabilities. As the first GPU with HBM3e, H200’s faster, larger memory fuels the acceleration of generative AI and LLMs while advancing scientific computing for HPC workloads.

The NVIDIA HGX H200 is now available in the Cirrascale AI Innovation Cloud. Experience the highest performance in generative AI and HPC applications.

Reserve Now

Authorized NVIDIA Cloud Service Provider

Cirrascale Offers the HGX 200 for Breakthrough Performance

Cirrascale offers the HGX H200 in its AI Innovation Cloud as an 8-GPU configuration giving you full GPU-to-GPU bandwidth through NVIDIA NVSwitch. Leveraging the power of H200 multi-precision TensorCores, an eight-way HGX H200 provides over 32 petaFLOPS of FP8 deep learning compute and over 1.1TB of aggregate HBM memory for the highest performance in generative AI and HPC applications.

The NVIDIA H200 is the world’s first GPU with HBM3e memory with 4.8TB/s of memory bandwidth, a 1.4X increase over H100. H200 also expands GPU memory capacity nearly 2X to 141 gigabytes (GB). The combination of faster and larger HBM memory accelerates performance of computationally intensive generative AI and HPC applications, while meeting the evolving demands of growing model sizes.

Industry-Leading Generative AI Training

The era of generative AI has arrived, and it requires billion-parameter models to take on the paradigm shift in business operations and customer experiences.

NVIDIA H200 GPUs feature the Transformer Engine with FP8 precision, which provides up to 5X faster training over A100 GPUs for large language models such as GPT-3 175B. The combination of fourth-generation NVlink, which offers 900GB/s of GPU-to-GPU interconnect, PCIe Gen5, and NVIDIA Magnum IO™ software, delivers efficient scalability from small enterprise to massive unified computing clusters of GPUs. These infrastructure advances, working in tandem with the NVIDIA AI Enterprise software suite, make the NVIDIA H200 the most powerful end-to-end generative AI and HPC data center platform.

Projected performance, subject to change. LLM LoRA fine-tuning : 64 GPUs | H200 vs. A100 | GPT3-175B.

Unlock Insights with High-Performance LLM Inference

In the ever-evolving landscape of AI, businesses rely on large language models to address a diverse range of inference needs. An AI inference accelerator must deliver the highest throughput at the lowest TCO when deployed at scale for a massive user base.

H200 doubles inference performance compared to H100 when handling LLMs such as Llama2 70B.

Preliminary measured performance, subject to change.
GPT-3 175B: ISL 80, OSL 200 | x8 H100 GPUs BS 64 | x8 H200 GPUs BS 128
Llama2 70B: ISL 2K, OSL 128 | Throughput | H100 1x GPU BS 8 | H200 1x GPU BS 32

Why Cirrascale?

We're proud of the fact that we have worked with cloud pioneers from the very start. We were the trusted cloud backbone that helped OpenAI meet their cloud compute needs early on, and we continue to engage with today's bleeding edge AI companies, like yours.

Access to the Latest AI Accelerators

The Cirrascale AI Innovation Cloud is the only cloud service where you can test and deploy on every leading AI Accelerator in one cloud.

Specialized Cloud and Managed Services

Work with us to tailor the right solution for you with our wide range of system configurations, optimized for your specific workload requirements.

Transparent, Budget Friendly Pricing

With our no-surprises billing, long-term discounts, and no data transfer fees; Cirrascale offers unmatched pricing that’s built around your needs.

Request More Information

Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

Pricing

NVIDIA H200 Instance Pricing

Instance
Processor Specs
System RAM
Local Storage
Network
Monthly Pricing
6-Month Pricing
Annual Pricing
8-GPU
NVIDIA H200
Dual 48-Core
2TB
960 NVMe
(4) 3.84TB NVMe
25Gb Bonded
(3200Gb Available)
$26,499
$23,849
$21,199
Cirrascale Cloud Services has one of the largest selections of NVIDIA GPUs available in the cloud.
The above represents our most popular instances, but check out our pricing page for more instance types.
Not seeing what you need? Contact us for a specialized cloud quote for the configuration you need.
OAM
4X AMD Instinct MI250
Processor Specs
Dual 64-Core
System RAM
1TB
Local Storage
1TB(1) 960 NVMe
(1) 3.84TB NVMe
Network
25Gb Bonded
Monthly Pricing
$4,679
6-Month Pricing
$4,211
Annual Pricing
$3,743
All pricing above is based on Cirrascale's No Surprises billing model. There are no hidden fees and discounts may apply for long-term commitments depending on the service requested. All pricing shown for servers are per server per month.

Ready To Get Started?

Ready to take advantage of our flat-rate monthly billing, no ingress/egress data fees, and fast multi-tiered storage?

Get Started