AMD Instinct Series Cloud

AMD Instinct™ Series

AMD Instinct Series accelerators, delivered in the Cirrascale AI Innovation Cloud, enable performance leadership that is uniquely well-suited to power even the most demanding AI and HPC workloads.

We've partnered with AMD to offer their AMD Instinct Series accelerators in the cloud for customers to test, utilize and fully deploy. These accelerators provide exceptional compute performance, large memory density, high bandwidth memory, and support for specialized data formats. AMD Instinct accelerators are built on AMD CDNA™ architecture, which features Matrix Core Technologies and supports a broad range of precision capabilities.

No items found.

AMD Benefits

AMD ROCm Software

AMD ROCm™ is an open software stack including drivers, development tools, and APIs that enable GPU programming from low-level kernel to end-user applications. ROCm is optimized for Generative AI and HPC applications, and is easy to migrate existing code into.

ROCm enables AI and HPC application development across a broad range of demanding workloads.

AMD Infinity Fabric Technology

Cirrascale-hosted AMD Instinct series accelerators with advanced peer- to-peer I/O connectivity through a maximum of eight AMD Infinity Fabric™ links deliver up to 800 GB/s I/O bandwidth performance. With a cache-coherent solution using optimized AMD EPYC™ CPUs and Instinct accelerators, Infinity Fabric unlocks the promise of unified computing, enabling a quick and simple on-ramp for CPU code to accelerated platforms.

No items found.

Discover the Benefits of AMD Instinct hosted by Cirracale

Optimal Performance at the Right Price

  • High performance with larger memory than other acceleration offerings
  • Optimized for leading Generative AI models, including LLMs

Ease of Use

  • AMD drivers pre-installed and configured by Cirrascale
  • ROCm software included for easy access to frameworks and tools
  • Hugging Face transformers supported out of the box

Ease of Use

  • Highly scalable for the most demanding training, tuning and inference workloads

Simple & Secure Cloud Operations

  • Simple onboarding – No DevOps required
  • SDKs, storage and network are configured and ready to go

AMD Products

AMD Instinct MI300X

AMD Instinct™ MI300X accelerators are uniquely well-suited to power even the most demanding AI and HPC workloads, offering exceptional compute performance, large memory density, high bandwidth memory, and support for specialized data formats.

AMD Instinct MI300X accelerators are built on AMD CDNA™ 3 architecture, which offers Matrix Core Technologies and support for a broad range of precision capabilities—from the highly efficient INT8 and FP8 (including sparsity support for AI) to the most demanding FP64 for HPC.

AMD Instinct MI250

The AMD Instinct MI250 accelerator brings customers the compute engine selected for the first U.S. Exascale supercomputer.

AMD Instinct MI250 accelerators are built on AMD CDNA™ architecture, which offers Matrix Core Technologies and support for a broad range of precision capabilities—from the highly efficient INT8 and FP8 to the most demanding FP64 for HPC.

Pricing

AMD Instinct Series Instance Pricing

OAM
Processor Specs
System RAM
Local Storage
Network
Monthly Pricing
6-Month Pricing
Annual Pricing
8X AMD Instinct MI300X
Dual 48-Core
2.3TB
(1) 960 NVMe
(4) 3.84TB NVMe
25Gb Bonded
(3200Gb Available)
$22,499
$20,249
$17,999
4X AMD Instinct MI250
Dual 64-Core
1TB
(1) 960 NVMe
(1) 3.84TB NVMe
25Gb Bonded
$4,679
$4,211
$3,743
OAM
8X AMD Instinct MI300X
4X AMD Instinct MI250
Processor Specs
Dual 48-Core
Dual 64-Core
System RAM
2.3TB
1TB
Local Storage
(1) 960 NVMe
(4) 3.84TB NVMe
(1) 960 NVMe
(1) 3.84TB NVMe
Network
25Gb Bonded
(3200Gb Available)
25Gb Bonded
Monthly Pricing
$22,499
$4,679
6-Month Pricing
$20,249
$4,211
Annual Pricing
$17,999
$3,743
All pricing above is based on Cirrascale's No Surprises billing model. There are no hidden fees and discounts may apply for long-term commitments depending on the service requested. All pricing shown for servers are per server per month.

Pricing

NVIDIA GPU Cloud

OAM
Processor Specs
System RAM
Local Storage
Network
Monthly Pricing
6-Month Pricing
Annual Pricing
8-GPU
NVIDIA H200
Dual 48-Core
2TB
960 NVMe
(4) 3.84TB NVMe
25Gb Bonded
(3200Gb Available)
$26,499
$23,849
$21,199
8-GPU
NVIDIA H100
Dual 48-Core
2TB
(1) 960 NVMe
(4) 3.84TB NVMe
25Gb Bonded
(3200Gb Available)
$24,999
$22,499
$19,999

Cirrascale Cloud Services has one of the largest selections of NVIDIA GPUs available in the cloud.
The above represents our most popular instances, but check out our pricing page for more instance types.
Not seeing what you need? Contact us for a specialized cloud quote for the configuration you need.

OAM
8-GPU NVIDIA H200
8-GPU NVIDIA H100
Processor Specs
Dual 48-Core
Dual 48-Core
System RAM
2TB
2TB
Local Storage
(1) 960 NVMe
(4) 3.84TB NVMe
(1) 960 NVMe
(4) 3.84TB NVMe
Network
25Gb Bonded
(3200Gb Available)
25Gb Bonded
(3200Gb Available)
Monthly Pricing
$26,499
$24,999
6-Month Pricing
$23,849
$22,499
Annual Pricing
$21,199
$19,999
All pricing above is based on Cirrascale's No Surprises billing model. There are no hidden fees and discounts may apply for long-term commitments depending on the service requested. All pricing shown for servers are per server per month.

Pricing

Qualcomm Cloud AI 100 Series Pricing

Config
vCPUs
System RAM
Local Storage
Monthly Pricing
Annual Pricing
8X AI 100 Ultra
128
512GB
(2) 3.84TB NVMe
$4,699
$3,759
Octo AI 100 Pro
64
384GB
1TB NVMe
$2,499
$2,019
Quad AI 100 Pro
48
182GB
1TB NVMe
$1,259
$1,009
Dual AI 100 Pro
24
48GB
1TB NVMe
$629
$519
Single AI 100 Pro (128)
32
128GB
1TB NVMe
$549
$439
Single AI 100 Pro (64)
32
64GB
1TB NVMe
$369
$289
Single AI 100 Pro (48)
12
48GB
1TB NVMe
$329
$259
Config
8X AI 100 Ultra
Octo AI 100 Pro
Quad AI 100 Pro
Dual AI 100 Pro
Single AI 100 Pro (128)
Single AI 100 Pro (64)
Single AI 100 Pro (48)
vCPUs
128
64
48
24
32
32
12
System RAM
512GB
384GB
182GB
48GB
64GB
64GB
48GB
Local Storage
(2) 3.84TB NVMe
1TB NVMe
1TB NVMe
1TB NVMe
1TB NVMe
1TB NVMe
1TB NVMe
Monthly Pricing
$4,699
$2,499
$1,259
$629
$549
$369
$329
Annual Pricing
$3,759
$2,019
$1,009
$519
$439
$289
$259
All pricing above is based on Cirrascale's No Surprises billing model. There are no hidden fees and discounts may apply for long-term commitments depending on the service requested. All pricing shown for servers are per server per month.

Qualcomm branded products are products of Qualcomm Technologies, Inc. and/or its subsidiaries.

Pricing

The Cerebras AI Model Studio

Fine-Tuning - Standard Offering Pricing
Model
Parameters
Fine-tuning price per 1K tokens
Fine-tuning price per example (MSL 2048)
Fine-tuning price per example (MSL 4096)
Cerebras time to 10B tokens (h)**
Cerebras time to 10B tokens (h)**
Eleuther GPT-J
6
$0.00055
$0.0011
$0.0023
17
132
Eleuther GPT-NeoX
20
$0.00190
$0.0039
$0.0078
56
451
CodeGen* 350M
0.35
$0.00003
$0.00006
$0.00013
1
8
CodeGen* 2.7B
2.7
$0.00026
$0.0005
$0.0027
8
61
CodeGen* 6.1B
6.1
$0.00065
$0.0013
$0.0030
19
154
CodeGen* 16.1B
16.1
$0.00147
$0.0030
$0.011
44
350
Model
Eleuther GPT-J
Eleuther GPT-NeoX
CodeGen* 350M
CodeGen* 2.7B
CodeGen* 6.1B
CodeGen* 16.1B
Parameters
6
20
0.35
2.7
6.1
16.1
Fine-tuning price per 1K tokens
$0.00055
$0.00190
$0.00003
$0.00026
$0.00065
$0.00147
Fine-tuning price per example (MSL 2048)
$0.0011
$0.0039
$0.00006
$0.0005
$0.0013
$0.0030
Fine-tuning price per example (MSL 4096)
$0.0023
$0.0078
$0.00013
$0.0027
$0.0030
$0.011
Cerebras time to 10B tokens (h)**
17
56
1
8
19
44
AWS p4d (8xA100) time to 10B tokens (h)
132
451
8
61
154
350
* T5 tokens to train from the original T5 paper. Chinchilla scaling laws not applicable.

** Note that GPT-J was pre-trained on ~400B tokens. Fine-tuning jobs can employ a wide range of dataset sizes, but often use order 1-10% of the pre-training tokens. As such, one might fine-tune a model like GPT-J with ~4-40B tokens. We provide estimated wall clock time to fine-tune train the model checkpoints above with 10B tokens on Cerebras AI Model Studio and an AWS p4d instance in the table above to give you a sense of how much time jobs of this scale could take.
Fixed-Price Production Model Training
Model
Parameters
Tokens to Train to Chinchilla Point (B)
Cerebras AI Model Studio CS-2 Days to Train
Cerebras AI Model Studio Price to Train
GPT3-XL
1.3
26
0.4
$2,500
GPT-J
6
120
8
$45,000
GPT-3 6.7B
6.7
134
11
$40,000
T-5 11B
11
34*
9
$60,000
GPT-3 13B
13
260
39
$150,000
GPT NeoX
20
400
47
$525,000
GPT 70B
70
1,400
Contact For Quote
Contact For Quote
GPT 175B
175
3,500
Contact For Quote
Contact For Quote
Model
GPT3-XL
GPT-J
GPT-3 6.7B
T-5 11B
GPT-3 13B
GPT NeoX
GPT 70B
GPT 175B
Parameters
1.3
6
6.7
11
13
20
70
175
Tokens to Train to Chinchilla Point (B)
26
120
134
34*
260
400
1,400
3,500
Cerebras AI Model Studio CS-2 Days to Train
0.4
8
11
9
39
47
Contact For Quote
Contact For Quote
Cerebras AI Model Studio Price to Train
$2,500
$45,000
$40,000
$60,000
$150,000
$525,000
Contact For Quote
Contact For Quote
* T5 tokens to train from the original T5 paper. Chinchilla scaling laws not applicable.

** Expected number of days, based on training experience to date, using a 4-node Cerebras Wafer-Scale Cluster.  Actual training of model may take more or less time.

Ready To Get Started?

Ready to take advantage of our flat-rate monthly billing, no ingress/egress data fees, and fast multi-tiered storage?

Get Started