Skip to main content

Supported Cloud Instances

Learn about the instance types we support


Jump to a cloud provider:


We offer a range of instance types designed to handle a variety of machine learning workloads. These cloud instances vary in their CPU, RAM (Random Access Memory), and GPU configurations, which allow you to orchestrate the right balance of performance and cost for your use case.

The instances listed on this page are available on demand — they can be provisioned immediately when you create a nodepool. Clarifai also offers the latest NVIDIA, AMD, and Google accelerators (including B300, B200, H100, MI355X, and TPU v7) that require reservations and are available on request.

info
  • See the pricing page to learn more about pricing for each instance type.

  • Contact us to access GPU types not listed here, including instances that require dedicated provisioning.

Amazon Web Services (AWS) Instances

G7E Instances

The AWS G7E series provides the latest generation of high-performance NVIDIA RTX PRO 6000 GPUs, scaling from single-GPU setups to large multi-GPU clusters. These instances are designed for the most demanding deep learning training and large-scale inference workloads.

Instance TypeGPUGPU MEMORYCPUPrice/hr
g7e.48xlarge8x NVIDIA-RTX-PRO-6000 (96Gi)768Gi191.1 cores (1,887.96Gi)$33.16
g7e.24xlarge4x NVIDIA-RTX-PRO-6000 (96Gi)384Gi95.3 cores (939.79Gi)$16.52
g7e.12xlarge2x NVIDIA-RTX-PRO-6000 (96Gi)192Gi47.4 cores (466.19Gi)$10.37
g7e.8xlarge1x NVIDIA-RTX-PRO-6000 (96Gi)96Gi31.5 cores (230.74Gi)$5.26
g7e.4xlarge1x NVIDIA-RTX-PRO-6000 (96Gi)96Gi15.5 cores (112.34Gi)$5.00

Key Features

  • NVIDIA RTX PRO 6000 GPU — 96 GiB of GPU memory per card, delivering high memory capacity for very large models and long-context inference.

  • Scalable multi-GPU configurations — From a single GPU up to 8 GPUs (768 GiB combined GPU memory), suitable for distributed training of frontier models.

  • High CPU & RAM — Up to 191 cores and ~1.9 TiB RAM in the 8-GPU configuration, ensuring data pipelines match GPU throughput.

Example Use Cases

  • Fine-tuning or running inference on large language models that require more than 48 GiB of GPU memory per card (e.g., 70B+ parameter models on a single GPU).
  • Multi-GPU distributed training of foundation models where high per-GPU memory is critical.
Looking for the latest GPUs or TPUs?

The newest accelerators — including NVIDIA B300, B200, H100, AMD MI355X, MI300X, and Google TPU v7 — are available with reservations and are not automatically provisioned on demand. Contact us to discuss availability and reserve capacity.

G6 Instances

The AWS G6 series introduces next-generation NVIDIA GPUs for the most demanding machine learning and simulation workloads. These instances scale from single-GPU mid-tier setups to multi-GPU, high-memory configurations capable of handling large-scale model training.

Instance TypeGPUGPU MEMORYCPUPrice/hr
g6e.12xlarge4x NVIDIA-L40S (44.99Gi)179.95Gi47.4 cores (351.44Gi)$13.10
g6e.2xlarge1x NVIDIA-L40S (44.99Gi)44.99Gi7.5 cores (57.95Gi)$2.81
g6e.xlarge1x NVIDIA-L40S (44.99Gi)44.99Gi3.5 cores (28.35Gi)Contact Us
g6.2xlarge1x NVIDIA-L4 (22.49Gi)22.49Gi7.5 cores (28.35Gi)$1.22
g6.xlarge1x NVIDIA-L4 (22.49Gi)22.49Gi3.5 cores (13.55Gi)$1.01

Key Features

  • Next-Gen GPUs — NVIDIA L4 GPUs target efficient inference and fine-tuning, while L40S GPUs deliver high throughput for large-scale training.

  • Scalable GPU memory — From 22.49 GiB (L4) to nearly 180 GiB (multi-L40S), supporting workloads from mid-sized tasks to multi-modal foundation models.

  • High vCPU & RAM options — Up to 47.4 cores and 351 GiB RAM in g6e.12xlarge, enabling massive parallelism and data-heavy preprocessing.

  • Flexible tiers — Ranges from cost-efficient single-GPU instances to powerful multi-GPU setups.

Example Use Cases

  • G6 (L4 instances) support mid-tier workloads such as fine-tuning BERT-large, or computer vision tasks like text-to-image generation and object recognition.

  • G6e (L40S instances) support advanced training workloads, including large-scale language models (e.g., GPT-4, T5-XL) or multi-modal tasks requiring both vision and language.

G5 Instances

The AWS G5 series provides high-performance GPU capabilities for workloads that demand more memory and compute power. These instances are optimized for deep learning training, large-scale inference, and advanced computer vision tasks.

Instance TypeGPUGPU MEMORYCPUPrice/hr
g5.2xlarge1x NVIDIA-A10G (22.49Gi)22.49Gi7.5 cores (28.35Gi)$1.51
g5.xlarge1x NVIDIA-A10G (22.49Gi)22.49Gi3.5 cores (13.55Gi)$1.26

Key Features

  • NVIDIA A10G GPU — High compute throughput and memory bandwidth, enabling faster training for deep learning and support for more complex models compared to T4 GPUs.

  • Scalable CPU & memory — From ~3.5 to ~7.5 vCPUs and 13.55 to 28.35 GiB of RAM, supporting data-heavy preprocessing, augmentation, and orchestration alongside GPU tasks.

  • Balanced design — Efficient for both training and inference, bridging the gap between lightweight GPU instances (like G4dn) and specialized multi-GPU clusters.

Example Use Cases

  • Training mid-sized NLP models like GPT-2 or T5 for text generation, or training image segmentation models like UNet or Mask R-CNN for medical imaging.

  • Running object tracking, pose estimation, or other GPU-accelerated pipelines for video analytics.

G4DN Instances

The AWS G4dn series is built for GPU-accelerated workloads at a moderate scale. These instances combine NVIDIA T4 GPUs with balanced CPU and memory resources, making them well-suited for small-to-medium machine learning and inference tasks.

Instance TypeGPUGPU MEMORYCPUPrice/hr
g4dn.xlarge1x NVIDIA-T4 (15Gi)15Gi3.5 cores (13.86Gi)$0.65

Key Features

  • NVIDIA T4 GPU — Designed for inference and light training, offering strong efficiency for workloads at a lower cost compared to heavier GPU families.

  • vCPUs and RAM — Provides ~3.5 vCPUs (baseline) and ~13.86 GiB memory, giving enough capacity to manage GPU-accelerated tasks, preprocessing, and orchestration.

  • Balanced performance — Ideal when you need GPU acceleration without the overhead of large, expensive GPU instances.

Example Use Cases

  • Inference workloads, such as running NLP models such as BERT-base for summarization, classification, or question answering.

  • Light training smaller models or experimenting with prototypes before scaling to larger GPU families.

T3A Instances

The AWS T3A series is intended for cost‑effective, general‑purpose workloads that do not require GPU acceleration. It provides a balanced mix of CPU and memory, making it suitable for lightweight use cases.

Instance TypeGPUGPU MEMORYCPUPrice/hr
t3a.2xlarge--7.5 cores (28.35Gi)$0.36
t3a.xlarge--3.5 cores (13.55Gi)$0.18
t3a.large--1.5 cores (6.4Gi)$0.11
t3a.medium--1.5 cores (2.89Gi)$0.07

Key Features

  • vCPU (virtual CPUs) performance — Burstable performance that adapts to workload spikes. For example, t3a.medium provides ~1.5 vCPUs, while t3a.2xlarge scales up to ~7.5 vCPUs.

  • Memory — Ranges from 2.89 GiB to 28.35 GiB, enabling efficient in-memory data handling for lightweight to moderately intensive workloads.

  • Efficiency — Optimized for cost savings compared to other instance families, making them budget-friendly for everyday use.

Example Use Case

  • Running simple models such as for classification or regression tasks.

Note: The CPU values (e.g., 1.5 cores) are baseline vCPU allocations expressed as fractional units. The instance can burst up to its full vCPU count (e.g., 2 vCPUs for t3a.medium) by consuming CPU credits.

T4G Instances

The AWS T4G series provides ARM-based (AWS Graviton2) CPU-only instances for cost-effective general-purpose workloads. These instances offer a strong performance-per-dollar ratio for lightweight inference and preprocessing tasks that do not require GPU acceleration.

Instance TypeGPUGPU MEMORYCPUPrice/hr
t4g.xlarge--3.5 cores (13.55Gi)Contact Us
t4g.large--1.5 cores (6.4Gi)Contact Us
t4g.medium--1.5 cores (2.89Gi)Contact Us

Key Features

  • AWS Graviton2 ARM processors — Deliver up to 40% better price performance compared to comparable x86 instances, making them cost-efficient for CPU-bound workloads.

  • Memory — Ranges from 2.89 GiB to 13.55 GiB, suitable for lightweight model serving and data processing.

  • Cost efficiency — A budget-friendly alternative to T3A for workloads that are compatible with ARM64 architecture.

Example Use Case

  • Running lightweight models or preprocessing pipelines where ARM compatibility is confirmed and cost savings are a priority.

Note: Ensure your model container is built for linux/arm64 when using T4G instances.

Google Cloud Platform (GCP) Instances

G4 Instances

The GCP G4 series features NVIDIA RTX PRO 6000 GPUs, offering 96 GiB of GPU memory per card — the highest per-GPU memory available in the GCP lineup. These instances scale from a single GPU to 8-GPU configurations and are well-suited for large model inference and training workloads that demand high GPU memory capacity.

Instance TypeGPUGPU MEMORYCPUPrice/hr
g4-standard-3848x NVIDIA-RTX-PRO-6000 (96Gi)768Gi382.4 cores (1,403.83Gi)$45.00
g4-standard-1924x NVIDIA-RTX-PRO-6000 (96Gi)384Gi190.9 cores (698.23Gi)$22.50
g4-standard-962x NVIDIA-RTX-PRO-6000 (96Gi)192Gi95.1 cores (345.43Gi)$11.27
g4-standard-481x NVIDIA-RTX-PRO-6000 (96Gi)96Gi47.2 cores (169.03Gi)$5.65

Key Features

  • NVIDIA RTX PRO 6000 GPU — 96 GiB HBM per card, ideal for workloads requiring more GPU memory than H100/A100 instances provide.

  • Multi-GPU scaling — From 1 to 8 GPUs (up to 768 GiB combined GPU memory), supporting distributed inference and training at scale.

  • High CPU & RAM — Up to 382 cores and ~1.4 TiB RAM in the 8-GPU variant.

Example Use Cases

  • Inference for very large language models (70B–400B+ parameters) where maximizing GPU memory per card reduces the need for model parallelism.
  • Multi-GPU distributed training where per-GPU memory is the primary bottleneck.

A2 & A3 High-Performance Instances

The A2 and A3 series are GCP's flagship high-performance GPU instances, designed for large-scale deep learning, high-performance inference, and real-time AI workloads. With NVIDIA A100 and H100 GPUs, they scale from single-GPU setups to multi-GPU powerhouse configurations capable of training foundation models.

Instance TypeGPUGPU MEMORYCPUPrice/hr
a3-highgpu-8g8x NVIDIA-H100 (79.65Gi)637.18Gi206.8 cores (1,827.19Gi)$110.63
a3-highgpu-1g1x NVIDIA-H100 (79.65Gi)79.65Gi25.3 cores (221.95Gi)$13.82
a2-ultragpu-1g1x NVIDIA-A100 (80Gi)80Gi11.3 cores (159.23Gi)$7.13

Key Features

  • Next-generation GPUs — A100 (80 GiB) excels at large-scale training with strong throughput and memory bandwidth. H100 (80 GiB) delivers significant improvements for transformer-based models, enabling faster training and inference. Multi-GPU (a3-highgpu-8g) configurations scale this power dramatically.

  • Massive CPU & RAM scaling — From 11.3 cores / 159 GiB RAM in a2-ultragpu-1g to 206.8 cores / 1.8 TiB RAM in a3-highgpu-8g, ensuring parallel data pipelines can keep pace with GPU compute.

  • Flexible tiers — Options for single-GPU tasks or multi-GPU clusters, matching workloads of different scales and budgets.

Example Use Cases

  • Single-GPU (A2 / A3-1g) can be used for training or fine-tuning mid-to-large language models (e.g., GPT-3, T5-XL) or advanced vision models.

  • Multi-GPU (A3-8g) can be used for training large-scale, next-generation foundation models (e.g., GPT-4, PaLM, multi-modal transformers) where scale and GPU memory aggregation are critical.

  • Deploying video analytics, autonomous systems, or robotics pipelines that demand real-time, ultra-low latency.

G2-Standard Instances

The GCP G2-Standard series provides GPU acceleration with NVIDIA L4 GPUs, designed for moderate machine learning and inference workloads. These instances scale from small setups to larger configurations, balancing cost with performance for small-to-medium tasks.

Instance TypeGPUGPU MEMORYCPUPrice/hr
g2-standard-321x NVIDIA-L4 (22.49Gi)22.49Gi31.3 cores (118.07Gi)$2.16
g2-standard-161x NVIDIA-L4 (22.49Gi)22.49Gi15.3 cores (57.75Gi)$1.44
g2-standard-121x NVIDIA-L4 (22.49Gi)22.49Gi11.3 cores (42.71Gi)$1.26
g2-standard-81x NVIDIA-L4 (22.49Gi)22.49Gi7.3 cores (27.67Gi)$1.08
g2-standard-41x NVIDIA-L4 (22.49Gi)22.49Gi3.4 cores (12.63Gi)$0.90

Key Features

  • NVIDIA L4 GPUs — Optimized for inference and light training, delivering strong efficiency for vision and NLP tasks at lower cost compared to heavier GPU families.

  • CPU & memory scaling — From ~3.4 to ~31.3 cores and 12.63 GiB to 118.07 GiB RAM, allowing smooth orchestration of preprocessing, data loading, and GPU-bound tasks.

  • Cost-performance balance — A versatile option for teams that need GPU acceleration without the expense of A100/H100-based instances.

Example Use Cases

  • Running transformer-based models like BERT-base for summarization, classification, or Q&A.
  • Fine-tuning smaller computer vision models for object detection or image classification.

TPU v6e & v7 High-Performance Instances

Google's TPU v6e and v7 instances represent the latest generations of purpose-built AI accelerators, offering significant performance improvements over v5 for large-scale training and inference.

Instance TypeAcceleratorsAccelerator MemoryCPUPrice/hr
tpu7x-standard-4t4x GOOGLE-TPU-v7-222.8 cores (933.43Gi)Contact Us
ct6e-standard-4t4x GOOGLE-TPU-v6e-178.9 cores (698.23Gi)Contact Us
tpu7x-standard-1t1x GOOGLE-TPU-v7-55.2 cores (227.83Gi)Contact Us
ct6e-standard-1t1x GOOGLE-TPU-v6e-43.3 cores (165.11Gi)Contact Us

Key Features

  • TPU v6e (Trillium) — Google's sixth-generation TPU, delivering ~4x the compute performance per chip compared to v5e. Optimized for both training and serving large transformer models.

  • TPU v7 — Google's latest-generation TPU, designed for frontier-scale training workloads with improved interconnect bandwidth and compute throughput over v6e.

  • No exposed accelerator memory — TPU memory is managed by the TPU runtime rather than exposed as a configurable VRAM value.

  • High CPU & RAM — From 43 cores / 165 GiB RAM (v6e 1-chip) to 222 cores / 933 GiB RAM (v7 4-chip), supporting large data pipelines alongside TPU compute.

Example Use Cases

  • TPU v6e can be used for cost-efficient, high-throughput serving of large language models (e.g., Gemma, LLaMA) and training of mid-to-large scale transformer architectures.
  • TPU v7 can be used for training frontier models at scale where maximum TPU throughput and interconnect performance are required.

TPU v5e & v5p High-Performance Instances

Google's cloud TPU v5e and v5p instances are purpose-built accelerators optimized for deep learning training and inference. Unlike GPUs, TPUs (Tensor Processing Units) are specialized for matrix-heavy tensor operations, making them ideal for transformer-based models and large-scale distributed training.

Instance TypeAcceleratorsAccelerator MemoryCPUPrice/hr
ct5p-hightpu-4t4x GOOGLE-TPU-v5p-206.8 cores (431.67Gi)Contact Us
ct5lp-hightpu-4t4x GOOGLE-TPU-v5e-111.1 cores (180.79Gi)Contact Us
ct5lp-hightpu-1t1x GOOGLE-TPU-v5e-23.3 cores (42.71Gi)Contact Us

Key Features

  • Specialized Tensor Processing Units (TPUs) — TPU v5e provides a balanced design optimized for cost-efficiency in large-scale training and inference, which is great for productionizing ML workloads where throughput matters. TPU v5p provides higher-performance generation with faster interconnects and larger scaling potential, designed for frontier model training.

  • Scalable CPU & memory — From 23.3 cores / 42.71 GiB RAM in the 1-core v5e instance to 206.8 cores / 431.67 GiB RAM in the 4-core v5p, ensuring sufficient orchestration power for massive training workloads.

  • No exposed GPU memory — TPU memory is not presented like GPU VRAM but is instead managed by the TPU runtime for high-efficiency tensor operations.

Example Use Cases

  • TPU v5e (1t, 4t) can be used for cost-efficient training of language models (e.g., BERT, T5-small/XL) and vision transformers (ViT).
  • TPU v5p (4t) can be used for training large foundation models such as PaLM, Gemini-like multi-modal architectures, or massive LLMs where performance and throughput at scale are critical.

N2-Standard Instances

The GCP N2-Standard series offers cost-effective, general-purpose compute for workloads that don't require GPU acceleration. These instances balance CPU and memory, making them well-suited for lightweight applications, preprocessing, and small-scale model deployment.

Instance TypeGPUGPU MEMORYCPUPrice/hr
n2-standard-16--15.3 cores (57.75Gi)$1.08
n2-standard-8--7.3 cores (27.67Gi)$0.54
n2-standard-4--3.4 cores (12.63Gi)$0.29
n2-standard-2--1.4 cores (5.42Gi)$0.14

Key Features

  • vCPUs — Baseline performance scales from ~1.4 to ~15.3 cores, with the ability to burst to the full allocation (2 to 16 vCPUs). Optimized for CPU-intensive tasks, such as running traditional models.

  • Memory (RAM) — From 5.42 GiB to 57.75 GiB, supporting in-memory data handling for lightweight to moderately intensive workloads.

  • Cost efficiency — Designed to deliver consistent performance at a lower cost, ideal for everyday compute tasks without GPU requirements.

Example Use Case

  • Running small-scale machine learning models or serving simple inference workloads.

Vultr Cloud Servers Instances

GH200 & MI300X High-Performance GPU Instances

Vultr offers high-performance GPU instances powered by NVIDIA and AMD accelerators. These instances are built for AI training, inference, and HPC (high-performance computing) workloads that demand extreme compute and memory bandwidth.

Instance TypeGPUGPU MEMORYCPUPrice/hr
vbm-256c-2048gb-8-mi300x-gpu8x AMD-MI300X (127.82Gi)1,022.53Gi255.4 cores (1,945.34Gi)$39.92
vbm-72c-480gb-gh200-gpu1x NVIDIA-GH200 (95.58Gi)95.58Gi71.4 cores (455.74Gi)$2.99

Key Features

  • NVIDIA GH200 superchip — Combines Hopper GPU architecture with Grace CPU integration, delivering ultra-fast memory bandwidth and low-latency compute, ideal for training massive AI models and real-time inference.

  • AMD MI300X GPUs — Designed for frontier AI workloads, offering huge HBM3 memory (128 GiB per GPU) and scaling efficiency with 8 GPUs per instance. Excellent for distributed training of very large models.

  • High CPU & RAM configurations — From 71 cores / 455 GiB RAM (GH200) up to 255 cores / 1.9 TiB RAM (MI300X cluster), ensuring orchestration and preprocessing don't bottleneck GPU performance.

Example Use Cases

  • NVIDIA GH200 (single GPU instance) can be used for training and inference for large language models (e.g., LLaMA 2–70B, GPT-3 scale). It can also be used for real-time multi-modal use cases requiring tight GPU-CPU integration, such as speech-to-speech AI assistants or interactive robotics.

  • AMD MI300X (8-GPU cluster instance) can be used for training frontier LLMs and multi-modal models (GPT-4, Gemini-class, or open LLMs at >100B parameters).

VCG Instances

The Vultr VCG series provides instances with dedicated NVIDIA GPUs, enabling acceleration for deep learning, inference, and GPU-intensive applications. These instances scale from entry-level GPU setups to multi-GPU clusters, making them versatile for workloads ranging from experimentation to frontier AI model training.

Instance TypeGPUGPU MEMORYCPUPrice/hr
vcg-b200-248c-2826g-1536vram8x NVIDIA-B200 (179.06Gi)1,432.49Gi255.4 cores (1,945.34Gi)$79.92
vcg-a100-12c-120g-80vram1x NVIDIA-A100 (80Gi)80Gi11.4 cores (113.74Gi)$2.99
vcg-l40s-16c-180g-48vram1x NVIDIA-L40S (44.99Gi)44.99Gi15.4 cores (170.74Gi)$2.09
vcg-a16-6c-64g-16vram1x NVIDIA-A16 (16Gi)16Gi5.4 cores (60.54Gi)$0.58

Key Features

  • Range of NVIDIA GPUs — From the A16 (lightweight inference) to the A100 (high-performance training), L40S (next-gen accelerated workloads), and B200 clusters (frontier-scale AI with 8 GPUs).

  • High vCPU and RAM configurations — Scales from 5.4 cores / 60 GiB RAM in entry-level instances to 255 cores / 1.9 TiB RAM in multi-GPU setups, ensuring GPU workloads are matched with sufficient CPU and memory.

  • Scalable GPU memory — Ranges from 16 GiB (A16) for smaller tasks up to 1.4 TiB (8 × B200) for extreme AI training.

Example Use Cases

  • High-performance training and inference for large-scale deep learning models.
  • Running AI inference workloads with optimized GPU acceleration.

VC2 Instances

The Vultr VC2 series provides general-purpose compute instances optimized for workloads that do not require GPU acceleration. With a balance of CPU and memory, these instances are best suited for lightweight use cases.

Instance TypeGPUGPU MEMORYCPUPrice/hr
vc2-24c-96gb--23.4 cores (90.94Gi)$1.08
vc2-16c-64gb--15.4 cores (60.54Gi)$0.54
vc2-8c-32gb--7.4 cores (30.14Gi)$0.29
vc2-6c-16gb--5.4 cores (14.94Gi)$0.14
vc2-4c-8gb--3.4 cores (7.34Gi)$0.07
vc2-2c-4gb--1.4 cores (3.54Gi)$0.04

Key Features

  • Scalable CPU and RAM — Configurations range from 1.4 cores / 3.54 GiB RAM (vc2-2c-4gb) up to 23.4 cores / 90.94 GiB RAM (vc2-24c-96gb).

  • Cost-effective — Optimized for environments where GPU acceleration is unnecessary, making them a good fit for traditional compute workloads.

  • Flexibility — Suitable for a broad range of general-purpose tasks, with instance sizes that scale from small testing environments to larger services.

Example Use Cases

  • Suitable for lightweight applications as well as development and testing environments.

Oracle Distributed Cloud Instances

AMD-MI300X GPU Instances

The Oracle MI300X family provides cutting-edge AMD Instinct MI300X GPUs, purpose-built for large-scale AI training and HPC (high-performance computing) workloads. These instances feature massive GPU memory and CPU scaling, supporting frontier use cases.

Instance TypeGPUGPU MemoryCPUPrice/hr
BM.GPU.MI300X.88 × AMD-MI300X (127.82Gi)1,022.53Gi111.5 cores (1,945.01Gi)Contact Us

Key Features

  • 8 × AMD MI300X GPUs, each with 128 GiB HBM3 memory (over 1 TiB GPU memory total).
  • Extremely high CPU scaling, 111.5 cores, and nearly 2 TiB of system RAM.
  • Suited for frontier-scale AI training and supercomputing-class workloads.

Example Use Cases

  • Training LLMs and multi-modal models (>100B parameters).
  • High-performance computing simulations, such as weather forecasting and scientific modeling.

NVIDIA-A10G GPU Instances

The Oracle A10G series provides NVIDIA GPUs optimized for inference, visualization, and moderate ML training tasks. These instances combine consistent vCPU allocations with scalable GPU configurations, from single-GPU VMs to multi-GPU bare metal nodes.

Instance TypeGPUGPU MemoryCPUPrice/hr
BM.GPU.A10.44 × NVIDIA-A10G (22.49Gi)89.95Gi14.8 cores (227.41Gi)Contact Us
VM.GPU.A10.11 × NVIDIA-A10G (22.49Gi)22.49Gi14.8 cores (227.41Gi)Contact Us

Key Features

  • NVIDIA A10G GPUs with 22.49 GiB memory each.
  • Same CPU allocation (14.8 cores / 227 GiB RAM) across single- and four-GPU configurations.

Example Use Cases

  • VM.A10.1 (single GPU) can be used for small-scale ML inference, 3D rendering, and graphics-heavy applications.
  • BM.A10.4 (four GPUs) can be used for larger inference workloads, distributed graphics rendering, or multi-GPU training of medium-sized models.

Standard CPU-Only Instances

The Oracle E6 Flex series provides CPU-only instances for workloads that do not require GPU acceleration.

Instance TypeGPUGPU MemoryCPUPrice/hr
VM.Standard.E6.Flex0.8 cores (14.61Gi)Contact Us

Key Features

  • No GPU, pure vCPU, and memory resources.
  • 0.8 cores and 14.61 GiB RAM, suitable for supporting tasks.
  • Cost-efficient for non-GPU workloads.

Example Use Cases

  • Running control-plane services or lightweight applications alongside GPU clusters.
  • Workloads requiring basic compute and memory without acceleration.