Skip to content

GPU Accelerator Configuration Guide

Enabling GPU Accelerators

GPU accelerator configuration enables you to specify dedicated GPU resources for your SaaS Products. This feature allows your services to leverage hardware acceleration for AI/ML workloads, high-performance computing, graphics processing, and other GPU-intensive tasks.

This guide covers two ways to use GPUs with your SaaS Products:

  1. External GPU attachment using acceleratorConfiguration (GCP N1 instances + Tesla GPUs only)
  2. Built-in GPU instances across AWS, GCP (G2, A2, A3, A4 families), and Azure (modern GPUs pre-installed)

External GPU Attachment with acceleratorConfiguration

The acceleratorConfiguration feature is part of the configurationOverrides section in your service compute configuration. It provides a declarative way to specify the type and count of GPU accelerators that should be attached to your compute instances.

Critical Limitations

External GPU attachment via acceleratorConfiguration is only supported with:

- **Instance Family**: N1 instances only
- **GPU Types**: Tesla series GPUs only (T4, V100, P100, P4)
- **Cloud Provider**: Google Cloud Platform only

All other instance families (E2, C2, M1, N2, etc.) **cannot** attach external GPUs.

Configuration Properties

Each accelerator configuration requires:

  • Type: The Tesla GPU accelerator type (required)
  • Count: Number of GPU accelerators to attach (required, minimum: 1)

Supported Accelerator Types

These are the only GPU types that can be attached to N1 instances:

  • nvidia-tesla-t4 - NVIDIA Tesla T4 (16GB GDDR6)
  • nvidia-tesla-v100 - NVIDIA Tesla V100 (16GB/32GB HBM2)
  • nvidia-tesla-p100 - NVIDIA Tesla P100 (16GB HBM2)
  • nvidia-tesla-p4 - NVIDIA Tesla P4 (8GB GDDR5)
  • nvidia-tesla-p4-vws - NVIDIA Tesla P4 Virtual Workstation (8GB GDDR5)
  • nvidia-tesla-t4-vws - NVIDIA Tesla T4 Virtual Workstation (16GB GDDR6)
  • nvidia-tesla-p100-vws - NVIDIA Tesla P100 Virtual Workstation (16GB HBM2)

Examples

Basic Tesla T4 configuration (Docker Compose Style)

x-omnistrate-compose-spec:
  services:
    gpu-service:
      x-omnistrate-compute:
        instanceTypes:
          - name: n1-standard-8
            cloudProvider: gcp
            configurationOverrides:
              acceleratorConfiguration:   
                type: "nvidia-tesla-t4"
                count: 1

High-performance Tesla V100 configuration (Omnistrate Spec)

services:
  - name: gpu-v100-service
    compute:
      instanceTypes:
        - name: n1-standard-4
          cloudProvider: gcp
          configurationOverrides:
            acceleratorConfiguration:   
              type: "nvidia-tesla-v100"
              count: 1

Multi-GPU setup with Tesla T4

services:
  - name: gpu-cluster
    compute:
      instanceTypes:
        - name: n1-standard-16
          cloudProvider: gcp
          configurationOverrides:
            acceleratorConfiguration:   
              type: "nvidia-tesla-t4"
              count: 4

Built-in GPU Instances (Alternative Approach)

If you need modern GPUs (L4, A100, H100), you should use built-in GPU instance families instead of acceleratorConfiguration. These instances come with GPUs pre-installed and offer better performance. Built-in GPU instances are available across AWS, GCP, and Azure.

No Configuration Required

Built-in GPU instances do not use acceleratorConfiguration. Simply specify the instance type and cloud provider - the GPUs are already included.

GCP built-in GPU instance families

G2 family - NVIDIA L4 GPUs

  • GPU: NVIDIA L4 (24GB GDDR6)
  • Use Cases: AI inference, machine learning, graphics workloads
  • Instance Types: g2-standard-4, g2-standard-8, g2-standard-16, etc.

A2 family - NVIDIA A100 GPUs

  • GPU: NVIDIA A100 (40GB HBM2e)
  • Use Cases: High-performance training, large-scale inference
  • Instance Types: a2-highgpu-1g, a2-highgpu-2g, a2-highgpu-4g, etc.

A3 family - NVIDIA H100 GPUs

  • GPU: NVIDIA H100 (80GB HBM3)
  • Use Cases: Advanced AI training, large language models
  • Instance Types: a3-highgpu-8g, a3-megagpu-8g, etc.

A4 family - NVIDIA L4 GPUs

  • GPU: NVIDIA L4 (24GB GDDR6)
  • Use Cases: AI inference, video processing
  • Instance Types: a4-highgpu-1g, a4-highgpu-2g, etc.

GCP built-in GPU examples

Using G2 with built-in L4 GPUs

x-omnistrate-compose-spec:
  services:
    modern-gpu-service:
      x-omnistrate-compute:
        instanceTypes:
          - name: g2-standard-8
            cloudProvider: gcp
            # No acceleratorConfiguration needed - L4 GPU is built-in

Using A2 with built-in A100 GPUs

services:
  - name: training-service
    compute:
      instanceTypes:
        - name: a2-highgpu-1g
          cloudProvider: gcp
          # No acceleratorConfiguration needed - A100 GPU is built-in

Using A3 with built-in H100 GPUs

services:
  - name: llm-service
    compute:
      instanceTypes:
        - name: a3-highgpu-8g
          cloudProvider: gcp
          # No acceleratorConfiguration needed - H100 GPUs are built-in

AWS built-in GPU instance families

G4dn family - NVIDIA T4 GPUs

  • GPU: NVIDIA T4 (16GB GDDR6)
  • Use Cases: AI inference, graphics-intensive applications, machine learning
  • Instance Types: g4dn.xlarge, g4dn.2xlarge, g4dn.4xlarge, g4dn.8xlarge, g4dn.12xlarge, g4dn.16xlarge

G5 family - NVIDIA A10G GPUs

  • GPU: NVIDIA A10G (24GB GDDR6)
  • Use Cases: AI inference, machine learning training, graphics workloads
  • Instance Types: g5.xlarge, g5.2xlarge, g5.4xlarge, g5.8xlarge, g5.12xlarge, g5.16xlarge, g5.48xlarge

G6 family - NVIDIA L4 GPUs

  • GPU: NVIDIA L4 (24GB GDDR6)
  • Use Cases: AI inference, video transcoding
  • Instance Types: g6.xlarge, g6.2xlarge, g6.4xlarge, g6.8xlarge, g6.12xlarge, g6.16xlarge, g6.48xlarge

P4d family - NVIDIA A100 GPUs

  • GPU: NVIDIA A100 (40GB HBM2e)
  • Use Cases: High-performance training, large-scale inference, HPC
  • Instance Types: p4d.24xlarge

P5 family - NVIDIA H100 GPUs

  • GPU: NVIDIA H100 (80GB HBM3)
  • Use Cases: Advanced AI training, large language models, generative AI
  • Instance Types: p5.48xlarge

AWS built-in GPU examples

Using G4dn with built-in T4 GPUs

services:
  - name: inference-service
    compute:
      instanceTypes:
        - name: g4dn.xlarge
          cloudProvider: aws
          # T4 GPU is built-in

Using G5 with built-in A10G GPUs

services:
  - name: ml-training-service
    compute:
      instanceTypes:
        - name: g5.2xlarge
          cloudProvider: aws
          # A10G GPU is built-in

Azure built-in GPU instance families

NC T4 v3 series - NVIDIA T4 GPUs

  • GPU: NVIDIA T4 (16GB GDDR6)
  • Use Cases: AI inference, machine learning, graphics
  • Instance Types: Standard_NC4as_T4_v3, Standard_NC8as_T4_v3, Standard_NC16as_T4_v3, Standard_NC64as_T4_v3

NCv3 series - NVIDIA V100 GPUs

  • GPU: NVIDIA V100 (16GB HBM2)
  • Use Cases: Deep learning training, HPC simulations
  • Instance Types: Standard_NC6s_v3, Standard_NC12s_v3, Standard_NC24s_v3

NC A100 v4 series - NVIDIA A100 GPUs

  • GPU: NVIDIA A100 (80GB HBM2e)
  • Use Cases: Large-scale AI training, inference, HPC
  • Instance Types: Standard_NC24ads_A100_v4, Standard_NC48ads_A100_v4, Standard_NC96ads_A100_v4

ND H100 v5 series - NVIDIA H100 GPUs

  • GPU: NVIDIA H100 (80GB HBM3)
  • Use Cases: Advanced AI training, large language models, generative AI
  • Instance Types: Standard_ND96isr_H100_v5

Azure built-in GPU examples

Using NC T4 v3 with built-in T4 GPUs

services:
  - name: inference-service
    compute:
      instanceTypes:
        - name: Standard_NC4as_T4_v3
          cloudProvider: azure
          # T4 GPU is built-in

Using NC A100 v4 with built-in A100 GPUs

services:
  - name: training-service
    compute:
      instanceTypes:
        - name: Standard_NC24ads_A100_v4
          cloudProvider: azure
          # A100 GPU is built-in

Multi-Cloud GPU Configuration

You can configure GPU instances across multiple cloud providers in the same Plan specification. The platform automatically selects the appropriate instance type based on the customer's deployment region and cloud provider.

services:
  - name: gpu-inference
    compute:
      instanceTypes:
        - name: g4dn.xlarge
          cloudProvider: aws
        - name: g2-standard-4
          cloudProvider: gcp
        - name: Standard_NC4as_T4_v3
          cloudProvider: azure