Skip to content

GitHub Actions GPU Runners | NVIDIA GPU Specifications

Machine provides a variety of GPU runners for GitHub Actions to match your specific CI/CD and machine learning workload requirements. Each GitHub Actions GPU runner type comes with pre-installed NVIDIA Device Drivers 555.58, CUDA 12.1.0 and cuDNN 9.2.1, allowing you to start using GPU acceleration immediately without any configuration.

You are always free to install additional drivers, CUDA, or cuDNN versions, or even build your own from source.

GitHub Actions GPU Runners - Available Types

Machine supports all GPU instances currently available on AWS, including NVIDIA GPUs and AWS Inferentia accelerators.

NVIDIA GPU Runners for GitHub Actions

GPU TypeGPU MemoryCUDA CoresTensor CoresUse Cases
T4G16GB2,560320Entry-level ML training, inference
T416GB2,560320General-purpose ML, computer vision
L424GB7,680240Balanced training/inference, mid-range ML
A10G24GB9,216288Advanced training, larger models
L40S48GB18,176568Large model training, high-performance ML

AWS AI Accelerators

Accelerator TypevCPURAMAccelerator MemoryUse Cases
TRAINIUM832GB32GBHigh-performance training
INFERENTIA2416GB32GBNext-gen inference optimization

Pre-installed GPU Software for CI/CD

Each runner comes with the following software pre-installed:

GitHub Actions Runner Specifications

Besides the GPU, Machine runners offer configurable CPU and RAM options to match your specific workload requirements.

All GPU Runner Configurations and Pricing

Every GPU type supports multiple CPU/RAM configurations. Default configurations (used when cpu and ram labels are omitted) are marked with bold pricing.

NVIDIA T4G (ARM64, 16GB VRAM)

vCPURAM$/Min (Spot)$/Min (On-Demand)
48 GB$0.0037$0.0140
816 GB$0.0027$0.0185
1632 GB$0.0031$0.0276

NVIDIA T4 (X64, 16GB VRAM)

vCPURAM$/Min (Spot)$/Min (On-Demand)
416 GB$0.0048$0.0175
832 GB$0.0069$0.0251
1664 GB$0.0110$0.0401

NVIDIA L4 (X64, 24GB VRAM)

vCPURAM$/Min (Spot)$/Min (On-Demand)
416 GB$0.0057$0.0268
832 GB$0.0036$0.0326
1664 GB$0.0046$0.0441

NVIDIA A10G (X64, 24GB VRAM)

vCPURAM$/Min (Spot)$/Min (On-Demand)
416 GB$0.0126$0.0335
832 GB$0.0110$0.0404
1664 GB$0.0174$0.0541

NVIDIA L40S (X64, 48GB VRAM)

vCPURAM$/Min (Spot)$/Min (On-Demand)
432 GB$0.0161$0.0620
864 GB$0.0179$0.0747
16128 GB$0.0153$0.1001

AWS AI Accelerators

AcceleratorvCPURAMAccelerator RAM$/Min (Spot)$/Min (On-Demand)
Inferentia2416 GB32 GB$0.0025$0.0253
Inferentia232128 GB32 GB$0.0098$0.0656
Trainium832 GB32 GB$0.0052$0.0448

Prices shown are the best available rates across all regions and are subject to change. Additional EBS storage charges apply. See Pricing for full details.

All runners include a 100GB gp3 root volume by default, with configurable size, IOPS, and throughput via runner labels. See Configuration Options for storage label details.

Instance Metrics

Machine runners collect CloudWatch metrics by default, providing real-time visibility into resource utilization for every job. After a job completes, metrics are displayed as sparkline charts on the Machine dashboard.

Collected GPU metrics:

  • GPU utilization percentage
  • GPU memory utilization and usage
  • GPU temperature
  • GPU power draw

Collected system metrics:

  • CPU utilization
  • Memory usage
  • Disk read/write operations
  • Network bytes in/out

You can control metrics collection per job using runner labels:

runs-on:
- machine
- gpu=A10G
- metrics=true # Enable metrics (default)
- metrics_interval=10 # Collect every 10 seconds (default: 60)

To disable metrics for a job, set metrics=false. See Configuration Options for full details.

Next Steps