Skip to main content
Koyeb logo

Koyeb

Serverless GPUs with per-second billing and global deployment

Inference specialist🇫🇷 FR

Last reviewed May 8, 2026

Koyeb is a serverless platform offering on-demand NVIDIA GPUs from RTX-class up through B200 with per-second billing, scale-to-zero and global edge deployment.

We're actively tracking prices for Koyeb. Check back soon, or browse other providers with current pricing.

Pros & Cons

Advantages

  • Per-second billing across the full GPU range
  • Multi-GPU configurations (2x, 4x, 8x) available on-demand
  • Serverless model removes idle-instance costs

Limitations

  • Smaller per-cluster scale than dedicated training neoclouds
  • Some GPU SKUs require requesting access
  • Long-running fixed-instance training is not the primary use case

Key Features

Serverless GPU Instances

Per-second billing with scale-to-zero across the GPU catalog

Wide GPU Range

From RTX-4000-SFF-ADA and L4 through L40S, A100, H100, H200 and B200

Global Edge Deployment

Deploy applications close to users across multiple regions from a single push

Pricing Options

OptionDetails
Per-Second GPU BillingCharged per second of GPU runtime, with scale-to-zero when idle
Multi-GPU ConfigurationsPre-configured 2x, 4x and 8x GPU instances for larger workloads

Availability & Support

Regions

Global edge presence across North America, Europe and Asia

Support

Documentation, community forum and paid support tiers

Getting Started

  1. 1

    Create an account

    Sign up via the Koyeb console using email or GitHub

  2. 2

    Pick a GPU instance

    Select an instance type from the GPU catalog and configure your service

  3. 3

    Deploy from Git or container

    Push from a GitHub repo or Docker image and Koyeb handles the rest