Loading...
NVIDIA T4 · RTX 4090 · A100

GPU Computing,
from training to inference.

GPU instances with NVIDIA Tesla T4, RTX 4090 and A100. CUDA stack pre-installed, NVLink for multi-GPU, up to 2.5 PFLOPS.

~/ml/training — zsh
PyTorchTensorFlowJupyter
SDKs →
02 · Platform

NVIDIA GPUs, ready for production AI.

Complete GPU isolation per VM. Pre-configured CUDA stack, NVLink for multi-GPU training, MIG support on A100. Your models train on dedicated hardware — no shared resources, no performance variance.

01/ gpu hardware

From inference to large-scale training.

4 GPU models across 3 NVIDIA architectures. Pick the right card for your workload.

Tesla T4
16 GB GDDR665 TFLOPS
RTX 4090
24 GB GDDR6X82 TFLOPS
A100 40GB
40 GB HBM2312 TFLOPS
A100 80GB
80 GB HBM2e312 TFLOPS
02/ ml stack

Pre-installed. Start training now.

Every instance ships with the full NVIDIA + ML stack. No setup, no driver hell.

CUDA12.x
cuDNN8.9.x
TensorRT8.6 LTS
PyTorch2.x
TensorFlow2.x
Jupyter LabML ext.
03/ performance
FP16 Peak
312TFLOPS
A100 Tensor
Memory BW
2TB/s
HBM2e
NVLink
600GB/s
bidirectional
Deployment
< 2min
GPU ready
03 · Pricing

GPU instances for every scale.

All GPUs active
Hourly billing availableExcl. VAT (HT)
04 · Regions

GPU clusters across Europe.

3 datacenters with dedicated GPU racks. Each region offers full NVIDIA GPU isolation, NVLink-capable nodes and GDPR-compliant data residency. Your training data stays in the EU.

ParisFrance
PAR1
Tier IV
< 5ms
T4 · RTX 4090 · A100
Primary GPU region · NVLink & NVSwitch
AmsterdamNetherlands
AMS1
Tier IV
< 8ms
T4 · RTX 4090 · A100
AMS-IX peering · Multi-GPU training
FrankfurtGermany
FRA1
Tier III+
< 10ms
T4 · RTX 4090 · A100
DE-CIX connected · Central EU coverage
05 · Observability

Every GPU metric, live.

GPU utilization, VRAM, temperature, TFLOPS throughput and power draw — all streamed to your console. Compatible with Prometheus, Grafana and custom exporters.

gpu-7b43e·par1-a·A100 40GB
live · 1s
GPU Util87%
VRAM34.2 / 40 GiB
GPU Temp67°C
TFLOPS289 FP16
06 · Developer experience

GPU infra as code.

Provision GPU clusters with Terraform, automate training pipelines with our SDKs. Your ML infrastructure lives in Git.

main.tf
resource "vmcloud_gpu" "training" {
count = 2
flavor = "gpu-enterprise"
image = "pytorch-24.04-cuda12"
region = "par1-a"
ssh_keys = ["ml-key"]
}
› terraform apply
Apply complete! Resources: 2 A100 ready in 1m42s
Terraform
Python SDK
REST API
Jupyter
Compliance & sovereignty

Your training data stays in Europe.

ISO 27001RGPDAES-256GPU isolation24/7 monitoring
07 · Use Cases

Built for compute-intensive workloads.

From model training to real-time inference — our GPU instances deliver dedicated NVIDIA hardware with predictable performance and transparent pricing.

3 EU datacenters·CUDA 12.x·NVLink
AI Training

LLM fine-tuning & training

Multi-GPU A100 setups with NVLink for fast gradient sync. Train LLaMA, Mistral or custom models with full HBM2e memory.

GPU-ENTERPRISE · A100 40GB · €2,323 HT/mo
Inference

Model serving at scale

Tesla T4 instances optimized for inference with TensorRT. Low latency, high throughput, cost-efficient for production APIs.

GPU-STARTER · Tesla T4 · €468 HT/mo
3D Rendering

Blender, V-Ray, Unreal Engine

RTX 4090 with Ada Lovelace ray tracing cores. Scale rendering farms on demand, pay by the hour.

GPU-PRO · RTX 4090 · €1.50 HT/h
Scientific Computing

Molecular simulation & HPC

A100 clusters with NVSwitch and InfiniBand for massively parallel workloads. CUDA-optimized for maximum throughput.

GPU-CLUSTER · 4× A100 · €9,274 HT/mo
05 · FAQ

Frequently asked questions.

Need help choosing the right GPU? Our ML engineers can advise.

Yes. Each GPU is exclusively allocated to your VM. No sharing, no time-slicing. A100 GPUs also support MIG (Multi-Instance GPU) if you want to partition a single GPU into isolated sub-instances.

All instances ship with CUDA 12.x, cuDNN 8.9, TensorRT 8.6, PyTorch 2.x, TensorFlow 2.x, and Jupyter Lab with ML extensions. Custom images with your own stack are also supported.

GPU-QUANTUM (2× A100) and above include NVLink for fast GPU-to-GPU communication at 600 GB/s bidirectional. GPU-CLUSTER and GPU-SUPERCOMPUTE use NVSwitch for full-bandwidth any-to-any GPU communication.

GPU instances are provisioned in under 2 minutes. This includes GPU allocation, driver initialization, CUDA stack verification and NVMe warming. The image is pre-cached for instant availability.

Yes. All GPU plans support hourly billing. A refundable deposit is required: 50% of the equivalent monthly price (all GPU plans exceed the €50 threshold). You only pay actual usage at end of month.

Your GPU in 2 minutes.

CUDA stack ready, drivers loaded, NVMe warm. Start training from … HT/hour with a Tesla T4.

Hourly billingCUDA pre-installed3 EU datacenters