GPU

GPU L40S

Dedicated L40S GPU for private inference and production AI workloads.

Private Inference

Run LLM models locally on L40S with 48 GB VRAM — your data never leaves the VM.

Ollama Pre-installed

Download and serve models instantly. No setup, no configuration, just inference.

Full Agent Capabilities

Shell commands, file editing, code execution, and any LLM provider — all in a secure sandbox.

Specifications

CPU

16 vCPU

Memory

180 GB

Storage

1200 GB NVMe

Workspaces

5

GPU

L40S

VRAM

48 GB

GPU L40S

GPU

Dedicated L40S GPU for private inference and production AI workloads.

$1499/month
CPU: 16 vCPU
RAM: 180 GB RAM
Disk: 1200 GB NVMe
BW: 12 TB Transfer
GPU: L40S — 48 GB VRAM
  • Everything in GPU A100
  • NVIDIA L40S — 48 GB GDDR6
  • Run 70B+ parameter models
  • 5 workspaces
  • Hourly backups
  • Firewall management (50 rules)
  • SSH access

Other plans in this category