GPU
GPU L40S
Dedicated L40S GPU for private inference and production AI workloads.
Private Inference
Run LLM models locally on L40S with 48 GB VRAM — your data never leaves the VM.
Ollama Pre-installed
Download and serve models instantly. No setup, no configuration, just inference.
Full Agent Capabilities
Shell commands, file editing, code execution, and any LLM provider — all in a secure sandbox.
Specifications
CPU
16 vCPU
Memory
180 GB
Storage
1200 GB NVMe
Workspaces
5
GPU
L40S
VRAM
48 GB
GPU L40S
GPUDedicated L40S GPU for private inference and production AI workloads.
$1499/month
CPU: 16 vCPU
RAM: 180 GB RAM
Disk: 1200 GB NVMe
BW: 12 TB Transfer
GPU: L40S — 48 GB VRAM
- Everything in GPU A100
- NVIDIA L40S — 48 GB GDDR6
- Run 70B+ parameter models
- 5 workspaces
- Hourly backups
- Firewall management (50 rules)
- SSH access