vAquila Icon

vAquila Control Center

Operate your local vLLM runtime with confidence. Launch models, monitor VRAM, inspect logs, and validate inference from one reliable control surface.

Get Started
FastAPI local UI for Docker + vLLM workflows
vAquila Dashboard Runtime

Managed containers

1

All vAquila containers

Running now

1

Currently serving requests

Cached models

2

Ready from local HF cache

CAPACITY OVERVIEW

GPU utilization

GPU 0 • NVIDIA GeForce RTX 5070 Ti Laptop GPU75.6% used
Qwen/Qwen3-4B-Instruct-2507-FP85.37 GiB

Deployment

Launch a vLLM container with explicit runtime knobs. Configure ports, context lengths, timeouts, and more.

🧬

Background Jobs

Track async launches and deeply inspect both task initialization logs and raw container outputs in real-time.

📊

Host Metrics

View detailed breakdown of CPU usage, RAM allocations, and logical core distributions per model.

COMING SOON

vAquila Enterprise

Scale your local AI infrastructure across teams. Advanced security, compliance, and orchestration built for production.

🛡️

SSO & SAML

Secure authentication integrating directly with your corporate identity providers.

🔑

Role-Base Access (RBAC)

Granular permissions: control who can launch, view, or stop specific models.

🌐

Multi-Node Clusters

Deploy and orchestrate vLLM instances across multiple remote GPU servers simultaneously.