Skip to content
 PRODUCT — Manage

Control Plane for Every AI System

The Prediction Guard admin console allows you to manage every AI system in your environment(s). Track inventories of AI assets (models and MCP servers), access kill switches, configure API keys, and manage throughput in one place.

 
 
 
systems
System Management

Track and control your inventory of AI assets

Organize disparate AI assets (models and MCP servers) into single-tenant "AI systems". You can tailor these various systems according to your needs configuring API keys, model settings, MCP server connections, etc.

Each AI system has its own API gateway where you can manage keys and throughput/ quota, regardless of whether you choose to access remote, third-party models & MCP servers or host those inside your environment. 

 
 
 
Screenshot from 2026-05-13 11-04-10
Model Integration

Any Model Vendor. Integrate Self-hosted or via Endpoints.

Prediction Guard is model vendor agnostic. AI systems that are managed in Prediction Guard can include both self-hosted, cloud VPC endpoint, and external/ third party models. 

 
 
 
Screenshot from 2026-05-13 11-09-40
Tool Configuration

Manage the full inventory of tools used by your agents.

AI agents are much more than a single AI model. They are powered by a distributed system of tools connections and AI model calls. Prediction Guard let's you organize your AI models and MCP servers into single-tenant AI systems for consistent enforcement of access and governance.

Both AI models and assets like MCP servers are available to developers behind an OpenAI and Anthropic compatible APIs for rapid agent development (without losing visibility or control).

 
 
 
Screenshot from 2026-05-13 13-36-39
Manage your AI Systems

Complete Control and Optionality to Manage "Locked down" AI Systems

 
Self-Hosted Control Plane
Regardless of which AI assets you configure in your systems (self-hosted, cloud or third-party), the control plane (including gateway, governance enforcement, and supply chain management) lives inside your security boundary
 
Support for any Environment
Our Kubernetes-based deployment of the Prediction Guard control plane can be hosted on-prem, hybrid, air-gapped, or in your cloud VPC. These services are lightweight and only require CPU-based instances.
 
Support for any AI Model Vendors
Self-host any popular model family (Qwen, Llama, Gemma, GPT, etc.) or connect popular pay-as-you-go model endpoints from Azure, AWS Bedrock, Vertex, OpenAI, Anthropic, etc.
 
API Key & Throughput Settings
Create and manage your own set of API keys per "AI System" regardless of the mix of underlying model providers. This way you can maintain a limited set of accounts on AI vendors and manage internal API, throughput, quotas, cost, etc. centrally via Prediction Guard.
 
AI Bill of Materials
Because you are able to manage all of your AI assets, you can generate exportable AIBOMs per system in CycloneDX format. Maintain a full inventory for compliance, governance, and risk management.
 
Seamless Configuration Updates
Regardless of where your AI systems are hosted (on-prem, hybrid, or cloud), Prediction Guard's Admin Console allows you to configure model settings, MCP updates, model kill switches, etc. centrally. AI systems poll this admin API for configuration updates, allowing your team to manage increasingly complicated ecosystems of AI tools. 

Ready to Own Your AI Stack?

See how Prediction Guard gives you full sovereignty over every model, agent, and API key in your organization.