PaiRun is the platform AI builders use to ship agents to production — with the primitives Kubernetes gives services. Declarative specs, identity, budgets, audit. No glue code.
Integrated capability planes purpose-built for the demands of agentic AI workloads — from lifecycle and auto-scaling to zero-trust security and full observability.
Declarative agent specs, controller-driven reconciliation, and metric-driven autoscaling. Scale on token rate, JIRA backlog depth, or any HTTP endpoint — no HPA required.
Token consumption and tool calls logged per workload. Tamper-evident cryptographic audit chain queryable via the CLI. Budget enforcement with hard daily cutoffs per agent.
Non-root execution, seccomp syscall filtering, Landlock filesystem confinement, and a cryptographic audit chain on every agent. AI-specific threat detection built in.
Every agent gets a cryptographic SPIFFE identity. Secrets injected at runtime, agents never hold credentials. Platform-brokered OAuth to third-party services.
A unified, OpenAI-compatible endpoint for all agents. Route across OpenRouter, Anthropic, OpenAI, and local models with fallback chains, cost limits, and semantic caching.
Connect agents to Jira, GitHub, Slack, and more. Each binding carries its own access control policy. Agents get exactly the permissions they need, nothing more.
Bridge your local Chrome to a remote agent over an encrypted WebSocket relay. Agents can browse, screenshot, and interact with any page — without a headless browser running in the cluster.
Run agents without a container image. Define a system prompt, tools, and skills in an Agent template — the platform harness handles the runtime. Each task gets an isolated, ephemeral container with a full event stream.
No scripts, no manual wiring. One YAML for persistent long-running agents — bring your own image. Or use task mode: declare a system prompt and tools in an Agent template, skip the container entirely.
apiVersion: pai.io/v1 kind: Agent metadata: name: research-agent spec: type: service image: registry.internal/openclaw:v1.2.0 modelBindings: [claude-sonnet-4-6, gemini-flash] providers: [jira-read-comment, github-pr-writer] autoscaling: minReplicas: 1 maxReplicas: 5 metrics: - type: tokenRate targetValuePerReplica: 500 tokens: maxPerDay: 50000 maxPerRequest: 8192 filesystem: readOnlyPaths: [/etc, /home/node/.config]
apiVersion: pai.io/v1 kind: Agent metadata: name: code-reviewer spec: modelBindings: [claude-sonnet-4-6] system: | You are a senior code reviewer. Review PRs for correctness and security. tools: - type: bash - type: read - type: web_fetch providers: [github-pr-writer] packages: pip: [pygithub] --- kind: Agent type: task spec: agentDefinition: code-reviewer task: "Review PR #142 for security issues"
Four domains, zero gaps. From continuous risk assessment through runtime, identity, and data — including AI-specific threats no existing platform understands.
Agents target a ModelBinding, not a provider. The gateway handles routing, fallback, cost enforcement, and credential isolation.
API keys live in the platform secret store. Agents have zero knowledge of provider credentials.
Token consumption tracked and budgeted per agent. Hard cutoffs prevent runaway spend before it hits your bill.
Every LLM call logged with model, tokens, latency, cost, and the agent identity that made it.
Platform-aware failover chains. If quota is exceeded or a provider is down, route to the next option seamlessly.
Switch models or providers without touching agent code. Your agents target an abstraction — the platform handles the rest.
Every binding to an external system carries its own access control policy. Agents get exactly the permissions they need. Nothing more.
Each binding declares an explicit allow/deny list of actions. An agent can have read-only access to Jira and write access to GitHub simultaneously — no shared policy surface.
API keys, OAuth tokens, and app credentials live in the platform secret store. The agent process never sees them. All requests are proxy-brokered using short-lived scoped tokens.
Go beyond action policies — restrict a GitHub binding to specific repositories, a Jira binding to specific projects, or a Slack binding to specific channels.
Every API call through a binding is logged: calling agent identity, timestamp, action, target resource, and outcome. Compliant by default.
Disable a binding platform-wide in one operation. All agents using it lose access immediately — no redeployment required.
Pai extends Kubernetes with AI-native CRDs and a control plane that understands agent semantics, not just containers.
Install Pai, connect your LLM provider, and deploy an AI agent with full observability and zero-trust security.