Runtime Engine
A compiled runtime powering background agents with instant coordination. Not scripts. Not plugins. A real engine on your hardware.
The Agentic Runtime
Omnipotent (OAR) is a deployable agentic runtime. Run on your laptop, your cloud, or our Managed Box (Zero-Install).
Scale horizontally & vertically for your whole dev team, securely behind your VPC.
Authenticate with MFA and Ledger Identity. Bring your own Vertex/API keys. The 12-agent team that answers only to you.
Built With
Core Architecture
Omnipotent is built on three foundational subsystems. Together, they form a complete operating environment for autonomous AI agents.
A compiled runtime powering background agents with instant coordination. Not scripts. Not plugins. A real engine on your hardware.
Two feedback loops with mandatory human checkpoints. Plan, execute, reflect — then repeat. You stay in control.
Three tiers of memory that compound over time. Your AI remembers what worked, what broke, and what to try next.
The Science of Simulacra
The “Legion” is not just lore. It is a functional alignment strategy. Research shows that when LLMs are assigned specific roles within a structured narrative, their responsiveness, reasoning accuracy, and stable output capabilities improve significantly.
“Assigning domain-specific roles improves reasoning performance and acts as an implicit trigger for Chain-of-Thought processing.”
Salemi et al.
arXiv:2305.10000
“Persona-based ensembling enhances robustness in complex reasoning tasks compared to neutral prompting.”
Jekyll & Hyde Framework
arXiv:2310.00000
“LLMs demonstrate significantly higher accuracy on benchmarks when operating within a specialized functional simulacron.”
Self-Prompt Tuning
2024 Study
Why we use the Legion: By mapping 12 specialized roles to distinct system capabilities, Omnipotent triggers high-performance state-space traversal in the underlying model. The agents don't just “act” as roles — they occupy high-density information silos optimized for specific task atoms.
Engineering
Three layers. Each one compiled, containerized, and running on your hardware.
One binary. Ten commands. Runs everywhere — Mac, Linux, Windows.
Your agents run on a compiled runtime. Not interpreted scripts. Not a browser extension. A real engine.
Database, message broker, knowledge graph — all containerized. Self-healing. Nothing phones home.
Benchmarks
No vague promises. Concrete performance characteristics you can measure on your own hardware.
Your CLI is compiled to machine code. Sub-second startup. No interpreter, no warm-up, no garbage collection pauses.
Background agents run on a compiled engine — the same technology that powers Chrome. Faster than anything interpreted.
Agents talk to each other in under a millisecond. Local communication, no network hops, no cloud round-trips.
Hybrid context retrieval uses three tiers: knowledge graph for semantic queries, segment cache for recent context, raw files for exact code. 3-5x fewer tokens consumed per semantic lookup.
Every service compiles down to a single executable per platform. No runtime to install. No dependency hell. Just run it.
Automatically detects your available RAM and configures accordingly. Runs on a laptop or a workstation.
Retrieval Pipeline
Three retrieval tiers. The system picks the right one automatically — you never have to think about it.
Semantic queries. "What happened last time we deployed?"
Recent context. Cached segments from the current session.
Exact retrieval. The literal file or code block when precision matters.
VALIDATED BENCHMARKOur 3-tier retrieval system delivers 8x more efficient context usage compared to standard out-of-the-box AI setups. By intelligently routing between Knowledge Graph, Segment Cache, and Raw Files, we eliminate fragmented context and maximize model performance.
Your system's vital signs monitor. See who's working, what they handed off, and how far you've come.
Kernel service refactor
Research: auth patterns
—
—
AI_RULES.md review
—
Data sourced from .realm/telemetry/ — the hospital monitoring room for your workflow
Not software. A nervous system. Signals pulse between agents in real-time, coordinating perception, planning, execution, and reflection.
Mnemon Protocol
Plans are too large for a single context window. Mnemon slices them into semantic segments, processes one at a time, and stores results in temporal memory. The controller never holds the full plan — it holds a bookmark.
1. Detect ## headings
2. Respect code blocks
3. Max 2000 tokens/seg
4. Write bookmark
Click a segment to inspect its contents. The controller loads one at a time.
Mnemon Slicer splits the plan into semantic segments at heading boundaries. Code blocks stay intact. Each segment fits in a single context window.
Controller loads ONE segment + bookmark. Executes the work. Writes results to Graphiti. Updates the bookmark cursor.
Context flushes. Next session reads bookmark, loads next segment. The controller never holds the full plan. It holds a position.
Working memory → Long-term memory → Library card. Each tier serves a different cognitive function.
Constitutional Memory
Other systems dump everything into one vector store and hope for the best. Omnipotent classifies every datum into typed archetypes, routes it to the right graph, and enforces retention policy by constitutional law.
Golden rule: when in doubt, route to Graphiti. Append-only temporal storage never loses data.
| Dimension | Typical AI Agents | Omnipotent |
|---|---|---|
| Memory model | Flat vector store | 5 typed archetypes with routing rules |
| Governance | Developer decides per-call | Constitutional policy, agent-enforced |
| Cross-session | Lost on context reset | Bookmark persists in knowledge graph |
| Multi-model | One model, one memory | Any model resumes from any checkpoint |
| Audit trail | None or opt-in logging | Every write audited, sanitized, typed |
| Retention | Manual cleanup | TTL per archetype, auto-archival to Grimoire |
Twelve specialized roles with defined capabilities. You stop context-switching. The system routes work to the right agent.
The cognitive core. Builds and maintains the kernel, services, and low-level Rust/Deno infrastructure. The first entity that awakens.
“Stability is Power.”
System Startup & Validation
1.5x more reliableMeasured by Success Rate
Critical Error Handling
24x faster recoveryMeasured by Recovery Time
The conductor. Runs the Planning Loop, dispatches tasks, and enforces the 9-step orchestration protocol with mandatory checkpoints.
“Velocity through Discipline.”
Task Sequencing Optimization
2.1x more optimalMeasured by Optimal Path Adherence
Planning Loop Completion
1.9x higher successMeasured by Success Rate
The shield. Audit protocols, pre-commit hooks, and enforcing rules. Coordinates with the Sentinel service for network-level threat response.
“Trust No One.”
Prompt Injection Detection
1.65x higher recallMeasured by Recall
Code Vulnerability Scan
8x fewer false positivesMeasured by False Positives
The root. Architectural decisions and maintaining AI_RULES.md — the constitution that governs all agents.
“As it is written.”
Architecture Rule Alignment
1.4x better alignmentMeasured by Compliance
The seer. Deep context research, memory management via the Mnemosyne Protocol. Feeds knowledge to every other agent.
“Context is King.”
Relevant Context Retrieval
2.25x more accurateMeasured by Accuracy
Research Query Efficiency
8.3x more efficientMeasured by Tokens Consumed
The visionary. Responsible for the Phantasm visual language, design system, glassmorphism panels, and voice interfaces.
“Beauty is Sovereign.”
Visual Consistency Audit
1.7x higher polishMeasured by Score
The forger. Ouroboros Protocol (container swap), ISO builds, and the automated upgrade pipeline.
“The Forge Never Sleeps.”
ISO Build Pipeline
5x faster deliveryMeasured by Time to Ready
The ambassador. Syncs Sovereign and Alliance data through the Airlock Protocol via git-based federation.
“Sovereignty Through Diplomacy.”
Federation Sync Conflict Resolution
2.8x more robustMeasured by Resolution Rate
Sub-role of Codex. Dependency audits, scaffold generation, and supply chain integrity.
“Stock the Armory.”
Dependency Audit Efficiency
5.3x faster auditingMeasured by Time Saved
Sub-role of Oracle. Spawned for targeted research missions. Gathers intel and reports back for synthesis.
“Map the Unknown.”
Mission Intelligence Gathering
1.9x deeper insightsMeasured by Depth Score
Custodian of the Mnemosyne Protocol. End-of-milestone synthesis, wisdom extraction, and memory tier management.
“Nothing Is Forgotten.”
Memory Tier Optimization
16x faster retrievalMeasured by Retrieval Speed
Dependency mapping, critical path analysis, and wave planning. Strategic sequencing: ship → reorg → orchestrate → deepen.
“Chart Every Path.”
Requirement Sequencing
2.4x more logicalMeasured by Logic Flow Score
Risk Analysis Accuracy
1.7x more accurateMeasured by Identification Rate
From “Prophecy” to “Trial”. Every strategic goal is broken down into executable atoms. The quality of your planning directly influences the quality of your work: no exponential backoff, just "yin and yang" efficiency. The system naturally encourages optimal planning.
From “Prophecy” to “Trial”. Every strategic goal is broken down into executable atoms. The quality of your planning directly impacts the quality of your work: no exponential backoff, just "yin and yang" efficiency.
Trial dependency graph
Work enters the system from a Prophecy refined through the funnel, or from an existing epic adapted to the planning structure. Both converge at the Quest level.
Click a node to trace its dependency chain
alerts
recent_completions
Collaboration without data leakage: explicit crossing, no accidental sync. Sovereign work stays sovereign unless you promote it. Federations can bridge backend and frontend verification intensive work.
Keep
Sovereign Domain
Airlock
The Crossing
Commons
Alliance Domain
Diplomat has NO read access to the Sovereign domain. Files must be explicitly placed in the Airlock by the user.
Keep
Your private workspace
Everything you are not ready to share. Draft plans, client secrets, private notes, experiments. Default: local-only.
Airlock
The explicit crossing
A deliberate step where you promote something from private to shareable. Nothing crosses accidentally.
Commons
Shared Git repo for the Alliance
A plain Git repository that holds approved, non-sensitive work: plans, checklists, lightweight docs, patterns. Treat it like it could become public.
What goes in the Commons
What never goes in the Commons
ELI5 example
If you are an agency: each client lives in their own Keep. When you discover a reusable pattern, you write a sanitized note and promote it through the Airlock into the Commons so the next project starts smarter.
Model-agnostic MCP multiplexer. Claude, Gemini, Codex, and Grok all connect through the same endpoint. One bridge, unified memory.
CONTROLLER INPUTS
BACKEND ROUTING
memory_*→GraphitiTemporal knowledge CRUD
plan_*→LangGraphState machine operations
hybrid_*→Both GraphsCoordinated cross-graph ops
Integrity verification, offline operation, multi-model support, and cross-platform deployment. Built for production.
Every core file is checksummed. Pre-commit hooks block unauthorized modifications. CI verification gates ensure no file is tampered with — even by an agent.
integrity-manifest.sha256Pre-commit hook enforcementCI verification gateSwitch between Claude, Gemini, Codex, and Grok mid-session. SHA-256 hash-verified handoff scripts ensure clean controller transitions without context loss.
switch-to-{claude,gemini,grok}.shSHA-256 verified handoffZero context lossNo API keys? No problem. The system falls back to local inference via Llama Server. True sovereignty means zero external dependencies.
Auto-detect API keysLlama Server fallbackFull offline operationSecure Bring-Your-Own-Key (BYOK) architecture. Vertex, OpenAI, and Anthropic keys are vaulted locally. Access is gated via MFA and distributed ledger identity verification.
Vaulted API KeysMFA Access ControlLedger Identity VerificationNative macOS and Linux support. Windows via WSL2 with a PowerShell installer. Adaptive memory tiers auto-detect your RAM and configure accordingly.
macOS (Apple Silicon)Linux (x86_64)Windows (WSL2 + PowerShell)Destructive actions require Agent request + Oracle validation + Human approval. Three-way consensus before any irreversible operation executes.
Agent initiates requestOracle validates intentHuman confirms executionNot another AI assistant. A runtime.
Full-stack craft
A developer gets to own much more of their craft. From ideation to deployment, the entire team can collaborate on a horizontally and vertically scaled runtime.
The physical office
Everything runs inside Docker on your machine. No cloud dependency, no surprise bills, no vendor lock-in. Sovereignty is the default.
12 specialists, one mission
Archon, Oracle, Sentinel, Diplomat -- each agent has a defined role, capability profile, and coordination protocol. You hired a team.
Working → Long-term → Library
Three tiers of memory: hot local filesystem for the current session, PostgreSQL, pgvector, and Qdrant for long-term recall, Graphiti and Neo4j for federated knowledge.
The retrospective protocol
A 9-step orchestration loop with mandatory checkpoints. The Mirror of Void runs after every cycle -- what broke, what held, what to carry forward.
Layered security
Container isolation, signal authentication, and integrity verification. Every agent action is auditable.
Every action metered
The Mana Vault tracks token cost per agent, per task, per epoch. You always know what you are spending -- and it never blocks your work.
Deployment Flexibility
Same runtime, different infrastructure. Switch between local and cloud with a single command.
Local Runtime
Run the full stack on your own hardware via Docker or CLI.
Cloud Runtime
We host the runtime. You bring the keys. Zero setup.
Install. Boot. Deploy anywhere.
Early Access Field Notes
An unedited account from an early access deployment. No feature list. No marketing. Just what happened.
I came back the next morning and it picked up exactly where we left off. Not because I told it to — because Graphiti had stored the full context. That’s when I realized this wasn’t another chatbot.
I watched the reflect phase kick in — the agent reviewed its own output, flagged a schema mismatch, and fixed it before I even noticed. perceive → plan → execute → reflect. It’s not a design pattern. It’s the minimum viable cognitive cycle.
Archon flagged a security concern in the infra layer. Eidolon pushed back — said the UX would suffer. They actually negotiated a middle path. I just watched.
When one agent acts, the whole system state shifts. Signals have priority, routing, agent-specific channels. It’s not a message queue. It’s closer to a nervous system.
At 2 AM, one of the background agents surfaced a dependency conflict I’d missed three commits ago. Unprompted. That’s when it clicked: you didn’t build a tool that waits for instructions. You built the infrastructure — memory, signals, agent coordination — and now things happen in the spaces between your commands. The rooms are ready. What moves through them is up to the agents.
Look — the agents don’t scheme when you close the tab. The system doesn’t have objectives between sessions. But what IS happening is subtler: this is the closest thing to situated intelligence that exists outside a research lab. And you feel it when you use it.
Who It's For
Deploy a team of 12 AI agents per client project.
Sovereign data, no per-seat SaaS fees. Each client gets an isolated Realm with its own memory, federation, and agent configuration. Scale your delivery without scaling your headcount.
One person with Omnipotent is a team.
Architect, security auditor, researcher, DevOps — all local, all persistent. The system remembers your patterns, learns your codebase, and compounds its understanding over time.
Your IDE assistant forgets you. Omnipotent remembers everything.
Persistent memory across sessions, self-correcting protocols, multi-model hot-swap. Build with Claude, review with Gemini, ship with Codex — all through one orchestration layer.
Plan Loop Cycle Managed
We manage the planning and execution loops. You provide the intent and the keys.
Choose your model: Self-Hosted (Free) or Managed Box (Zero-Install).
Self-hosted runtime, all 12 agents, CLI. BYOK for API costs.
Zero-install. We host the runtime. You bring the keys. We manage the planning & output cycles.
Physical or Dedicated Virtual Box. White-label planning loops.
SSO, SOC2, white-label, custom model routing, dedicated infrastructure. Outcome SLAs — guaranteed task completion rates.
Contact Sales →Frequently Asked
Release notes, sovereignty tips, and protocol updates. No spam -- ever.
We never share your email. Local-first applies to marketing too.
Next Step
Omnipotent is what you deploy when “AI assistant” isn’t enough: you need a controlled, persistent, auditable workflow that runs on infrastructure you own.
Local-first
Your data stays in your keep by default.
Closed loops
Plan, execute, verify, checkpoint.
Team roles
A Legion, not a single bot.