The AI agent harness with enough nerve to act like an operating system for software work.
Single binary. Native Rust. Real memory. Parallel worktrees. Design and spec lifecycles built in.
Omegon is not a chatbot in your terminal. It is a systems engineering harness for operators who build: a terminal-native agent that can read and edit code, run commands, manage project memory across sessions, decompose work into parallel children, and track design intent as a first-class artifact.
Most “coding agents” are thin wrappers around an API call, a file picker, and some hopeful marketing.
Omegon takes the opposite position.
Software work is a live system with:
- state
- memory
- lifecycle artifacts
- source control boundaries
- multiple inference backends
- cost and quota pressure
- operator intent that must survive more than one session
So Omegon treats the agent as one subsystem inside a larger harness.
That means you get:
- a real terminal UI instead of a transcript dump
- provider honesty instead of vague “smart routing” slogans
- persistent project memory instead of amnesia between runs
- design-tree and OpenSpec lifecycles instead of planning in random markdown scraps
- parallel git worktree execution instead of praying one giant prompt does the right thing
curl -fsSL https://omegon.styrene.io/install.sh | shPreview channel one-shot:
CHANNEL=rc curl -fsSL https://omegon.styrene.dev/install.sh | shOr with Homebrew:
brew tap styrene-lab/tap
brew install omegon # stable
brew install styrene-lab/tap/omegon-rc # preview / rcStable docs: https://omegon.styrene.io/docs/install Preview docs: https://omegon.styrene.dev/docs/install
omegon login openai-codex
omOmegon now installs two standard entrypoints from the same binary:
om— slim, copy-friendly, familiar terminal modeomegon— full harness mode
You can move between them interactively at runtime:
/warp— toggle slim ↔ full/shackle— force slim (om) mode/unshackle— force full (omegon) mode
Flags still override the entrypoint default when you need the opposite posture:
om --full
omegon --slimexport ANTHROPIC_API_KEY=sk-ant-...
omollama pull qwen3:32b
omIf you want the shortest happy path instead of a full tour, start here:
- https://omegon.styrene.io/docs/get-started
- Preview / staging guidance: https://omegon.styrene.dev/docs/get-started
Omegon is a Rust-native binary with no Node.js runtime dependency.
It gives you:
- structured conversation rendering
- collapsible tool cards
- wrapped multiline editor
- live engine, memory, and system telemetry
- primary browser surface launch via
/auspex open - local browser compatibility/debug surface via
/dash
Omegon distinguishes concrete runtime backends instead of hiding them behind mushy branding.
Examples:
- Anthropic/Claude
- OpenAI API
- OpenAI/Codex
- Ollama (Local)
- Ollama Cloud
The footer separates:
- provider → concrete backend identity
- model → selected runtime model
- limit → upstream quota/bucket telemetry when available
That sounds small until you’ve spent an hour debugging a system that lied about what model path it was actually using.
Omegon stores durable facts under typed sections:
- Architecture
- Decisions
- Constraints
- Known Issues
- Patterns & Conventions
- Specs
This is not “chat history with search.” It is a project memory system with recall, persistence, and lifecycle-aware usage.
Omegon ships with two durable lifecycle systems:
- Design Tree — explore questions, record research, dependencies, and decisions
- OpenSpec — write Given/When/Then specs, generate plans, verify implementation, archive change history
This makes the agent better at long-running work because intent is written down in forms the harness can query.
For larger tasks, Omegon can assess complexity and split work into child tasks with isolated scopes.
That means:
- separate worktrees
- concrete file boundaries
- merge and conflict detection
- better operator control over risky multi-file changes
Omegon now ships with an in-repo token-efficiency comparison harness under scripts/benchmark_harness.py and ai/benchmarks/.
Use it for two things:
- compare harnesses honestly — same task, same acceptance, different agent surface
- shape Omegon’s signal profile with evidence — inspect totals, wall clock, and Omegon’s
sys/tools/conv/mem/hist/thinkbuckets before changing prompt, history, or tool-surface behavior
Current stance:
omis the de-facto comparison profile for mainstream CLI coding agents- default
omegonis the premium harness mode when richer systems-engineering behavior is worth extra token cost omegon --slimandom --fullremain valid overrides when you want the opposite posture from the entrypoint default- the benchmark harness stays in-repo for now because it is still tightly coupled to Omegon internals (
--usage-json,omegon_context, auth/provider behavior, and clean-room runtime mechanics)
Design notes:
- [[docs/design/evidence-driven-signal-shaping|Evidence-Driven Signal Shaping]]
- [[docs/design/signal-shaping-profiles|Signal Shaping Profiles]]
- [[docs/design/signal-classes-and-retention-policy|Signal Classes and Retention Policy]]
Launch Omegon in a repo:
omThen prompt it normally:
Read README.md and summarize the architecture.
Or make it do real systems work:
Inspect the auth flow, identify the weakest boundary, propose a minimal fix, add tests, and commit it.
Or lean into lifecycle mode:
We need to refactor the session model. Explore the design space, surface assumptions, write a spec, and then implement it.
Omegon treats inference as three separate controls:
- Capability tier —
local→retribution→victory→gloriana - Thinking level —
off/minimal/low/medium/high - Context class —
squad/maniple/clan/legion
This is the right model because “which model?” is not the only question that matters. Capability, reasoning effort, and context budget are different levers.
Omegon exposes structured tools for:
- file reads/writes/edits
- shell execution
- web search
- git operations
- date/time resolution
- memory management
- design-tree queries and mutations
- OpenSpec lifecycle management
- codebase retrieval
- model control
- background services
Current surfaces include:
- Anthropic/Claude
- OpenAI API
- OpenAI/Codex
- OpenRouter
- Groq
- xAI
- Mistral
- Cerebras
- Hugging Face
- Ollama (Local)
- Ollama Cloud
/tutorial is an interactive overlay, not a static lesson reader.
It can read code, store memory, create lifecycle artifacts, and walk an operator through real work.
core/ Rust workspace
crates/
omegon/ Main binary — TUI, agent loop, tools, web surface
omegon-git/ Git and worktree operations
omegon-memory/ Project memory runtime
omegon-secrets/ Secret resolution and redaction
omegon-traits/ Shared protocol and event types
site/ Public docs site
openspec/ Spec-driven lifecycle artifacts
docs/ Durable architecture and design docs
skills/ Markdown skill packs
themes/ Alpharius theme assets
cd core
cargo build --releaseRelease binary:
core/target/release/omegon
Type-check and test path for TypeScript and Rust changes lives in repo-local commands and release validation.
Mutable user state lives under:
~/.config/omegon/
Common files:
auth.json— provider credentialssettings.json— user settingsAGENTS.md— global directives
Project-local control surfaces include:
AGENTS.md.omegon-versionai/memory/facts.jsonlopenspec/docs/
Omegon ships signed releases with:
- SHA-256 checksums
- cosign signatures
- CycloneDX SBOMs
- GitHub build attestations
Install/update docs:
- https://omegon.styrene.io/docs/install
- Preview / RC docs: https://omegon.styrene.dev/docs/install
Start here:
- https://omegon.styrene.io/docs/
- https://omegon.styrene.io/docs/quickstart
- https://omegon.styrene.io/docs/providers
- https://omegon.styrene.io/docs/tui
- https://omegon.styrene.io/docs/tutorial
Preview / staging docs:
BSL 1.1 — © 2024–2026 Black Meridian, LLC
In practical terms: if you are not trying to offer Omegon itself as a competing hosted agent service, the license is unlikely to be the part that ruins your day.