The central nervous system for DevOps. Kortex senses every event in your stack, reasons about it with real AI, takes safe corrective action on its own, and remembers every incident forever.
Kortex is organized around five capabilities that mirror how a biological brain handles its body.
Logs, metrics, traces, health checks, crashes, deploys — from any source via HTTP, gRPC, OTLP, or Unix socket. No vendor lock-in.
Statistical baselines catch the obvious. Stack-trace clustering catches the recurring. LLM reasoning catches the hidden root cause.
Restart workers, drain slots, autoscale, roll back deploys, isolate sick backends — automatically, with explicit opt-in and a kill switch.
Every incident is stored, indexed, and learned from. Auto-generated playbooks. MTTR tracked per category.
Compose actions across multiple Koder products and external systems. Drain, migrate, reload, verify, mark deploy successful — one workflow.
Concrete capabilities, not buzzwords.
Accept OpenTelemetry logs, metrics, and traces out of the box. Bring your existing instrumentation.
Z-score over moving windows on p50/p95/p99 latency, error rates, throughput, saturation. Cheap, fast, no LLM required.
Stack traces normalized and grouped. First occurrences flagged. Never debug the same error twice.
Anthropic Claude, OpenAI, Ollama (local), or koder-ai. Switch via config. Tool use unified across providers.
Ask "why is service X slow?" and watch the AI fetch logs, compare baselines, check deploys, and respond with citations.
Simple TOML rules: when X then do Y. Cooldowns, dry-run mode, audit log, kill switch.
Resolved incidents become reusable runbooks. Same problem next time? Kortex already knows the fix.
Direct consumer of the koder-jet event bus. No exporters, no scrapers, no intermediate format.
OIDC login, role-based access (viewer, operator, admin), full audit trail of every action.
From raw event to autonomous fix in three stages.
Point your services at Kortex over OTLP, gRPC, HTTP, or a local socket. Events stream into ClickHouse continuously.
Statistical detectors spot anomalies in real time. The LLM brain investigates by querying logs, metrics, deploys, and history — then proposes a root cause with evidence.
Approve a fix in chat, or let a reflex rule fire it automatically. Kortex executes the action, verifies success, and writes the incident to memory.
Kortex is the first observability product designed around autonomy from day one.
| Capability | Kortex | Datadog | New Relic | PagerDuty |
|---|---|---|---|---|
| Universal event ingest (OTLP) | ✓ | ✓ | ✓ | — |
| Statistical anomaly detection | ✓ | ✓ | ✓ | — |
| AI-native root cause analysis | ✓ | add-on | add-on | — |
| Conversational debugging with tool use | ✓ | — | — | — |
| Autonomous remediation built in | ✓ | — | — | runbooks |
| Cross-system workflow coordination | ✓ | — | — | limited |
| Pluggable LLM providers (incl. local) | ✓ | — | — | — |
| Self-hosted, single binary | ✓ | — | — | — |
| Auto-generated playbooks from incidents | ✓ | — | — | — |
Kortex is in early development. Sign in with Koder ID to follow progress and join the early access list.
Sign In to Get Early Access