Release Note

Session-Aware Context: The OMNI v0.5.4 Engine Overhaul

There is a critical flaw hiding in plain sight across every AI context filtering tool on the market: the complete lack of memory. Compressing raw text blindly — line by line, buffer by buffer — systematically destroys the continuity between commands. Your agent runs cargo build, gets a cryptic linker error, then runs cargo test, and the context engine treats these as two completely unrelated events. The reasoning chain is shattered before it even begins. OMNI v0.5.4 solves this problem at its architectural root with our most ambitious release yet: Session-Aware Distillation.

From Passive Filter to Stateful Engine

The headline feature of this release is the injection of live tracking state directly into the distillation pipeline. Every distiller now inherently “remembers” the toolchain context of your current session. If you ran a Docker build three commands ago and now you are inspecting logs, the engine understands the relationship and preserves cross-command continuity in its output. This is not caching — this is genuine contextual intelligence built into the core processing loop.

Three New Domain Distillers

Alongside the session layer, we shipped three entirely new engine distillers, each purpose-built for a high-noise domain. The CloudDistiller handles Docker, Kubernetes, and Terraform output with surgical precision, knowing exactly which infrastructure state transitions matter and which are repetitive boilerplate. The SystemOpsDistiller targets everyday shell tools like ls, env, grep, and tree — commands that produce massive output but where only a fraction carries diagnostic value. Finally, the JsTsDistiller is specifically tuned for ESLint and TypeScript Compiler output, preserving error chains while stripping the verbose configuration dumps that plague JavaScript monorepos.

The OMNI Filter Pack Expansion

We also migrated and enhanced twelve new TOML-based filters for modern development tools. Playwright, Ruff, golangci-lint, .NET, Prisma, Bun, Cypress, Jest, mypy, Black, pnpm, and PHPUnit all received dedicated, hand-tuned semantic extraction rules. For JavaScript developers working in polyglot monorepos, this update alone can eliminate thousands of wasted tokens per session.

A Modular Pipeline You Can Reason About

Under the hood, we undertook a significant refactoring of the monolithic processing pipeline into cleanly separated abstractions: Classify → Score → Compose → Distill → Deliver. Each stage now operates independently with well-defined boundaries, making the engine dramatically easier to extend, test, and debug. This modular architecture is the foundation that will carry OMNI through its next generation of features.

Zero-Cost ANSI Processing

Performance engineering received dedicated attention in this release. The strip_ansi function — called on every single line of input — was refactored to leverage Rust’s Cow<str> memory strategy. The result is zero heap allocations for clean text that contains no ANSI escape codes, which accounts for roughly eighty percent of all input. For the remaining twenty percent, the allocation is minimal and precisely scoped. In benchmarks, this change alone reduced per-line processing overhead by forty percent.

Lessons Learned: Silent Exits and Security Hardening

Several important bugs were resolved in this release. OMNI now terminates silently on completely blank piped inputs, eliminating the stderr pollution that was confusing automated CI pipelines. The security guardrail layer was updated to ensure all denylist environment variable queries are strictly case-insensitive — preventing a subtle bypass vector where OMNI_PASSTHROUGH could be evaded by setting omni_passthrough. We also expanded Windows CI coverage, matrix-testing across operating systems to catch platform-specific regressions before they ship.

The Eradication of Undefined Behavior

Perhaps the most technically significant fix in this release is the eradication of std::env::set_var from the test suite. In Rust, mutating environment variables across threads is undefined behavior, and our parallel test runner was silently relying on it. We replaced every instance with proper dependency injection, completely eliminating a class of intermittent test failures that had been haunting the CI pipeline for weeks. This is the kind of invisible, deeply unglamorous work that separates production-grade infrastructure from prototype-quality code.

What Comes Next

With session awareness embedded in the engine’s DNA, OMNI is no longer a stateless text filter — it is a contextual reasoning layer for your AI agent. The next frontier is predictive context: using session patterns to pre-fetch and pre-score relevant data before your agent even submits its next request. The stateless era is over.