Turning private work files and code into reliable, evidence-backed context for AI.
Systems Architect | 15+ years building mission-critical infrastructure | AI Harnessing & Context Engineering
Reasoning is improving fast. Reliable context is still the bottleneck.
- Now: building DocMason, a local-first, evidence-first knowledge base for AI-assisted deep research over private work files.
- Before: built code-intelligence systems across llama-github, LlamaPReview, and repo-graph-rag.
- Direction: the Mason ecosystem—moving from deep document analysis to multi-stage pipelines that generate editable, consulting-grade native deliverables.
Over 15 years of architecting mission-critical systems, the recurring failure mode is always the same: in high-stakes environments, being "almost right" is useless. The bottleneck to reliable output—whether from humans or AI—is rarely raw reasoning capacity. It is context precision. That constraint drives everything I build.
DocMason is my current open-source focus: a local-first, provenance-first knowledge base for AI-assisted deep research over private work files. It is not a document chatbot. It compiles unstructured artifacts into knowledge infrastructure that agents can actually use. Its native operating pattern is simple: the repo is the app, and Codex is the runtime.
Core architectural priorities:
- Deterministic ingestion: parsing PDFs, decks, spreadsheets, emails, and repository-native text without silent failures.
- Reliable outputs: provenance-first retrieval instead of vague, hallucination-prone document chat.
- Actionable output: extending the Mason ecosystem beyond extraction. The next step is a deterministic, multi-stage pipeline separating semantic narratives from visual layout DOMs, turning deep document analysis directly into editable, consulting-grade native presentations (PPTX) for serious white-collar work.
I came to document intelligence through code intelligence.
- llama-github: the retrieval substrate, built to give LLMs GitHub-native context instead of raw repository dumps.
- LlamaPReview: field validation for that thesis. It achieved a measured 61% signal-to-noise ratio in AI code review across 4,000+ active repositories (35K+ combined stars).
- repo-graph-rag: the Code Mesh research artifact, exploring deterministic repository graphs and traversal-first retrieval.
- llamapreview-context-research: formalizing the exact failure mode of Context Instability.
This path started with helping AI understand code diffs, but led to a broader conclusion: the real computing frontier is shifting toward understanding full knowledge environments and generating high-stakes output from them. Code Mesh was the logical end of one inquiry, but not the final product surface.
By late 2025, the fundamental physics of software engineering began shifting beneath our feet.
It became clear that reactive code review would not remain the terminal surface of AI engineering. Even advanced code-graph traversals—which I explored via repo-graph-rag a year before the current industry hype wave—were solving a problem that was rapidly moving upstream. As autonomous agents and "vibe coding" accelerated, the sheer generative volume made analyzing post-facto diffs feel like patching a leaky pipe while the plumbing was being replaced.
The scarce problem was no longer analyzing code after it was written. The real frontier shifted toward commanding entire knowledge environments, preventing architectural collapse, and ensuring agents operate under deterministic constraints before generation happens. This is why I chose not to commercialize my early graph-based intelligence tools, and instead pivoted my focus toward proactive context infrastructure and deterministic visual output pipelines.
Whether building traditional software or complex multi-agent systems, open-ended "vibe coding" hits a scaling wall. The bottleneck isn't generating code; it is preventing architectural collapse as AI-driven mutations accumulate.
To solve this, I formalized a universal paradigm for AI harnessing engineering—the Dao / Fa / Qi / Shu of agentic coding. It shifts AI from an open-ended conversational copilot to a strictly harnessed actor within a deterministic system:
- Dao (Direction & Value): Defining the invariant product boundaries. Without Dao, AI optimizes for local illusions of progress, building features that demo well but corrupt the long-term architecture.
- Fa (Runtime Law & State): Governing identity, state transitions, and truth surfaces. Without Fa, AI silently hallucinates state, conflating generated projections with canonical authored truth.
- Qi (Machinery & Control): The actual subsystems (controllers, commit barriers, projection layers) that enforce the laws. Without Qi, the rules only exist on paper, and the system relies on human vigilance to prevent AI drift.
- Shu (Execution & Phasing): The deterministic sequence of implementation and validation. Without Shu, AI coding devolves into an endless loop of patching symptoms instead of shipping structural phases.
For a deep dive into building sustainable agent operating surfaces, see my framework on Vibe Coding and Governance.
- I Built repo-graph-rag Before the Code Graph Wave
- Vibe Coding Is Not Prompting. It Is Governance.
- Why I Killed My AI Code Review SaaS (4,000+ Repos) Right Before Monetization
- Drowning in AI Code Review Noise? A Framework to Measure Signal vs. Noise
- Beyond the Diff: How LlamaPReview Catches Critical Bugs
I build systems that move agents from merely reading documents to executing serious knowledge work end-to-end.


