Writing Hub
AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.
Project Topics
After Auditing 10 Bio-AI Repositories, I Think We're Scaling the Wrong Layer
After auditing 10 open-source Bio-AI repositories, one pattern stood out: the field is scaling packaging faster than verification. Here is what that gap actually costs.
Everyone Was Talking About Context Engineering. Nobody Had Solved Governance.
Everyone Was Talking About Context Engineering. Nobody Had Solved Governance.
The Model Already Read the README. MICA v0.1.8 Made It a Protocol
v0.1.7 made scoring a contract with fail-closed gates. v0.1.8 recognized that README-first behavior could serve as invocation — and formalized it as a schema-level protocol. This article uses simplified examples to show how the invocation gap that had existed since v0.0.1 was finally closed
From Fail-Closed Blocking to Reproducible PASS/BLOCK Separation (EXP-032B)
A validation study showing how EXP-032B achieved reproducible PASS/BLOCK separation across A/B/C control arms by patching false-blocking causes, improving observability, and measuring replay drift under observer-shadow conditions.
LOGOS LawBinder: From Governed Reasoning to Audit-Grade Execution
This article explains how LOGOS v1.4.1 improves production AI reasoning with multi-engine orchestration, complexity-aware governance, and audit-friendly failure tracing.
Turning a Research Paper into a Runnable System
Turn a research paper into a runnable system. This article shows how HRPO’s core equations were implemented with bounded policy lag, KL rejection, and execution checks to test real-world fidelity.
Why AI Dismisses Your Best Work in One Second
Why do AI models dismiss original work in seconds? This essay explores the hidden mechanics of AI skimming—shortcut learning, probabilistic safety, fast-thinking defaults, and why depth requires time.
AGI Is Not a Destination — It Is a Promise
From Death Star hype to a compass of meaning: AGI is not a weapon of scale, but a promise of reasoning. Our experiment reveals the hinge.

Sailing the Sea of AI Lies & Hallucinations — Navigating Truth with SR9/DI2
An in-depth exploration of why AI lies and hallucinates, and how the SR9/DI2 framework detects and corrects ethical drift, ensuring AI remains aligned and trustworthy over time.