Writing Hub
AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.
Project Topics

I Audited 10 Open-Source Bio-AI Repos. Most Could Produce Outputs. Few Could Establish Trust.
I audited 10 visible repositories. Most could produce outputs. Very few could establish what those outputs meant.

Medical AI Repositories Need More Than Benchmarks. We Built STEM-AI to Audit Trust
STEM-AI is a governance audit framework for public medical AI repositories. It scores README integrity, cross-platform consistency, and code infrastructure — because benchmarks alone don't tell you if a bio-AI tool is safe to trust.

My LLM Kept Forgetting My Project. So I Built a Governance Schema.
Session loss isn't a UX inconvenience — it's a structural failure with compounding consequences for long-running AI projects. This post defines the problem precisely and introduces MICA, a governance schema for AI context management.

95% of AI Businesses Will Die. Here’s How to Not Be One of Them.
What the data, a founder’s confession, and 70 years of tech history tell us about who actually survives.

Is MCP Really Dead? A History of AI Hype — Told Through the Rise and Fall of a Protocol
When a protocol doesn’t die — it just stops being interesting. A forensic look at MCP, OpenClaw, and the psychology of AI hype cycles.

Prompt, Pray & Push: Why Your AI Agent Keeps Failing You
The one concept that turns expensive spaghetti into great agentic engineering.

The Pull Request Illusion: How AI Is Hollowing Out Software’s Last Line of Defense
GitHub Just Added a Switch to Turn Off Pull Requests. That’s Not a Feature. It’s a Warning.

Your Agentic Stack Has Two Layers. It Needs Three.
Most agentic stacks cover tools and skills, but miss intent governance. Learn why a third layer is needed to stop AI drift, scope creep, and technically correct systems heading in the wrong direction.

AI Agents Are Poisoning Your Codebase From the Inside
Explore how AI-generated code can silently degrade software quality through weakened tests, rising code churn, and duplication—and how teams can prevent it with better governance.

AI Isn’t Killing Your Expertise. It’s Just Moving the Paywall.
Why ‘Writing Faster’ Is Worthless When Nobody Can Verify What’s True

LawBinder v1.3.0: Governance as a Kernel (Not a Guardrail)
LawBinder v1.3.0 shows how AI governance can run like a kernel, using deterministic Rust-based enforcement, replayable audit signatures, and bounded-latency policy checks in the critical path.

Why I Stopped Treating Complexity as a Bug
On intent, governance, and why “clean code” heuristics fail in AI-generated systems
Showing page 1 of 2 · 21 matching posts