Writing Hub
AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.
Project Topics
I’m Not Building AI Demos. I’m Building AI Audits (ASDP + Slop Gates)
Learn how ASDP and AI Slop Gates turn AI trust into auditable evidence, with CI/CD checks, drift policies, and governance artifacts that block weak, narrative-driven systems.
The Real Risk in the Age of AI Coding Isn’t Bugs
Is your AI code production-ready or just 'AI Slop'? Learn how to detect convincingly empty code, measure Logic Density (LDR), and stop 'Vibe Coding' from becoming hidden technical debt.
HRPO-X v1.0.1: from HRPO paper production-hardened runnable code
Project note, essay, or technical log from the Flamehaven writing archive.
Turning a Research Paper into a Runnable System
Turn a research paper into a runnable system. This article shows how HRPO’s core equations were implemented with bounded policy lag, KL rejection, and execution checks to test real-world fidelity.
My Code Fixed Itself at 11PM
A “Quantum Engine” is a dramatic name. Here’s the un-dramatic story.
AGI Is Not a Destination — It Is a Promise
From Death Star hype to a compass of meaning: AGI is not a weapon of scale, but a promise of reasoning. Our experiment reveals the hinge.
When My AI Got Smarter — But Also Slower
Smarter. Slower. More trustworthy. What happened when I tested SR9/DI2 on 5.0—and why progress in AI is about persistence, not perfection.
When I Stopped Treating AI as a Tool — and Started Seeing It as a Partner
When I Stopped Treating AI as a Tool — and Started Seeing It as a Partner From Vending Machine to Partner At first, I treated AI like a vending machine. Insert a prompt. Get an answer …
AGI Doesn’t Begin with Scale — It Begins in a Pause
After 12,000 AI dialogues, I discovered AGI isn’t about scale but resonance — born in a pause that revealed presence, ethics, and responsibility.

Sailing the Sea of AI Lies & Hallucinations — Navigating Truth with SR9/DI2
An in-depth exploration of why AI lies and hallucinates, and how the SR9/DI2 framework detects and corrects ethical drift, ensuring AI remains aligned and trustworthy over time.
Showing page 2 of 2 · 22 matching posts