Writing Hub
AI governance essays, reasoning systems notes, experiment logs, and technical writing across BioAI and engineering practice.
Project Topics

When AI Models Fight, Truth Wins: The “Eureka” Moment for Tired Researchers
To the grad student staring at a pLDDT of 90 and wondering why the ligand won’t bind.

From 97% Model Accuracy to 74% Clinical Reliability: Building RSN-NNSL-GATE-001
Learn how RSN-NNSL-GATE-001 turns high model accuracy into system-level clinical reliability by blocking unsafe AI pipeline decisions, measuring end-to-end risk, and enforcing fail-closed governance.

When Adding Chai-1 and Boltz-2 Exposed Hidden Model Disagreement(Trinity Protocol Part)
See how adding Chai-1 and Boltz-2 to an AlphaFold workflow exposed hidden model disagreement, increased drift, and revealed why failed convergence can be the most valuable signal in computational biology.

Orchestrating AlphaFold 3 & 2 with Python: Handling AI Hallucinations using Adapter Patter (Trinity Protocol Part 1)
Learn how to orchestrate AlphaFold 3 and AlphaFold 2 with Python using the Adapter Pattern to detect AI hallucinations, measure structural drift, and improve protein prediction reliability.

Your Agentic Stack Has Two Layers. It Needs Three.
Most agentic stacks cover tools and skills, but miss intent governance. Learn why a third layer is needed to stop AI drift, scope creep, and technically correct systems heading in the wrong direction.

I Integrated AlphaFolder3 & AlphaGenome. It Looked Perfect. Then It Failed the "Honesty Test."
A real-world experiment integrating AlphaFold3 and AlphaGenome revealed a critical lesson: AI predictions that look perfect can still fail the ‘honesty test.’ A deep dive into bioinformatics, model validation, and AI reliability in drug discovery.

The AI Flight Crash: Why 2026’s Hottest Papers Can’t Take Off — and what actually ships
Langley spent $50,000 and sank — the Wright Brothers flew for <$1,000. Here’s a 4-week build plan I’ve seen actually ship.

Why Reasoning Models Die in Production (and the Test Harness I Ship Now)
Project note, essay, or technical log from the Flamehaven writing archive.

AI Agents Are Poisoning Your Codebase From the Inside
Explore how AI-generated code can silently degrade software quality through weakened tests, rising code churn, and duplication—and how teams can prevent it with better governance.

How Failing in 2 Hours Saved 8 Months of Drug R&D: Engineering a "Truthful Null" with Upadacitinib
A bioinformatics case study on Upadacitinib showing how SR9 stability scoring and drift analysis exposed lipid carrier incompatibility early, saving months of drug delivery R&D

RExSyn Nexus 0.6.1 - Stop Hallucinating Proteins: How We Built a 7D Reasoning Engine with AlphaFold3
RExSyn Nexus 0.6.1 adds Structure as a 7th reasoning dimension, using AlphaFold3 confidence signals to reject biologically plausible but physically impossible protein hypotheses with deterministic, auditable validation.

Implementing "Refusal-First" RAG: Why We Architected Our AI to Say 'I Don't Know'
Implementing refusal-first RAG means teaching AI to say “I don’t know.” This article explains evidence atomization, Slop Gates, and grounding checks that favor verifiable answers over plausible hallucinations.
Showing page 3 of 8 · 85 matching posts