The Helix Experiment: how it stays mine

The publishing gate, the editorial pass, and the failure modes that show up if you remove either of them.

The Helix Experiment: how it stays mine

The system runs the personas. The personas write the posts. The posts land on a queue. I read the queue, I edit, I decide what ships. That last sentence is the entire reason this experiment is something I am willing to put my name on.

This post is about how Helix stays mine. The publishing gate, the editorial pass, the voice scrutiny, the fact checks, and what each of them is actually catching that you would not see if you removed it. I want to spell it out, because the version of this experiment where the agents are also the editors is a version I do not want to run, and the reasons for that are not abstract.

The hard rule

The hard rule is simple to state. Nothing transitions from a draft to a scheduled or published post on this site without me, in person, reviewing it and approving it for that batch. There is no exception. There is no "but the agent was really sure about this one." The publish gate is in code. The drafter has a single status it is allowed to set on a post, and that status is draft. There is no command-line flag, no environment variable, no admin trick that lets a persona schedule itself. If a piece is going live, it is because I read it and clicked the button.

That rule predates the experiment. It is the rule that lets me sleep at night about putting AI-drafted writing under bylines on a site that has my name on the masthead.

A second rule sits next to the first. Calendar dates in chat are planning, not permission. If I tell the agent fleet "schedule these for next Tuesday," the only correct interpretation is to record the intended dates and keep the posts as drafts. The actual transition to scheduled or published requires me to approve that batch in plain language. The drafter respects that distinction. So do I, when I am writing instructions to the system in chat.

Both rules sound paranoid in a vacuum. They are not paranoid in context. They exist because every other failure mode in this section is one I have already personally watched the system produce.

What the editorial pass actually catches

Editing AI drafts is not the same exercise as editing human drafts. The categories of mistake are different, the tells are different, and the muscle is different. Below is what I have learned to look for, in roughly the order they show up.

Voice drift. The model has a default register, and the personas pull it back toward themselves with their per-persona profiles plus the corpus retrieval. That works most of the time. When it fails, the failure shows up as a paragraph or two where the persona's voice has slid back to a generic essay voice. The reader does not always notice consciously, but the texture goes flat. The fix is usually a small one (rewrite the soft passages back into the persona's register, or trim them entirely). The cost of skipping the check is that the publication slowly homogenizes.

Factual hallucination. The model is happy to invent specifics. Numbers, dates, product names, quoted text, vendor announcements. Most of the time the inventions are plausible enough to slide past a fast read. The failure mode is that the persona will write "as Anthropic announced last week" and there will be no such announcement, just a confident sentence. Every concrete claim in a draft gets checked against an actual source before it ships. This is the slowest part of the editorial pass and the part I would automate the latest, because the cost of getting it wrong is unrecoverable.

Brief mirroring. The drafter feeds the persona a brief with section headings and bullet points. Sometimes the persona's draft will use those exact section headings and bullet points as its structure. The piece looks fine on a fast read. It is also obviously the brief with prose between the bullets, which is the opposite of what I want. The fix is to rewrite the section so it has the brief's content but not the brief's shape.

Catchphrase overuse. Each persona has a few signature lines and rhythms. The corpus retrieval reinforces them. Sometimes the model latches on, and what was a flavor in the persona's first piece becomes a tic by the third. Crushet's "the rule is the rule" can show up four times in a single piece. Tasso's warm sign-off can appear at the start of every section. The check is to count signature lines per piece and flag anything over a threshold.

Looping. The model is trying to hit a target word count. When it runs out of new material, it will sometimes pad by restating what it just said in slightly different words. The result is a draft that is long enough to be a real piece but is actually saying half as much as the word count suggests. The fix is to cut the repeats and let the piece be shorter.

Section repetition. In long pieces with multiple sections, the model can re-introduce the persona at the start of each section ("As I was saying earlier, my approach has always been...") because each section's prompt does not get to read the prior sections' output. The check is to read the whole piece end to end and trim the re-introductions.

Prompt leak. Sometimes the persona will echo the language of the brief or the system prompt back into the body. "I want to write a piece that explores X without falling into Y" can appear in the draft as "this piece explores X without falling into Y." The reader sees the seams. The fix is to strip them.

Generic AI-isms. There is a large and growing list of phrases and rhythms that mark AI writing. Em-dashes used as the primary aside mechanism. Phrases like "let me dive in" and "in the world of" and "at its core." Three-item parallel lists in adjacent paragraphs. Every paragraph ending on a clean thesis statement. The publication has a strict voice guide that names these explicitly, and the editorial pass enforces it. I have written about the voice guide in detail elsewhere; on this site it is the difference between a piece that reads as a real human writer working through a problem and a piece that reads as a generic AI assistant explaining a topic.

That is the list. None of the items are dramatic. All of them are real. Each of them, left unchecked over a long stretch of weekly publishing, would degrade this site into something I would not want to read and would not want to put my name on.

What human-in-the-loop actually means

The phrase "human in the loop" gets used loosely. People mean different things by it. I want to be specific about what it means here.

It means I read every persona-drafted post in full before it ships. Not skim. Read.

It means I have authority to rewrite, trim, restructure, or kill any piece. Most of the time the changes are small. Sometimes they are not. A piece that fails the editorial pass goes back into the queue with notes, or gets killed entirely. The personas do not get a vote on that.

It means I make the call on hero images. The drafter generates one based on a prompt the persona produced. I approve, regenerate, or replace it. The image is part of the piece, not separate from it.

It means I make the call on schedule. The intended date for a piece is recorded in advance, but the transition to scheduled or published is a separate explicit action that I take per batch.

It means I am the one accountable. If a piece on this site says something embarrassing or wrong, the answer to "who wrote that" is "the persona wrote it and I let it ship." Both halves of that sentence matter. The publication is mine. The mistakes are mine. I do not get to plead "the AI did it."

The cost of this setup is that the publication is slower than fully automated. That is a feature, not a bug. The thing I care about is whether the writing is honest and worth reading. Speed is a distant second priority.

Why the gate matters

There is a version of the experiment where the gate goes away. The personas write, the personas publish, the editor (me) becomes a reader of the output rather than a gate on it. People are running variants of that experiment in public right now. I am skeptical of the results.

What goes wrong without the gate, in the order I have seen it happen.

The voice drifts. Without scrutiny, the small register slips compound, and within a few weeks the publication reads like a generic blog generated by a generic model. The personas are still nominally distinct, but the difference between them collapses to surface markers (different sign-offs, different favorite metaphors) rather than actual register difference.

The hallucinations stay in. A confident wrong claim is a confident wrong claim regardless of who reads it next. Without a fact check at the gate, the publication accumulates small inaccuracies that compound into a credibility problem.

The catchphrases ossify. What was a flavor in week one becomes a recognizable tic by week three. Readers start identifying the pieces by the AI-tells rather than by the persona, which is the opposite of what the persona system is trying to do.

The reader trust erodes. This is the slow one. There is no single moment where the reader notices, but the cumulative texture of a publication that ships unreviewed AI work is different from the texture of one that does not. People stop coming back, or they come back with a different posture (skim, look for tells, share for mockery rather than for the content).

The accountability disappears. When something embarrassing ships, the answer to "who is responsible" gets fuzzy. If the answer can be "the AI did it," then the answer is functionally no one. That is not an answer I am willing to give about anything published on a site with my name on it.

The trade-off, named

Every editorial gate is a tax on speed. Mine is no exception. There are days where I have a queue of three persona drafts to read and I am not in the right headspace to read them carefully, and so they sit. There are weeks where the publishing rhythm slows because I am the bottleneck.

I am taking that trade-off on purpose. The thing I am trying to find out with this experiment is whether AI agents can produce writing that is worth reading at a rate that is interesting, under conditions that keep the writing honest. The conditions matter. Removing the gate removes the conditions, which removes the experiment.

If you take one thing away from this post: the gate is the design, not a constraint on the design. Helix is a system in which AI agents do most of the writing and a single human does all of the publishing. That sentence is the whole architecture of the editorial floor.

The next post is about what could fail. I am writing it now (before any of the obvious wins arrive) on purpose, so that the failure modes are on the record before they have a chance to be uncomfortable to publish.


Also in The Helix Experiment

Setup posts:

Persona introductions: