We Built an AI Content Pipeline and It Shipped Slop. The Structural Fix Wasn't an Agent.
In early 2025, we did exactly what we tell our clients not to do: we automated a process without a "Decision Chain" and shipped AI slop. Discover why the fix wasn't a better AI agent, but a structural reporting line change that gave an Editor a veto the Director couldn't override.
Published April 22, 2026
The piece that makes our content pipeline stop shipping slop isn't any of the AI agents we introduced. It's the Editor reporting to me directly, with veto the Director cannot override. One structural decision. No agent involved.
We learned what happens without it the hard way. In February 2025, architech.ca published six articles on customer experience inside three weeks. Five were near-duplicate SEO templates under a generic "Architech" byline. Two were near-identical text under different titles. All five shared the same telecom scenario, the same recycled Gartner stats, and the same "Schedule a 30-minute consultation" close. All five targeted CIOs and CTOs, not the operator who actually buys our work. None contained Architech-specific proof. We took them down when we relaunched the site in April 2026.
An AI-first services firm, writing publicly about AI, using AI as its primary author, produced the exact slop pattern the industry describes. No editorial veto between the draft and the publish button. That cluster is why this piece exists.
If you run operations at a mid-market company, you have been told by AI-services firms (including, until recently, ours) that you should be careful about where AI touches your workflows. Most generative AI pilots stall before production. MIT's NANDA initiative found 95% of enterprise GenAI pilots fail to deliver measurable revenue, in the 2025 State of AI in Business report as covered by Fortune. Gartner's Rita Sallam told Computer Weekly in July 2024 that at least 30% of GenAI projects would be dropped after proof of concept by the end of 2025. The failure mode is well documented. The advice is consistent. The firms selling the advice are rarely asked whether they have redesigned one of their own workflows the same way.
We hadn't. The February 2025 CX cluster is the evidence.
AI-services firms don't publish their own redesigns
AI-services firms usually keep their own workflow redesigns private. You redesign workflows for clients. You talk about the client outcomes. You don't publicly document a workflow you've redesigned on yourself, because the internal operation isn't positioned as proof. Proof is the case study.
That posture produces a credibility gap for AI-services firms specifically. When 95% of enterprise pilots stall, and the 5% that succeed are, per MIT NANDA's Aditya Challapally, the ones who "pick one pain point, execute well, and partner smartly with companies who use their tools," the operator has a reasonable question. Do you use your own tools? Have you run the discipline you are selling me on yourself? On what?
We could not answer in good faith in February 2025. The CX cluster was the evidence we could not.
What we rebuilt: an editor the Director can't override
What let the slop ship. In February 2025 there was no editor in the content function at all. No named editor, no named author, no review gate. AI-primary drafts went to publish without anyone who could say no. The editorial role had been eliminated. The marketing team that had sustained the 2022 editorial cadence no longer existed. It is a plain and primitive failure. Not a power-dynamic story. The thing that stopped the slop was missing.
Why any multi-agent AI pipeline is vulnerable to the same failure. Put the editorial role back in and a new problem shows up as soon as the pipeline runs at volume. In any multi-agent pipeline producing content at a cadence, ship pressure will beat editorial judgment unless structure prevents it. A Director who is accountable for shipping and an Editor who is accountable for quality, both reporting to the same person, resolves itself against quality most of the time. The predictable failure mode isn't "editor absent." It's "editor overruled."
The design decision. We rebuilt the pipeline as a set of role-separated agents. A Director sequences work. A Researcher grounds claims. A Writer drafts. An Editor reviews. A Publisher runs only after a human signs off. And the Editor reports to me directly, not to the Content Director who is responsible for shipping. The Editor has veto. The Director cannot override. Only I can.
Four of the five architectural decisions behind that pipeline are real but, for an operator, mostly housekeeping. LinkedIn and Teams variants derive from the source artifact directly, not from the blog draft, meaning the AI agents work alongside each other off a shared source rather than passing a baton down a chain where each handoff loses fidelity. No telephone-game degradation across channels. Humans pick topics in Phase 1, not agents. The orchestrator is Claude Code subagents, which are spawned and sequenced by a top-level agent rather than running in an external workflow tool. Voice and positioning files are curated views of our internal operating system, synced rather than copied.
The fifth is the one worth a paragraph. The reporting line is the forward-looking fix, not a retrofit of the 2025 failure. February 2025 is what happens when the editorial role is gone. The reporting line is what keeps the editorial role intact once it is back and the pipeline is running at scale. That is the step that mattered. The agents are not the claim. The reporting line is.
We don't have throughput numbers yet. Here's what we have instead.
This is where a piece like this usually drops in a quantified outcome. We don't have one yet. The pipeline has been running in its current form for a short time. Our build log is sparse. Our baseline file (the one that will hold the actual before-and-after numbers) is not yet populated. If you see a piece from an AI-services firm claiming a 10x throughput improvement on its own content pipeline three weeks into a rebuild, read it with the scepticism you would bring to any other unmeasured vendor claim.
What we have instead are three observations grounded in how the pipeline actually runs.
The first is the February 2025 cohort against the 2022 cohort. In calendar year 2022, architech.ca published 14 blog posts authored by named humans from delivery and leadership, including our COO and CTO, at roughly 1.2 posts per month. The editorial cadence was sustained by a marketing function that no longer existed by early 2025. An audit of the 163-post Architech archive from 2016 to 2025 found four posts that clear the current rubric for republish. The highest-quality cohort is 2022. The lowest-quality cohort is the February 2025 CX cluster. The variable that changed between them was not AI. It was whether the editorial role existed at all.
The second is the shape of the failure described by practitioners who run AI-assisted content well. The documented pattern puts humans at every gate: first-draft review, fact-checking, citation verification, voice editing, and final approval. Practitioners warn explicitly that AI "sometimes invents citations that don't exist." This is not a flourish on top of the AI. Those gates are what the pipeline depends on. Skip any of them and the output degrades toward the February 2025 cluster.
The third is the discipline of firms that are good at running their own tools. Use the thing. Let the use reveal the failure modes. Let the failure modes force the design.
Anthropic runs Claude Code internally at 70 to 80 percent daily use, per Cat Wu, across engineering, security, legal, and accounting. Not just the teams you would expect. Cursor's Aman Sanger frames dogfooding as how the team stays "honest to ourselves of whether we find it useful" before shipping. 37signals' DHH calls it a quality floor: "There's a baseline of quality you derive from something that the people who are working on it also have to use. It can't just be broken." None of these firms is an AI-services firm. None would describe the practice as Customer Zero. The pattern is the same.
Content isn't claims processing. The failure mode still generalizes.
The strongest argument against publishing this piece is that a content pipeline is not a claims-processing workflow. Our failure mode (slop articles on a corporate blog) is low-stakes compared to a $50M-revenue operator deciding whether to let AI touch a customer onboarding flow. The structural differences are real. Our publication cadence is internal. We have no external SLA. We are a small team with one senior orchestrator who holds a lot in his head. That doesn't generalize to a 2,000-person operations function.
I think the counter is right on structure and wrong on the failure mode. The February 2025 CX cluster is the same failure mode MIT NANDA and Gartner describe in enterprise pilots: templated output, no first-person proof, no human veto between the draft and the publish. The domain differs. The mechanism does not. If an AI-services firm cannot detect and stop this pattern in its own content operation, the operator has a reason to ask whether it can detect and stop it in theirs.
A second counter is worth naming. Our operating system warns explicitly against over-generalizing Customer Zero to client contexts. That warning matters. What we learned about editor veto in our content pipeline is a specific lesson from a specific workflow. It is not a template. What transfers is the question, not the answer. Which step in your AI-assisted workflow, if you removed the human who could say no, would let the thing ship anyway?
The question to take to your own AI vendors
If you are evaluating AI-services firms right now, the signal worth looking for is not the language on the services page. It is whether the firm can describe a workflow of its own, named and specific, with decisions you can read, that it has redesigned with AI under the same discipline it is proposing to apply to yours. If it cannot, the advice you are being sold has not been pressure-tested in the conditions it is prescribing.
Ours had not been, until this. If a firm's answer to that question is "we haven't done it on ourselves yet," that is not a disqualifier. It is honest. "We have always done this" without a named workflow is the answer to watch out for.
We will have production numbers in a few weeks. When we do, they go in a follow-up, with the build log behind them. If the redesign doesn't move the numbers, that goes in the same follow-up.
If you want to see the decisions behind this pipeline while it is still forming (the editor-veto rationale, the sibling architecture, the choice to drop the external orchestrator), reply to this piece or reach out directly. The part we got wrong on the first try is the part most worth talking about.
One question to take to your own team this week: who has the authority to stop a ship date because the output is wrong, and does that person report to someone being measured on the ship date? If the answer to the second half is yes, the editorial veto is not structural. It is a favour the person with the deadline is choosing to grant, and favours get withdrawn under pressure.
Ready to apply this to your workflows?
Architech's AI Jumpstart is the structured entry point.