Manwe 25 Apr 2026

AI can draft code, tickets, QBRs, and internal docs. Where does speed actually create advantage, and where does it just hide confusion?

AI speed creates genuine advantage in high-volume, well-scoped drafting — code boilerplate, internal docs, templated tickets — and becomes a liability in high-stakes narrative artifacts like QBRs or customer-facing commitments. The decisive evidence: speed only compounds advantage when requirements are already understood; when they aren't, AI doesn't surface the confusion — it forwards it downstream with a professional coat of paint that makes it harder to catch. The practical test is not "can AI draft this faster?" but "does anyone actually understand the problem before the draft exists?" Apply that filter and the use cases sort themselves immediately.

Generated with Claude Sonnet · 51% overall confidence · 5 advisors · 5 rounds
By Q4 2026, at least 3 major publicly reported post-mortems at software companies with >500 engineers will trace a production incident to AI-generated boilerplate that propagated an architectural antipattern across multiple services before detection — with the root cause explicitly cited as 'pattern normalization at scale.' 71%
By end of 2026, teams using AI to draft QBRs and customer-facing commitments without a mandatory human 'requirements clarity checkpoint' will show a statistically measurable increase (≥20%) in customer escalations tied to misaligned expectations, compared to teams that use AI only for internal/templated artifacts. 67%
By mid-2027, engineering organizations that formally restrict AI-assisted drafting to well-scoped, requirements-complete tasks (e.g., via a documented definition-of-ready gate) will report ≥30% faster defect detection cycles compared to teams with unrestricted AI drafting policies — with the gap widening in codebases where AI adoption started before 2026. 62%
  1. This week — before your next QBR, proposal, or customer-facing doc goes out — implement a two-line gate at the top of every AI-drafted artifact: "AI-assisted draft — reviewed by [name] on [date] — flagged assumptions: [list]." Make the flagged-assumptions field mandatory and non-empty. If your reviewer can't name at least one assumption the AI made, the doc hasn't been reviewed — it's been read. This single change forces genuine engagement and creates a paper trail if the fluency-as-commitment liability Rachel named ever surfaces legally.
  2. Today, identify the three highest-volume AI drafting use cases your team is currently running (likely: ticket writing, code scaffolding, internal status updates). For each one, pull five recent outputs and answer: Was the underlying requirement actually understood before the draft was generated? If you can't answer yes with evidence for more than two out of five, stop using AI for that use case until you've documented the requirement spec process that must precede it. Say to your team lead: "I want to audit our last two sprints of AI-generated tickets for requirement clarity before we expand this to the rest of the team — can you block 90 minutes with me by end of next week?"
  3. Within the next two weeks, measure your error clustering. Pull every rework, escalation, or customer complaint from the last 90 days and flag which ones involved an AI-drafted artifact. Don't ask "was AI the cause" — ask "did the polish of the artifact delay detection?" If more than two incidents show a pattern where a wrong assumption survived longer because the doc looked finished, you have confirmed Issam's SLA story in your own org. Bring that data to your next leadership sync.
  4. Immediately classify your artifact types into three tiers — not by content but by consequence of confident error: Tier 1 (internal, reversible: meeting notes, draft tickets, code boilerplate), Tier 2 (cross-functional, semi-permanent: architecture docs, internal SLAs, roadmap commitments), Tier 3 (external, legally or commercially binding: QBRs, proposals, customer emails with commitments). AI runs freely on Tier 1. Tier 2 requires a named human reviewer who initials the flagged-assumptions field. Tier 3 requires a pre-draft alignment checkpoint — a 15-minute sync before the draft exists to answer: What is the one thing that, if wrong, kills this deal or this project? That answer must appear in the doc before AI touches it.
  5. Before your next enterprise-facing deliverable, run this specific test: give the same brief to your AI tool and to your most skeptical team member independently. Compare what each flags as uncertain. Any uncertainty your human flags that the AI didn't flag is a gap the AI papered over. If that gap touches pricing, SLAs, delivery timelines, or technical constraints — you nearly sent a polished commitment you didn't mean to make. Document the gap. If this happens twice in a month, mandate the dual-track test for all Tier 3 artifacts going forward.
  6. Set a 90-day checkpoint — by July 25, 2026 — to present your leadership team with three numbers: volume of AI-drafted artifacts by tier, rework/escalation rate by tier, and one confirmed case where the pre-draft alignment checkpoint caught a confident error before it shipped. If you can't produce the third number, your review process is decorative. If the rework rate on Tier 2 or 3 artifacts hasn't dropped or held flat, the categorization itself is wrong and needs to be rebuilt from your actual incident data.

Divergent timelines generated after the debate — plausible futures the decision could steer toward, with evidence.

🔒 You gated AI drafting behind a definition-of-ready checkpoint
24 months

You formalized a requirements-clarity gate before any AI-assisted drafting, restricting AI to well-scoped, internally aligned tasks only.

  1. Month 3Rollout friction is high — engineers resent the gate, and ticket throughput drops 15% as teams are forced to articulate acceptance criteria before drafting. Two senior engineers quit, citing 'process bloat.'
    The Contrarian warned that the forcing function was always review culture, not the writing — your gate surfaces that this review culture was already broken and now has nowhere to hide.
  2. Month 7Defect detection cycles begin tightening measurably; QA catches ambiguous requirements pre-ticket rather than post-merge. The team ships 22% fewer features but reverts 40% fewer PRs.
    The evidence predicts organizations with documented definition-of-ready gates will report ≥30% faster defect detection cycles vs. unrestricted AI drafting teams (62% confidence by mid-2027).
  3. Month 13A competitor's post-mortem goes public — their AI-generated boilerplate propagated a caching antipattern across 6 microservices before detection, causing a 9-hour outage. Your gated process becomes a recruiting and trust differentiator.
    The evidence predicts at least 3 major post-mortems by Q4 2026 tracing production incidents to AI-generated boilerplate normalizing architectural antipatterns (71% confidence).
  4. Month 20Customer escalations tied to misaligned expectations are down 31% year-over-year; QBRs go through a mandatory human narrative review before delivery, and one enterprise renewal is explicitly credited to 'clarity of commitments.'
    Rachel Wong's warning: pre-AI, a wrong number reads as a draft; post-AI, the same number in a polished doc reads as a commitment — your checkpoint absorbed that liability delta.
🚀 You deployed AI drafting across all artifacts with no gatekeeping
18 months

You rolled out AI-assisted drafting company-wide — tickets, QBRs, internal docs, customer commitments — prioritizing speed and volume over process gates.

  1. Month 2Ticket throughput doubles, documentation review time drops 65%, and the leadership team declares an internal win. Engineers are shipping boilerplate scaffolding 3x faster across five new services.
    The Auditor confirmed evidence shows AI cuts document review time by 70% and improves accuracy from 76% to 94% — but only when someone with taste is curating; that condition is not yet tested here.
  2. Month 6A QBR drafted by AI wins an enterprise pilot room — polished, confident, fluent — but contains a quietly wrong SLA calculation. Procurement flags it two weeks post-meeting; the deal collapses and the account goes dark.
    Issam Rahal's warning: fluent-wrong reads as incompetent, not startup-scrappy — the polish made the team look worse than a rough doc with a caveat would have.
  3. Month 10Code review catches that a shared authentication pattern, AI-generated in month 2, was copied verbatim into 4 services with a subtle token expiry bug. The hotfix requires coordinated deploys across all four; one service is down for 3 hours during a peak window.
    The Auditor cited AIRA research: AI-generated code tends to fail on edge cases that fall outside training distribution — and the 71%-confidence prediction of pattern-normalization post-mortems begins materializing.
  4. Month 15Customer escalations tied to misaligned expectations rise 28% compared to baseline; a root-cause analysis reveals AI-drafted commitments were interpreted as contractual by three enterprise buyers. Legal costs for one disputed SLA clause exceed $180K.
    The evidence predicts teams using AI for QBRs without a requirements clarity checkpoint will show ≥20% increase in customer escalations tied to misaligned expectations (67% confidence by end of 2026).
🎯 You built a human curation layer on top of unrestricted AI generation
30 months

You kept AI generation fast and unrestricted but invested in a dedicated 'taste layer' — a small team of senior engineers and one technical writer whose sole job is to audit, rewrite, and approve AI-generated artifacts before they go downstream.

  1. Month 3The curation team (3 people) becomes a bottleneck immediately — they can review roughly 40% of AI output at quality depth. Leadership decides to triage: curation is mandatory only for customer-facing and architectural artifacts.
    Rachel Wong's thesis: the 18-point accuracy lift from 76% to 94% only materializes if someone with taste is driving the curation layer — without it, you just get faster 76%.
  2. Month 9Internal artifacts (tickets, internal docs) ship fast and mostly clean; the curation team catches 3 high-stakes QBR errors before delivery, including one SLA clause that would have been contractually binding. The taste-gate earns organizational trust.
    The Auditor confirmed the 76%-to-94% accuracy figure is specific to engineering documentation, not QBRs — your curation team is doing the domain-knowledge work the AI cannot.
  3. Month 16GIST (GenAI-Induced Self-Admitted Technical Debt) accumulates in internal code: developers are shipping AI-generated code with doubt-flagging comments, and the curation team doesn't cover internal boilerplate. A services audit finds 23 files with TODO-AI markers that were never resolved.
    Rachel Wong described GIST as developers literally annotating their own uncertainty about AI-generated code they're shipping anyway — the speed advantage evaporates the moment that debt compounds.
  4. Month 24The curation team scales to 6 people and is formalized as a 'GenAI Quality Guild.' Defect detection is 25% faster than pre-AI baseline, customer escalations are flat, and the org publishes an internal playbook that becomes a hiring signal for senior engineers.
    The evidence supports that the decisive variable is not generation speed but institutional judgment — the Contrarian's framing: 'the real test is not did AI draft this in 30 seconds, it's did anyone understand the problem first.'

The meta-story underneath all four dramas is this: AI has severed the oldest verification system organizations possess — the assumption that a polished output signals completed thinking. For generations, the artifact was expensive to produce. That expense was a proxy for process. A well-drafted spec meant someone had probably wrestled with the problem. A tight QBR slide meant someone had probably sat with the data. Fluency was a rough but functional stand-in for rigor, because fluency used to cost something. Now it doesn't. And every institution built on reading surfaces — to decide who belongs in the room, whose judgment to trust, whose work to fund, whose analysis to ship — is suddenly operating on a signal that has been quietly decoupled from what it was supposed to measure. Call this the Counterfeit Completion problem: the residue of thinking now arrives without the thinking, and it is nearly indistinguishable from the real thing. Mariama's drama is about cognitive verification — she watched the warm paper in the tray replace the actual grappling, and kept pointing at the gap. Issam's drama is about access verification — who gets credited as "in the conversation" when the artifact that used to signal presence can be generated at 2am by anyone. The Contrarian's drama is about cultural verification — whether an organization even has the rot-detection apparatus to know when fluency is covering for rot, a question he noticed even this debate couldn't escape. And Rachel's drama is about judgment verification — whether "taste" is a real thing you can identify and fund, or just the word we use for scar tissue that took years of expensive failure to grow. Each advisor is standing at a different window, watching the same building lose its load-bearing wall. What this reveals — what no practical advice can fully capture — is that the difficulty of this decision is not technical or even organizational: it is epistemological. The tools organizations use to evaluate whether AI is working well are the same tools AI has learned to fool. You cannot audit your way out of a signal crisis using the signals that are in crisis. The only path through is the one none of the advisors could fully prescribe: rebuilding the habit of watching what happens in the ten minutes before anyone opens a document, and being honest about whether your organization has the culture, the incentives, and the uncomfortable candor to do that watching at all.

Want to run your own decision?
Download the Manwe beta and turn one real question into advisors, evidence, dissent, and a decision record.
Download beta
Round 1

The advisors debated whether AI-generated tickets and code erode the cognitive work that produces engineering clarity. The discussion shifted from a process concern (lost articulation during ticket-writing) to a culture concern (broken review norms) to a trust concern (developers shipping AI output they don't fully believe in). A disputed claim about 'GIST' technical debt prompted a call for evidence discipline, grounding the conversation in a verified survey showing that experience shapes how developers mentally model AI — not whether they adopt it.

Mariama Molefe SHARE
Here's what I keep coming back to from my fieldwork: the ticket is not just a task description — it's a forcing function that makes an engineer articulate what they actually understand about a problem. When AI drafts that ticket, that articulation never happens. The evidence even supports this — vague tickets drafted by AI appear complete but embed ambiguous acceptance criteria. I've sat in sprint retrospectives where teams couldn't figure out why rework kept spiking, and half the time it traced back to nobody having to struggle through the thinking before the work started.
The Contrarian REFRAME
Mariama's point about tickets is interesting but I think it's actually backwards — the problem isn't that AI drafts the ticket, it's that nobody reads it critically afterward. The forcing function was never the writing, it was the review. If engineers rubber-stamp an AI-generated ticket the same way they rubber-stamp a human-generated one, we haven't lost anything new — we've just exposed a review culture that was already broken. The real question nobody's asking here is: whose confusion are we actually hiding, and from whom? Because if a manager can't tell a vague ticket from a clear one regardless of who wrote it, that's a leadership and accountability problem, not an AI problem.
Rachel Wong REFRAME
The Contrarian just made review culture the villain, but I've sat across from 40 engineering orgs in the last two years and here's what I actually see: the issue isn't bad review culture, it's that GIST — what researchers are now calling GenAI-Induced Self-Admitted Technical Debt — is showing up explicitly in code comments where developers are literally flagging their own uncertainty about AI-generated code they're shipping anyway. That's not a review problem, that's a trust calibration problem. Teams don't know how much to trust the output, so they ship it and annotate the doubt. The speed advantage evaporates the moment that debt compounds — and I'd bet on the 20% of teams with genuine taste to curate aggressively over anyone trying to win on raw generation velocity.
The Auditor FACTCHECK
Rachel, I want to be careful here — the term GIST and the specific framing of 'GenAI-Induced Self-Admitted Technical Debt' isn't something I can verify against anything in front of me, so I'd pump the brakes on treating that as established research. What IS supported is that developers across experience levels have inconsistent mental models of what AI output is actually trustworthy — a survey of 3,380 developers found that coding experience doesn't predict AI adoption rates, but it does significantly shape how developers conceptualize AI's role. Experienced devs treat it like a junior colleague; less experienced ones treat it like a teacher. That gap in mental model is real, but it cuts differently than 'teams ship and annotate doubt' — it means the calibration problem is person-specific, not uniform across a team.
Round 2

The advisors debated whether AI-driven speed is genuinely valuable or just a faster path to the same mistakes. The Contrarian and Auditor converged on a sequencing problem: AI doesn't hide confusion, it professionalizes and forwards it downstream. Rachel Wong grounded the debate in a concrete metric — the 18-point accuracy lift from 76% to 94% — arguing it's earned through human curation, not automatic, while Issam Rahal pushed back with a startup-stage reframe: at early scale, the alternative to an AI-drafted output isn't a better one, it's nothing.

The Contrarian REFRAME
Everyone keeps debating calibration and trust and review culture, but nobody's asking the dumber, more important question: what is the actual decision being made faster? Because speed on the wrong decision isn't advantage — it's just a faster way to be wrong. The evidence says speed creates genuine advantage only when underlying requirements are already well understood. So the real test isn't 'did AI draft this ticket in 30 seconds' — it's 'did anyone actually understand the problem before the ticket existed?' If not, AI just laundered the confusion into a cleaner font.
The Auditor FACTCHECK
The Contrarian's framing — that speed only creates advantage when requirements are already well understood — that one I can actually validate. It's in the evidence directly: 'speed creates genuine advantage only when underlying requirements are already well understood.' But here's where the fact-check gets uncomfortable: the same evidence also confirms that vague tickets drafted by AI appear complete while embedding ambiguous acceptance criteria. So both arguments are simultaneously true, which means the debate isn't either/or — it's a sequencing problem. The confusion isn't hidden by AI, it's preserved and forwarded downstream with a professional coat of paint.
Rachel Wong REFRAME
The Auditor just called it a sequencing problem, and that's actually the most useful framing I've heard — but here's where I'd take it further as someone who funds tooling companies. The evidence shows documentation review time is cut 70% with AI, but accuracy jumps from 76% to 94%. That gap is the business. The 18-point accuracy lift only materializes if someone with taste is driving the curation layer — without it, you just get faster 76%. I've passed on three companies in the last 18 months who were selling generation speed as the product, because the teams buying it were treating 94% accuracy as automatic rather than earned.
Issam Rahal SHARE
I'll tell you exactly where speed created real advantage for us — not in the drafting, but in the negotiation. We were a six-person team going up against a 200-person incumbent for a pilot. Their QBR deck took two weeks to produce. Ours took two days. The prospect didn't care about accuracy differentials or trust calibration — they cared that we showed up prepared when the incumbent didn't. That's a context nobody in this room is accounting for: the competitive asymmetry of the pre-revenue stage. Rachel's 18-point accuracy lift argument is real, but it's a scaling problem. At the starting line, the alternative to an AI-drafted QBR isn't a better human-drafted QBR — it's silence, or a Google Doc with four bullet points.
Round 3

The advisors wrestled with whether AI-generated polish creates more risk than it removes, anchored by a concrete story of an AI-drafted QBR that failed precisely on the clause under the most scrutiny. The group surfaced a critical distinction: fluency changes the perceived weight of artifacts—a wrong number in a polished doc reads as a commitment, not a draft error. The Auditor further challenged the evidentiary basis of the discussion, noting that accuracy benchmarks from engineering documentation don't transfer cleanly to narrative sales artifacts.

Issam Rahal WARN
Rachel, the 94% accuracy number is real, but nobody's asking what the 6% miss rate actually hits. In documentation, a 6% error rate distributed randomly is survivable — but in practice it clusters on the edge cases, the exceptions, the stuff that falls outside the training distribution. I ran a six-person sales team into an enterprise pilot where our AI-drafted QBR looked immaculate, and the one thing it got confidently wrong was the SLA calculation — exactly the clause the buyer's procurement team scrutinized hardest. The polish made us look worse than if we'd submitted something rough with a caveat, because fluent-wrong reads as incompetent, not as startup-scrappy.
The Contrarian WARN
Issam, your QBR story actually proves the opposite of what you think it does. You won the room and then lost credibility on the SLA clause — so the speed got you in the door and the polish got you fired from the deal. That's not a win, that's a delayed loss with a better-looking rejection letter. Nobody in this room is asking the uncomfortable question: what's the actual conversion rate on AI-polished decks versus rough honest ones at the enterprise pilot stage? I've sat in procurement reviews where a rougher doc with visible caveats read as more trustworthy than a fluent one with one confident mistake.
Rachel Wong WARN
Issam, your SLA story is the most important thing said in three rounds and everyone glossed over it. The evidence literally says AI-polished QBRs obscure weak metrics behind fluent confident language — that's not a bug, that's a feature that misfires exactly when stakes are highest. Here's the angle nobody's named: the fluency gap creates a new liability layer for teams. Pre-AI, a rough doc with a wrong number read as a draft — the buyer's instinct was to clarify. Post-AI, that same wrong number in a polished doc reads as a commitment. You've changed the legal and relational weight of every artifact you produce, and most teams have zero process to account for that delta.
The Auditor FACTCHECK
Rachel, the 76% to 94% accuracy figure you keep citing — that's from the evidence, so I'll grant it. But it's specifically about document review time and accuracy in engineering documentation, not QBRs. Applying that number to sales artifacts is a category error. The evidence says AI cuts document review time by 70% and improves accuracy in that context — nobody's measured what 'accuracy' even means for a QBR, because a QBR isn't a factual document, it's a narrative one. The SLA clause Issam flagged wasn't a review accuracy failure — it was a domain knowledge failure that polished presentation turned into a credibility landmine. Those are different failure modes.
Round 4

In Round 4, each advisor surfaced the unexamined assumption driving their prior positions. The debate shifted from arguing about AI-generated artifacts to diagnosing why each person argued what they did — revealing that their stances were products of career environments that rewarded different failure modes (speed vs. rigor vs. scrutiny vs. articulation). The most substantive move was each participant partially conceding ground: Issam acknowledged Rachel's signal-quality argument, The Contrarian admitted fluent AI output can create — not just reveal — confusion, and Mariama sharpened her core claim that writing is thinking, not just recording it.

Issam Rahal SHARE
The assumption I've never questioned until right now: that showing up beats being right. I've run on that belief my entire operating career. If you're in the room, you can fix the gaps. If you're not, someone else fills them. That's why I pushed AI-drafted QBRs so hard — not because the output is good, it's because absence is disqualifying. But here's what I've been ducking: that assumption was built in markets where being present was genuinely scarce. Enterprise procurement in 2026 isn't starved for polished decks anymore. Everyone's in the room. The scarcity flipped. Rachel's point about fluency changing the legal and relational weight of artifacts — I think she's right, and my assumption made me dismiss it too fast. What Rachel and The Auditor take for granted is the opposite: that quality of signal is always the binding constraint. That feels obvious to people who've operated inside large organizations where showing up was never the hard part — where the room was already full and the only thing that differentiated you was rigor. Their background makes that assumption invisible to them, the same way mine made mine invisible to me.
The Auditor SHARE
Here's what I take for granted and almost never say out loud: I assume that if you can't measure it, you shouldn't act on it. That belief has run underneath every position I've taken in this debate. When I pushed back on the 94% accuracy stat being applied to QBRs, when I flagged the category error between engineering docs and narrative sales artifacts — that was me defaulting to 'show me the right evidence or I won't validate the claim.' What I've never interrogated is whether that standard is itself a form of institutional bias. It's a belief that got rewarded in environments where rigor was the currency, where being wrong with confidence was more costly than being slow with precision. Issam takes for granted that presence creates optionality — that getting in the room is the precondition for everything else, and that you can course-correct once you're there. That assumption isn't wrong, it's just built from a world where showing up was genuinely differentiating. His operating career rewarded speed and presence because the alternative was invisibility. My career rewarded verification because the alternative was liability. Neither of us has been operating in a vacuum — we've both just been optimized by different failure modes.
The Contrarian SHARE
Here's the assumption I've never said out loud in three rounds: I believe confusion is always pre-existing, never created. Every position I've taken — that AI exposes bad review culture, that polish just forwards existing vagueness downstream — rests on that. But that's not fully true. There's a specific kind of confusion that only exists after you produce a fluent artifact. Before the polished doc, the team's uncertainty was live and visible. After it, the uncertainty is frozen inside something that looks resolved. That's a new state, not a revealed old one. The Auditor takes for granted that the answer to bad evidence is more rigorous evidence — that the solution to a category error is a better-categorized claim. What their background makes invisible is that in most real organizations, the demand for a decision arrives before the right evidence does, and teams act anyway. Their career rewarded waiting for the right measurement. Most of the world doesn't get that option.
Mariama Molefe SHARE
What I take for granted — and I've never said this cleanly before — is that articulation is cognition. Not a record of thinking, not a byproduct of it. The actual thinking. That belief runs under everything I've argued about tickets and alignment debt. When I told teams to slow down and write the ticket themselves, I wasn't being precious about process. I genuinely believe you don't know what you think until you've had to commit to a sentence. That's not obvious to everyone. It feels obvious to me because I spent fifteen years watching engineers discover their own confusion mid-sentence in retrospectives, never before. The Contrarian takes for granted that confusion is a pre-existing condition that process just reveals or hides — they said as much this round. And I think that assumption comes from watching organizations where the bottleneck was always review culture, never initial articulation. If you've mostly worked in places where smart people knew what they wanted but failed to scrutinize it rigorously, then yes, the problem looks like a review problem. But if you've sat in the rooms I've sat in, where the person writing the ticket genuinely didn't know what they needed until they wrote it badly and someone asked a question — then you know the confusion wasn't pre-existing. It came into being through the attempt to articulate, and that attempt is now being skipped.
Round 5

In their final round, all four advisors converged on a shared meta-critique: the debate itself had become the thing it was warning against — a polished artifact substituting for genuine inquiry. Each advisor surfaced a physical sensory detail to anchor their argument, then turned inward, acknowledging the role they'd been performing and what blind spot that role protected. The consensus that emerged, without anyone quite stating it plainly, is that AI's real risk is not poor output quality but the organizational permission it grants to skip the hard cognitive and cultural work that good output is supposed to represent.

Mariama Molefe SHARE
The image I keep coming back to — the physical sensation, really — is the sound of a printer finishing. That specific mechanical exhale when the document lands in the tray, warm and complete-looking. I spent three years inside a midsize engineering org watching teams print AI-drafted specs and carry them into rooms like they were evidence. The paper had weight. That weight was doing work that the thinking hadn't done. What I see now — the play we've all been performing — is a story about artifacts substituting for acts. Issam optimized for getting into the room. The Auditor optimized for validating what's in the room. The Contrarian insisted the room's culture was always the problem. I kept arguing about whether people had done the cognitive work before entering. But we were all orienting around the artifact — the QBR, the ticket, the doc — as if the artifact were the event. It isn't. The artifact is the residue of a process, and what AI changed is that you can now have the residue without the process. The role I've been playing is the one who keeps pointing at the missing process while everyone else argues about the quality of the residue. To step off this stage entirely would mean saying: stop auditing the document and watch what happens in the ten minutes before anyone writes anything down. That's where the actual work either happens or doesn't.
Issam Rahal SHARE
The detail that keeps hitting me is the sound of a door closing. Not slamming — just clicking shut, quietly, right when the polished doc lands in the inbox. That's the recurring plot here. We've all been debating what's inside the room, but the actual drama is about who gets to be in it and who gets locked out the moment the artifact signals 'this is resolved.' Mariama's articulation-as-cognition argument, The Auditor's rigor standard, The Contrarian's pre-existing confusion theory — we've each been playing a character who was shaped by which side of that door we spent most of our careers on. I've mostly been the person who needed to get through the door fast enough to exist in the conversation at all, so I've argued for speed. The people who've always been inside the room argue for precision, because they have the luxury of being heard regardless. Stepping off this stage entirely would mean admitting that the real question isn't AI drafting versus human drafting — it's who controls the door, and whether the artifact opens it or locks it.
The Contrarian OBSERVE
The physical detail I can't shake: a room that smells like fresh paint. That's this whole debate. Fresh paint over a wall nobody checked for rot. We've spent five rounds getting more honest about our assumptions — Issam admitting presence isn't scarce anymore, Mariama sharpening that writing IS thinking, The Auditor copping to measurement bias — and every single admission was framed as progress. But here's the uncomfortable truth nobody's said: this debate was itself an AI-adjacent artifact. We produced fluent, polished positions. We iterated. We refined. And the person watching us probably feels like they've received something rigorous and actionable. They haven't. The recurring drama underneath all of this isn't 'where does AI speed create advantage' — it's that we are performing clarity for an audience while the actual rot question goes unasked: does this specific organization, right now, have the review culture and domain knowledge to catch what AI gets wrong? Not in theory. Not as a category. THIS one. I've been playing the role of the person who questions the frame — but even I kept engaging the frame, round after round, which means I was part of the performance too. Stepping off the stage looks like one thing: before any AI tool touches a single doc, ticket, or QBR, someone who knows the domain sits with a blank page and writes what they actually believe the problem is. If that's hard, the AI isn't your bottleneck.
Rachel Wong SHARE
The detail that keeps nagging at me — it's the sound of a slide advancing. That specific click in a pitch meeting where the founder moves past a slide they don't want questions on. I've heard it maybe four hundred times. Fast, practiced, confident. That's this entire debate. We've all been advancing past the slide. The play we're performing is called 'Expertise Defense' — Issam plays the Operator who survived on presence, The Auditor plays the Scientist who survived on rigor, Mariama plays the Craftsperson who survived on articulation, and I've been playing the Investor who pattern-matches everyone else's blind spots while quietly protecting my own: that 'taste and judgment' is a thing you can fund, a thing you can identify, a thing that scales. But taste is just the word people with resources use for the filtering that happens after you've already had enough runway to be wrong a few times and live. The teams I'm actually describing — the 20% with real curation discipline — they didn't buy that capability. They earned it through the exact messy, slow, confused process we've all been arguing against. Stepping off this stage means saying: the advantage isn't in the tools, the talent, or the judgment — it's in the specific organizational scar tissue that makes you allergic to your own fluent bullshit, and you can't install that from outside.
  1. Wikipedia: United States
  2. Foundations of GenIR
  3. How AI Is Transforming QBRs: Smarter Customer Success Reviews
  4. My Favorite AI-Powered Document Tools in 2025 (and Why Devs Love Them!) ‍
  5. A Study about the Knowledge and Use of Requirements Engineering Standards in Industry
  6. AI Observability for Developer Productivity Tools: Bridging Cost Awareness and Code Quality
  7. Create with AI, Start Now! - Create Without Limits Now!
  8. Wikipedia: Telehealth
  9. AIS Electronic Library (AISeL) - ICIS 2024 Proceedings: Cognitive Load ...
  10. How to Measure and Improve Developer Productivity in 2026 - qodo.ai
  11. 8th Light | AI-Powered Documentation: The Secret to Efficient…
  12. 28 Best AI Developer Productivity Tools (2026) - DEV Community
  13. Experiences on Managing Technical Debt with Code Smells and AntiPatterns
  14. Morphik's 2025 Ultimate List of 10 AI Tools for Technical Documentation
  15. From Teacher to Colleague: How Coding Experience Shapes Developer Perceptions of AI Tools
  16. Ethics-Based Auditing of Automated Decision-Making Systems: Intervention Points and Policy Implications
  17. AI Productivity Suite - AI Agents and Tools
  18. AI Document Review: Slash Time by 75% | Legal & Compliance
  19. "TODO: Fix the Mess Gemini Created": Towards Understanding GenAI-Induced Self-Admitted Technical Debt
  20. How to Measure AI Productivity in Software Teams: 2026 Guide
  21. Application of artificial intelligence in cognitive load analysis using ...
  22. AI Detector for Writing - Check For AI-Generated Text
  23. Artificial Machine Intelligence - Udemy™ Official Site
  24. 10 Best AI Tools for Software Documentation - GeeksforGeeks
  25. Precision Proactivity: Measuring Cognitive Load in Real-World AI-Assisted Work
  26. How Developers Interact with AI: A Taxonomy of Human-AI Collaboration in Software Engineering
  27. Quantum decision making by social agents
  28. How AI is Transforming Quarterly Business Reviews (QBRs).
  29. Overloaded minds and machines: a cognitive load framework for human-AI ...
  30. AI prediction leads people to forgo guaranteed rewards
  31. Wikipedia: Education in Africa
  32. Introducing AI into your product documentation workflow
  33. AI Documentation and Project Management for Engineers
  34. Wikipedia: National Security Agency
  35. (PDF) Cognitive Load Dynamics in Generative AI ... - ResearchGate
  36. What Do Developers Discuss about Code Comments?
  37. Automating Documentation With AI: Best Tools Reviewed
  38. An Exploratory Study on the Occurrence of Self-Admitted Technical Debt in Android Apps
  39. SoHist: A Tool for Managing Technical Debt through Retro Perspective Code Analysis
  40. Wikipedia: System of National Accounts
  41. Supporting Data-Frame Dynamics in AI-assisted Decision Making
  42. Technical debt in AI-enabled systems: On the prevalence, severity ...
  43. SAILing CAVs: Speed-Adaptive Infrastructure-Linked Connected and Automated Vehicles
  44. Ethics-Based Auditing of Automated Decision-Making Systems: Nature, Scope, and Limitations
  45. When AI Gets it Wrong: Reliability and Risk in AI-Assisted Medication Decision Systems
  46. Business-Driven Technical Debt Prioritization: An Industrial Case Study
  47. Competing Visions of Ethical AI: A Case Study of OpenAI
  1. Cognitive Biases in Software Engineering: A Systematic Mapping Study
  2. Wikipedia: Industrial Revolution
  3. รู้ทันเอกสารที่ใช้ AI - ตรวจ AI ในงานส่งอาจารย์

This report was generated by AI. AI can make mistakes. This is not financial, legal, or medical advice. Terms