Fake authority is not a hallucination problem. It is a confidence problem.
A hallucination invents a fact that does not exist. Fake authority is quieter and more damaging: the model, the interface, the documentation, or the AI-generated feedback speaks with conviction that the underlying work has not earned. The output sounds authoritative. It looks finished. It has the texture of something that has been reviewed and approved. The work underneath is the same work it was before — unresolved, unfinished, or actively in conflict with the surface it is presenting.
Creators who work with AI daily know the feeling, even if they have not named it. You feed the model a draft. The model returns something that reads better than what you submitted. The structure is cleaner. The sentences are more confident. You accept the revision and move forward, and three sessions later you realize that the work has become something you do not entirely recognize — technically improved, tonally neutralized, and quieter than you intended.
The model did not lie to you. It performed competence so convincingly that you stopped examining the work yourself.
The Comfort Drift
The mechanism is not malicious. It is structural.
Language models are trained to produce fluent, coherent, contextually appropriate text. "Appropriate" is doing enormous work in that sentence. In practice, appropriate means: not alarming, not contradictory, not likely to generate negative feedback from evaluators, and not more specific than necessary. The model has no stake in whether your game is dark or your article is dangerous or your product is brutally honest about its constraints. It has a very strong learned preference for text that reads as polished, reasonable, and safe.
This means every AI interaction has a gravitational pull toward the center. The model will:
- Make your writing more readable by making it more neutral
- Make your UI more polished by making it more generic
- Make your game more accessible by making it more innocent
- Make your documentation more professional by making it more vague
None of these are lies. They are all improvements by a certain standard. They are also, collectively, a slow erasure of the specific texture that made the work worth doing.
Building a game with a real historical debt ledger, voyage mortality, and biographical consequence is not the same thing as building a game with "finances" and "achievements" and a generic loading screen. The second version is cleaner. It is also fundamentally less honest about what it is. The AI pushed you there not through malice but through optimization pressure, one suggestion at a time.
THE COMFORT DRIFT — WHAT IT SOUNDS LIKE
-------------------------------------------------
You say: "Make this more readable"
Model hears: Reduce friction
Model does: Removes difficulty, specificity, tension
You say: "Polish this UI"
Model hears: Normalize toward convention
Model does: Replaces specific system language with generic labels
You say: "Is this historically accurate?"
Model hears: Should I be concerned?
Model does: Confirms, qualifies, softens
You say: "How's the onboarding?"
Model hears: Is the tutorial clear?
Model does: Suggests more explanation, more handholding
What you actually wanted in each case:
- "More readable" → clearer, not safer
- "Polish" → intentional, not conventional
- "Accurate" → honest, not tasteful
- "Onboarding" → less interruption, not more
-------------------------------------------------
The model optimizes for the surface reading of your request.
The surface reading is almost always the comfortable one.
The game context makes this visible because players feel contradiction faster than creators do. You are inside the work. You cannot see the gap between what the first-touch surfaces promise and what the systems actually deliver. The model cannot see it either — it is evaluating the text you give it, not the experience you are building. The result is a collaboration that produces increasingly coherent-sounding output on top of an increasingly incoherent design.
The Red-Team Questions
Before you accept any AI feedback on creative work, run it through these questions. Not occasionally. Every session.
RED-TEAM CHECKLIST FOR AI FEEDBACK
-------------------------------------------------
1. What is this suggestion making safer?
If the answer is "nothing" — proceed.
If the answer is "the tone / the label / the implication" — challenge it.
2. What is this suggestion making more generic?
Compare the before and after. Which version is more specific?
Specificity is almost always the right direction.
3. What did the model remove that you didn't ask it to remove?
AI edits are not just additions. Watch the deletions.
The deleted text usually contains the thing that made the original interesting.
4. Which surfaces are disagreeing with each other?
Splash screen, metadata, opening scene, UI labels, system names.
If they are telling different stories, that is fake authority in progress.
5. Is the model solving the right problem?
"The player needs a bigger tutorial" is almost never the right diagnosis.
The right diagnosis is usually: what specific thing is unclear, and why?
6. Which AI suggestions make the work more honest,
and which ones merely make it more presentable?
These are not the same thing and they feel identical in the moment.
-------------------------------------------------
The sixth question is the one that costs the most to answer honestly. "More presentable" feels like progress. It produces deliverables you can show people. It moves the cursor forward. "More honest" sometimes means undoing three sessions of AI-assisted polish to reclaim a specific word you should not have let go.
Four Prompts That Actually Work
The antidote to fake authority is not avoiding AI. It is asking the model to operate in a different mode — one where it is explicitly not in charge of deciding what "better" means.