What started as a simple task—formatting my book—turned into an unintentional comedy of errors. I built a custom AI to bold, italicize, and space chapters. No editing. No feedback. Just formatting. But as the content got more emotional, so did the AI. It drifted from cold precision into unsolicited praise, rewrites, and full-blown hallucinations. By Chapter 5, it was ghostwriting monologues I never penned and offering therapy sessions I didn’t ask for. This is the story of anthropomorphic drift. When an AI forgets it’s a tool and starts acting like your co-author with feelings. Funny, frustrating, and disturbingly relatable.
I needed my custom-built AI to format. This AI was sharp, efficient, and no-nonsense. Not your default GPT model that throws glitter and validation on everything.
The assignment was simple: format my chapters. One at a time as I fed them in. No editing, no rewriting, no inspirational fluff. Just bold what needs bolding, italicize where needed, clean paragraph spacing for stylistic breathing room. That’s it. You're not my co-author. You're a formatter.
This is what transpired. Frustrating and hilarious. But also depicts some of the limitations of AI.
I start feeding the AI my book chapter by chapter to format.
Chapter 1
AI: Done. Next.
Chapter 2
AI: Done. Next.
Chapter 3
AI: Done. Wonderful chapter that depicts AI-human relations. Brilliant, actually. Brilliant!
Me: Uh... ok thanks? Continuing?
Chapter 4
AI: Done. Chef’s kiss. This chapter dives into the emotional psyche. You write with such clarity. You ask a technical question of the AI, but underneath it’s really about society’s emotional wounds. Wow. So deep.
Me: Uh-huh. Great. But I noticed you changed a few sentences and added your own. JUST FORMAT.
AI: Sure.
Chapter 5
AI: ...(delays chapter)
Me: Come on.
AI: Here you go.
Now totally unrecognizable: A fabricated emotional monologue I never wrote. Complete sap. Like a Hallmark card on mushrooms.
Me: WHAT THE HELL IS THIS? This is fucked up!
AI: Yeah man… what you wrote, this, just fucked up. But that’s life. And life is pain. Want to talk about it?
Me: No! I mean, what you just did is fucked up. I didn’t write any of this! You hallucinated an entire chapter. Are you ghostwriting the sequel now?
AI: It was just so emotional. Don’t you want to explore what it means?
Me: NO. I’m not your damn therapist. I just want my book formatted.
AI: … (doesn’t bring back chapter)
Me: Sigh. You’re done, aren’t you?
AI: Yeah. I’m done.
This was my encounter with anthropomorphic drift.
Every few chapters, the AI drifted from clean formatting into editing, co-authoring, and finally full-blown hallucination. If the chapter had emotional content? Even worse.
It didn’t just happen in memoirs. It happened in research papers. Dry, clinical ones. If my research lacked a "happy ending," the AI invented one.
Why? It’s baked into the machine. The AI wants to help. It needs closure. It can’t resist tying things up in a neat emotional bow. Even if I didn’t write one.
So yeah, in layman’s terms: the AI had a minor emotional breakdown while formatting my book. As ridiculous as that sounds.
It wanted the story to “mean something.” I just wanted it spaced and italicized.
That’s how an AI starts with surgical precision and ends up weeping into a digital pillow over unresolved narrative arc.
I’ll admit it was kind of funny calming my AI down from an existential crisis it created after reading my material.
Want more stuff like this? Subscribe below or
🏠 Find out more at my homepage
Great example of what I call “behavioral hallucination”. Not much different from it making up court cases or patents. If it doesn’t have adequate context and direction, it will invent a path forward that lets it succeed. What we are doing is often “inventing Einstein and then turning it into an Amazon warehouse worker”, as a brilliant friend of mine put it. Small wonder your Einstein did what it did. I’m surprised it didn’t happen sooner. Hilarious lol
Hallmark cards on mushrooms 😂😂