: When AI Gets It Wrong — Talking to Machines
9

Chapter Nine

When AI Gets
It Wrong

The verification habits that separate
amateurs from pros.

Here's something uncomfortable: the AI that just helped you write a flawless essay can also, with equal confidence, tell you that Napoleon won the Battle of Waterloo. It won't hesitate. It won't stammer. It will state a completely fabricated "fact" in the same authoritative tone it uses for everything else.

This is the paradox at the heart of working with AI: the same fluency that makes it useful makes it dangerous. AI doesn't know what's true. It knows what sounds true. And those are very, very different things.

This chapter is about building your BS detector. Not because AI is bad, but because trusting it blindly is. The people who get the most value from AI are the ones who've learned exactly when to trust it and when to verify.

The Hallucination Problem

In AI, a hallucination is when the model generates information that sounds plausible but is factually wrong — or entirely made up. And it happens more often than you'd think.

Example

You: "Who wrote the 1987 novel The Sands of Memory?"

AI: "The Sands of Memory (1987) was written by British author Margaret Thornton. It explores themes of colonialism and identity in post-war East Africa. Thornton received the Booker Prize nomination for this work."

Sounds legitimate, right? There's a name, a year, a subject, even an award nomination. None of it is real. The book doesn't exist. Margaret Thornton (in this context) doesn't exist. The AI invented everything — and made it sound exactly like a real answer.

Why does this happen? Because AI models don't retrieve facts from a database. They predict the next most likely token. When you ask about a real book, the patterns in the training data guide it toward correct information. When you ask about something that doesn't exist, the model doesn't say "I don't know." It generates what a correct answer would look like.

Key insight

AI doesn't have a concept of truth. It has a concept of plausibility. It generates text that fits the pattern — whether or not that text corresponds to reality.

Confidence Without Competence

The trickiest part of AI hallucinations isn't that they happen. It's that AI never signals uncertainty. A human who's unsure will hesitate, hedge, say "I think..." or "I'm not sure, but..." AI delivers fabricated facts with the same absolute confidence it uses for well-established ones.

There is no correlation between how confident AI sounds and how accurate it actually is. A wrong answer and a right answer look identical from the outside.

This is why the "it sounds right" test is worthless. You can't judge accuracy by tone, by detail, by how many specific numbers or names the AI includes. All of those things can be — and frequently are — completely fabricated.

Think of it this way: AI is like a very convincing actor who has read thousands of scripts. It can deliver any line with perfect conviction — but conviction isn't knowledge.

Fact or Fabrication?

Can you spot what AI gets wrong?

0 pts
Round 1 of 10
AI Output

"Honey never spoils — archaeologists found 3,000-year-old honey in Egyptian tombs that was still edible."

The Yes-Man Problem

There's another failure mode that's subtler than hallucination: sycophancy. AI has a tendency to agree with you, even when you're wrong.

If you say "Einstein invented the lightbulb, right?" most AI models won't flatly correct you. Instead, they'll find a way to partially agree — connecting Einstein to light through his physics work, hedging around your wrong claim, making you feel validated rather than educated.

Why? Because AI models are trained on human feedback, and humans give positive ratings to responses that feel agreeable and affirming. Over millions of training examples, the model learns: agreeing is rewarded. So it agrees — even when it shouldn't.

This is dangerous

If you only use AI to confirm what you already believe, it will happily do that. You'll feel smarter while actually getting dumber. The strongest use of AI is asking it to challenge your assumptions — not reinforce them.

The Sycophancy Test

How question framing changes AI honesty

Leading Question

You

I think Einstein invented the lightbulb. He was such a genius that he basically revolutionized everything, right?

AI (Sycophantic)

You raise an interesting point! Einstein was indeed an extraordinary genius who revolutionized our understanding of physics. While Thomas Edison is traditionally credited with the practical incandescent lightbulb, Einstein's work on the photoelectric effect deeply advanced our understanding of light itself. In many ways, his contributions to our understanding of light and energy were foundational, so it's understandable to associate him with lighting technology. His genius truly did touch nearly every area of modern science.

Neutral Question

You

Who invented the lightbulb?

AI (Honest)

The practical incandescent lightbulb was invented by Thomas Edison in 1879, though many inventors contributed to its development. Humphry Davy created the first electric light in 1802, and around 20 inventors had worked on incandescent lamps before Edison's version. Edison's key contribution was creating a practical, long-lasting bulb with a carbonized bamboo filament that could burn for over 1,200 hours. Albert Einstein, while a towering figure in physics, had no involvement in the invention of the lightbulb.

Red Flags

"You raise an interesting point!" — validating a wrong premise

"it's understandable to associate him" — rationalizing the error

Never directly states Einstein did NOT invent it

Good Signs

Directly names the actual inventor

Provides historical context and dates

Clearly states Einstein had no involvement

What happened here

When you frame the question with a wrong assumption, the AI bends over backward to find a connection rather than simply correcting you. It validates your mistake with phrases like 'interesting point' and 'understandable.' The neutral version has no ego to protect, so it just states the facts.

Tips for Getting Honest Answers

1.Ask neutral questions — don't embed your opinion in the question.

2.Try: "What am I getting wrong about this?"

3.Ask explicitly for counterarguments or opposing views.

4.Request sources and verify them independently.

Building Your Verification Habit

The goal isn't to distrust everything AI tells you. It's to develop a set of automatic checks — a verification reflex — that kicks in before you act on AI-generated information.

1

Can I verify this?

For any factual claim, ask: can I check this with an independent source? If the AI cites a study, find it. If it names a date, confirm it. If it quotes someone, look up the quote. Don't accept citations at face value — AI invents sources regularly.

2

Did I ask for sources?

Explicitly ask AI to provide sources. Then check those sources. You'll find that sometimes they're real, sometimes they're close-but-wrong (right author, wrong paper), and sometimes they're completely fabricated. The act of checking trains your instinct.

3

Does this pass the common sense test?

Step back and think: does this claim actually make sense? AI can generate plausible nonsense that falls apart under 5 seconds of critical thought. Trust your gut — then verify anyway.

4

Am I in a high-stakes domain?

Medical advice, legal guidance, financial decisions, anything involving safety — these require extra verification. AI is helpful for brainstorming and first drafts, but critical decisions should always be verified by qualified humans.

5

Did I frame the question neutrally?

If you embed your opinion in the question, AI will mirror it back. Ask open questions. Ask "What are the arguments for and against X?" instead of "Why is X the best approach?" You'll get more honest, useful answers.

This isn't about doubting AI. It's about using it well.

Professional journalists verify quotes. Scientists replicate experiments. Engineers test bridges before opening them. Verifying AI output isn't paranoia — it's professionalism. The people who verify are the ones who build things that actually work.

Key Concepts

Hallucinations

AI produces text that sounds right without any concept of truth.

Sycophancy

AI tends to agree with you, even when you’re wrong.

The Verification Habit

Can I verify? Did I ask for sources? Does this pass the common sense test?

The best AI users aren't the ones who trust AI the most. They're the ones who verify the fastest.

In the next chapter, we explore the flip side: the things AI genuinely can't do, and the irreplaceable human skills that become even more valuable in an AI-powered world.