AI Mistakes & How to Catch Them
AI will confidently lie to your face — here's how to spot it, prevent it, and protect yourself.
A lawyer cited a court case that never existed — and a judge wasn't amused
In 2023, a New York attorney named Steven Schwartz used ChatGPT to research legal precedents for a case. The AI gave him six relevant court cases with names, citations, and even short summaries of the rulings. They sounded perfect. He cited them in his filing.
There was one problem: none of those cases existed. Not a single one. The AI had invented case names, fabricated judges' names, made up rulings, and assigned fake citation numbers — all with absolute confidence.
The judge discovered the fraud. Schwartz, his partner Peter LoDuca, and their firm Levidow, Levidow & Oberman were jointly sanctioned $5,000 by Judge P. Kevin Castel — payable jointly and severally. (The case was Mata v. Avianca, May 2023.) The story made international headlines.
And here's the part that should make you pay attention: Schwartz wasn't lazy. He wasn't cutting corners. He asked AI a question, got an answer that looked completely real, and trusted it. That's exactly what most people do with AI every day.
This module teaches you to never be Steven Schwartz.
The three ways AI will fail you
AI doesn't make random errors like a typo or a math mistake. It makes systematic errors — patterns of failure that are predictable once you know what to look for.
| Failure type | What happens | Why it happens | How dangerous it is |
|---|---|---|---|
| Hallucination | AI invents facts, sources, statistics, people, or events that don't exist | It predicts plausible-sounding text, not truthful text | Very high — you can't tell by looking |
| Bias | AI reflects unfair patterns from its training data | It learned from human-written text, which contains human biases | High — can cause real harm to real people |
| Confidentiality leaks | Your sensitive data gets stored, trained on, or exposed | Data you paste into AI may not stay private | Critical — legal and financial consequences |
Think of AI as a brilliant intern on their first day
You just hired someone who graduated top of their class. They're articulate, fast, and eager to help. But it's their first day. They've never seen your company before. They've never met your clients. They don't know your industry's unwritten rules.
Would you let this person:
- File a legal brief without checking it? No.
- Send an email to a client without reviewing it? No.
- Make decisions about someone's job or loan application? Absolutely not.
But would you let them draft the brief, write the first version of the email, or compile the data for the decision? Yes — and then you check their work.
That's the right relationship with AI. Use it for the first draft. Verify the final output.
✗ Without AI
- ✗Copy AI output directly into work
- ✗Trust confident-sounding claims
- ✗Skip fact-checking because 'AI knows'
- ✗Blame AI when things go wrong
✓ With AI
- ✓Use AI output as a first draft
- ✓Verify any specific facts or numbers
- ✓Spot-check claims against sources
- ✓Own the final output regardless of how it was produced
Hallucination: when AI invents facts with a straight face
Hallucination isn't a bug — it's a structural feature of how language models work. The model predicts what words should come next based on patterns. Sometimes the most plausible-sounding next word is wrong. The model doesn't know the difference because it doesn't "know" anything — it generates text.
What hallucinations look like
| Category | Example hallucination | Why it's convincing |
|---|---|---|
| Fake citations | "According to Smith v. Johnson (2019), 547 U.S. 312..." | Has the exact format of a real citation |
| Invented statistics | "Studies show 73% of remote workers are more productive" | Sounds specific and authoritative |
| Fake people | "Dr. Elaine Martinez, a leading AI researcher at Stanford..." | Uses a plausible name, real university |
| Wrong dates | "The company was founded in 2008 by..." | Close to the real date, sounds right |
| Plausible but wrong explanations | "This happens because the TCP protocol uses..." | Technically coherent but factually wrong |
When hallucinations are most likely
High-risk situations:
- Asking for specific statistics, studies, or research papers
- Asking about obscure topics, small companies, or recent events
- Asking for legal, medical, or financial facts
- Asking about people (especially non-famous individuals)
Lower-risk situations:
- Asking for general explanations of well-known concepts
- Brainstorming and ideation
- Editing and rewriting text you provide
- Summarizing documents you paste in (AI can't hallucinate about text it's looking at)
Common AI output failures (in rough order of frequency in practice):
| Failure type | What it looks like |
|---|---|
| Hallucinated facts | Confident claims about things that aren't true |
| Outdated information | Correct at training time, wrong now |
| Wrong tone or voice | Technically accurate but inappropriate for context |
| Missed nuance | Oversimplified answers to complex questions |
| Privacy exposure | Including sensitive information that shouldn't appear in output |
There Are No Dumb Questions
"If I paste a document and ask AI to summarize it, can it still hallucinate?"
Much less likely, but yes. AI can occasionally misinterpret the document, combine two separate points into one that wasn't made, or state a conclusion more strongly than the source supports. The risk drops dramatically when AI is working from source material rather than its memory — but always spot-check the summary against the original.
"Why can't they just fix hallucination?"
Because hallucination comes from the same mechanism that makes AI useful. The model generates new text by predicting patterns — that's what lets it write, summarize, translate, and create. If you made it only output text it was 100% certain about, it would barely produce anything. The industry is making progress (newer models hallucinate less), but it's a fundamental tradeoff, not a bug to patch.
Make AI hallucinate on purpose
50 XPBias: when AI reflects unfair patterns
AI learns from human text. Human text contains biases. Therefore AI outputs contain biases. It's math, not malice.
Real examples of AI bias
| Scenario | What happened | The bias at play |
|---|---|---|
| Resume screening | AI ranked male candidates higher for engineering roles | Training data reflected historical hiring patterns |
| Image generation | "CEO" almost always generated a white man in a suit | Training images over-represented this demographic |
| Loan applications | AI flagged certain zip codes as higher risk | Zip codes correlated with race, creating a proxy for discrimination |
| Translation | "The doctor... she" was translated to "he" in some languages | Gendered assumptions baked into language patterns |
How to spot bias in your AI outputs
Ask yourself these questions every time AI makes a recommendation, classification, or decision about people:
- Would this output change if I swapped the person's name, gender, or background? If yes, there's bias.
- Does this recommendation match a stereotype? If AI is confirming every expectation, it might be reflecting bias, not reality.
- Who is missing from the output? If AI's "list of experts" has no diversity, that's a signal.
- Is this too convenient? If AI's analysis perfectly supports what you already believe, ask it to argue the opposite.
There Are No Dumb Questions
"Am I responsible for AI bias in my work?"
Yes, in the same way you're responsible for any tool you use. If you use a calculator and enter the wrong numbers, you can't blame the calculator. If you use AI output that contains bias and it harms someone, the fact that AI generated it doesn't shield you from accountability. You're the human in the loop. That's why you're here.
"Is there a bias-free AI?"
No. Every AI model has biases because every training dataset has biases. The question isn't "is this AI biased?" (it is) but "what biases does this AI have, and how do they affect my specific use case?" Some biases are harmless for your task. Others could be devastating. You need to know the difference.
The bias detection exercise
25 XPConfidentiality: what you should NEVER paste into AI
This is the mistake that gets people fired. Not hallucination, not bias — pasting sensitive information into an AI tool.
The red list: never paste these
| Category | Examples | Why it's dangerous |
|---|---|---|
| Company secrets | Unreleased product plans, financial projections, board meeting notes | May be stored, trained on, or breached |
| Personal data | Employee SSNs, customer emails, patient records | Violates GDPR, HIPAA, and other regulations |
| Passwords & credentials | API keys, login credentials, security tokens | Can be logged and exposed |
| Legal documents | Contracts under NDA, privileged attorney-client communications | Breaks privilege and confidentiality obligations |
| Customer data | Individual transaction records, support tickets with names | May violate your company's data agreements |
When it goes wrong: the Samsung incident
The difference between them and you: now you know.
The green list: safe to paste
| Category | Why it's safe |
|---|---|
| Public information | It's already out there |
| Your own original writing | You own it, no confidentiality concerns |
| Anonymized data | No way to identify individuals |
| Generic templates and examples | No sensitive specifics |
| Questions about general topics | Not revealing anything proprietary |
The decision flowchart
Pro tip: Many companies now offer enterprise AI tools (like ChatGPT Enterprise, Azure OpenAI, or Claude for Enterprise) where your data isn't used for training and stays within your organization's boundaries. Check with your IT department — you might already have access.
The confidentiality audit
25 XPThe "trust but verify" workflow
Here's the system that protects you from all three failure types:
The 60-second verification checklist
Before you use any AI output for anything important, run through this:
- Facts check: Did I verify any specific claims, statistics, or citations?
- Source check: Can I trace each fact back to a real, primary source?
- Bias check: Would this output change if I swapped demographics?
- Confidentiality check: Did I avoid pasting anything sensitive?
- Smell test: Does anything feel "too perfect" or suspiciously convenient?
This takes 60 seconds. Skipping it can cost you your reputation.
Build your personal verification workflow
50 XPBack to Steven Schwartz
He wasn't lazy. He wasn't cutting corners. He asked a question, got an answer that looked completely real, and trusted it — the same thing most people do with AI every day. The difference is that his mistake was filed with a federal court, where it couldn't be quietly fixed.
What he was missing wasn't intelligence or diligence. It was the 60-second habit: verify the facts, trace the sources, check the output before it goes out the door.
You now have that habit. The next time AI gives you a citation, a statistic, or a legal precedent — you'll know to check it. That's the entire point.
Key takeaways
- AI hallucinates — it invents facts, sources, and statistics that don't exist. The more obscure the topic and the more specific the claim, the higher the risk. Always verify factual claims against primary sources.
- AI reflects biases from its training data. When AI makes recommendations about people (hiring, lending, evaluating), check for bias by mentally swapping demographics.
- Never paste sensitive information into AI tools. Company secrets, personal data, passwords, and legal documents should never go into a consumer AI tool. Use enterprise tools with proper data handling, or anonymize first.
- Trust but verify is not optional — it's your job. The 60-second verification checklist protects your reputation. A brilliant intern's draft still needs your review before it goes out the door.
Knowledge Check
1.Why does AI hallucinate?
2.Which of the following is SAFE to paste into a consumer AI tool?
3.When is AI MOST likely to hallucinate?
4.Who is responsible when AI output contains bias that harms someone?