Session 4 — When Data Deceives Grade 5 Data Science · Ages 10–11 ← → or Space to navigate · F = fullscreen
1 / 15
Session 4 of 8

When Data
Deceives

The data can be 100% accurate — and the conclusion can still be completely wrong. Today we learn how.

🔍 Data Science for Young Minds · Grade 5 — Data Detective
2 / 15
Today's Plan

What We're Doing Today

3 / 15
Opening Hook

Same Data — Two Stories

Imagine a bar chart showing a company's sales performance. Both versions below use identical numbers.

Graph A — Y-axis starts at 95%

  • The bar for "this year" looks THREE TIMES taller
  • Looks like a massive improvement
  • Headline: "Sales EXPLODE upward!"

Graph B — Y-axis starts at 0%

  • The bars look nearly identical in height
  • The actual change: 96% → 98%
  • Headline: "Modest 2% improvement"

Both graphs are accurate. But Graph A is designed to mislead. A Data Detective always checks the Y-axis first.

4 / 15
Lesson 1

Cherry-Picking

Cherry-picking means selecting only the data that supports your conclusion — and ignoring everything else.

The Trick

  • "Our school's math scores improved last year!"
  • (Shows only math — reading, science, and writing all dropped)
  • The cherry-picker is telling the truth — about one cherry

The Detective Question

  • "What data was NOT shown here?"
  • "What happened to the other subjects?"
  • "Over what time period?"
  • "Who chose which data to include?"

Cherry-picking is especially dangerous because each individual data point is true. You need to ask what's missing.

5 / 15
Lesson 1

Survivorship Bias

Survivorship bias happens when we only study the people or things that "survived" a process — missing all those that didn't.

Detective question: "Who is NOT in this dataset — and why might they be missing?"

6 / 15
Lesson 2

Framing Effects

A framing effect occurs when the same data is presented differently to create a different emotional response.

Same statistic — different frame:

"This surgery has a 90% survival rate"

vs.

"This surgery has a 10% mortality rate"

More examples:

"9 out of 10 dentists recommend this toothpaste"

vs.

"1 in 10 dentists does NOT recommend this toothpaste"

Detective question: "What is the full number? Can I restate this statistic a different way?"

7 / 15
Lesson 3

Simpson's Paradox — The Mind-Bender

A trend appears in separate groups — but completely reverses when the groups are combined. Even trained statisticians find this hard.

"How can Treatment A be better for mild patients AND better for severe patients — but Treatment B appear better overall? Let's see..."

The key is group sizes. When groups are very different in size, combining them distorts the picture. The combined number hides what's really happening inside each group.

8 / 15
Lesson 3

Simpson's Paradox — Example 1

Two schools. Which is doing better?

Student GroupSchool A pass rateSchool B pass rateWho's better?
Strong students90% (90 out of 100)85% (17 out of 20)School A ✓
Struggling students30% (6 out of 20)20% (20 out of 100)School A ✓
Overall combined80% (96/120)31% (37/120)School A wins... obviously?

Wait — School A is better in BOTH groups, AND has a higher overall rate. So where's the paradox? Now look at the group sizes: School A has 100 strong + 20 struggling. School B has 20 strong + 100 struggling. School A has MORE easy cases — that inflates its overall rate.

9 / 15
Lesson 3

Simpson's Paradox — Example 2

Two medical treatments. Which should you choose?

Patient TypeTreatment A successTreatment B successWho's better?
Mild cases81% (81/100)87% (234/270)Treatment B
Severe cases73% (192/263)69% (55/80)Treatment A ✓
Overall combined78% (273/363)83% (289/350)Treatment B looks better!

Treatment A is better for severe cases. Treatment B is only used on mild (easier) cases more often. The combined number favors B — but for your health, A is the better choice. Always look at subgroups.

10 / 15
🧠
Brain Break — Name That Trick!

Your teacher will describe a scenario. Call out the deception technique: Cherry-Picking, Survivorship Bias, Framing Effect, or Simpson's Paradox!

"A magazine only publishes success stories from people who used their diet plan" · "A graph's Y-axis starts at 94%" · "The drug is 95% safe!" said instead of "5% of users had serious side effects"

Naming the trick is the first step to defeating it!

11 / 15
Activity Time!

Data Detectives — 4 Case Studies

Each case study contains accurate data — but draws a misleading conclusion. Your job:

⏱ You have 15 minutes in your groups. Then we share out. Remember: the data itself is real — the interpretation is the problem.

12 / 15
Debrief

What Did the Data Detectives Find?

Case 1 — Cherry-Picking

Only the 3 best sales months shown. The other 9 were declining. Detective Q: "What data was left out?"

Case 2 — Survivorship Bias

"Successful athletes train 6hrs/day." Missing: the many who trained equally hard and didn't succeed. Detective Q: "Who's NOT in this data?"

Case 3 — Framing Effect

Drug A: "20% side effects." Drug B: "80% side-effect free." Same drug. Detective Q: "Can I restate this differently?"

Case 4 — Simpson's Paradox

Overall reading scores rose — but both groups dropped. Change in student mix caused it. Detective Q: "What do subgroups show?"

13 / 15
Write Time

Reflection Write

"Which deception technique was hardest for you to spot — and why? What question would a Data Detective ask to catch it?"

✍️ 6 minutes. Use your worksheet — Part 4. Name the specific technique and propose your detective question.

14 / 15
Vocabulary Check

Session 4 Key Terms

Cherry-Picking
Selecting only favorable data; ignoring contradictory evidence
Survivorship Bias
Only studying "survivors" of a process; missing those who didn't make it
Framing Effect
Same data, different presentation, different emotional impact
Simpson's Paradox
A trend reverses when groups are combined due to different group sizes
Confirmation Bias
Noticing data that agrees with you; overlooking data that doesn't
The Detective Question
"What data is missing here? Who chose what to show?"
15 / 15
Session Close

The Data Detective's Creed

"Accurate data + misleading presentation = misinformation.
Ask what's missing. Look at the subgroups. Check the axis."

Next session: Even honest data has randomness built in. We learn about probability — and why 10 coin flips can look very misleading.