Texas Sharpshooter Fallacy

aka Clustering Illusion Fallacy · Post-Designation Fallacy · Data Dredging Fallacy

Cherry-picking a data cluster and building a story around it after the fact, ignoring everything that doesn't fit.

Illustration: Texas Sharpshooter Fallacy
WHAT IT IS

The glitch, explained plainly.

Imagine you throw a whole bag of darts at a wall with your eyes closed. Then you walk up, find the three darts that happen to be close together, draw a circle around them, and tell your friends you're an amazing dart player. You're not lying that those darts are close — but you're hiding all the other darts that went everywhere. That's what this fallacy is: picking the part that looks good and pretending the messy rest doesn't exist.

The Texas Sharpshooter Fallacy occurs when someone sifts through a large or complex dataset, identifies an apparent pattern or cluster after the fact, and then presents it as though it were a predicted or meaningful finding — while discarding all the surrounding data that fails to support the conclusion. Unlike forming a genuine hypothesis before examining evidence, this fallacy reverses the scientific process: the 'target' is drawn around whatever already looks interesting. It is especially pernicious because the patterns discovered are often real in a superficial statistical sense — they just aren't meaningful, having emerged from random variation in sufficiently large data. The fallacy is a core driver of the replication crisis in science, where researchers engage in 'HARKing' (Hypothesizing After Results are Known) and selectively report only the variables that yielded statistical significance.

SOUND FAMILIAR?

Where it shows up.

  1. 01 A marketing analyst runs a campaign across 30 demographic segments. Only one segment — women aged 35-40 in suburban zip codes — shows a significant uptick in conversions. She presents this to leadership as proof they've found their ideal customer profile and recommends reallocating the entire budget to target that segment, without mentioning the 29 segments that showed no effect.
  2. 02 A school board notices that three schools in the eastern district all scored above the state average on math tests. They publicly credit the eastern district's teaching methods as superior and propose expanding them district-wide, without examining whether the clustering is simply random variation among dozens of schools.
  3. 03 A nutritional researcher measures 50 different biomarkers in participants following a new diet. One biomarker — serum folate — shows a statistically significant improvement. The researcher publishes a paper titled 'Novel Diet Significantly Improves Folate Levels,' omitting the 49 biomarkers that showed no change and not correcting for multiple comparisons.
  4. 04 After a plane crash, a journalist discovers that the pilot had visited a particular airport three times in the past year, just like two other pilots involved in recent incidents. She writes an investigative piece suggesting that airport may have a dangerous influence on pilot safety, without examining how many thousands of pilots visited that airport with no incidents.
  5. 05 A venture capitalist reviews the portfolios of the top 10 performing funds and notices that 7 of them invested early in companies with female CEOs. He constructs a thesis that female-led startups outperform, basing a new investment strategy on this observation — without accounting for the hundreds of funds that also invested in female-led companies and performed unremarkably.
IN DIFFERENT DOMAINS

Where it shows up at work.

The same glitch looks different depending on the terrain. Finance, medicine, a relationship, a team — same mechanism, different costume.

Finance & investing

Investors and analysts frequently back-test trading strategies across decades of market data, then highlight the specific combination of indicators that would have yielded the highest returns — without accounting for the thousands of indicator combinations that failed. This leads to overfitted models that perform well historically but collapse on new data.

Medicine & diagnosis

Epidemiologists face constant pressure from communities reporting apparent disease clusters. When boundaries are drawn around observed cases after the fact rather than defined in advance, nearly any region can appear to have an abnormally high rate of illness, leading to false alarms, wasted resources, and public fear about non-existent environmental hazards.

Education & grading

Administrators may examine student performance data across many variables — teacher, classroom, time of day, curriculum version — and highlight whichever combination shows the best results as a 'proven approach,' ignoring the many other combinations that showed no effect and failing to validate the finding with new data.

Relationships

People sometimes review a partner's past behavior after discovering infidelity, selectively stringing together small moments that 'now make sense' as evidence of a pattern of deception, while ignoring the overwhelming number of interactions that were genuinely honest and loving.

Tech & product

Product teams running multivariate A/B tests across many user segments, features, and time windows will inevitably find some combination that shows a statistically significant result. Treating that finding as a validated insight without replication leads to feature changes that don't actually improve the product.

Workplace & hiring

HR departments may identify that several top performers share a particular trait — such as attending the same university or having a hobby in common — and build hiring criteria around it, ignoring the many employees with the same trait who performed poorly and the top performers who lack it entirely.

Politics Media

Pundits cherry-pick economic indicators, crime statistics, or polling data from specific time windows or regions that support their preferred political narrative, while ignoring the broader dataset that tells a more ambiguous or contradictory story.

HOW TO SPOT IT

Ask yourself…

  • Did I define what I was looking for before I looked at the data, or did I find this pattern after browsing?
  • How many other patterns, variables, or clusters did I examine and discard before landing on this one?
  • Would this pattern hold up if I tested it on a completely new, independent dataset?
HOW TO DEFEND AGAINST IT

The playbook.

  • Pre-register hypotheses: Before analyzing data, write down exactly what you expect to find and what would count as evidence for or against it.
  • Apply the 'how many barns?' test: Ask how many variables, subgroups, or time windows you examined before finding this pattern. The more you looked, the less meaningful any single cluster is.
  • Demand replication on fresh data: Never trust a pattern found in exploratory analysis until it has been validated on an independent dataset that was not used to discover it.
  • Use statistical corrections for multiple comparisons (e.g., Bonferroni correction) when testing many hypotheses simultaneously.
  • Actively seek the 'misses': Force yourself to look at all the data points that don't fit the pattern, not just the ones that do.
FAMOUS CASES

In history.

  • The Erin Brockovich case: Hexavalent chromium in a California town's water was blamed for a cancer cluster, resulting in a $333 million settlement, but subsequent epidemiological analysis showed the cancer rate was no higher than — and actually slightly below — the general population.
  • Nostradamus prophecy interpretations: Over 1,000 vague quatrains are retroactively matched to modern events by selecting the few that seem to fit while ignoring the vast majority that don't.
  • The 1980s Texas cancer cluster scare: Elevated cancer rates in certain Texas counties triggered public alarm and investigation, but rigorous statistical analysis found no consistent environmental cause — the clusters were consistent with random variation.
  • The Bible Code phenomenon: Researchers claimed to find hidden prophetic messages in the Torah by selecting specific letter-spacing patterns, ignoring the fact that similarly 'meaningful' patterns can be found in any sufficiently long text.
WHERE IT COMES FROM
Academic origin

Epidemiologist Seymour Grufferman, 1977. First appeared in a paper on Hodgkin's disease clustering (Cancer, 39: 1829–1833), using the sharpshooter metaphor to warn against post-hoc interpretation of disease clusters. The concept gained broader traction in critical thinking and statistics during the 1990s.

Evolutionary origin

Ancestral humans who over-detected patterns — seeing a predator in rustling grass even when it was just wind — survived more often than those who under-detected. This asymmetry in the cost of errors (Type I errors being cheaper than Type II errors) selected for brains that aggressively find clusters and assign meaning, even at the expense of frequent false positives.

IN AI SYSTEMS

How the machines inherit it.

Machine learning models are highly susceptible to this fallacy during training. When models are evaluated on the same data used to find patterns (overfitting), they 'draw the bullseye' around noise in the training set. Feature selection algorithms that test thousands of variables and report only the significant ones without correction replicate the fallacy at computational scale. In LLMs, cherry-picked benchmark results or selectively reported evaluation metrics can create a misleading picture of model capability.

Read more on Wikipedia
FREE FIELD ZINE

10 glitches quietly running your life.

A free field-zine PDF — ten cognitive glitches named, illustrated, with a defense move for each. Plus the weekly Glitch Report on Fridays — one bias named, two spotted in the wild, one defense move. Unsubscribe any time.

EXPLORE MORE

Related glitches.

LAUNCH PRICE

Train against your blindspots.

50 cards are free to preview. Buyers unlock the rest of the deck plus the interactive training — Spot-the-Bias Quiz unlimited, Swipe Deck with spaced repetition, My Blindspots, Decision Pre-Flight, the Printable Deck + Cheat Sheets, and the Field Guide e-book. $29.50$59.

Unlock the full deck

Everything below — yours forever. Pay once, use across every device.

Half-off launch — limited to the first 100 readers. Auto-applied at checkout.
$59 $29.50
one-time payment · lifetime access
  • All interactive digital cards — search, filter, flip, shuffle on any device
  • Five training modes — Spot-the-Bias Quiz, Swipe Deck, Pre-Flight, Blindspots, Journal
  • Curated Lenses + Decision Templates + Defense Playbook
  • Printable Deck PDFs + Field Guide e-book + Cheat Sheets + Anki Export
  • Every future improvement, included
Unlock  $29.50

30-day refund · no questions asked

Unlock the full deck

Everything below — yours forever. Pay once, use across every device.

Half-off launch — limited to the first 100 readers. Auto-applied at checkout.
$59 $29.50
one-time payment · lifetime access
  • All interactive digital cards — search, filter, flip, shuffle on any device
  • Five training modes — Spot-the-Bias Quiz, Swipe Deck, Pre-Flight, Blindspots, Journal
  • Curated Lenses + Decision Templates + Defense Playbook
  • Printable Deck PDFs + Field Guide e-book + Cheat Sheets + Anki Export
  • Every future improvement, included
Unlock  $29.50

30-day refund · no questions asked