Files
2025-11-30 08:30:14 +08:00

11 KiB

Common Biases in Scientific Research

Cognitive Biases Affecting Researchers

1. Confirmation Bias

Description: Tendency to search for, interpret, and recall information that confirms preexisting beliefs.

Manifestations:

  • Designing studies that can only support the hypothesis
  • Interpreting ambiguous results as supportive
  • Remembering hits and forgetting misses
  • Selectively citing literature that agrees

Mitigation:

  • Preregister hypotheses and analysis plans
  • Actively seek disconfirming evidence
  • Use blinded data analysis
  • Consider alternative hypotheses

2. Hindsight Bias (I-Knew-It-All-Along Effect)

Description: After an event, people perceive it as having been more predictable than it actually was.

Manifestations:

  • HARKing (Hypothesizing After Results are Known)
  • Claiming predictions that weren't made
  • Underestimating surprise at results

Mitigation:

  • Document predictions before data collection
  • Preregister studies
  • Distinguish exploratory from confirmatory analyses

3. Publication Bias (File Drawer Problem)

Description: Positive/significant results are more likely to be published than negative/null results.

Manifestations:

  • Literature appears to support effects that don't exist
  • Overestimation of effect sizes
  • Inability to estimate true effects from published literature

Mitigation:

  • Publish null results
  • Use preregistration and registered reports
  • Conduct systematic reviews with grey literature
  • Check for funnel plot asymmetry in meta-analyses

4. Anchoring Bias

Description: Over-reliance on the first piece of information encountered.

Manifestations:

  • Initial hypotheses unduly influence interpretation
  • First studies in a field set expectations
  • Pilot data biases main study interpretation

Mitigation:

  • Consider multiple initial hypotheses
  • Evaluate evidence independently
  • Use structured decision-making

5. Availability Heuristic

Description: Overestimating likelihood of events based on how easily examples come to mind.

Manifestations:

  • Overemphasizing recent or dramatic findings
  • Neglecting base rates
  • Anecdotal evidence overshadowing statistics

Mitigation:

  • Consult systematic reviews, not memorable papers
  • Consider base rates explicitly
  • Use statistical thinking, not intuition

6. Bandwagon Effect

Description: Adopting beliefs because many others hold them.

Manifestations:

  • Following research trends without critical evaluation
  • Citing widely-cited papers without reading
  • Accepting "textbook knowledge" uncritically

Mitigation:

  • Evaluate evidence independently
  • Read original sources
  • Question assumptions

7. Belief Perseverance

Description: Maintaining beliefs even after evidence disproving them.

Manifestations:

  • Defending theories despite contradictory evidence
  • Finding ad hoc explanations for discrepant results
  • Dismissing replication failures

Mitigation:

  • Explicitly consider what evidence would change your mind
  • Update beliefs based on evidence
  • Distinguish between theories and ego

8. Outcome Bias

Description: Judging decisions based on outcomes rather than the quality of the decision at the time.

Manifestations:

  • Valuing lucky guesses over sound methodology
  • Dismissing good studies with null results
  • Rewarding sensational findings over rigorous methods

Mitigation:

  • Evaluate methodology independently of results
  • Value rigor and transparency
  • Recognize role of chance

Experimental and Methodological Biases

9. Selection Bias

Description: Systematic differences between those selected for study and those not selected.

Types:

  • Sampling bias: Non-random sample
  • Attrition bias: Systematic dropout
  • Volunteer bias: Self-selected participants differ
  • Berkson's bias: Hospital patients differ from general population
  • Survivorship bias: Only examining "survivors"

Detection:

  • Compare characteristics of participants vs. target population
  • Analyze dropout patterns
  • Consider who is missing from the sample

Mitigation:

  • Random sampling
  • Track and analyze non-responders
  • Use strategies to minimize dropout
  • Report participant flow diagrams

10. Observer Bias (Detection Bias)

Description: Researchers' expectations influence observations or measurements.

Manifestations:

  • Measuring outcomes differently across groups
  • Interpreting ambiguous results based on group assignment
  • Unconsciously cueing participants

Mitigation:

  • Blinding of observers/assessors
  • Objective, automated measurements
  • Standardized protocols
  • Inter-rater reliability checks

11. Performance Bias

Description: Systematic differences in care provided to comparison groups.

Manifestations:

  • Treating experimental group differently
  • Providing additional attention to one group
  • Differential adherence to protocols

Mitigation:

  • Standardize all procedures
  • Blind participants and providers
  • Use placebo controls
  • Monitor protocol adherence

12. Measurement Bias (Information Bias)

Description: Systematic errors in how variables are measured.

Types:

  • Recall bias: Systematic differences in accuracy of recall
  • Social desirability bias: Responding in socially acceptable ways
  • Interviewer bias: Interviewer's characteristics affect responses
  • Instrument bias: Measurement tools systematically err

Mitigation:

  • Use validated, objective measures
  • Standardize data collection
  • Blind participants to hypotheses
  • Verify self-reports with objective data

13. Confounding Bias

Description: Effect of extraneous variable mixed with the variable of interest.

Examples:

  • Age confounding relationship between exercise and health
  • Socioeconomic status confounding education and outcomes
  • Indication bias in treatment studies

Mitigation:

  • Randomization
  • Matching
  • Statistical adjustment
  • Stratification
  • Restriction

14. Reporting Bias

Description: Selective reporting of results.

Types:

  • Outcome reporting bias: Selectively reporting outcomes
  • Time-lag bias: Delayed publication of negative results
  • Language bias: Publishing positive results in English
  • Citation bias: Preferentially citing positive studies

Mitigation:

  • Preregister all outcomes
  • Report all planned analyses
  • Distinguish primary from secondary outcomes
  • Use study registries

15. Spectrum Bias

Description: Test performance varies depending on the spectrum of disease severity in the sample.

Manifestations:

  • Diagnostic tests appearing more accurate in extreme cases
  • Treatment effects differing by severity

Mitigation:

  • Test in representative samples
  • Report performance across disease spectrum
  • Avoid case-control designs for diagnostic studies

16. Lead-Time Bias

Description: Apparent survival benefit due to earlier detection, not improved outcomes.

Example:

  • Screening detecting disease earlier makes survival seem longer, even if death occurs at same age

Mitigation:

  • Measure mortality, not just survival from diagnosis
  • Use randomized screening trials
  • Consider length-time and overdiagnosis bias

17. Length-Time Bias

Description: Screening disproportionately detects slower-growing, less aggressive cases.

Example:

  • Slow-growing cancers detected more often than fast-growing ones, making screening appear beneficial

Mitigation:

  • Randomized trials with mortality endpoints
  • Consider disease natural history

18. Response Bias

Description: Systematic pattern in how participants respond.

Types:

  • Acquiescence bias: Tendency to agree
  • Extreme responding: Always choosing extreme options
  • Neutral responding: Avoiding extreme responses
  • Demand characteristics: Responding based on perceived expectations

Mitigation:

  • Mix positive and negative items
  • Use multiple response formats
  • Blind participants to hypotheses
  • Use behavioral measures

Statistical and Analysis Biases

19. P-Hacking (Data Dredging)

Description: Manipulating data or analyses until significant results emerge.

Manifestations:

  • Collecting data until significance reached
  • Testing multiple outcomes, reporting only significant ones
  • Trying multiple analysis methods
  • Excluding "outliers" to reach significance
  • Subgroup analyses until finding significance

Detection:

  • Suspiciously perfect p-values (just below .05)
  • Many researcher degrees of freedom
  • Undisclosed analyses
  • Fishing expeditions

Mitigation:

  • Preregister analysis plans
  • Report all analyses conducted
  • Correct for multiple comparisons
  • Distinguish exploratory from confirmatory

20. HARKing (Hypothesizing After Results are Known)

Description: Presenting post hoc hypotheses as if they were predicted a priori.

Why problematic:

  • Inflates apparent evidence
  • Conflates exploration with confirmation
  • Misrepresents the scientific process

Mitigation:

  • Preregister hypotheses
  • Clearly label exploratory analyses
  • Require replication of unexpected findings

21. Base Rate Neglect

Description: Ignoring prior probability when evaluating evidence.

Example:

  • Test with 95% accuracy in rare disease (1% prevalence): positive result only 16% likely to indicate disease

Mitigation:

  • Always consider base rates/prior probability
  • Use Bayesian reasoning
  • Report positive and negative predictive values

22. Regression to the Mean

Description: Extreme measurements tend to be followed by less extreme ones.

Manifestations:

  • Treatment effects in extreme groups may be regression artifacts
  • "Sophomore slump" in high performers

Mitigation:

  • Use control groups
  • Consider natural variation
  • Don't select based on extreme baseline values without controls

23. Texas Sharpshooter Fallacy

Description: Selecting data after seeing patterns, like shooting arrows then drawing targets around clusters.

Manifestations:

  • Finding patterns in random data
  • Subgroup analyses selected post hoc
  • Geographic clustering studies without correction

Mitigation:

  • Prespecify hypotheses
  • Correct for multiple comparisons
  • Replicate findings in independent data

Reducing Bias: Best Practices

Study Design

  1. Randomization
  2. Blinding (single, double, triple)
  3. Control groups
  4. Adequate sample size
  5. Preregistration

Data Collection

  1. Standardized protocols
  2. Validated instruments
  3. Objective measures when possible
  4. Multiple observers/raters
  5. Complete data collection

Analysis

  1. Intention-to-treat analysis
  2. Prespecified analyses
  3. Appropriate statistical tests
  4. Multiple comparison corrections
  5. Sensitivity analyses

Reporting

  1. Complete transparency
  2. CONSORT, PRISMA, or similar guidelines
  3. Report all outcomes
  4. Distinguish exploratory from confirmatory
  5. Share data and code

Meta-Level

  1. Adversarial collaboration
  2. Replication studies
  3. Open science practices
  4. Peer review
  5. Systematic reviews