Files
2025-11-30 08:38:26 +08:00

329 lines
13 KiB
Markdown
Raw Permalink Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Research Claim Map: Advanced Methodologies
## Table of Contents
1. [Triangulation Techniques](#1-triangulation-techniques)
2. [Source Verification Methods](#2-source-verification-methods)
3. [Evidence Synthesis Frameworks](#3-evidence-synthesis-frameworks)
4. [Bias Detection and Mitigation](#4-bias-detection-and-mitigation)
5. [Confidence Calibration Techniques](#5-confidence-calibration-techniques)
6. [Advanced Investigation Patterns](#6-advanced-investigation-patterns)
## 1. Triangulation Techniques
### Multi-Source Verification
**Independent corroboration**:
- **Minimum 3 independent sources** for high-confidence claims
- Sources are independent if: different authors, organizations, funding, data collection methods
- Example: Government report + Academic study + Industry analysis (all using different data)
**Detecting circular citations**:
- Trace back to original source - if A cites B, B cites C, C cites A → circular, invalid
- Check publication dates - later sources should cite earlier, not reverse
- Use citation indexes (Google Scholar, Web of Science) to map citation networks
**Convergent evidence**:
- Different methodologies reaching same conclusion (surveys + experiments + observational)
- Different populations/contexts showing same pattern
- Example: Lab studies + field studies + meta-analyses all finding same effect
### Cross-Checking Strategies
**Fact-checking databases**:
- Snopes, FactCheck.org, PolitiFact for public claims
- Retraction Watch for scientific papers
- OpenSecrets for political funding claims
- SEC EDGAR for financial claims
**Domain-specific verification**:
- Medical: PubMed, Cochrane Reviews, FDA databases
- Technology: CVE databases, vendor security advisories, benchmark repositories
- Business: Crunchbase, SEC filings, earnings transcripts
- Historical: Primary source archives, digitized records
**Temporal consistency**:
- Check if claim was true at time stated (not just currently)
- Verify dates in citations match narrative
- Look for anachronisms (technology/events cited before they existed)
## 2. Source Verification Methods
### CRAAP Test (Currency, Relevance, Authority, Accuracy, Purpose)
**Currency**: When was it published/updated?
- High: Within last year for fast-changing topics, within 5 years for stable domains
- Medium: Dated but still applicable
- Low: Outdated, context has changed significantly
**Relevance**: Does it address your specific claim?
- High: Directly addresses claim with same scope/context
- Medium: Related but different scope (e.g., different population, timeframe)
- Low: Tangentially related, requires extrapolation
**Authority**: Who is the author/publisher?
- High: Recognized expert, peer-reviewed publication, established institution
- Medium: Knowledgeable but not top-tier, some editorial oversight
- Low: Unknown author, self-published, no credentials
**Accuracy**: Can it be verified?
- High: Data/methods shared, replicable, other sources corroborate
- Medium: Some verification possible, mostly consistent with known facts
- Low: Unverifiable claims, contradicts established knowledge
**Purpose**: Why was it created?
- High: Inform/educate, transparent about limitations
- Medium: Persuade but with evidence, some bias acknowledged
- Low: Sell/propagandize, misleading framing, undisclosed conflicts
### Domain Authority Assessment
**Academic sources**:
- Journal impact factor (higher = more rigorous peer review)
- H-index of authors (citation impact)
- Institutional affiliation (R1 research university > teaching-focused college)
- Funding source disclosure (NIH grant > pharmaceutical company funding for drug study)
**News sources**:
- Editorial standards (corrections policy, fact-checking team)
- Awards/recognition (Pulitzer, Peabody, investigative journalism awards)
- Ownership transparency (independent > owned by entity with vested interest)
- Track record (history of accurate reporting vs retractions)
**Technical sources**:
- Benchmark methodology disclosure (reproducible specs, public data)
- Vendor independence (third-party testing > vendor self-reporting)
- Community verification (open-source code, peer reproduction)
- Standards compliance (IEEE, NIST, OWASP standards)
## 3. Evidence Synthesis Frameworks
### GRADE System (Grading of Recommendations Assessment, Development and Evaluation)
**Start with evidence type**:
- Randomized controlled trials (RCTs): Start HIGH quality
- Observational studies: Start LOW quality
- Expert opinion: Start VERY LOW quality
**Downgrade for**:
- Risk of bias (methodology flaws, conflicts of interest)
- Inconsistency (conflicting results across studies)
- Indirectness (different population/intervention than claim)
- Imprecision (small sample, wide confidence intervals)
- Publication bias (only positive results published)
**Upgrade for**:
- Large effect size (strong signal)
- Dose-response gradient (more X → more Y)
- All plausible confounders would reduce effect (conservative estimate)
**Final quality rating**:
- **High**: Very confident true effect is close to estimate
- **Moderate**: Moderately confident, true effect likely close
- **Low**: Limited confidence, true effect may differ substantially
- **Very Low**: Very little confidence, true effect likely very different
### Meta-Analysis Interpretation
**Effect size + confidence intervals**:
- Large effect + narrow CI = high confidence
- Small effect + narrow CI = real but modest effect
- Any effect + wide CI = uncertain
- Example: "10% improvement (95% CI: 5-15%)" vs "10% improvement (95% CI: -5-25%)"
**Heterogeneity (I² statistic)**:
- I² < 25%: Low heterogeneity, studies agree
- I² 25-75%: Moderate heterogeneity, some variation
- I² > 75%: High heterogeneity, studies conflict (be skeptical of pooled estimate)
**Publication bias detection**:
- Funnel plot asymmetry (missing small negative studies)
- File drawer problem (unpublished null results)
- Check trial registries (ClinicalTrials.gov) for unreported studies
## 4. Bias Detection and Mitigation
### Common Cognitive Biases in Claim Evaluation
**Confirmation bias**:
- **Symptom**: Finding only supporting evidence, ignoring contradictions
- **Mitigation**: Actively search for "why this might be wrong", assign someone to argue against
- **Example**: Believing vendor claim because you want product to work
**Availability bias**:
- **Symptom**: Overweighting vivid anecdotes vs dry statistics
- **Mitigation**: Prioritize data over stories, ask "how representative?"
- **Example**: Fearing plane crashes (vivid news) over car crashes (statistically riskier)
**Authority bias**:
- **Symptom**: Accepting claims because source is prestigious (Nobel Prize, Harvard, etc.)
- **Mitigation**: Evaluate evidence quality independently, check if expert in this specific domain
- **Example**: Believing physicist's medical claims (out of domain expertise)
**Anchoring bias**:
- **Symptom**: First number heard becomes reference point
- **Mitigation**: Seek base rates, compare to industry benchmarks, gather range of estimates
- **Example**: Vendor says "saves 50%" → anchor on 50%, skeptical of analyst saying 10%
**Recency bias**:
- **Symptom**: Overweighting latest information, dismissing older evidence
- **Mitigation**: Consider full timeline, check if latest is outlier or trend
- **Example**: One bad quarter → ignoring 5 years of growth
### Source Bias Indicators
**Financial conflicts of interest**:
- Study funded by company whose product is being evaluated
- Author owns stock, serves on board, receives consulting fees
- Disclosure: Look for "Conflicts of Interest" section in papers, FDA disclosures
**Ideological bias**:
- Think tank with known political lean
- Advocacy organization with mission-driven agenda
- Framing: Watch for loaded language, cherry-picked comparisons
**Selection bias in studies**:
- Participants not representative of target population
- Dropout rate differs between groups
- Outcomes measured selectively (dropped endpoints with null results)
**Reporting bias**:
- Positive results published, negative results buried
- Outcomes changed after seeing data (HARKing: Hypothesizing After Results Known)
- Subsetting data until significance found (p-hacking)
## 5. Confidence Calibration Techniques
### Bayesian Updating
**Start with prior probability** (before seeing evidence):
- Base rate: How often is this type of claim true?
- Example: "New product will disrupt market" - base rate ~5% (most fail)
**Update with evidence** (likelihood ratio):
- How much more likely is this evidence if claim is true vs false?
- Strong evidence: Likelihood ratio >10 (evidence 10× more likely if claim true)
- Weak evidence: Likelihood ratio <3
**Calculate posterior probability** (after evidence):
- Use Bayes theorem or intuitive updating
- Example: Prior 5%, strong evidence (LR=10) → Posterior ~35%
### Fermi Estimation for Sanity Checks
**Decompose claim into estimable parts**:
- Claim: "Company has 10,000 paying customers"
- Decompose: Employees × customers per employee, or revenue ÷ price per customer
- Cross-check: Do the numbers add up?
**Example**:
- Claim: Startup has 1M users
- Check: Founded 2 years ago → 1,370 new users/day → 57/hour (24/7) or 171/hour (8hr workday)
- Reality check: Plausible for viral product? Need marketing spend estimate.
### Confidence Intervals and Ranges
**Avoid point estimates** ("70% confident"):
- Use ranges: "60-80% confident" acknowledges uncertainty
- Ask: What would make me 90% confident? What's missing?
**Sensitivity analysis**:
- Best case scenario (all assumptions optimistic) → upper bound confidence
- Worst case scenario (all assumptions pessimistic) → lower bound confidence
- Most likely scenario → central estimate
## 6. Advanced Investigation Patterns
### Investigative Journalism Techniques
**Paper trail following**:
- Follow money: Who benefits financially from this claim being believed?
- Follow timeline: Who said what when? Any story changes over time?
- Follow power: Who has authority/incentive to suppress contradicting evidence?
**Source cultivation**:
- Insider sources (whistleblowers, former employees) for claims companies hide
- Expert sources (academics, consultants) for technical evaluation
- Documentary sources (contracts, emails, internal memos) for ground truth
**Red flags in interviews**:
- Vague answers to specific questions
- Defensiveness or hostility when questioned
- Inconsistencies between different tellings
- Refusal to provide documentation
### Legal Evidence Standards
**Burden of proof levels**:
- **Beyond reasonable doubt** (criminal): 95%+ confidence
- **Clear and convincing** (civil high stakes): 75%+ confidence
- **Preponderance of evidence** (civil standard): 51%+ confidence (more likely than not)
**Hearsay rules**:
- Firsthand testimony > secondhand ("I saw X" > "Someone told me X")
- Exception: Business records, public records (trustworthy hearsay)
- Watch for: Anonymous sources, "people are saying", "experts claim"
**Chain of custody**:
- Document handling: Who collected, stored, analyzed evidence?
- Tampering risk: Could evidence have been altered?
- Authentication: How do we know this document/photo is genuine?
### Competitive Intelligence Validation
**HUMINT (Human Intelligence)**:
- Customer interviews: "Do you use competitor's product? How does it work?"
- Former employees: Glassdoor reviews, LinkedIn networking
- Conference presentations: Technical details revealed publicly
**OSINT (Open Source Intelligence)**:
- Public filings: SEC 10-K, patents, trademarks
- Job postings: What skills are they hiring for? (reveals technology stack, strategic priorities)
- Social media: Employee posts, company announcements
- Web archives: Wayback Machine to see claim history, website changes
**TECHINT (Technical Intelligence)**:
- Reverse engineering: Analyze product directly
- Benchmarking: Test performance claims yourself
- Network analysis: DNS records, API endpoints, infrastructure footprint
### Scientific Reproducibility Assessment
**Replication indicator**:
- Has anyone reproduced the finding? (Strong evidence)
- Did replication attempts fail? (Evidence against)
- Has no one tried to replicate? (Unknown, be cautious)
**Pre-registration check**:
- Was study pre-registered (ClinicalTrials.gov, OSF)? Reduces p-hacking risk
- Do results match pre-registered outcomes? If different, why?
**Data/code availability**:
- Can you access raw data to re-analyze?
- Is code available to reproduce analysis?
- Are materials specified to replicate experiment?
**Robustness checks**:
- Do findings hold with different analysis methods?
- Are results sensitive to outliers or specific assumptions?
- Do subsample analyses show consistent effects?
---
## Workflow Integration
**When to use advanced techniques**:
**Triangulation** → Every claim (minimum requirement)
**CRAAP Test** → When assessing unfamiliar sources
**GRADE System** → Medical/health claims, policy decisions
**Bayesian Updating** → When you have prior knowledge/base rates
**Fermi Estimation** → Quantitative claims that seem implausible
**Investigative Techniques** → High-stakes business decisions, fraud detection
**Legal Standards** → Determining action thresholds (e.g., firing employee, lawsuit)
**Reproducibility Assessment** → Scientific/technical claims
**Start simple, add complexity as needed**:
1. Quick verification: CRAAP test + Google fact-check
2. Moderate investigation: Triangulate 3 sources + basic bias check
3. Deep investigation: Full methodology above + expert consultation