Initial commit
This commit is contained in:
328
skills/research-claim-map/resources/methodology.md
Normal file
328
skills/research-claim-map/resources/methodology.md
Normal file
@@ -0,0 +1,328 @@
|
||||
# Research Claim Map: Advanced Methodologies
|
||||
|
||||
## Table of Contents
|
||||
1. [Triangulation Techniques](#1-triangulation-techniques)
|
||||
2. [Source Verification Methods](#2-source-verification-methods)
|
||||
3. [Evidence Synthesis Frameworks](#3-evidence-synthesis-frameworks)
|
||||
4. [Bias Detection and Mitigation](#4-bias-detection-and-mitigation)
|
||||
5. [Confidence Calibration Techniques](#5-confidence-calibration-techniques)
|
||||
6. [Advanced Investigation Patterns](#6-advanced-investigation-patterns)
|
||||
|
||||
## 1. Triangulation Techniques
|
||||
|
||||
### Multi-Source Verification
|
||||
|
||||
**Independent corroboration**:
|
||||
- **Minimum 3 independent sources** for high-confidence claims
|
||||
- Sources are independent if: different authors, organizations, funding, data collection methods
|
||||
- Example: Government report + Academic study + Industry analysis (all using different data)
|
||||
|
||||
**Detecting circular citations**:
|
||||
- Trace back to original source - if A cites B, B cites C, C cites A → circular, invalid
|
||||
- Check publication dates - later sources should cite earlier, not reverse
|
||||
- Use citation indexes (Google Scholar, Web of Science) to map citation networks
|
||||
|
||||
**Convergent evidence**:
|
||||
- Different methodologies reaching same conclusion (surveys + experiments + observational)
|
||||
- Different populations/contexts showing same pattern
|
||||
- Example: Lab studies + field studies + meta-analyses all finding same effect
|
||||
|
||||
### Cross-Checking Strategies
|
||||
|
||||
**Fact-checking databases**:
|
||||
- Snopes, FactCheck.org, PolitiFact for public claims
|
||||
- Retraction Watch for scientific papers
|
||||
- OpenSecrets for political funding claims
|
||||
- SEC EDGAR for financial claims
|
||||
|
||||
**Domain-specific verification**:
|
||||
- Medical: PubMed, Cochrane Reviews, FDA databases
|
||||
- Technology: CVE databases, vendor security advisories, benchmark repositories
|
||||
- Business: Crunchbase, SEC filings, earnings transcripts
|
||||
- Historical: Primary source archives, digitized records
|
||||
|
||||
**Temporal consistency**:
|
||||
- Check if claim was true at time stated (not just currently)
|
||||
- Verify dates in citations match narrative
|
||||
- Look for anachronisms (technology/events cited before they existed)
|
||||
|
||||
## 2. Source Verification Methods
|
||||
|
||||
### CRAAP Test (Currency, Relevance, Authority, Accuracy, Purpose)
|
||||
|
||||
**Currency**: When was it published/updated?
|
||||
- High: Within last year for fast-changing topics, within 5 years for stable domains
|
||||
- Medium: Dated but still applicable
|
||||
- Low: Outdated, context has changed significantly
|
||||
|
||||
**Relevance**: Does it address your specific claim?
|
||||
- High: Directly addresses claim with same scope/context
|
||||
- Medium: Related but different scope (e.g., different population, timeframe)
|
||||
- Low: Tangentially related, requires extrapolation
|
||||
|
||||
**Authority**: Who is the author/publisher?
|
||||
- High: Recognized expert, peer-reviewed publication, established institution
|
||||
- Medium: Knowledgeable but not top-tier, some editorial oversight
|
||||
- Low: Unknown author, self-published, no credentials
|
||||
|
||||
**Accuracy**: Can it be verified?
|
||||
- High: Data/methods shared, replicable, other sources corroborate
|
||||
- Medium: Some verification possible, mostly consistent with known facts
|
||||
- Low: Unverifiable claims, contradicts established knowledge
|
||||
|
||||
**Purpose**: Why was it created?
|
||||
- High: Inform/educate, transparent about limitations
|
||||
- Medium: Persuade but with evidence, some bias acknowledged
|
||||
- Low: Sell/propagandize, misleading framing, undisclosed conflicts
|
||||
|
||||
### Domain Authority Assessment
|
||||
|
||||
**Academic sources**:
|
||||
- Journal impact factor (higher = more rigorous peer review)
|
||||
- H-index of authors (citation impact)
|
||||
- Institutional affiliation (R1 research university > teaching-focused college)
|
||||
- Funding source disclosure (NIH grant > pharmaceutical company funding for drug study)
|
||||
|
||||
**News sources**:
|
||||
- Editorial standards (corrections policy, fact-checking team)
|
||||
- Awards/recognition (Pulitzer, Peabody, investigative journalism awards)
|
||||
- Ownership transparency (independent > owned by entity with vested interest)
|
||||
- Track record (history of accurate reporting vs retractions)
|
||||
|
||||
**Technical sources**:
|
||||
- Benchmark methodology disclosure (reproducible specs, public data)
|
||||
- Vendor independence (third-party testing > vendor self-reporting)
|
||||
- Community verification (open-source code, peer reproduction)
|
||||
- Standards compliance (IEEE, NIST, OWASP standards)
|
||||
|
||||
## 3. Evidence Synthesis Frameworks
|
||||
|
||||
### GRADE System (Grading of Recommendations Assessment, Development and Evaluation)
|
||||
|
||||
**Start with evidence type**:
|
||||
- Randomized controlled trials (RCTs): Start HIGH quality
|
||||
- Observational studies: Start LOW quality
|
||||
- Expert opinion: Start VERY LOW quality
|
||||
|
||||
**Downgrade for**:
|
||||
- Risk of bias (methodology flaws, conflicts of interest)
|
||||
- Inconsistency (conflicting results across studies)
|
||||
- Indirectness (different population/intervention than claim)
|
||||
- Imprecision (small sample, wide confidence intervals)
|
||||
- Publication bias (only positive results published)
|
||||
|
||||
**Upgrade for**:
|
||||
- Large effect size (strong signal)
|
||||
- Dose-response gradient (more X → more Y)
|
||||
- All plausible confounders would reduce effect (conservative estimate)
|
||||
|
||||
**Final quality rating**:
|
||||
- **High**: Very confident true effect is close to estimate
|
||||
- **Moderate**: Moderately confident, true effect likely close
|
||||
- **Low**: Limited confidence, true effect may differ substantially
|
||||
- **Very Low**: Very little confidence, true effect likely very different
|
||||
|
||||
### Meta-Analysis Interpretation
|
||||
|
||||
**Effect size + confidence intervals**:
|
||||
- Large effect + narrow CI = high confidence
|
||||
- Small effect + narrow CI = real but modest effect
|
||||
- Any effect + wide CI = uncertain
|
||||
- Example: "10% improvement (95% CI: 5-15%)" vs "10% improvement (95% CI: -5-25%)"
|
||||
|
||||
**Heterogeneity (I² statistic)**:
|
||||
- I² < 25%: Low heterogeneity, studies agree
|
||||
- I² 25-75%: Moderate heterogeneity, some variation
|
||||
- I² > 75%: High heterogeneity, studies conflict (be skeptical of pooled estimate)
|
||||
|
||||
**Publication bias detection**:
|
||||
- Funnel plot asymmetry (missing small negative studies)
|
||||
- File drawer problem (unpublished null results)
|
||||
- Check trial registries (ClinicalTrials.gov) for unreported studies
|
||||
|
||||
## 4. Bias Detection and Mitigation
|
||||
|
||||
### Common Cognitive Biases in Claim Evaluation
|
||||
|
||||
**Confirmation bias**:
|
||||
- **Symptom**: Finding only supporting evidence, ignoring contradictions
|
||||
- **Mitigation**: Actively search for "why this might be wrong", assign someone to argue against
|
||||
- **Example**: Believing vendor claim because you want product to work
|
||||
|
||||
**Availability bias**:
|
||||
- **Symptom**: Overweighting vivid anecdotes vs dry statistics
|
||||
- **Mitigation**: Prioritize data over stories, ask "how representative?"
|
||||
- **Example**: Fearing plane crashes (vivid news) over car crashes (statistically riskier)
|
||||
|
||||
**Authority bias**:
|
||||
- **Symptom**: Accepting claims because source is prestigious (Nobel Prize, Harvard, etc.)
|
||||
- **Mitigation**: Evaluate evidence quality independently, check if expert in this specific domain
|
||||
- **Example**: Believing physicist's medical claims (out of domain expertise)
|
||||
|
||||
**Anchoring bias**:
|
||||
- **Symptom**: First number heard becomes reference point
|
||||
- **Mitigation**: Seek base rates, compare to industry benchmarks, gather range of estimates
|
||||
- **Example**: Vendor says "saves 50%" → anchor on 50%, skeptical of analyst saying 10%
|
||||
|
||||
**Recency bias**:
|
||||
- **Symptom**: Overweighting latest information, dismissing older evidence
|
||||
- **Mitigation**: Consider full timeline, check if latest is outlier or trend
|
||||
- **Example**: One bad quarter → ignoring 5 years of growth
|
||||
|
||||
### Source Bias Indicators
|
||||
|
||||
**Financial conflicts of interest**:
|
||||
- Study funded by company whose product is being evaluated
|
||||
- Author owns stock, serves on board, receives consulting fees
|
||||
- Disclosure: Look for "Conflicts of Interest" section in papers, FDA disclosures
|
||||
|
||||
**Ideological bias**:
|
||||
- Think tank with known political lean
|
||||
- Advocacy organization with mission-driven agenda
|
||||
- Framing: Watch for loaded language, cherry-picked comparisons
|
||||
|
||||
**Selection bias in studies**:
|
||||
- Participants not representative of target population
|
||||
- Dropout rate differs between groups
|
||||
- Outcomes measured selectively (dropped endpoints with null results)
|
||||
|
||||
**Reporting bias**:
|
||||
- Positive results published, negative results buried
|
||||
- Outcomes changed after seeing data (HARKing: Hypothesizing After Results Known)
|
||||
- Subsetting data until significance found (p-hacking)
|
||||
|
||||
## 5. Confidence Calibration Techniques
|
||||
|
||||
### Bayesian Updating
|
||||
|
||||
**Start with prior probability** (before seeing evidence):
|
||||
- Base rate: How often is this type of claim true?
|
||||
- Example: "New product will disrupt market" - base rate ~5% (most fail)
|
||||
|
||||
**Update with evidence** (likelihood ratio):
|
||||
- How much more likely is this evidence if claim is true vs false?
|
||||
- Strong evidence: Likelihood ratio >10 (evidence 10× more likely if claim true)
|
||||
- Weak evidence: Likelihood ratio <3
|
||||
|
||||
**Calculate posterior probability** (after evidence):
|
||||
- Use Bayes theorem or intuitive updating
|
||||
- Example: Prior 5%, strong evidence (LR=10) → Posterior ~35%
|
||||
|
||||
### Fermi Estimation for Sanity Checks
|
||||
|
||||
**Decompose claim into estimable parts**:
|
||||
- Claim: "Company has 10,000 paying customers"
|
||||
- Decompose: Employees × customers per employee, or revenue ÷ price per customer
|
||||
- Cross-check: Do the numbers add up?
|
||||
|
||||
**Example**:
|
||||
- Claim: Startup has 1M users
|
||||
- Check: Founded 2 years ago → 1,370 new users/day → 57/hour (24/7) or 171/hour (8hr workday)
|
||||
- Reality check: Plausible for viral product? Need marketing spend estimate.
|
||||
|
||||
### Confidence Intervals and Ranges
|
||||
|
||||
**Avoid point estimates** ("70% confident"):
|
||||
- Use ranges: "60-80% confident" acknowledges uncertainty
|
||||
- Ask: What would make me 90% confident? What's missing?
|
||||
|
||||
**Sensitivity analysis**:
|
||||
- Best case scenario (all assumptions optimistic) → upper bound confidence
|
||||
- Worst case scenario (all assumptions pessimistic) → lower bound confidence
|
||||
- Most likely scenario → central estimate
|
||||
|
||||
## 6. Advanced Investigation Patterns
|
||||
|
||||
### Investigative Journalism Techniques
|
||||
|
||||
**Paper trail following**:
|
||||
- Follow money: Who benefits financially from this claim being believed?
|
||||
- Follow timeline: Who said what when? Any story changes over time?
|
||||
- Follow power: Who has authority/incentive to suppress contradicting evidence?
|
||||
|
||||
**Source cultivation**:
|
||||
- Insider sources (whistleblowers, former employees) for claims companies hide
|
||||
- Expert sources (academics, consultants) for technical evaluation
|
||||
- Documentary sources (contracts, emails, internal memos) for ground truth
|
||||
|
||||
**Red flags in interviews**:
|
||||
- Vague answers to specific questions
|
||||
- Defensiveness or hostility when questioned
|
||||
- Inconsistencies between different tellings
|
||||
- Refusal to provide documentation
|
||||
|
||||
### Legal Evidence Standards
|
||||
|
||||
**Burden of proof levels**:
|
||||
- **Beyond reasonable doubt** (criminal): 95%+ confidence
|
||||
- **Clear and convincing** (civil high stakes): 75%+ confidence
|
||||
- **Preponderance of evidence** (civil standard): 51%+ confidence (more likely than not)
|
||||
|
||||
**Hearsay rules**:
|
||||
- Firsthand testimony > secondhand ("I saw X" > "Someone told me X")
|
||||
- Exception: Business records, public records (trustworthy hearsay)
|
||||
- Watch for: Anonymous sources, "people are saying", "experts claim"
|
||||
|
||||
**Chain of custody**:
|
||||
- Document handling: Who collected, stored, analyzed evidence?
|
||||
- Tampering risk: Could evidence have been altered?
|
||||
- Authentication: How do we know this document/photo is genuine?
|
||||
|
||||
### Competitive Intelligence Validation
|
||||
|
||||
**HUMINT (Human Intelligence)**:
|
||||
- Customer interviews: "Do you use competitor's product? How does it work?"
|
||||
- Former employees: Glassdoor reviews, LinkedIn networking
|
||||
- Conference presentations: Technical details revealed publicly
|
||||
|
||||
**OSINT (Open Source Intelligence)**:
|
||||
- Public filings: SEC 10-K, patents, trademarks
|
||||
- Job postings: What skills are they hiring for? (reveals technology stack, strategic priorities)
|
||||
- Social media: Employee posts, company announcements
|
||||
- Web archives: Wayback Machine to see claim history, website changes
|
||||
|
||||
**TECHINT (Technical Intelligence)**:
|
||||
- Reverse engineering: Analyze product directly
|
||||
- Benchmarking: Test performance claims yourself
|
||||
- Network analysis: DNS records, API endpoints, infrastructure footprint
|
||||
|
||||
### Scientific Reproducibility Assessment
|
||||
|
||||
**Replication indicator**:
|
||||
- Has anyone reproduced the finding? (Strong evidence)
|
||||
- Did replication attempts fail? (Evidence against)
|
||||
- Has no one tried to replicate? (Unknown, be cautious)
|
||||
|
||||
**Pre-registration check**:
|
||||
- Was study pre-registered (ClinicalTrials.gov, OSF)? Reduces p-hacking risk
|
||||
- Do results match pre-registered outcomes? If different, why?
|
||||
|
||||
**Data/code availability**:
|
||||
- Can you access raw data to re-analyze?
|
||||
- Is code available to reproduce analysis?
|
||||
- Are materials specified to replicate experiment?
|
||||
|
||||
**Robustness checks**:
|
||||
- Do findings hold with different analysis methods?
|
||||
- Are results sensitive to outliers or specific assumptions?
|
||||
- Do subsample analyses show consistent effects?
|
||||
|
||||
---
|
||||
|
||||
## Workflow Integration
|
||||
|
||||
**When to use advanced techniques**:
|
||||
|
||||
**Triangulation** → Every claim (minimum requirement)
|
||||
**CRAAP Test** → When assessing unfamiliar sources
|
||||
**GRADE System** → Medical/health claims, policy decisions
|
||||
**Bayesian Updating** → When you have prior knowledge/base rates
|
||||
**Fermi Estimation** → Quantitative claims that seem implausible
|
||||
**Investigative Techniques** → High-stakes business decisions, fraud detection
|
||||
**Legal Standards** → Determining action thresholds (e.g., firing employee, lawsuit)
|
||||
**Reproducibility Assessment** → Scientific/technical claims
|
||||
|
||||
**Start simple, add complexity as needed**:
|
||||
1. Quick verification: CRAAP test + Google fact-check
|
||||
2. Moderate investigation: Triangulate 3 sources + basic bias check
|
||||
3. Deep investigation: Full methodology above + expert consultation
|
||||
Reference in New Issue
Block a user