Files
2025-11-30 09:07:22 +08:00

18 KiB

Documentation Management Skill

Systematic documentation methodology for Claude Code projects using empirically validated templates, patterns, and automation.


Frontmatter

name: documentation-management
version: 1.0.0
status: validated
domain: Documentation
tags: [documentation, writing, templates, automation, quality]
validated_on: meta-cc
convergence_iterations: 4
total_development_time: 20-22 hours
value_instance: 0.82
value_meta: 0.82
transferability: 93%

Validation Evidence:

  • V_instance = 0.82: Accuracy 0.75, Completeness 0.85, Usability 0.80, Maintainability 0.85
  • V_meta = 0.82: Completeness 0.75, Effectiveness 0.70, Reusability 0.85, Validation 0.80
  • Retrospective Validation: 90% structural match, 93% transferability, -3% adaptation effort across 3 diverse documentation types
  • Dual Convergence: Both layers exceeded 0.80 threshold in Iteration 3

Quick Start

1. Understand Your Documentation Need

Identify which documentation type you need:

  • Tutorial: Step-by-step learning path (use templates/tutorial-structure.md)
  • Concept: Explain technical concept (use templates/concept-explanation.md)
  • Example: Demonstrate methodology (use templates/example-walkthrough.md)
  • Reference: Comprehensive command/API guide (use templates/quick-reference.md)
  • Troubleshooting: Problem-solution guide (use templates/troubleshooting-guide.md)

2. Start with a Template

# Copy the appropriate template
cp .claude/skills/documentation-management/templates/tutorial-structure.md docs/tutorials/my-guide.md

# Follow the template structure and guidelines
# Fill in sections with your content

3. Apply Core Patterns

Use these validated patterns while writing:

  • Progressive Disclosure: Start simple, add complexity gradually
  • Example-Driven: Pair every concept with concrete example
  • Problem-Solution: Structure around user problems, not features

4. Automate Quality Checks

# Validate all links
python .claude/skills/documentation-management/tools/validate-links.py docs/

# Validate command syntax
python .claude/skills/documentation-management/tools/validate-commands.py docs/

5. Evaluate Quality

Use the quality checklist in each template to self-assess:

  • Accuracy: Technical correctness
  • Completeness: All user needs addressed
  • Usability: Clear navigation and examples
  • Maintainability: Modular structure and automation

Core Methodology

Documentation Lifecycle

This methodology follows a 4-phase lifecycle:

Phase 1: Needs Analysis

  • Identify target audience and their questions
  • Determine documentation type needed
  • Gather technical details and examples

Phase 2: Strategy Formation

  • Select appropriate template
  • Plan progressive disclosure structure
  • Identify core patterns to apply

Phase 3: Writing/Execution

  • Follow template structure
  • Apply patterns (progressive disclosure, example-driven, problem-solution)
  • Create concrete examples

Phase 4: Validation

  • Run automation tools (link validation, command validation)
  • Review against template quality checklist
  • Test with target audience if possible

Value Function (Quality Assessment)

V_instance (Documentation Quality) = (Accuracy + Completeness + Usability + Maintainability) / 4

Component Definitions:

  • Accuracy (0.0-1.0): Technical correctness, working links, valid commands
  • Completeness (0.0-1.0): User needs addressed, edge cases covered
  • Usability (0.0-1.0): Navigation, clarity, examples, accessibility
  • Maintainability (0.0-1.0): Modular structure, automation, version tracking

Target: V_instance ≥ 0.80 for production-ready documentation

Example Scoring:

  • 0.90+: Exceptional (comprehensive, validated, highly usable)
  • 0.80-0.89: Excellent (production-ready, all needs met)
  • 0.70-0.79: Good (functional, minor gaps)
  • 0.60-0.69: Fair (usable, notable gaps)
  • <0.60: Poor (significant issues)

Templates

This skill provides 5 empirically validated templates:

1. Tutorial Structure (templates/tutorial-structure.md)

  • Purpose: Step-by-step learning path for complex topics
  • Size: ~300 lines
  • Validation: 100% match with Installation Guide
  • Best For: Onboarding, feature walkthroughs, methodology guides
  • Key Sections: What/Why/Prerequisites/Concepts/Workflow/Examples/Troubleshooting

2. Concept Explanation (templates/concept-explanation.md)

  • Purpose: Explain single technical concept clearly
  • Size: ~200 lines
  • Validation: 100% match with JSONL Reference
  • Best For: Architecture docs, design patterns, technical concepts
  • Key Sections: Definition/Why/When/How/Examples/Edge Cases/Related

3. Example Walkthrough (templates/example-walkthrough.md)

  • Purpose: Demonstrate methodology through concrete example
  • Size: ~250 lines
  • Validation: Validated in Testing and Error Recovery examples
  • Best For: Case studies, success stories, before/after demos
  • Key Sections: Context/Setup/Execution/Results/Lessons/Transferability

4. Quick Reference (templates/quick-reference.md)

  • Purpose: Comprehensive command/API reference
  • Size: ~350 lines
  • Validation: 70% match with CLI Reference (85% transferability)
  • Best For: CLI tools, APIs, configuration options
  • Key Sections: Overview/Common Tasks/Commands/Parameters/Examples/Troubleshooting

5. Troubleshooting Guide (templates/troubleshooting-guide.md)

  • Purpose: Problem-solution structured guide
  • Size: ~550 lines
  • Validation: Validated with 3 BAIME issues
  • Best For: FAQ, debugging guides, error resolution
  • Key Sections: Problem Categories/Symptoms/Diagnostics/Solutions/Prevention

Retrospective Validation Results:

  • 90% structural match across 3 diverse documentation types
  • 93% transferability (templates work with <10% adaptation)
  • -3% adaptation effort (net time savings)
  • 9/10 template fit quality

Patterns

1. Progressive Disclosure

Problem: Users overwhelmed by complex topics presented all at once. Solution: Structure content from simple to complex, general to specific.

Implementation:

  • Start with "What is X?" before "How does X work?"
  • Show simple examples before advanced scenarios
  • Use hierarchical sections (Overview → Details → Edge Cases)
  • Defer advanced topics to separate sections

Validation: 4+ uses across BAIME guide, iteration docs, FAQ, examples

See: patterns/progressive-disclosure.md for comprehensive guide

2. Example-Driven Explanation

Problem: Abstract concepts hard to understand without concrete examples. Solution: Pair every concept with concrete, realistic example.

Implementation:

  • Define concept briefly
  • Immediately show example
  • Explain how example demonstrates concept
  • Show variations (simple → complex)

Validation: 3+ uses across BAIME concepts, templates, examples

See: patterns/example-driven-explanation.md for comprehensive guide

3. Problem-Solution Structure

Problem: Documentation organized around features, not user problems. Solution: Structure around problems users actually face.

Implementation:

  • Identify user pain points
  • Group by problem category (not feature)
  • Format: Symptom → Diagnosis → Solution → Prevention
  • Include real error messages and outputs

Validation: 3+ uses across troubleshooting guides, error recovery

See: patterns/problem-solution-structure.md for comprehensive guide


Automation Tools

Purpose: Detect broken internal/external links, missing files Usage: python tools/validate-links.py docs/ Output: List of broken links with file locations Speedup: 30x faster than manual checking Tested: 13/15 links valid in meta-cc docs

2. Command Validation (tools/validate-commands.py)

Purpose: Validate code blocks for correct syntax, detect typos Usage: python tools/validate-commands.py docs/ Output: Invalid commands with line numbers Speedup: 20x faster than manual testing Tested: 20/20 commands valid in BAIME guide

Both tools are production-ready and integrate with CI/CD for automated quality gates.


Examples

Example 1: BAIME Usage Guide (Tutorial)

Context: Create comprehensive guide for BAIME methodology Template Used: tutorial-structure.md Result: 1100-line tutorial with V_instance = 0.82

Key Decisions:

  • Two domain examples (Testing + Error Recovery) to demonstrate transferability
  • FAQ section for quick answers (11 questions)
  • Troubleshooting section with concrete examples (3 issues)
  • Progressive disclosure: What → Why → How → Examples

Lessons Learned:

  • Multiple examples prove universality (single example insufficient)
  • Comparison table synthesizes insights
  • FAQ should be added early (high ROI)

Example 2: CLI Reference (Quick Reference)

Context: Document meta-cc CLI commands Template Used: quick-reference.md Result: Comprehensive command reference with 70% template match

Adaptations:

  • Added command categories (MCP tools vs CLI)
  • Emphasized output format (JSONL/TSV)
  • Included jq filter examples
  • More example-heavy than template (CLI needs concrete usage)

Lessons Learned:

  • Quick reference template adapts well to CLI tools
  • Examples more critical than structure for CLI docs
  • ~15% adaptation effort for specialized domains

Example 3: Retrospective Validation Study

Context: Test templates on existing meta-cc documentation Approach: Applied templates to 3 diverse docs (CLI, Installation, JSONL)

Results:

  • 90% structural match: Templates matched existing high-quality docs
  • 93% transferability: <10% adaptation needed
  • -3% adaptation effort: Net time savings
  • Independent evolution: 2/3 docs evolved same structure naturally

Insight: Templates extract genuine universal patterns (descriptive, not prescriptive)


Quality Standards

Production-Ready Criteria

Documentation is production-ready when:

  • V_instance ≥ 0.80 (all components)
  • All links valid (automated check)
  • All commands tested (automated check)
  • Template quality checklist complete
  • Examples concrete and realistic
  • Reviewed by domain expert (if available)

Quality Scoring Guide

Accuracy Assessment:

  • All technical details correct?
  • Links valid?
  • Commands work as documented?
  • Examples realistic and tested?

Completeness Assessment:

  • All user questions answered?
  • Edge cases covered?
  • Prerequisites clear?
  • Examples sufficient?

Usability Assessment:

  • Navigation intuitive?
  • Examples concrete?
  • Jargon defined?
  • Progressive disclosure applied?

Maintainability Assessment:

  • Modular structure?
  • Automated validation?
  • Version tracked?
  • Easy to update?

Transferability

Cross-Domain Validation

This methodology has been validated across:

  • Tutorial Documentation: BAIME guide, Installation guide
  • Reference Documentation: CLI reference, JSONL reference
  • Concept Documentation: BAIME concepts (6 concepts)
  • Troubleshooting: BAIME issues, error recovery

Transferability Rate: 93% (empirically measured) Adaptation Effort: -3% (net time savings) Domain Independence: Universal (applies to all documentation types)

Adaptation Guidelines

When adapting templates to your domain:

  1. Keep Core Structure (90% match is ideal)

    • Section hierarchy
    • Progressive disclosure
    • Example-driven approach
  2. Adapt Content Depth (10-15% variation)

    • CLI tools need more examples
    • Concept docs need more diagrams
    • Troubleshooting needs real error messages
  3. Customize Examples (domain-specific)

    • Use your project's terminology
    • Show realistic use cases
    • Include actual outputs
  4. Follow Quality Checklist (from template)

    • Ensures consistency
    • Prevents common mistakes
    • Validates completeness

Usage Guide

For New Documentation

  1. Identify Documentation Type

    • What is the primary user need? (learn, understand, reference, troubleshoot)
    • Select matching template
  2. Copy Template

    cp templates/[template-name].md docs/[your-doc].md
    
  3. Follow Template Structure

    • Read "When to Use" section
    • Follow section guidelines
    • Apply quality checklist
  4. Apply Core Patterns

    • Progressive disclosure (simple → complex)
    • Example-driven (concept + example)
    • Problem-solution (if applicable)
  5. Validate Quality

    python tools/validate-links.py docs/[your-doc].md
    python tools/validate-commands.py docs/[your-doc].md
    
  6. Self-Assess

    • Calculate V_instance score
    • Review template checklist
    • Iterate if needed

For Existing Documentation

  1. Assess Current State

    • Calculate V_instance (current quality)
    • Identify gaps (completeness, usability)
    • Determine target V_instance
  2. Select Improvement Strategy

    • Structural: Apply template structure (if V_instance < 0.60)
    • Incremental: Add missing sections (if V_instance 0.60-0.75)
    • Polish: Apply patterns and validation (if V_instance > 0.75)
  3. Apply Template Incrementally

    • Don't rewrite from scratch
    • Map existing content to template sections
    • Fill gaps systematically
  4. Validate Improvements

    • Run automation tools
    • Recalculate V_instance
    • Verify gap closure

Best Practices

Writing Principles

  1. Empirical Validation Over Assumptions

    • Test examples before documenting
    • Validate links and commands automatically
    • Use real user feedback when available
  2. Multiple Examples Demonstrate Universality

    • Single example shows possibility
    • Two examples show pattern
    • Three examples prove universality
  3. Progressive Disclosure Reduces Cognitive Load

    • Start with "What" and "Why"
    • Move to "How"
    • End with "Advanced"
  4. Problem-Solution Matches User Mental Model

    • Users come with problems, not feature requests
    • Structure guides around solving problems
    • Include symptoms, diagnosis, solution
  5. Automation Enables Scale

    • Manual validation doesn't scale
    • Invest in automation tools early
    • Integrate into CI/CD
  6. Template Creation Is Infrastructure

    • First template takes time (~2 hours)
    • Subsequent uses save 3-4 hours each
    • ROI is multiplicative

Common Mistakes

  1. Deferring Quick Wins

    • FAQ sections take 30-45 minutes but add significant value
    • Add FAQ early (Iteration 1, not later)
  2. Single Example Syndrome

    • One example doesn't prove transferability
    • Add second example to demonstrate pattern
    • Comparison table synthesizes insights
  3. Feature-Centric Structure

    • Users don't care about features, they care about problems
    • Restructure around user problems
    • Use problem-solution pattern
  4. Abstract-Only Explanations

    • Abstract concepts without examples don't stick
    • Always pair concept with concrete example
    • Show variations (simple → complex)
  5. Manual Validation Only

    • Manual link/command checking is error-prone
    • Create automation tools early
    • Run in CI for continuous validation

Integration with BAIME

This methodology was developed using BAIME and can be used to document other BAIME experiments:

When Creating BAIME Documentation

  1. Use Tutorial Structure for methodology guides

    • What is the methodology?
    • When to use it?
    • Step-by-step workflow
    • Example applications
  2. Use Example Walkthrough for domain examples

    • Show concrete BAIME application
    • Include value scores at each iteration
    • Demonstrate transferability
  3. Use Troubleshooting Guide for common issues

    • Structure around actual errors encountered
    • Include diagnostic workflows
    • Show recovery patterns
  4. Apply Progressive Disclosure

    • Start with simple example (rich baseline)
    • Add complex example (minimal baseline)
    • Compare and synthesize

Extraction from BAIME Experiments

After BAIME experiment converges, extract documentation:

  1. Patterns → pattern files in skill
  2. Templates → template files in skill
  3. Methodology → tutorial in docs/methodology/
  4. Examples → examples/ in skill
  5. Tools → tools/ in skill

This skill itself was extracted from a BAIME experiment (Bootstrap-Documentation).


Maintenance

Version: 1.0.0 (validated and converged) Created: 2025-10-19 Last Updated: 2025-10-19 Status: Production-ready

Validated On:

  • BAIME Usage Guide (Tutorial)
  • CLI Reference (Quick Reference)
  • Installation Guide (Tutorial)
  • JSONL Reference (Concept)
  • Error Recovery Example (Example Walkthrough)

Known Limitations:

  • No visual aid generation (diagrams, flowcharts) - manual process
  • No maintenance workflow (focus on creation methodology)
  • Spell checker not included (link and command validation only)

Future Enhancements:

  • Add visual aid templates (architecture diagrams, flowcharts)
  • Create maintenance workflow documentation
  • Develop spell checker with technical term dictionary
  • Add third domain example (CI/CD or Knowledge Transfer)

Changelog:

  • v1.0.0 (2025-10-19): Initial release from BAIME experiment
    • 5 templates (all validated)
    • 3 patterns (all validated)
    • 2 automation tools (both working)
    • Retrospective validation complete (93% transferability)

References

Source Experiment: /home/yale/work/meta-cc/experiments/documentation-methodology/ Convergence: 4 iterations, ~20-22 hours, V_instance=0.82, V_meta=0.82 Methodology: BAIME (Bootstrapped AI Methodology Engineering)

Related Skills:

  • testing-strategy: Systematic testing methodology
  • error-recovery: Error handling patterns
  • knowledge-transfer: Onboarding methodologies

External Resources: