Files
gh-cskiro-claudex-claude-co…/skills/claude-md-auditor/reference/research-insights.md
2025-11-29 18:16:51 +08:00

2.9 KiB

Research-Based Optimization

Academic findings on LLM context utilization and attention patterns.

Lost in the Middle Phenomenon

Source: Liu et al., 2023 - "Lost in the Middle: How Language Models Use Long Contexts"

Key Finding

LLMs perform best when relevant information is positioned at the beginning or end of context, not the middle.

Performance Curve

Performance
    ^
100%|*                              *
    |  *                          *
 75%|    *                      *
    |      *                  *
 50%|        *    ****    *
    |          **      **
    +----------------------------> Position
    Start    Middle         End

Implications for CLAUDE.md

  • Critical information: First 20% of file
  • Reference material: Last 20% of file
  • Supporting details: Middle sections

Optimal Structure

Based on research findings:

# Project Name

## CRITICAL (Top 20%)
- Build commands
- Breaking patterns to avoid
- Security requirements

## IMPORTANT (Next 30%)
- Core architecture
- Main conventions
- Testing requirements

## SUPPORTING (Middle 30%)
- Detailed patterns
- Edge cases
- Historical context

## REFERENCE (Bottom 20%)
- Links and resources
- Version history
- Maintenance notes

Token Efficiency Research

Context Window Utilization

  • Diminishing returns after ~4K tokens of instructions
  • Optimal range: 1,500-3,000 tokens
  • Beyond 5K: Consider splitting into imports

Information Density

  • Prefer lists over paragraphs (better attention)
  • Use code blocks (higher signal-to-noise)
  • Avoid redundancy (wastes attention budget)

Attention Calibration (MIT/Google Cloud AI, 2024)

Finding

Recent models (Claude 3.5+) show improved but not eliminated middle-position degradation.

Recommendations

  1. Chunking: Group related information together
  2. Explicit markers: Use headers and formatting
  3. Repetition: Critical items can appear twice (top and bottom)

Claude-Specific Performance Data

Context Awareness in Claude 4/4.5

  • Better at tracking multiple requirements
  • Still benefits from positional optimization
  • Explicit priority markers help attention allocation

Effective Markers

**CRITICAL**: Must follow exactly
**IMPORTANT**: Strongly recommended
**NOTE**: Additional context

Practical Applications

Audit Criteria Based on Research

Check positioning of:

  • Security requirements (should be top)
  • Build commands (should be top)
  • Error-prone patterns (should be top or bottom)
  • Reference links (should be bottom)

Flag as issues:

  • Critical info buried in middle
  • Long unstructured paragraphs
  • Missing headers/structure
  • No priority markers

References

  • Liu et al. (2023). "Lost in the Middle: How Language Models Use Long Contexts"
  • MIT/Google Cloud AI (2024). "Attention Calibration in Large Language Models"
  • Anthropic (2024). "Claude's Context Window Behavior"