Files
gh-jamesrochabrun-skills/skills/anthropic-architect/references/decision_rubric.md
2025-11-29 18:48:55 +08:00

16 KiB

Architectural Decision Rubric

Comprehensive framework for choosing the right Anthropic architecture for your project.

Overview

This rubric evaluates seven key dimensions to recommend the optimal combination of Skills, Agents, Prompts, and SDK primitives.

The Seven Dimensions

1. Task Complexity

2. Reusability Requirements

3. Context Management

4. Security & Control

5. Performance Needs

6. Maintenance Burden

7. Time to Market


1. Task Complexity Analysis

Low Complexity (Score: 1-3)

Characteristics:

  • Single operation or simple workflow
  • Clear input → output mapping
  • No dependencies or minimal dependencies
  • Linear execution (no branching)
  • 1-5 steps maximum
  • Can be described in one sentence

Recommendation: Direct Prompts

Examples:

  • "Explain this function"
  • "Fix this typo"
  • "Add a comment to this class"
  • "Format this JSON"
  • "Translate this error message"

Implementation:

Simple, clear instruction to Claude without additional structure

Medium Complexity (Score: 4-6)

Characteristics:

  • Multiple related operations
  • Structured workflow with steps
  • Some dependencies between steps
  • Requires reference materials or examples
  • 5-20 steps
  • Benefits from organization and guidance
  • Reusable pattern

Recommendation: Skills

Examples:

  • Generate comprehensive code review
  • Create design system component
  • Write technical documentation with examples
  • Analyze codebase for patterns
  • Generate test suite following guidelines

Implementation:

skill/
├── SKILL.md (main documentation)
├── references/
│   ├── patterns.md
│   ├── examples.md
│   └── guidelines.md
└── scripts/ (optional)
    └── helper.sh

High Complexity (Score: 7-9)

Characteristics:

  • Multi-step autonomous workflow
  • Needs isolated context
  • Parallel execution beneficial
  • Different phases with different requirements
  • 20+ steps or multiple parallel tracks
  • Requires exploration, planning, execution, validation
  • Different tool permissions per phase

Recommendation: Agents/Subagents

Examples:

  • Full codebase refactoring
  • Comprehensive security audit
  • Multi-file feature implementation
  • Documentation generation across entire project
  • Performance optimization analysis and fixes

Implementation:

Main Agent (orchestrator)
├── Explore Subagent (read-only, analysis)
├── Plan Subagent (planning, no execution)
├── Execute Subagent (write permissions)
└── Validate Subagent (testing, verification)

Custom Complexity (Score: 10)

Characteristics:

  • Unique workflow requirements
  • System integration needed
  • Custom tools required
  • Specific feedback loops
  • Production deployment
  • Fine-grained control necessary

Recommendation: SDK Primitives

Examples:

  • Custom CI/CD integration
  • Proprietary system automation
  • Domain-specific code analysis
  • Production AI features
  • Specialized agent behaviors

Implementation:

import { Agent, Tool } from 'claude-agent-sdk';

const customAgent = new Agent({
  tools: [customTool1, customTool2],
  workflow: customFeedbackLoop,
  integrations: [ciSystem, deploySystem]
});

2. Reusability Requirements

Single Use (Score: 1-2)

Characteristics:

  • One-time task
  • Project-specific
  • No future reuse expected
  • Temporal need

Recommendation: Direct Prompts

Examples:

  • Debug this specific bug
  • Update this particular file
  • Answer question about this code

Personal Reuse (Score: 3-4)

Characteristics:

  • You'll use it multiple times
  • Personal workflow optimization
  • Not shared with team

Recommendation: Skills (Personal)

Examples:

  • Your personal code review checklist
  • Your preferred refactoring patterns
  • Your documentation template

Storage: Local .claude/skills/ directory


Team Reuse (Score: 5-7)

Characteristics:

  • Multiple team members benefit
  • Team-wide patterns
  • Shared knowledge
  • Collaboration value

Recommendation: Skills (Team Plugin)

Examples:

  • Team coding standards
  • Project-specific patterns
  • Shared workflows
  • Team documentation templates

Storage: Team repository plugin


Organization Reuse (Score: 8-9)

Characteristics:

  • Cross-team benefit
  • Company-wide standards
  • Multiple projects
  • Organization knowledge

Recommendation: Skills (Organization Marketplace)

Examples:

  • Company coding standards
  • Security review guidelines
  • Architecture patterns
  • Compliance requirements

Distribution: Internal marketplace


Product Feature (Score: 10)

Characteristics:

  • End-user facing
  • Production deployment
  • Product differentiation
  • Revenue impact

Recommendation: SDK Primitives

Examples:

  • AI-powered product feature
  • Customer-facing automation
  • Production workflow
  • SaaS feature

Implementation: Custom SDK integration


3. Context Management Needs

Minimal Context (Score: 1-3)

Characteristics:

  • Self-contained task
  • No external references
  • All info in prompt
  • < 1000 tokens

Recommendation: Direct Prompts

Example:

Explain this function:
[paste function]

Structured Context (Score: 4-6)

Characteristics:

  • Reference materials needed
  • Organized information
  • Progressive disclosure beneficial
  • 1K-10K tokens

Recommendation: Skills with Progressive Disclosure

Example:

skill/
├── SKILL.md
└── references/
    ├── quick_reference.md (loaded first)
    ├── detailed_patterns.md (loaded on demand)
    └── examples.md (loaded when needed)

Pattern:

  • Start with minimal context
  • Load more as needed
  • Query-based retrieval

Isolated Context (Score: 7-9)

Characteristics:

  • Separate concerns
  • Avoid context pollution
  • Parallel execution
  • Different contexts per phase
  • 10K+ tokens per context

Recommendation: Agents/Subagents

Example:

Explore Agent: Codebase context (read-only)
Plan Agent: Planning context (insights from explore)
Code Agent: Implementation context (plan + target files)
Review Agent: Validation context (changes + tests)

Benefits:

  • No context pollution
  • Clear boundaries
  • Parallel execution
  • Optimal token usage

Custom Context (Score: 10)

Characteristics:

  • Specific context handling
  • Integration requirements
  • Custom context sources
  • Dynamic context loading

Recommendation: SDK Primitives

Example:

const context = await customContextLoader({
  source: proprietarySystem,
  filter: taskSpecific,
  transform: domainFormat
});

4. Security & Control Requirements

Basic Safety (Score: 1-3)

Characteristics:

  • Read-only operations
  • No sensitive data
  • Standard guardrails sufficient
  • Low risk

Recommendation: Direct Prompts + Skills

Controls:

  • Standard Claude safety features
  • No additional restrictions needed

Controlled Access (Score: 4-6)

Characteristics:

  • Write operations
  • Specific tool permissions needed
  • Some sensitive operations
  • Medium risk

Recommendation: Agents with Tool Restrictions

Controls:

Explore Agent: [Read, Grep, Glob]  // Read-only
Plan Agent: [TodoWrite]             // Planning only
Code Agent: [Read, Edit, Write]     // Code changes
Review Agent: [Bash, Read]          // Testing

Pattern:

  • Allowlist approach
  • Minimal permissions
  • Explicit grants

High Security (Score: 7-9)

Characteristics:

  • Sensitive operations
  • Compliance requirements
  • Audit logging needed
  • High risk

Recommendation: Agents with Confirmations

Controls:

Agent {
  tools: allowlistOnly,
  confirmations: [
    'git push',
    'deployment',
    'data deletion',
    'sensitive operations'
  ],
  audit: true,
  denylist: dangerousCommands
}

Pattern:

  • Deny-all default
  • Explicit confirmations
  • Audit all operations
  • Block dangerous commands

Maximum Security (Score: 10)

Characteristics:

  • Production systems
  • Financial/medical data
  • Regulatory compliance
  • Critical infrastructure

Recommendation: SDK Primitives + Custom Security

Controls:

const secureAgent = new Agent({
  security: {
    denyAll: true,
    allowlist: minimalTools,
    mfa: true,
    audit: comprehensiveLogger,
    encryption: true,
    rateLimits: strict,
    monitoring: realtime,
    rollback: automatic
  }
});

5. Performance Needs

Standard Performance (Score: 1-3)

Characteristics:

  • User can wait
  • Not time-sensitive
  • Occasional use

Recommendation: Any approach


Fast Response (Score: 4-6)

Characteristics:

  • Quick feedback expected
  • Interactive use
  • Multiple requests

Recommendation: Skills with Progressive Disclosure

Optimization:

  • Load minimal context initially
  • Query additional info on demand
  • Cache frequent queries

High Performance (Score: 7-9)

Characteristics:

  • Real-time or near real-time
  • Parallel execution beneficial
  • Resource optimization critical

Recommendation: Agents (Parallel Execution)

Optimization:

Parallel Subagents:
├── Agent 1: File 1-10
├── Agent 2: File 11-20
├── Agent 3: File 21-30
└── Agent 4: Aggregation

Execution: All run simultaneously

Maximum Performance (Score: 10)

Characteristics:

  • Production SLA requirements
  • Sub-second responses
  • High throughput
  • Resource limits

Recommendation: SDK Primitives + Custom Optimization

Optimization:

const optimizedAgent = new Agent({
  caching: aggressive,
  parallelization: maximum,
  contextCompression: true,
  earlyTermination: true,
  resourceLimits: optimized
});

6. Maintenance Burden

Low Maintenance (Score: 1-3)

Characteristics:

  • Set and forget
  • Stable requirements
  • Minimal updates

Recommendation: Direct Prompts (no maintenance)


Medium Maintenance (Score: 4-6)

Characteristics:

  • Periodic updates
  • Evolving patterns
  • Team contributions

Recommendation: Skills (easy to update)

Maintenance:

  • Update reference docs
  • Add new examples
  • Version control friendly
  • Clear documentation

High Maintenance (Score: 7-9)

Characteristics:

  • Regular updates
  • Multiple contributors
  • Evolving requirements

Recommendation: Skills + Version Control

Maintenance:

skill/
├── CHANGELOG.md
├── VERSION
├── SKILL.md
└── references/
    └── (versioned docs)

Custom Maintenance (Score: 10)

Characteristics:

  • Custom codebase
  • Breaking changes
  • Integration updates
  • Production support

Recommendation: SDK Primitives (with CI/CD)

Maintenance:

// Automated testing
// Version management
// Deployment pipeline
// Monitoring and alerts

7. Time to Market

Immediate (Score: 1-3)

Characteristics:

  • Need it now
  • No setup time
  • Quick win

Recommendation: Direct Prompts

Time: Seconds to minutes


Quick (Score: 4-6)

Characteristics:

  • Hours to days
  • Some setup acceptable
  • Reusability valuable

Recommendation: Skills

Time: 1-4 hours to create


Planned (Score: 7-9)

Characteristics:

  • Days to weeks
  • Proper planning
  • Complex requirements

Recommendation: Agents/Subagents

Time: 1-3 days to design and implement


Strategic (Score: 10)

Characteristics:

  • Weeks to months
  • Product feature
  • Full development cycle

Recommendation: SDK Primitives

Time: 1+ weeks to build and deploy


Decision Matrix

Quick Reference Table

Dimension Prompts Skills Agents SDK
Complexity Low (1-3) Medium (4-6) High (7-9) Custom (10)
Reusability Single (1-2) Team (5-7) Org (8-9) Product (10)
Context Minimal (1-3) Structured (4-6) Isolated (7-9) Custom (10)
Security Basic (1-3) Controlled (4-6) High (7-9) Max (10)
Performance Standard (1-3) Fast (4-6) High (7-9) Max (10)
Maintenance Low (1-3) Medium (4-6) High (7-9) Custom (10)
Time to Market Immediate (1-3) Quick (4-6) Planned (7-9) Strategic (10)

Scoring Your Project

Step 1: Score Each Dimension

Rate your project on each of the 7 dimensions (1-10).

Step 2: Calculate Weighted Average

Different dimensions may have different importance for your use case.

Default Weights:

  • Task Complexity: 25%
  • Reusability: 20%
  • Context Management: 15%
  • Security: 15%
  • Performance: 10%
  • Maintenance: 10%
  • Time to Market: 5%

Step 3: Interpret Score

Average Score 1-3: Direct Prompts

  • Simple, clear instructions
  • No additional structure

Average Score 4-6: Skills

  • Organized expertise
  • Progressive disclosure
  • Reference materials

Average Score 7-9: Agents/Subagents

  • Complex workflows
  • Isolated contexts
  • Parallel execution

Average Score 10: SDK Primitives

  • Custom implementation
  • Full control
  • Production deployment

Special Cases

Hybrid Architectures

When: Scores span multiple ranges

Solution: Combine approaches

  • Direct Prompts for simple tasks
  • Skills for reusable expertise
  • Agents for complex workflows

Example:

Project with scores:
- Complexity: 7 (Agents)
- Reusability: 5 (Skills)
- Context: 4 (Skills)
- Security: 6 (Skills/Agents)

Recommendation: Agents + Skills
- Use Agents for complex workflows
- Load Skills for domain expertise
- Direct Prompts for simple operations

Decision Tree

Start
│
├─ Is it a simple, one-time task?
│  └─ YES → Direct Prompts
│  └─ NO → Continue
│
├─ Do you need reusable expertise?
│  └─ YES → Continue
│  └─ NO → Continue
│     │
│     ├─ Is it complex with multiple phases?
│     │  └─ YES → Agents
│     │  └─ NO → Direct Prompts
│
├─ Is it complex with isolated contexts needed?
│  └─ YES → Agents
│  └─ NO → Skills
│
├─ Is it a production feature or unique workflow?
│  └─ YES → SDK Primitives
│  └─ NO → Agents or Skills
│
└─ Default → Skills (best balance)

Example Evaluations

Example 1: Code Review Automation

Scores:

  • Complexity: 5 (structured workflow)
  • Reusability: 7 (team-wide)
  • Context: 5 (reference materials)
  • Security: 4 (read operations)
  • Performance: 5 (interactive)
  • Maintenance: 6 (evolving standards)
  • Time to Market: 5 (hours to setup)

Average: 5.3

Recommendation: Skills

  • Create code-review skill
  • Include team standards
  • Progressive disclosure of guidelines
  • Reference materials for patterns

Example 2: Codebase Migration

Scores:

  • Complexity: 9 (multi-phase, autonomous)
  • Reusability: 3 (one-time migration)
  • Context: 8 (isolated per phase)
  • Security: 7 (write operations)
  • Performance: 8 (parallel execution)
  • Maintenance: 3 (temporary)
  • Time to Market: 7 (proper planning)

Average: 6.4

Recommendation: Agents

  • Despite low reusability, complexity demands Agents
  • Explore Agent: Analyze codebase
  • Plan Agent: Create migration strategy
  • Migrate Agent: Execute changes
  • Validate Agent: Run tests

Example 3: Quick Bug Fix

Scores:

  • Complexity: 2 (single fix)
  • Reusability: 1 (one-time)
  • Context: 2 (minimal)
  • Security: 3 (single file change)
  • Performance: 2 (can wait)
  • Maintenance: 1 (no maintenance)
  • Time to Market: 1 (immediate)

Average: 1.7

Recommendation: Direct Prompt

  • Simple instruction
  • Fast execution
  • No overhead

Example 4: AI-Powered Product Feature

Scores:

  • Complexity: 10 (custom workflow)
  • Reusability: 10 (product feature)
  • Context: 10 (custom handling)
  • Security: 9 (production)
  • Performance: 9 (SLA requirements)
  • Maintenance: 10 (ongoing support)
  • Time to Market: 10 (strategic)

Average: 9.7

Recommendation: SDK Primitives

  • Custom agent implementation
  • Production-grade security
  • Full monitoring and controls
  • Integration with product

Summary

Use this rubric to:

  1. Score your project on 7 dimensions
  2. Calculate weighted average
  3. Interpret score to get recommendation
  4. Validate with decision tree
  5. Review example evaluations

Key Principle: Start simple, scale complexity as needed.

Remember: The best architecture is the simplest one that meets your requirements.