Files
gh-dhofheinz-open-plugins-p…/commands/architect/assess.md
2025-11-29 18:20:21 +08:00

1060 lines
33 KiB
Markdown
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
# Architecture Health Assessment Operation
You are executing the **assess** operation using the 10x-fullstack-engineer agent to perform comprehensive architecture health assessment with scoring and trend analysis.
## Parameters
**Received**: `$ARGUMENTS` (after removing 'assess' operation name)
Expected format: `[scope:"system|service|component"] [focus:"dimension"] [baseline:"reference"]`
Parse the arguments to extract:
- **scope** (optional): Assessment scope - "system" (entire architecture), "service" (specific service), "component" (specific component) - defaults to "system"
- **focus** (optional): Specific dimension to assess - "tech-debt", "security", "performance", "scalability", "maintainability", "cost", or "all" (default: "all")
- **baseline** (optional): Baseline for comparison - ADR number, date (YYYY-MM-DD), or "previous" for last assessment
## Workflow
### Phase 1: Baseline Discovery
Identify baseline for comparison if specified:
1. **Parse Baseline Reference**:
- If `baseline:"ADR-XXXX"`: Read that ADR and extract metrics
- If `baseline:"YYYY-MM-DD"`: Find assessment from that date
- If `baseline:"previous"`: Find most recent assessment file
- If not specified: This is the initial baseline assessment
2. **Locate Previous Assessments**:
- Search for assessment files in `docs/assessments/`
- Naming convention: `architecture-assessment-YYYY-MM-DD.md`
- Read most recent assessment if baseline not specified
3. **Extract Baseline Metrics**:
- Previous scores for each dimension
- Identified issues and their resolution status
- Recommendations and implementation status
- Trends from previous assessments
Use available tools:
- `Glob` to find assessment files
- `Read` to examine previous assessments
- `Bash` to list and sort assessment files by date
### Phase 2: Dimensional Assessment
Assess architecture across six key dimensions:
#### Dimension 1: Technical Debt
**Assessment Areas**:
- Code quality and complexity
- Outdated dependencies and libraries
- Deprecated patterns and practices
- TODO comments and temporary workarounds
- Duplicated code and logic
- Missing tests and documentation
- Legacy code without clear ownership
**Metrics to Collect**:
- Code complexity (cyclomatic complexity average)
- Code duplication percentage
- Outdated dependency count and severity
- TODO/FIXME/HACK comment count
- Test coverage percentage
- Documentation completeness score
- Time to onboard new developers (survey data)
**Scoring Criteria** (0-10):
- **10**: No technical debt, excellent code quality, comprehensive tests and docs
- **8-9**: Minimal debt, well-maintained, minor improvements needed
- **6-7**: Moderate debt, manageable but growing, action needed soon
- **4-5**: Significant debt, impacting velocity, requires dedicated effort
- **2-3**: Severe debt, major maintainability issues, urgent action needed
- **0-1**: Critical debt, system nearly unmaintainable, major refactoring required
**Issues to Identify**:
- High-complexity functions (cyclomatic complexity > 10)
- Dependencies with known vulnerabilities
- Code duplication > 5%
- Test coverage < 70%
- Missing documentation for public APIs
- Components > 500 lines
- Files with > 10 TODO comments
#### Dimension 2: Security
**Assessment Areas**:
- Authentication and authorization mechanisms
- Data encryption (at rest and in transit)
- Input validation and sanitization
- Dependency vulnerabilities
- Security headers and configurations
- Secrets management
- Access control and permissions
- Audit logging and monitoring
- Compliance with security standards (OWASP Top 10)
**Metrics to Collect**:
- Critical/High/Medium/Low vulnerability count
- Outdated security-related dependencies
- Missing security headers count
- Hardcoded secrets found
- Endpoints without authentication
- Failed security scan count
- Time since last security audit
- Compliance gaps (GDPR, HIPAA, SOC2 as applicable)
**Scoring Criteria** (0-10):
- **10**: Zero vulnerabilities, security best practices throughout, regular audits
- **8-9**: Minor issues only, strong security posture, proactive monitoring
- **6-7**: Some gaps, no critical issues, improvements needed
- **4-5**: Notable vulnerabilities, security gaps, action required
- **2-3**: Critical vulnerabilities, major gaps, urgent remediation needed
- **0-1**: Severe security issues, imminent risk, immediate action required
**Issues to Identify**:
- Critical/High severity CVEs in dependencies
- Missing authentication on sensitive endpoints
- Hardcoded credentials or API keys
- SQL injection vulnerabilities
- XSS vulnerabilities
- Missing CSRF protection
- Insufficient input validation
- Weak password policies
- Missing encryption for sensitive data
- Overly permissive access controls
#### Dimension 3: Performance
**Assessment Areas**:
- API response times
- Database query performance
- Frontend load times and Web Vitals
- Resource utilization (CPU, memory, I/O)
- Caching effectiveness
- Network latency and optimization
- Bottleneck identification
**Metrics to Collect**:
- API response time (p50, p95, p99)
- Database query time (average, p95)
- Page load time
- Time to First Byte (TTFB)
- First Contentful Paint (FCP)
- Largest Contentful Paint (LCP)
- Time to Interactive (TTI)
- CPU utilization (average, peak)
- Memory utilization (average, peak)
- Cache hit rate
- Slow query count (> 100ms)
- Bundle size (JS, CSS)
**Scoring Criteria** (0-10):
- **10**: Exceptional performance, p95 < 100ms, Lighthouse score > 95
- **8-9**: Excellent performance, p95 < 200ms, minor optimization opportunities
- **6-7**: Good performance, p95 < 500ms, some bottlenecks identified
- **4-5**: Acceptable performance, p95 < 1s, notable improvements needed
- **2-3**: Poor performance, p95 > 1s, significant bottlenecks
- **0-1**: Unacceptable performance, frequent timeouts, critical issues
**Issues to Identify**:
- API endpoints with p95 > 500ms
- Database queries > 100ms
- N+1 query patterns
- Missing database indexes
- Large bundle sizes (> 500KB)
- Unoptimized images
- Lack of caching
- Synchronous blocking operations
- Memory leaks
- CPU-intensive operations on main thread
#### Dimension 4: Scalability
**Assessment Areas**:
- Horizontal scaling capability
- Stateless design principles
- Database scaling strategy
- Caching architecture
- Load balancing and distribution
- Auto-scaling configuration
- Resource limits and bottlenecks
- Geographic distribution capability
**Metrics to Collect**:
- Current concurrent user capacity
- Maximum requests per second
- Database connection pool utilization
- Cache memory utilization
- Auto-scaling trigger points and effectiveness
- Time to scale up/down
- Cost per user/transaction
- Geographic latency measurements
**Scoring Criteria** (0-10):
- **10**: Proven at scale, linear scaling, multi-region, excellent architecture
- **8-9**: Scales well, some limits identified, minimal refactoring needed
- **6-7**: Moderate scalability, known bottlenecks, improvements planned
- **4-5**: Limited scalability, approaching capacity, refactoring required
- **2-3**: Poor scalability, frequent capacity issues, major work needed
- **0-1**: Cannot scale, constant capacity problems, architectural redesign needed
**Issues to Identify**:
- Stateful services blocking horizontal scaling
- Database as single point of failure
- No read replica configuration
- Missing connection pooling
- No caching layer
- Hard-coded resource limits
- No auto-scaling configuration
- Single-threaded bottlenecks
- Shared state preventing distribution
- No sharding strategy for large datasets
#### Dimension 5: Maintainability
**Assessment Areas**:
- Code organization and structure
- Test coverage and quality
- Documentation completeness
- Development workflow efficiency
- Deployment frequency and success rate
- Debugging and troubleshooting ease
- Knowledge distribution across team
- Onboarding time for new developers
**Metrics to Collect**:
- Test coverage percentage
- Test execution time
- Cyclomatic complexity (average, max)
- Code duplication percentage
- Documentation pages/sections
- Time to deploy
- Deployment success rate
- Mean time to recovery (MTTR)
- Time to onboard new developer
- Number of known issues/bugs
- Average time to resolve bugs
**Scoring Criteria** (0-10):
- **10**: Excellent maintainability, comprehensive tests/docs, fast iterations
- **8-9**: Highly maintainable, good practices, minor improvements possible
- **6-7**: Maintainable, some technical debt, manageable complexity
- **4-5**: Moderate maintainability, growing complexity, refactoring needed
- **2-3**: Poor maintainability, high complexity, difficult to change
- **0-1**: Unmaintainable, cannot safely make changes, requires rewrite
**Issues to Identify**:
- Test coverage < 70%
- Functions with cyclomatic complexity > 10
- Code duplication > 5%
- Missing API documentation
- No architecture diagrams
- Inconsistent coding standards
- Long deployment times (> 30 minutes)
- Deployment failure rate > 5%
- Long MTTR (> 4 hours)
- Tribal knowledge (single person knows critical systems)
#### Dimension 6: Cost Efficiency
**Assessment Areas**:
- Infrastructure cost optimization
- Resource utilization efficiency
- Over-provisioning identification
- Cost per user/transaction
- Serverless vs server cost analysis
- Database cost optimization
- Storage cost efficiency
- Monitoring and tooling costs
**Metrics to Collect**:
- Total monthly infrastructure cost
- Cost per user
- Cost per transaction
- Resource utilization rates (CPU, memory, storage)
- Idle resource costs
- Data transfer costs
- Third-party service costs
- Cost growth rate vs user growth rate
**Scoring Criteria** (0-10):
- **10**: Highly optimized, minimal waste, excellent cost/value ratio
- **8-9**: Well optimized, minor savings possible, good efficiency
- **6-7**: Reasonable costs, optimization opportunities identified
- **4-5**: Higher than optimal, notable waste, improvements needed
- **2-3**: Excessive costs, significant waste, urgent optimization required
- **0-1**: Unsustainable costs, severe waste, immediate action critical
**Issues to Identify**:
- Resources with < 30% utilization
- Over-provisioned databases
- Expensive queries/operations
- Inefficient data storage
- Unnecessary data retention
- Lack of resource right-sizing
- Missing reserved instance opportunities
- High data transfer costs
- Expensive third-party services
- Lack of cost monitoring/alerting
### Phase 3: Comparative Analysis
If baseline is available, compare current vs baseline:
1. **Score Comparison**:
- Calculate score change for each dimension
- Identify improvements (score increased)
- Identify regressions (score decreased)
- Calculate overall trend
2. **Issue Tracking**:
- Match current issues to baseline issues
- Identify resolved issues
- Identify new issues
- Track issue aging (how long unresolved)
3. **Recommendation Progress**:
- Review baseline recommendations
- Assess implementation status
- Measure impact of implemented recommendations
- Identify unaddressed recommendations
4. **Trend Analysis**:
- Multi-assessment trend if multiple baselines exist
- Velocity of improvement/degradation
- Projected future state
- Risk trajectory
**Trend Indicators**:
- ↑↑ Rapid improvement (> 2 points increase)
- ↑ Steady improvement (0.5-2 points increase)
- → Stable (< 0.5 points change)
- ↓ Degradation (-0.5 to -2 points decrease)
- ↓↓ Rapid degradation (> 2 points decrease)
### Phase 4: Recommendations and Roadmap
Generate prioritized recommendations:
1. **Quick Wins** (High Impact, Low Effort):
- Issues fixable in < 1 week
- Significant improvement to scores
- Low risk changes
2. **Critical Fixes** (High Impact, Any Effort):
- Security vulnerabilities
- Performance bottlenecks affecting users
- Scalability blockers
- High-severity issues
3. **Strategic Improvements** (High Impact, High Effort):
- Architectural refactoring
- Major technology upgrades
- Comprehensive test suite development
- Large-scale optimization
4. **Technical Debt Paydown** (Medium Impact, Variable Effort):
- Code quality improvements
- Documentation updates
- Dependency updates
- Complexity reduction
5. **Future-Proofing** (Future Impact, Planning Required):
- Capacity planning
- Architecture evolution
- Technology modernization
- Team skill development
**Roadmap Timeline**:
- **Immediate (This Sprint)**: Critical fixes and quick wins
- **Short-Term (1-3 Months)**: Important improvements and security fixes
- **Medium-Term (3-6 Months)**: Strategic improvements and debt paydown
- **Long-Term (6-12 Months)**: Major refactoring and future-proofing
## Output Format
Provide a comprehensive architecture health assessment report:
```markdown
# Architecture Health Assessment
**Assessment Date**: [YYYY-MM-DD]
**Scope**: [System / Service / Component Name]
**Focus**: [All Dimensions / Specific Dimension]
**Baseline**: [Baseline Reference or "Initial Assessment"]
**Assessor**: 10x-fullstack-engineer agent
## Executive Summary
[2-3 paragraph summary of overall architecture health, key findings, trends, and critical recommendations]
**Overall Health Score**: [X.X]/10 ([Trend])
**Key Findings**:
- [Most significant finding 1]
- [Most significant finding 2]
- [Most significant finding 3]
**Critical Actions Required**:
1. [Top priority action with timeline]
2. [Second priority action with timeline]
3. [Third priority action with timeline]
**Health Trend**: [Improving / Stable / Degrading] ([Explanation])
## Architecture Health Scorecard
### Summary Scores
| Dimension | Score | Change | Trend | Status |
|-----------|-------|--------|-------|--------|
| Technical Debt | [X.X]/10 | [±X.X] | [↑↓→] | [Critical/Poor/Fair/Good/Excellent] |
| Security | [X.X]/10 | [±X.X] | [↑↓→] | [Critical/Poor/Fair/Good/Excellent] |
| Performance | [X.X]/10 | [±X.X] | [↑↓→] | [Critical/Poor/Fair/Good/Excellent] |
| Scalability | [X.X]/10 | [±X.X] | [↑↓→] | [Critical/Poor/Fair/Good/Excellent] |
| Maintainability | [X.X]/10 | [±X.X] | [↑↓→] | [Critical/Poor/Fair/Good/Excellent] |
| Cost Efficiency | [X.X]/10 | [±X.X] | [↑↓→] | [Critical/Poor/Fair/Good/Excellent] |
| **Overall** | **[X.X]/10** | **[±X.X]** | **[↑↓→]** | **[Status]** |
**Status Legend**:
- Excellent (9-10): Best practices, minimal improvements needed
- Good (7-8): Solid foundation, minor enhancements possible
- Fair (5-6): Acceptable but improvements needed
- Poor (3-4): Significant issues, action required
- Critical (0-2): Severe problems, urgent action needed
**Change** is compared to baseline: [Baseline Reference]
### Score Visualization
```
Technical Debt [████████░░] 8.0/10 ↑ (+0.5)
Security [██████░░░░] 6.0/10 → (0.0)
Performance [███████░░░] 7.0/10 ↑ (+1.0)
Scalability [█████░░░░░] 5.0/10 ↓ (-0.5)
Maintainability [████████░░] 8.0/10 ↑ (+1.5)
Cost Efficiency [██████░░░░] 6.0/10 → (+0.2)
─────────────────────────────
Overall [██████░░░░] 6.7/10 ↑ (+0.5)
```
## Dimension 1: Technical Debt ([X.X]/10)
### Summary
[Brief assessment of technical debt state]
**Trend**: [Trend symbol and explanation]
### Key Metrics
| Metric | Current | Baseline | Target | Status |
|--------|---------|----------|--------|--------|
| Code Complexity (avg) | [X.X] | [X.X] | < 5 | [✅/⚠️/❌] |
| Code Duplication | [X]% | [X]% | < 3% | [✅/⚠️/❌] |
| Test Coverage | [X]% | [X]% | > 80% | [✅/⚠️/❌] |
| Outdated Dependencies | [X] | [X] | 0 | [✅/⚠️/❌] |
| TODO Comments | [X] | [X] | < 20 | [✅/⚠️/❌] |
### Issues Identified
**Critical Issues** (affecting score significantly):
1. **[Issue Name]**
- **Location**: [Component/file]
- **Impact**: [Description of impact]
- **Effort**: [Estimate]
- **Priority**: [High/Medium/Low]
**Notable Issues**:
- [Issue description with severity]
- [Issue description with severity]
### Recommendations
1. [Top recommendation with expected improvement]
2. [Second recommendation]
3. [Third recommendation]
## Dimension 2: Security ([X.X]/10)
### Summary
[Brief security assessment]
**Trend**: [Trend symbol and explanation]
### Key Metrics
| Metric | Current | Baseline | Target | Status |
|--------|---------|----------|--------|--------|
| Critical Vulnerabilities | [X] | [X] | 0 | [✅/⚠️/❌] |
| High Vulnerabilities | [X] | [X] | 0 | [✅/⚠️/❌] |
| Medium Vulnerabilities | [X] | [X] | < 5 | [✅/⚠️/❌] |
| Hardcoded Secrets | [X] | [X] | 0 | [✅/⚠️/❌] |
| Unprotected Endpoints | [X] | [X] | 0 | [✅/⚠️/❌] |
| Days Since Security Audit | [X] | [X] | < 90 | [✅/⚠️/❌] |
### Security Posture
**OWASP Top 10 Compliance**:
- A01: Broken Access Control: [✅/⚠️/❌] [Notes]
- A02: Cryptographic Failures: [✅/⚠️/❌] [Notes]
- A03: Injection: [✅/⚠️/❌] [Notes]
- A04: Insecure Design: [✅/⚠️/❌] [Notes]
- A05: Security Misconfiguration: [✅/⚠️/❌] [Notes]
- A06: Vulnerable Components: [✅/⚠️/❌] [Notes]
- A07: Authentication Failures: [✅/⚠️/❌] [Notes]
- A08: Data Integrity Failures: [✅/⚠️/❌] [Notes]
- A09: Logging Failures: [✅/⚠️/❌] [Notes]
- A10: SSRF: [✅/⚠️/❌] [Notes]
### Critical Security Issues
1. **[Vulnerability Name]**
- **Severity**: Critical/High/Medium
- **Location**: [Where found]
- **CVE**: [If applicable]
- **Exploit Risk**: [Assessment]
- **Remediation**: [How to fix]
- **Effort**: [Estimate]
### Recommendations
1. [Critical security recommendation]
2. [Important security recommendation]
3. [Security hardening recommendation]
## Dimension 3: Performance ([X.X]/10)
### Summary
[Brief performance assessment]
**Trend**: [Trend symbol and explanation]
### Key Metrics
| Metric | Current | Baseline | Target | Status |
|--------|---------|----------|--------|--------|
| API Response (p50) | [X]ms | [X]ms | < 100ms | [✅/⚠️/❌] |
| API Response (p95) | [X]ms | [X]ms | < 200ms | [✅/⚠️/❌] |
| API Response (p99) | [X]ms | [X]ms | < 500ms | [✅/⚠️/❌] |
| DB Query Time (avg) | [X]ms | [X]ms | < 50ms | [✅/⚠️/❌] |
| Page Load Time | [X]s | [X]s | < 2s | [✅/⚠️/❌] |
| LCP | [X]s | [X]s | < 2.5s | [✅/⚠️/❌] |
| FCP | [X]s | [X]s | < 1.5s | [✅/⚠️/❌] |
| Bundle Size | [X]KB | [X]KB | < 300KB | [✅/⚠️/❌] |
### Performance Bottlenecks
1. **[Bottleneck Description]**
- **Impact**: [User experience / throughput impact]
- **Current Performance**: [Measurement]
- **Target Performance**: [Goal]
- **Root Cause**: [Analysis]
- **Solution**: [Optimization approach]
- **Expected Improvement**: [Estimate]
- **Effort**: [Estimate]
### Slow Operations
Top 10 slowest operations:
1. [Operation]: [Time] - [Frequency] - [Impact]
2. [Operation]: [Time] - [Frequency] - [Impact]
[...]
### Recommendations
1. [Performance optimization with highest impact]
2. [Second optimization]
3. [Third optimization]
## Dimension 4: Scalability ([X.X]/10)
### Summary
[Brief scalability assessment]
**Trend**: [Trend symbol and explanation]
### Key Metrics
| Metric | Current | Baseline | Target | Status |
|--------|---------|----------|--------|--------|
| Concurrent Users | [X] | [X] | [X] | [✅/⚠️/❌] |
| Requests/Second | [X] | [X] | [X] | [✅/⚠️/❌] |
| DB Connections Used | [X]% | [X]% | < 70% | [✅/⚠️/❌] |
| Cache Hit Rate | [X]% | [X]% | > 80% | [✅/⚠️/❌] |
| Auto-scaling Effectiveness | [X]% | [X]% | > 90% | [✅/⚠️/❌] |
| Cost per User | $[X] | $[X] | < $[X] | [✅/⚠️/❌] |
### Scalability Limits
**Current Capacity**:
- Maximum concurrent users: [X] (utilization: [X]%)
- Maximum requests/second: [X] (utilization: [X]%)
- Database capacity: [X]% utilized
**Scaling Bottlenecks**:
1. **[Bottleneck Name]**
- **Current Limit**: [What breaks and when]
- **Impact**: [Failure mode]
- **Solution**: [How to scale past this]
- **Effort**: [Estimate]
### Scalability Readiness
- ✅ Stateless application design
- ✅ Horizontal auto-scaling configured
- ❌ Database read replicas not configured
- ❌ No caching layer
- ⚠️ Limited connection pooling
- ✅ CDN for static assets
### Recommendations
1. [Top scalability improvement]
2. [Second scalability improvement]
3. [Third scalability improvement]
## Dimension 5: Maintainability ([X.X]/10)
### Summary
[Brief maintainability assessment]
**Trend**: [Trend symbol and explanation]
### Key Metrics
| Metric | Current | Baseline | Target | Status |
|--------|---------|----------|--------|--------|
| Test Coverage | [X]% | [X]% | > 80% | [✅/⚠️/❌] |
| Cyclomatic Complexity (avg) | [X.X] | [X.X] | < 5 | [✅/⚠️/❌] |
| Code Duplication | [X]% | [X]% | < 3% | [✅/⚠️/❌] |
| Deployment Success Rate | [X]% | [X]% | > 95% | [✅/⚠️/❌] |
| MTTR | [X]h | [X]h | < 2h | [✅/⚠️/❌] |
| Time to Deploy | [X]min | [X]min | < 15min | [✅/⚠️/❌] |
| Onboarding Time | [X]days | [X]days | < 7days | [✅/⚠️/❌] |
### Code Quality Issues
**High Complexity Components**:
1. [Component]: Complexity [X] (target: < 10)
2. [Component]: Complexity [X]
3. [Component]: Complexity [X]
**Code Duplication Hotspots**:
- [Location]: [X]% duplication
- [Location]: [X]% duplication
**Testing Gaps**:
- [Component]: [X]% coverage (below target)
- [Component]: No integration tests
- [Component]: No E2E tests
### Recommendations
1. [Maintainability improvement with highest impact]
2. [Second improvement]
3. [Third improvement]
## Dimension 6: Cost Efficiency ([X.X]/10)
### Summary
[Brief cost efficiency assessment]
**Trend**: [Trend symbol and explanation]
### Key Metrics
| Metric | Current | Baseline | Target | Status |
|--------|---------|----------|--------|--------|
| Monthly Infrastructure Cost | $[X] | $[X] | $[X] | [✅/⚠️/❌] |
| Cost per User | $[X] | $[X] | < $[X] | [✅/⚠️/❌] |
| Cost per Transaction | $[X] | $[X] | < $[X] | [✅/⚠️/❌] |
| CPU Utilization | [X]% | [X]% | 60-80% | [✅/⚠️/❌] |
| Memory Utilization | [X]% | [X]% | 60-80% | [✅/⚠️/❌] |
| Storage Utilization | [X]% | [X]% | < 80% | [✅/⚠️/❌] |
| Cost Growth Rate | [X]% | [X]% | < User Growth | [✅/⚠️/❌] |
### Cost Breakdown
| Category | Monthly Cost | % of Total | Trend |
|----------|--------------|------------|-------|
| Compute | $[X] | [X]% | [↑↓→] |
| Database | $[X] | [X]% | [↑↓→] |
| Storage | $[X] | [X]% | [↑↓→] |
| Network/CDN | $[X] | [X]% | [↑↓→] |
| Third-party Services | $[X] | [X]% | [↑↓→] |
| Monitoring/Tools | $[X] | [X]% | [↑↓→] |
| **Total** | **$[X]** | **100%** | **[↑↓→]** |
### Cost Optimization Opportunities
1. **[Optimization Opportunity]**
- **Current Cost**: $[X]/month
- **Potential Savings**: $[X]/month ([X]%)
- **Approach**: [How to optimize]
- **Risk**: [Low/Medium/High]
- **Effort**: [Estimate]
### Waste Identified
- **Idle Resources**: $[X]/month
- **Over-provisioned Resources**: $[X]/month
- **Unnecessary Services**: $[X]/month
- **Inefficient Operations**: $[X]/month
- **Total Potential Savings**: $[X]/month ([X]% of total)
### Recommendations
1. [Cost optimization with highest ROI]
2. [Second optimization]
3. [Third optimization]
## Trend Analysis
[If multiple assessments exist, show historical trend]
### Score History
| Date | Overall | Tech Debt | Security | Performance | Scalability | Maintainability | Cost |
|------|---------|-----------|----------|-------------|-------------|-----------------|------|
| [Date] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] |
| [Date] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] |
| [Date] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] | [X.X] |
### Trend Visualization
```
Overall Score Trend
10 ┤
9 ┤
8 ┤ ●───●
7 ┤
6 ┤ ● ●───●
5 ┤
4 ┤ ● ●
3 ┤
└────────────────────────────────────────
Q1 Q2 Q3 Q4 Q1 Q2 Q3 Q4
```
### Velocity of Change
- **Improving**: [List dimensions improving and rate]
- **Stable**: [List stable dimensions]
- **Degrading**: [List degrading dimensions and rate]
### Projected Future State
Based on current trends, in 6 months:
- Overall Score: [X.X]/10 (projected)
- Key Risks: [Risks if trends continue]
- Key Opportunities: [Opportunities if improvements continue]
## Issue Tracking
### Resolved Since Last Assessment
✅ [Issue description] - Resolved on [date]
✅ [Issue description] - Resolved on [date]
✅ [Issue description] - Resolved on [date]
### Persistent Issues
⚠️ [Issue description] - Open for [X] days
⚠️ [Issue description] - Open for [X] days
⚠️ [Issue description] - Open for [X] days
### New Issues Identified
🆕 [Issue description] - [Severity]
🆕 [Issue description] - [Severity]
🆕 [Issue description] - [Severity]
## Recommendation Implementation Status
### From Previous Assessment
| Recommendation | Status | Impact | Notes |
|----------------|--------|--------|-------|
| [Rec 1] | ✅ Completed | [Positive/Negative/Neutral] | [Outcome] |
| [Rec 2] | 🔄 In Progress | [Expected impact] | [Progress notes] |
| [Rec 3] | ❌ Not Started | [Why not started] | [Plan] |
## Prioritized Recommendations
### Immediate Actions (This Sprint)
**Priority**: CRITICAL - Must address immediately
1. **[Action Item]**
- **Dimension**: [Affected dimension]
- **Current Score Impact**: [X.X points]
- **Effort**: [Time estimate]
- **Risk if Not Addressed**: [Description]
- **Expected Improvement**: [Score increase expected]
### Quick Wins (Next 2-4 Weeks)
**Priority**: HIGH - High impact, low effort
1. **[Action Item]**
- **Dimension**: [Affected dimension]
- **Impact**: [Benefit description]
- **Effort**: [Time estimate]
- **Expected Improvement**: [Score increase]
### Important Improvements (1-3 Months)
**Priority**: HIGH - Significant value, moderate effort
1. **[Action Item]**
- **Dimension**: [Affected dimension]
- **Impact**: [Benefit description]
- **Effort**: [Time estimate]
- **Dependencies**: [Prerequisites]
- **Expected Improvement**: [Score increase]
### Strategic Initiatives (3-6 Months)
**Priority**: MEDIUM - Long-term value, high effort
1. **[Action Item]**
- **Dimension**: [Affected dimension]
- **Impact**: [Strategic benefit]
- **Effort**: [Time estimate]
- **ROI**: [Return on investment]
- **Expected Improvement**: [Score increase]
### Ongoing Maintenance
**Priority**: CONTINUOUS - Regular activities
1. [Maintenance activity with frequency]
2. [Maintenance activity with frequency]
3. [Maintenance activity with frequency]
## Implementation Roadmap
### Sprint Planning
**Current Sprint**:
- [ ] [Critical action 1]
- [ ] [Critical action 2]
- [ ] [Quick win 1]
- [ ] [Quick win 2]
**Next Sprint**:
- [ ] [Quick win 3]
- [ ] [Quick win 4]
- [ ] [Important improvement 1]
**Following Sprints** (prioritized backlog):
1. [Important improvement 2]
2. [Important improvement 3]
3. [Strategic initiative 1]
4. [Strategic initiative 2]
### Milestone Timeline
- **Month 1**: [Key deliverables]
- Target overall score: [X.X]/10
- Critical dimensions: [Focus areas]
- **Month 3**: [Key deliverables]
- Target overall score: [X.X]/10
- Expected improvements: [Areas of improvement]
- **Month 6**: [Key deliverables]
- Target overall score: [X.X]/10
- Strategic goals achieved: [List]
### Success Metrics
Track progress with these metrics:
- Overall health score: [Current] → [Target in 6mo]
- [Specific dimension]: [Current] → [Target]
- [Critical metric]: [Current] → [Target]
- [Business metric]: [Current] → [Target]
## Risk Assessment
### Risks If Recommendations Not Implemented
1. **[Risk Description]**
- **Likelihood**: High/Medium/Low
- **Impact**: Critical/High/Medium/Low
- **Timeline**: [When risk materializes]
- **Mitigation**: [If we do nothing, what's the fallback]
### Risks in Implementing Recommendations
1. **[Risk Description]**
- **Likelihood**: High/Medium/Low
- **Impact**: [Potential negative impact]
- **Mitigation Strategy**: [How to manage risk]
## Conclusion
[Summary paragraph on overall architecture health state]
**Overall Assessment**: [Narrative assessment with trend context]
**Critical Success Factors for Improvement**:
1. [What needs to happen for health improvement]
2. [Key factor 2]
3. [Key factor 3]
**Next Assessment**: Recommended in [timeframe] to track progress
**Immediate Next Steps**:
1. [First action to take]
2. [Second action to take]
3. [Third action to take]
## Appendices
### Appendix A: Detailed Metrics
[Raw data and detailed measurements]
### Appendix B: Comparison to Industry Benchmarks
[How this architecture compares to similar systems]
### Appendix C: Methodology
[How assessment was conducted, tools used]
### Appendix D: References
- [Related ADRs]
- [Previous assessments]
- [Industry standards referenced]
- [Tools and frameworks used]
```
## Assessment Storage
Save the assessment document:
1. **Ensure Directory Exists**: Create `docs/assessments/` if needed
2. **Generate File Name**: `architecture-assessment-YYYY-MM-DD.md`
3. **Write File**: Save complete assessment
4. **Update Index**: Update `docs/assessments/README.md` with new assessment entry
## Agent Invocation
This operation MUST invoke the **10x-fullstack-engineer** agent for expert architecture assessment.
**Agent context to provide**:
- Assessment scope and focus
- Baseline comparison if available
- Collected metrics and measurements
- Identified issues across dimensions
- Current architecture state
**Agent responsibilities**:
- Apply 15+ years of architectural assessment experience
- Provide industry benchmark comparisons
- Identify subtle issues and patterns
- Score dimensions accurately and consistently
- Generate actionable, prioritized recommendations
- Assess trends and project future state
- Consider business context in recommendations
**Agent invocation approach**:
Present comprehensive assessment data and explicitly request:
"Using your 15+ years of full-stack architecture experience, assess this system's architecture health across all dimensions. Score each dimension 0-10, identify critical issues, analyze trends if baseline exists, and provide prioritized recommendations for improvement. Consider both technical excellence and business value."
## Error Handling
### Invalid Scope
```
Error: Invalid scope: [scope]
Valid scopes:
- system Entire architecture (default)
- service Specific service or microservice
- component Specific component or module
Example: /architect assess scope:"system"
```
### Invalid Focus
```
Error: Invalid focus: [focus]
Valid focus dimensions:
- all All dimensions (default)
- tech-debt Technical debt assessment only
- security Security assessment only
- performance Performance assessment only
- scalability Scalability assessment only
- maintainability Maintainability assessment only
- cost Cost efficiency assessment only
Example: /architect assess focus:"security"
```
### Baseline Not Found
```
Error: Baseline not found: [baseline]
Could not find assessment for baseline: [baseline]
Available baselines:
- [Date 1]: architecture-assessment-YYYY-MM-DD.md
- [Date 2]: architecture-assessment-YYYY-MM-DD.md
Or omit baseline for initial assessment.
```
### No Metrics Available
```
Warning: Limited metrics available for comprehensive assessment.
To improve assessment quality, consider:
- Setting up application monitoring (APM)
- Enabling performance profiling
- Running security scans
- Collecting usage metrics
- Implementing logging and tracing
Proceeding with code-based assessment only.
```
## Examples
**Example 1 - Initial Comprehensive Assessment**:
```
/architect assess
```
Full system assessment across all dimensions, establishing baseline.
**Example 2 - Focused Security Assessment**:
```
/architect assess focus:"security"
```
Deep dive into security posture only.
**Example 3 - Comparison to Previous Assessment**:
```
/architect assess baseline:"previous"
```
Compare to most recent assessment, show trends and progress.
**Example 4 - Quarterly Review**:
```
/architect assess baseline:"2024-01-15"
```
Compare to Q1 assessment to track quarterly progress.
**Example 5 - Service-Specific Assessment**:
```
/architect assess scope:"service" focus:"performance"
```
Assess specific service's performance characteristics.
**Example 6 - Cost Optimization Focus**:
```
/architect assess focus:"cost" baseline:"previous"
```
Focus on cost efficiency, compare to previous to track savings.
**Example 7 - Technical Debt Review**:
```
/architect assess focus:"tech-debt"
```
Assess technical debt accumulation for planning debt paydown sprint.