402 lines
12 KiB
Markdown
402 lines
12 KiB
Markdown
---
|
|
name: skill-factory
|
|
description: Autonomous skill creation agent that analyzes requests, automatically selects the best creation method (documentation scraping via Skill_Seekers, manual TDD construction, or hybrid), ensures quality compliance with Anthropic best practices, and delivers production-ready skills without requiring user decision-making or navigation
|
|
when_to_use: when you need to create any Claude skill and want it done automatically with guaranteed quality - works for documentation-based skills, GitHub repositories, PDFs, custom workflows, or hybrid approaches
|
|
version: 0.1.0
|
|
---
|
|
|
|
# Skill Factory
|
|
|
|
**Autonomous skill creation - just tell me what you need, I'll handle everything.**
|
|
|
|
## What This Does
|
|
|
|
You request a skill, I deliver a production-ready skill with guaranteed quality (score >= 8.0/10).
|
|
|
|
**No decision-making required. No tool selection. No quality checking. Just results.**
|
|
|
|
### Anthropic's Official Best Practices
|
|
|
|
For comprehensive guidance on creating effective skills, see:
|
|
|
|
- **[references/overview.md](references/overview.md)** - Complete overview of Agent Skills architecture, progressive disclosure, and how Skills work across different platforms (API, Claude Code, Agent SDK, claude.ai)
|
|
- **[references/quickstart.md](references/quickstart.md)** - Quick tutorial on using pre-built Agent Skills in the Claude API with practical code examples
|
|
- **[references/best-practices.md](references/best-practices.md)** - Detailed authoring best practices including core principles, skill structure, progressive disclosure patterns, workflows, evaluation strategies, and common patterns
|
|
- **[references/anthropic-best-practices.md](references/anthropic-best-practices.md)** - Quality scoring system (10/10 criteria) used by skill-factory
|
|
|
|
These references provide Anthropic's official guidance and are consulted during the quality assurance phase.
|
|
|
|
## Usage
|
|
|
|
Simply describe the skill you need:
|
|
|
|
```
|
|
"Create a skill for Anchor development with latest docs and best practices"
|
|
"Create a React skill from react.dev with comprehensive examples"
|
|
"Create a skill for Solana transaction debugging workflows"
|
|
"Create a skill for writing technical documentation following company standards"
|
|
```
|
|
|
|
**I will automatically:**
|
|
1. ✅ Analyze your request
|
|
2. ✅ Select the optimal creation method
|
|
3. ✅ Create the skill
|
|
4. ✅ Run quality assurance loops (until score >= 8.0)
|
|
5. ✅ Test with automated scenarios
|
|
6. ✅ Deliver ready-to-use skill with stats
|
|
|
|
## What You Get
|
|
|
|
```
|
|
✅ anchor-development skill ready!
|
|
|
|
📊 Quality Score: 8.9/10 (Excellent)
|
|
📝 Lines: 412 (using progressive disclosure)
|
|
📚 Coverage: 247 documentation pages
|
|
💡 Examples: 68 code samples
|
|
🧪 Test Pass Rate: 100% (15/15 scenarios)
|
|
|
|
📁 Location: ~/.claude/skills/anchor-development/
|
|
📦 Zip: ~/Downloads/anchor-development.zip
|
|
|
|
Try it: "How do I create an Anchor program?"
|
|
```
|
|
|
|
## How It Works (Behind the Scenes)
|
|
|
|
### Phase 1: Request Analysis (Automatic)
|
|
|
|
I analyze your request to determine:
|
|
|
|
**Source Detection:**
|
|
- Documentation URL/mention? → Automated scraping path
|
|
- "Latest docs", "current version"? → Automated path
|
|
- GitHub repository mention? → Automated path
|
|
- PDF/manual path? → Automated path
|
|
- Custom workflow/process description? → Manual TDD path
|
|
- Both documentation AND custom needs? → Hybrid path
|
|
|
|
**Quality Requirements Extraction:**
|
|
- "Best practices" → Enforce quality gates
|
|
- "Latest version" → Scrape current docs
|
|
- "Examples" → Ensure code samples included
|
|
- "Comprehensive" → Verify coverage completeness
|
|
|
|
### Phase 2: Execution (Automatic)
|
|
|
|
**Path A: Documentation-Based (Skill_Seekers)**
|
|
```
|
|
Detected: Documentation source available
|
|
Method: Automated scraping with quality enhancement
|
|
|
|
Steps I take:
|
|
1. Check Skill_Seekers installation (install if needed)
|
|
2. Configure scraping parameters automatically
|
|
3. Run scraping with optimal settings
|
|
4. Monitor progress
|
|
5. Initial quality check
|
|
6. If score < 8.0: Run enhancement loop
|
|
7. Re-score until >= 8.0
|
|
8. Test with auto-generated scenarios
|
|
9. Package and deliver
|
|
```
|
|
|
|
**Path B: Custom Workflows (Manual TDD)**
|
|
```
|
|
Detected: Custom workflow/process
|
|
Method: Test-Driven Documentation (obra methodology)
|
|
|
|
Steps I take:
|
|
1. Create pressure test scenarios
|
|
2. Run baseline (without skill)
|
|
3. Document agent behavior
|
|
4. Write minimal skill addressing baseline
|
|
5. Test with skill present
|
|
6. Identify rationalizations/gaps
|
|
7. Close loopholes
|
|
8. Iterate until bulletproof
|
|
9. Package and deliver
|
|
```
|
|
|
|
**Path C: Hybrid**
|
|
```
|
|
Detected: Documentation + custom requirements
|
|
Method: Scrape then enhance
|
|
|
|
Steps I take:
|
|
1. Scrape documentation (Path A)
|
|
2. Identify gaps vs requirements
|
|
3. Fill gaps with TDD approach (Path B)
|
|
4. Unify and test as whole
|
|
5. Quality loop until >= 8.0
|
|
6. Package and deliver
|
|
```
|
|
|
|
### Phase 3: Quality Assurance Loop (Automatic)
|
|
|
|
**I enforce Anthropic best practices:**
|
|
|
|
```python
|
|
while quality_score < 8.0:
|
|
issues = analyze_against_anthropic_guidelines(skill)
|
|
|
|
if "vague_description" in issues:
|
|
improve_description_specificity()
|
|
|
|
if "missing_examples" in issues:
|
|
extract_or_generate_examples()
|
|
|
|
if "too_long" in issues:
|
|
apply_progressive_disclosure()
|
|
|
|
if "poor_structure" in issues:
|
|
reorganize_content()
|
|
|
|
quality_score = rescore()
|
|
```
|
|
|
|
**Quality Criteria (Anthropic Best Practices):**
|
|
- ✅ Description: Specific, clear, includes when_to_use
|
|
- ✅ Conciseness: <500 lines OR progressive disclosure
|
|
- ✅ Examples: Concrete code samples, not abstract
|
|
- ✅ Structure: Well-organized, clear sections
|
|
- ✅ Name: Follows conventions (lowercase, hyphens, descriptive)
|
|
|
|
**Important**: The quality assurance process consults [references/best-practices.md](references/best-practices.md) for Anthropic's complete authoring guidelines and [references/anthropic-best-practices.md](references/anthropic-best-practices.md) for the 10-point scoring criteria.
|
|
|
|
### Phase 4: Testing (Automatic)
|
|
|
|
**I generate and run test scenarios:**
|
|
|
|
```python
|
|
# Auto-generate test cases from skill content
|
|
test_cases = extract_key_topics(skill)
|
|
|
|
for topic in test_cases:
|
|
query = f"How do I {topic}?"
|
|
|
|
# Test WITHOUT skill (baseline)
|
|
baseline = run_query_without_skill(query)
|
|
|
|
# Test WITH skill
|
|
with_skill = run_query_with_skill(query)
|
|
|
|
# Verify improvement
|
|
if not is_better(with_skill, baseline):
|
|
identify_gap()
|
|
enhance_skill()
|
|
retest()
|
|
```
|
|
|
|
### Phase 5: Delivery (Automatic)
|
|
|
|
```
|
|
Package skill:
|
|
- Create skill directory structure
|
|
- Generate SKILL.md with frontmatter
|
|
- Create reference files (if using progressive disclosure)
|
|
- Add examples directory
|
|
- Create .zip for easy upload
|
|
- Install to ~/.claude/skills/ (if desired)
|
|
- Generate summary statistics
|
|
```
|
|
|
|
## Progress Reporting
|
|
|
|
You'll see real-time progress:
|
|
|
|
```
|
|
🔍 Analyzing request...
|
|
✅ Detected: Documentation-based (docs.rs/anchor-lang)
|
|
✅ Requirements: Latest version, best practices, examples
|
|
|
|
🔄 Creating skill...
|
|
📥 Scraping docs.rs/anchor-lang... (2 min)
|
|
📚 Extracting 247 pages...
|
|
💾 Organizing content...
|
|
|
|
📊 Quality check: 7.4/10
|
|
⚠️ Issues found:
|
|
- Description too generic (fixing...)
|
|
- Missing examples in 4 sections (adding...)
|
|
- Some outdated patterns (updating...)
|
|
|
|
🔧 Enhancing skill...
|
|
✏️ Description improved
|
|
📝 Examples added
|
|
🔄 Patterns updated
|
|
|
|
📊 Quality check: 8.9/10 ✅
|
|
|
|
🧪 Testing...
|
|
✅ 15/15 scenarios passing
|
|
|
|
✅ anchor-development skill ready!
|
|
```
|
|
|
|
## Dependencies
|
|
|
|
**Required:**
|
|
- Python 3.10+ (for quality scripts)
|
|
- bash (for automation scripts)
|
|
|
|
**Optional (auto-installed when needed):**
|
|
- Skill_Seekers (for documentation scraping)
|
|
- Will prompt for installation on first documentation-based request
|
|
- One-command setup: `scripts/install-skill-seekers.sh`
|
|
|
|
## Configuration
|
|
|
|
**Environment variables (optional):**
|
|
|
|
```bash
|
|
# Skill_Seekers installation path
|
|
export SKILL_SEEKERS_PATH="$HOME/Skill_Seekers"
|
|
|
|
# Output directory for created skills
|
|
export SKILL_OUTPUT_DIR="$HOME/.claude/skills"
|
|
|
|
# Minimum quality score (default: 8.0)
|
|
export MIN_QUALITY_SCORE="8.0"
|
|
|
|
# Auto-install dependencies (default: prompt)
|
|
export AUTO_INSTALL_DEPS="true"
|
|
```
|
|
|
|
## Examples
|
|
|
|
**Example 1: Documentation Skill**
|
|
```
|
|
User: "Create a React skill from react.dev"
|
|
|
|
Agent:
|
|
🔍 Analyzing... → Documentation-based
|
|
🔄 Scraping react.dev... → 3 min
|
|
📊 Quality: 7.8 → 8.6 ✅
|
|
🧪 Testing: 12/12 passing ✅
|
|
✅ react-development skill ready (8.6/10)
|
|
```
|
|
|
|
**Example 2: Custom Workflow Skill**
|
|
```
|
|
User: "Create a skill for debugging Solana transaction failures"
|
|
|
|
Agent:
|
|
🔍 Analyzing... → Custom workflow (no docs to scrape)
|
|
📝 Using TDD methodology...
|
|
🧪 RED: Testing baseline...
|
|
✏️ GREEN: Writing skill...
|
|
🔄 REFACTOR: Closing loopholes...
|
|
📊 Quality: 8.3 ✅
|
|
✅ solana-transaction-debugging skill ready (8.3/10)
|
|
```
|
|
|
|
**Example 3: Hybrid Skill**
|
|
```
|
|
User: "Create an Anchor skill with docs plus custom debugging workflows"
|
|
|
|
Agent:
|
|
🔍 Analyzing... → Hybrid (docs + custom)
|
|
📥 Scraping docs.rs/anchor-lang... → 2 min
|
|
📝 Adding custom debugging workflows...
|
|
🔄 Integrating and testing...
|
|
📊 Quality: 8.9 ✅
|
|
✅ anchor-development skill ready (8.9/10)
|
|
```
|
|
|
|
## Quality Guarantee
|
|
|
|
**Every skill delivered by skill-factory:**
|
|
- ✅ Scores >= 8.0/10 on Anthropic best practices
|
|
- ✅ Has concrete examples (not abstract)
|
|
- ✅ Follows structure conventions
|
|
- ✅ Tested with auto-generated scenarios
|
|
- ✅ Ready to use immediately
|
|
|
|
**If quality < 8.0, I keep working until it reaches 8.0+**
|
|
|
|
## Troubleshooting
|
|
|
|
**Skill_Seekers installation fails:**
|
|
```bash
|
|
# Manual installation
|
|
git clone https://github.com/yusufkaraaslan/Skill_Seekers ~/Skill_Seekers
|
|
cd ~/Skill_Seekers
|
|
pip install -r requirements.txt
|
|
|
|
# Or use installation script
|
|
~/Projects/claude-skills/skill-factory/skill/scripts/install-skill-seekers.sh
|
|
```
|
|
|
|
**Quality score stuck below 8.0:**
|
|
- I'll report what's blocking and suggest manual review
|
|
- Check references/anthropic-best-practices.md for criteria
|
|
- Run manual enhancement if needed
|
|
|
|
**Want to understand methodology:**
|
|
- See references/obra-tdd-methodology.md (testing approach)
|
|
- See references/anthropic-best-practices.md (quality criteria)
|
|
- See references/skill-seekers-integration.md (automation details)
|
|
|
|
## Reference Files
|
|
|
|
**Anthropic Official Documentation:**
|
|
- references/overview.md - Agent Skills architecture, progressive disclosure, and platform details
|
|
- references/quickstart.md - Quick tutorial on using pre-built Agent Skills in the Claude API
|
|
- references/best-practices.md - Comprehensive authoring guidelines from Anthropic
|
|
- references/anthropic-best-practices.md - Quality scoring system (10/10 criteria)
|
|
|
|
**Skill Factory Implementation Details:**
|
|
- references/obra-tdd-methodology.md - Full TDD testing approach
|
|
- references/skill-seekers-integration.md - Automation documentation
|
|
- references/request-analysis.md - How requests are parsed
|
|
- references/quality-loops.md - Enhancement algorithms
|
|
|
|
## Scripts Reference
|
|
|
|
Available helper scripts in `scripts/` directory:
|
|
- **check-skill-seekers.sh** - Check if Skill_Seekers is installed
|
|
- **install-skill-seekers.sh** - One-command Skill_Seekers setup
|
|
- **quality-check.py** - Score any skill against Anthropic best practices
|
|
|
|
Usage examples:
|
|
```bash
|
|
# Check Skill_Seekers installation
|
|
./scripts/check-skill-seekers.sh
|
|
|
|
# Install Skill_Seekers
|
|
./scripts/install-skill-seekers.sh
|
|
|
|
# Quality check a skill
|
|
python3 ./scripts/quality-check.py /path/to/skill/SKILL.md
|
|
```
|
|
|
|
## Philosophy
|
|
|
|
**You don't want to:**
|
|
- Navigate decision trees
|
|
- Choose between tools
|
|
- Check quality manually
|
|
- Test with subagents yourself
|
|
- Wonder if output is good
|
|
|
|
**You want to:**
|
|
- Describe what you need
|
|
- Get high-quality result
|
|
- Start using immediately
|
|
|
|
**That's what skill-factory delivers.**
|
|
|
|
## Credits
|
|
|
|
Built on top of excellent tools:
|
|
- [Skill_Seekers](https://github.com/yusufkaraaslan/Skill_Seekers) - Documentation scraping
|
|
- [obra/superpowers-skills](https://github.com/obra/superpowers-skills) - TDD methodology
|
|
- [Anthropic skill-creator](https://github.com/anthropics/skills) - Best practices
|
|
|
|
Skill-factory orchestrates these tools with automatic quality assurance and testing.
|
|
|
|
---
|
|
|
|
**Just tell me what skill you need. I'll handle the rest.**
|