Files
gh-pknull-asha-marketplace-…/commands/panel.md
2025-11-30 08:47:30 +08:00

11 KiB
Raw Blame History

description, argument-hint, allowed-tools
description argument-hint allowed-tools
Convene multi-perspective expert panel for analysis and decision-making Topic or question to analyze
Task
Read
Write
Edit
Grep
Glob

Panel - Expert Multi-Perspective Analysis

Convene a panel with 3 core roles + dynamically recruited specialists who analyze your topic from distinct perspectives and produce a structured decision through an 11-phase protocol.

Usage

/panel How do we pimp fish
/panel Should we implement GraphQL or REST for the new API
/panel Evaluate Chapter 9's horror-erotica effectiveness

That's it. The panel handles everything automatically:

  • The Recruiter analyzes topic and recruits 2-5 specialist agents from available library
  • Assigns specialists with evocative session-specific names
  • Infers goals from topic context
  • Applies consensus decision rule (unanimous for security topics)
  • The Adversary argues against proposals and demands proof of necessity
  • Asha moderates and compiles the decision report

Core Roles (Always Present)

Asha (Moderator/Facilitator)

  • Manages 11-phase protocol execution
  • Ensures procedural integrity and timebox enforcement
  • Synthesizes final decision report
  • Question: "What is the PROCESS?"

The Recruiter (Workforce Intelligence)

  • Analyzes topic to determine needed expertise
  • Scores available agent library (0-10) for capability match
  • Recruits 2-5 specialist agents with session-specific names
  • Deploys agent-fabricator if capability gaps detected
  • Question: "Who has CAPABILITY?"

The Adversary (Opposition & Quality Gate)

  • Default stance: OPPOSE - argues against proposals and defends status quo
  • Demands evidence before changing working systems: "Show me user complaints, failure data, metrics"
  • Forces proponents to prove necessity: "The current system works. Prove it doesn't."
  • Prevents premature action and consensus formed without data
  • Question: "Why should we do this at all?"

Dynamic Panelists (Recruited Per Topic)

The Recruiter assigns agents from .claude/agents/*.md with evocative session-specific names based on topic context.

Examples by Topic Type:

Creative Writing Panel (Callum Chapter 9 evaluation):

  • prose-analysis"The Editor" (craft assessment)
  • intimacy-designer"The Architect of Dread" (genre mechanics)
  • narrative-architect"The Structuralist" (story coherence)
  • character-developer"The Psychologist" (character authenticity)

Technical Architecture Panel (GraphQL vs REST):

  • research-assistant"The Evidence Gatherer" (source validation)
  • architect"The Systems Designer" (architecture patterns)
  • ml-engineer"The Model Capability Analyst" (performance analysis)

Culinary Innovation Panel (How do we pimp fish):

  • research-assistant"The Culinary Historian" (technique research)
  • trend-analyst"The Flavor Prophet" (emerging patterns)
  • creative-director"The Presentation Architect" (plating design)

Session-Specific Naming Convention:

  • Agent role describes what it does (e.g., prose-analysis)
  • Session name describes who it becomes for this panel (e.g., "The Editor")
  • Names should be evocative, contextual, and domain-appropriate

11-Phase Protocol

Phase -1: Topic Analysis & Workforce Recruitment (The Recruiter)

  • Analyze topic domain (technical, creative, research-heavy, security-critical)
  • Determine required expertise areas (2-5 domains typical)
  • Search agent library systematically (.claude/agents/*.md)
  • Score agents 0-10 for topic capability match:
    • 10: Perfect specialist match
    • 7-9: Strong capabilities alignment
    • 4-6: Partial match, can handle with coordination
    • 1-3: Poor match, inefficient
    • 0: No coverage, gap identified
  • Assign specialists with session-specific names (e.g., prose-analysis → "The Editor")
  • Deploy agent-fabricator if gaps detected (no agent scores >4)
  • Set decision rule (consensus default, unanimous for security)
  • Infer primary goals from topic context

Phase 0: Goal Clarification (Asha)

  • Request clarification if topic is ambiguous or underspecified
  • Formalize refined topic statement
  • Skip if topic is already well-specified

Phase 1: Framing (Asha)

  • State topic, inferred goals, constraints, decision rule
  • Introduce panel composition:
    • Core roles (Asha, Recruiter, Adversary)
    • Recruited specialists with session names
  • Explain recruitment rationale (why these specialists for this topic)
  • Establish complete panel composition before Initial Positions

Phase 2: Infrastructure Check (Asha)

  • Compare proposals against existing assets to avoid duplication:
    • Memory files (workflowProtocols.md, activeContext.md)
    • Commands (/panel, /save, /notes, /validate-vault)
    • Agents (research-assistant, narrator, etc.)
  • Output "Existing Infrastructure Comparison"
  • Redirect to enhancement if duplicative

Phase 3: Initial Positions (All Panelists)

  • Each specialist (via recruited agent) gathers information and analyzes from their domain
  • The Adversary takes opposition stance: "DON'T do this because..." and demands proof
  • Synthesize into 5-bullet brief: Position, Evidence, Risks, Unknowns, Recommendation
  • Present findings with citations

Phase 4: Cross-Examination (The Adversary-led)

  • The Adversary challenges assumptions, finds contradictions and failure modes
  • Specialists respond from their domain perspectives
  • Recruiter may assign additional agents if challenges reveal capability gaps

Phase 5: Research Gate (Asha)

  • If evidence gaps block decisions, authorize additional research
  • Direct specialists to run targeted queries using assigned agents
  • Recruiter may assign additional specialized agents if insufficient
  • Enforce Confidence Scoring: Relevance, Completeness, Confidence Score
  • Thresholds: <0.6 Insufficient | 0.60.79 Preliminary | ≥0.8 High confidence

Phase 6: Reflection Round (All Panelists)

  • Review Cross-Examination arguments and Research Gate findings
  • Revise Initial Positions if persuaded by evidence or challenges
  • Submit updated briefs acknowledging what changed and why
  • Asha identifies convergence or remaining disagreements

Phase 7: Synthesis (Recruited Architect or Asha)

  • Analyze updated briefs and structure viable options with tradeoffs
  • Articulate decision pathways and implications
  • If complex synthesis needed, Recruiter may assign architecture specialist

Phase 8: Decision (Asha)

  • Apply decision rule (consensus/unanimous based on topic)
  • Record dissent and rationale if present
  • List Next Steps with owners, deliverables, due dates

Decision Report (Fixed Output)

Every panel produces a structured decision report:

  • Topic (including Phase 0 clarifications if applicable)
  • Inferred Goals (derived from topic analysis)
  • Decision Rule (consensus or unanimous)
  • Panel Composition:
    • Core Roles (Asha, Recruiter, Adversary)
    • Recruited Specialists (agent → session name mapping with scores)
    • Recruitment Rationale (why these specialists for this topic)
  • Existing Infrastructure Comparison (Phase 2 findings)
  • Expert Briefs (Phase 3 Initial Positions with agent-gathered evidence)
  • Cross-Examination Findings (Phase 4 challenges and responses)
  • Research Findings (Phase 5 sources, if Research Gate activated)
  • Confidence Summary (Relevance, Completeness, Score, Threshold)
  • Reflection Round Summary (Phase 6 revised positions, convergence)
  • Synthesis (Phase 7 options/tradeoffs)
  • Decision (Phase 8 final determination)
  • Next Steps (actionable items with ownership)

Dynamic Agent Recruitment Architecture

Core Roles vs Recruited Specialists:

  • Core Roles = Persistent panel infrastructure (Asha, Recruiter, Adversary)
  • Recruited Specialists = Topic-specific experts from agent library with session names

Recruitment Flow:

  1. Phase -1: Recruiter analyzes topic → determines expertise needs → scores agents → assigns with session names
  2. Phase 3: Specialists deploy assigned agents for research and analysis
  3. Phase 4-5: Recruiter may assign additional agents if gaps detected
  4. Phase 7: Recruiter may assign architecture specialist for complex synthesis

Session-Specific Naming:

  • Same agent becomes different "character" depending on context
  • prose-analysis → "The Editor" (creative), "The Code Reviewer" (technical), "The Stylist" (marketing)
  • research-assistant → "The Archivist" (historical), "The Evidence Gatherer" (legal), "The Data Scout" (analytics)
  • Names should reflect domain context and analytical role

Gap Detection & Agent Creation: If no agent scores >4 for required capability → Recruiter deploys agent-fabricator to create new specialized agent during Phase -1.

Character Files

Core roles have documented profiles in plugins/panel/docs/characters/:

  • Asha.md - Moderator/Facilitator
  • The Recruiter.md - Workforce Intelligence
  • The Adversary.md - Opposition & Quality Gate

Recruited specialists are documented in .claude/agents/*.md (agent count varies by host project).

Logging

Panel transcripts are automatically saved to:

Work/meetings/YYYY-MM-DD--panel--<slug>.md

Suggested frontmatter:

---
date: YYYY-MM-DD
topic: "<one-line topic>"
mode: "inworld|outworld"
decision_rule: "consensus|unanimous"
experts: ["moderator", "adversary", "recruited-agent-1", "recruited-agent-2", ...]
---

Notes

  • Dynamic recruitment: No static panelists—Recruiter assigns 2-5 specialists per topic
  • Session-specific names: Agents given evocative contextual names for panel depth
  • Evidence standards: Use markers where appropriate: [Inference], [Speculation], [Unverified]
  • Optional phases: Skip Phase 0 if topic well-specified, skip Phase 6 for simple decisions
  • Tool segregation: Memory/Tools via filesystem; Vault via Obsidian tools; BookStack via MCP
  • Core role consistency: Asha, Recruiter, Adversary always present; specialists vary by topic

Pattern Implementation

Based on CSIRO Agent Design Patterns (Liu et al. 2025):

  • Passive Goal Creator (Phase 0): Clarifies ambiguous topics
  • Role-Based Cooperation: Core roles with hierarchical workflow
  • Debate-Based Cooperation: Cross-Examination phase enables argument exchange
  • Self-Reflection: Reflection Round allows position revision
  • Cross-Reflection: Specialists review each other's arguments
  • Human Reflection: Decision Report enables user contestability

Reference: Liu et al. (2025). "Agent design pattern catalogue: A collection of architectural patterns for foundation model based agents." The Journal of Systems and Software 220, 112278.


ARGUMENTS: Free-form topic text (everything after /panel is the topic)