Files
SuperClaude/docs/memory/WORKFLOW_METRICS_SCHEMA.md
kazuki nakai 882a0d8356 refactor: PM Agent complete independence from external MCP servers (#439)
* refactor: PM Agent complete independence from external MCP servers

## Summary
Implement graceful degradation to ensure PM Agent operates fully without
any MCP server dependencies. MCP servers now serve as optional enhancements
rather than required components.

## Changes

### Responsibility Separation (NEW)
- **PM Agent**: Development workflow orchestration (PDCA cycle, task management)
- **mindbase**: Memory management (long-term, freshness, error learning)
- **Built-in memory**: Session-internal context (volatile)

### 3-Layer Memory Architecture with Fallbacks
1. **Built-in Memory** [OPTIONAL]: Session context via MCP memory server
2. **mindbase** [OPTIONAL]: Long-term semantic search via airis-mcp-gateway
3. **Local Files** [ALWAYS]: Core functionality in docs/memory/

### Graceful Degradation Implementation
- All MCP operations marked with [ALWAYS] or [OPTIONAL]
- Explicit IF/ELSE fallback logic for every MCP call
- Dual storage: Always write to local files + optionally to mindbase
- Smart lookup: Semantic search (if available) → Text search (always works)

### Key Fallback Strategies

**Session Start**:
- mindbase available: search_conversations() for semantic context
- mindbase unavailable: Grep docs/memory/*.jsonl for text-based lookup

**Error Detection**:
- mindbase available: Semantic search for similar past errors
- mindbase unavailable: Grep docs/mistakes/ + solutions_learned.jsonl

**Knowledge Capture**:
- Always: echo >> docs/memory/patterns_learned.jsonl (persistent)
- Optional: mindbase.store() for semantic search enhancement

## Benefits
-  Zero external dependencies (100% functionality without MCP)
-  Enhanced capabilities when MCPs available (semantic search, freshness)
-  No functionality loss, only reduced search intelligence
-  Transparent degradation (no error messages, automatic fallback)

## Related Research
- Serena MCP investigation: Exposes tools (not resources), memory = markdown files
- mindbase superiority: PostgreSQL + pgvector > Serena memory features
- Best practices alignment: /Users/kazuki/github/airis-mcp-gateway/docs/mcp-best-practices.md

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore: add PR template and pre-commit config

- Add structured PR template with Git workflow checklist
- Add pre-commit hooks for secret detection and Conventional Commits
- Enforce code quality gates (YAML/JSON/Markdown lint, shellcheck)

NOTE: Execute pre-commit inside Docker container to avoid host pollution:
  docker compose exec workspace uv tool install pre-commit
  docker compose exec workspace pre-commit run --all-files

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* docs: update PM Agent context with token efficiency architecture

- Add Layer 0 Bootstrap (150 tokens, 95% reduction)
- Document Intent Classification System (5 complexity levels)
- Add Progressive Loading strategy (5-layer)
- Document mindbase integration incentive (38% savings)
- Update with 2025-10-17 redesign details

* refactor: PM Agent command with progressive loading

- Replace auto-loading with User Request First philosophy
- Add 5-layer progressive context loading
- Implement intent classification system
- Add workflow metrics collection (.jsonl)
- Document graceful degradation strategy

* fix: installer improvements

Update installer logic for better reliability

* docs: add comprehensive development documentation

- Add architecture overview
- Add PM Agent improvements analysis
- Add parallel execution architecture
- Add CLI install improvements
- Add code style guide
- Add project overview
- Add install process analysis

* docs: add research documentation

Add LLM agent token efficiency research and analysis

* docs: add suggested commands reference

* docs: add session logs and testing documentation

- Add session analysis logs
- Add testing documentation

* feat: migrate CLI to typer + rich for modern UX

## What Changed

### New CLI Architecture (typer + rich)
- Created `superclaude/cli/` module with modern typer-based CLI
- Replaced custom UI utilities with rich native features
- Added type-safe command structure with automatic validation

### Commands Implemented
- **install**: Interactive installation with rich UI (progress, panels)
- **doctor**: System diagnostics with rich table output
- **config**: API key management with format validation

### Technical Improvements
- Dependencies: Added typer>=0.9.0, rich>=13.0.0, click>=8.0.0
- Entry Point: Updated pyproject.toml to use `superclaude.cli.app:cli_main`
- Tests: Added comprehensive smoke tests (11 passed)

### User Experience Enhancements
- Rich formatted help messages with panels and tables
- Automatic input validation with retry loops
- Clear error messages with actionable suggestions
- Non-interactive mode support for CI/CD

## Testing

```bash
uv run superclaude --help     # ✓ Works
uv run superclaude doctor     # ✓ Rich table output
uv run superclaude config show # ✓ API key management
pytest tests/test_cli_smoke.py # ✓ 11 passed, 1 skipped
```

## Migration Path

-  P0: Foundation complete (typer + rich + smoke tests)
- 🔜 P1: Pydantic validation models (next sprint)
- 🔜 P2: Enhanced error messages (next sprint)
- 🔜 P3: API key retry loops (next sprint)

## Performance Impact

- **Code Reduction**: Prepared for -300 lines (custom UI → rich)
- **Type Safety**: Automatic validation from type hints
- **Maintainability**: Framework primitives vs custom code

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactor: consolidate documentation directories

Merged claudedocs/ into docs/research/ for consistent documentation structure.

Changes:
- Moved all claudedocs/*.md files to docs/research/
- Updated all path references in documentation (EN/KR)
- Updated RULES.md and research.md command templates
- Removed claudedocs/ directory
- Removed ClaudeDocs/ from .gitignore

Benefits:
- Single source of truth for all research reports
- PEP8-compliant lowercase directory naming
- Clearer documentation organization
- Prevents future claudedocs/ directory creation

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* perf: reduce /sc:pm command output from 1652 to 15 lines

- Remove 1637 lines of documentation from command file
- Keep only minimal bootstrap message
- 99% token reduction on command execution
- Detailed specs remain in superclaude/agents/pm-agent.md

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* perf: split PM Agent into execution workflows and guide

- Reduce pm-agent.md from 735 to 429 lines (42% reduction)
- Move philosophy/examples to docs/agents/pm-agent-guide.md
- Execution workflows (PDCA, file ops) stay in pm-agent.md
- Guide (examples, quality standards) read once when needed

Token savings:
- Agent loading: ~6K → ~3.5K tokens (42% reduction)
- Total with pm.md: 71% overall reduction

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactor: consolidate PM Agent optimization and pending changes

PM Agent optimization (already committed separately):
- superclaude/commands/pm.md: 1652→14 lines
- superclaude/agents/pm-agent.md: 735→429 lines
- docs/agents/pm-agent-guide.md: new guide file

Other pending changes:
- setup: framework_docs, mcp, logger, remove ui.py
- superclaude: __main__, cli/app, cli/commands/install
- tests: test_ui updates
- scripts: workflow metrics analysis tools
- docs/memory: session state updates

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactor: simplify MCP installer to unified gateway with legacy mode

## Changes

### MCP Component (setup/components/mcp.py)
- Simplified to single airis-mcp-gateway by default
- Added legacy mode for individual official servers (sequential-thinking, context7, magic, playwright)
- Dynamic prerequisites based on mode:
  - Default: uv + claude CLI only
  - Legacy: node (18+) + npm + claude CLI
- Removed redundant server definitions

### CLI Integration
- Added --legacy flag to setup/cli/commands/install.py
- Added --legacy flag to superclaude/cli/commands/install.py
- Config passes legacy_mode to component installer

## Benefits
-  Simpler: 1 gateway vs 9+ individual servers
-  Lighter: No Node.js/npm required (default mode)
-  Unified: All tools in one gateway (sequential-thinking, context7, magic, playwright, serena, morphllm, tavily, chrome-devtools, git, puppeteer)
-  Flexible: --legacy flag for official servers if needed

## Usage
```bash
superclaude install              # Default: airis-mcp-gateway (推奨)
superclaude install --legacy     # Legacy: individual official servers
```

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactor: rename CoreComponent to FrameworkDocsComponent and add PM token tracking

## Changes

### Component Renaming (setup/components/)
- Renamed CoreComponent → FrameworkDocsComponent for clarity
- Updated all imports in __init__.py, agents.py, commands.py, mcp_docs.py, modes.py
- Better reflects the actual purpose (framework documentation files)

### PM Agent Enhancement (superclaude/commands/pm.md)
- Added token usage tracking instructions
- PM Agent now reports:
  1. Current token usage from system warnings
  2. Percentage used (e.g., "27% used" for 54K/200K)
  3. Status zone: 🟢 <75% | 🟡 75-85% | 🔴 >85%
- Helps prevent token exhaustion during long sessions

### UI Utilities (setup/utils/ui.py)
- Added new UI utility module for installer
- Provides consistent user interface components

## Benefits
-  Clearer component naming (FrameworkDocs vs Core)
-  PM Agent token awareness for efficiency
-  Better visual feedback with status zones

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactor(pm-agent): minimize output verbosity (471→284 lines, 40% reduction)

**Problem**: PM Agent generated excessive output with redundant explanations
- "System Status Report" with decorative formatting
- Repeated "Common Tasks" lists user already knows
- Verbose session start/end protocols
- Duplicate file operations documentation

**Solution**: Compress without losing functionality
- Session Start: Reduced to symbol-only status (🟢 branch | nM nD | token%)
- Session End: Compressed to essential actions only
- File Operations: Consolidated from 2 sections to 1 line reference
- Self-Improvement: 5 phases → 1 unified workflow
- Output Rules: Explicit constraints to prevent Claude over-explanation

**Quality Preservation**:
-  All core functions retained (PDCA, memory, patterns, mistakes)
-  PARALLEL Read/Write preserved (performance critical)
-  Workflow unchanged (session lifecycle intact)
-  Added output constraints (prevents verbose generation)

**Reduction Method**:
- Deleted: Explanatory text, examples, redundant sections
- Retained: Action definitions, file paths, core workflows
- Added: Explicit output constraints to enforce minimalism

**Token Impact**: 40% reduction in agent documentation size
**Before**: Verbose multi-section report with task lists
**After**: Single line status: 🟢 integration | 15M 17D | 36%

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* refactor: consolidate MCP integration to unified gateway

**Changes**:
- Remove individual MCP server docs (superclaude/mcp/*.md)
- Remove MCP server configs (superclaude/mcp/configs/*.json)
- Delete MCP docs component (setup/components/mcp_docs.py)
- Simplify installer (setup/core/installer.py)
- Update components for unified gateway approach

**Rationale**:
- Unified gateway (airis-mcp-gateway) provides all MCP servers
- Individual docs/configs no longer needed (managed centrally)
- Reduces maintenance burden and file count
- Simplifies installation process

**Files Removed**: 17 MCP files (docs + configs)
**Installer Changes**: Removed legacy MCP installation logic

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* chore: update version and component metadata

- Bump version (pyproject.toml, setup/__init__.py)
- Update CLAUDE.md import service references
- Reflect component structure changes

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

---------

Co-authored-by: kazuki <kazuki@kazukinoMacBook-Air.local>
Co-authored-by: Claude <noreply@anthropic.com>
2025-10-17 05:43:06 +05:30

402 lines
10 KiB
Markdown

# Workflow Metrics Schema
**Purpose**: Token efficiency tracking for continuous optimization and A/B testing
**File**: `docs/memory/workflow_metrics.jsonl` (append-only log)
## Data Structure (JSONL Format)
Each line is a complete JSON object representing one workflow execution.
```jsonl
{
"timestamp": "2025-10-17T01:54:21+09:00",
"session_id": "abc123def456",
"task_type": "typo_fix",
"complexity": "light",
"workflow_id": "progressive_v3_layer2",
"layers_used": [0, 1, 2],
"tokens_used": 650,
"time_ms": 1800,
"files_read": 1,
"mindbase_used": false,
"sub_agents": [],
"success": true,
"user_feedback": "satisfied",
"notes": "Optional implementation notes"
}
```
## Field Definitions
### Required Fields
| Field | Type | Description | Example |
|-------|------|-------------|---------|
| `timestamp` | ISO 8601 | Execution timestamp in JST | `"2025-10-17T01:54:21+09:00"` |
| `session_id` | string | Unique session identifier | `"abc123def456"` |
| `task_type` | string | Task classification | `"typo_fix"`, `"bug_fix"`, `"feature_impl"` |
| `complexity` | string | Intent classification level | `"ultra-light"`, `"light"`, `"medium"`, `"heavy"`, `"ultra-heavy"` |
| `workflow_id` | string | Workflow variant identifier | `"progressive_v3_layer2"` |
| `layers_used` | array | Progressive loading layers executed | `[0, 1, 2]` |
| `tokens_used` | integer | Total tokens consumed | `650` |
| `time_ms` | integer | Execution time in milliseconds | `1800` |
| `success` | boolean | Task completion status | `true`, `false` |
### Optional Fields
| Field | Type | Description | Example |
|-------|------|-------------|---------|
| `files_read` | integer | Number of files read | `1` |
| `mindbase_used` | boolean | Whether mindbase MCP was used | `false` |
| `sub_agents` | array | Delegated sub-agents | `["backend-architect", "quality-engineer"]` |
| `user_feedback` | string | Inferred user satisfaction | `"satisfied"`, `"neutral"`, `"unsatisfied"` |
| `notes` | string | Implementation notes | `"Used cached solution"` |
| `confidence_score` | float | Pre-implementation confidence | `0.85` |
| `hallucination_detected` | boolean | Self-check red flags found | `false` |
| `error_recurrence` | boolean | Same error encountered before | `false` |
## Task Type Taxonomy
### Ultra-Light Tasks
- `progress_query`: "進捗教えて"
- `status_check`: "現状確認"
- `next_action_query`: "次のタスクは?"
### Light Tasks
- `typo_fix`: README誤字修正
- `comment_addition`: コメント追加
- `variable_rename`: 変数名変更
- `documentation_update`: ドキュメント更新
### Medium Tasks
- `bug_fix`: バグ修正
- `small_feature`: 小機能追加
- `refactoring`: リファクタリング
- `test_addition`: テスト追加
### Heavy Tasks
- `feature_impl`: 新機能実装
- `architecture_change`: アーキテクチャ変更
- `security_audit`: セキュリティ監査
- `integration`: 外部システム統合
### Ultra-Heavy Tasks
- `system_redesign`: システム全面再設計
- `framework_migration`: フレームワーク移行
- `comprehensive_research`: 包括的調査
## Workflow Variant Identifiers
### Progressive Loading Variants
- `progressive_v3_layer1`: Ultra-light (memory files only)
- `progressive_v3_layer2`: Light (target file only)
- `progressive_v3_layer3`: Medium (related files 3-5)
- `progressive_v3_layer4`: Heavy (subsystem)
- `progressive_v3_layer5`: Ultra-heavy (full + external research)
### Experimental Variants (A/B Testing)
- `experimental_eager_layer3`: Always load Layer 3 for medium tasks
- `experimental_lazy_layer2`: Minimal Layer 2 loading
- `experimental_parallel_layer3`: Parallel file loading in Layer 3
## Complexity Classification Rules
```yaml
ultra_light:
keywords: ["進捗", "状況", "進み", "where", "status", "progress"]
token_budget: "100-500"
layers: [0, 1]
light:
keywords: ["誤字", "typo", "fix typo", "correct", "comment"]
token_budget: "500-2K"
layers: [0, 1, 2]
medium:
keywords: ["バグ", "bug", "fix", "修正", "error", "issue"]
token_budget: "2-5K"
layers: [0, 1, 2, 3]
heavy:
keywords: ["新機能", "new feature", "implement", "実装"]
token_budget: "5-20K"
layers: [0, 1, 2, 3, 4]
ultra_heavy:
keywords: ["再設計", "redesign", "overhaul", "migration"]
token_budget: "20K+"
layers: [0, 1, 2, 3, 4, 5]
```
## Recording Points
### Session Start (Layer 0)
```python
session_id = generate_session_id()
workflow_metrics = {
"timestamp": get_current_time(),
"session_id": session_id,
"workflow_id": "progressive_v3_layer0"
}
# Bootstrap: 150 tokens
```
### After Intent Classification (Layer 1)
```python
workflow_metrics.update({
"task_type": classify_task_type(user_request),
"complexity": classify_complexity(user_request),
"estimated_token_budget": get_budget(complexity)
})
```
### After Progressive Loading
```python
workflow_metrics.update({
"layers_used": [0, 1, 2], # Actual layers executed
"tokens_used": calculate_tokens(),
"files_read": len(files_loaded)
})
```
### After Task Completion
```python
workflow_metrics.update({
"success": task_completed_successfully,
"time_ms": execution_time_ms,
"user_feedback": infer_user_satisfaction()
})
```
### Session End
```python
# Append to workflow_metrics.jsonl
with open("docs/memory/workflow_metrics.jsonl", "a") as f:
f.write(json.dumps(workflow_metrics) + "\n")
```
## Analysis Scripts
### Weekly Analysis
```bash
# Group by task type and calculate averages
python scripts/analyze_workflow_metrics.py --period week
# Output:
# Task Type: typo_fix
# Count: 12
# Avg Tokens: 680
# Avg Time: 1,850ms
# Success Rate: 100%
```
### A/B Testing Analysis
```bash
# Compare workflow variants
python scripts/ab_test_workflows.py \
--variant-a progressive_v3_layer2 \
--variant-b experimental_eager_layer3 \
--metric tokens_used
# Output:
# Variant A (progressive_v3_layer2):
# Avg Tokens: 1,250
# Success Rate: 95%
#
# Variant B (experimental_eager_layer3):
# Avg Tokens: 2,100
# Success Rate: 98%
#
# Statistical Significance: p = 0.03 (significant)
# Recommendation: Keep Variant A (better efficiency)
```
## Usage (Continuous Optimization)
### Weekly Review Process
```yaml
every_monday_morning:
1. Run analysis: python scripts/analyze_workflow_metrics.py --period week
2. Identify patterns:
- Best-performing workflows per task type
- Inefficient patterns (high tokens, low success)
- User satisfaction trends
3. Update recommendations:
- Promote efficient workflows to standard
- Deprecate inefficient workflows
- Design new experimental variants
```
### A/B Testing Framework
```yaml
allocation_strategy:
current_best: 80% # Use best-known workflow
experimental: 20% # Test new variant
evaluation_criteria:
minimum_trials: 20 # Per variant
confidence_level: 0.95 # p < 0.05
metrics:
- tokens_used (primary)
- success_rate (gate: must be ≥95%)
- user_feedback (qualitative)
promotion_rules:
if experimental_better:
- Statistical significance confirmed
- Success rate ≥ current_best
- User feedback ≥ neutral
→ Promote to standard (80% allocation)
if experimental_worse:
→ Deprecate variant
→ Document learning in docs/patterns/
```
### Auto-Optimization Cycle
```yaml
monthly_cleanup:
1. Identify stale workflows:
- No usage in last 90 days
- Success rate <80%
- User feedback consistently negative
2. Archive deprecated workflows:
- Move to docs/patterns/deprecated/
- Document why deprecated
3. Promote new standards:
- Experimental → Standard (if proven better)
- Update pm.md with new best practices
4. Generate monthly report:
- Token efficiency trends
- Success rate improvements
- User satisfaction evolution
```
## Visualization
### Token Usage Over Time
```python
import pandas as pd
import matplotlib.pyplot as plt
df = pd.read_json("docs/memory/workflow_metrics.jsonl", lines=True)
df['date'] = pd.to_datetime(df['timestamp']).dt.date
daily_avg = df.groupby('date')['tokens_used'].mean()
plt.plot(daily_avg)
plt.title("Average Token Usage Over Time")
plt.ylabel("Tokens")
plt.xlabel("Date")
plt.show()
```
### Task Type Distribution
```python
task_counts = df['task_type'].value_counts()
plt.pie(task_counts, labels=task_counts.index, autopct='%1.1f%%')
plt.title("Task Type Distribution")
plt.show()
```
### Workflow Efficiency Comparison
```python
workflow_efficiency = df.groupby('workflow_id').agg({
'tokens_used': 'mean',
'success': 'mean',
'time_ms': 'mean'
})
print(workflow_efficiency.sort_values('tokens_used'))
```
## Expected Patterns
### Healthy Metrics (After 1 Month)
```yaml
token_efficiency:
ultra_light: 750-1,050 tokens (63% reduction)
light: 1,250 tokens (46% reduction)
medium: 3,850 tokens (47% reduction)
heavy: 10,350 tokens (40% reduction)
success_rates:
all_tasks: ≥95%
ultra_light: 100% (simple tasks)
light: 98%
medium: 95%
heavy: 92%
user_satisfaction:
satisfied: ≥70%
neutral: ≤25%
unsatisfied: ≤5%
```
### Red Flags (Require Investigation)
```yaml
warning_signs:
- success_rate < 85% for any task type
- tokens_used > estimated_budget by >30%
- time_ms > 10 seconds for light tasks
- user_feedback "unsatisfied" > 10%
- error_recurrence > 15%
```
## Integration with PM Agent
### Automatic Recording
PM Agent automatically records metrics at each execution point:
- Session start (Layer 0)
- Intent classification (Layer 1)
- Progressive loading (Layers 2-5)
- Task completion
- Session end
### No Manual Intervention
- All recording is automatic
- No user action required
- Transparent operation
- Privacy-preserving (local files only)
## Privacy and Security
### Data Retention
- Local storage only (`docs/memory/`)
- No external transmission
- Git-manageable (optional)
- User controls retention period
### Sensitive Data Handling
- No code snippets logged
- No user input content
- Only metadata (tokens, timing, success)
- Task types are generic classifications
## Maintenance
### File Rotation
```bash
# Archive old metrics (monthly)
mv docs/memory/workflow_metrics.jsonl \
docs/memory/archive/workflow_metrics_2025-10.jsonl
# Start fresh
touch docs/memory/workflow_metrics.jsonl
```
### Cleanup
```bash
# Remove metrics older than 6 months
find docs/memory/archive/ -name "workflow_metrics_*.jsonl" \
-mtime +180 -delete
```
## References
- Specification: `superclaude/commands/pm.md` (Line 291-355)
- Research: `docs/research/llm-agent-token-efficiency-2025.md`
- Tests: `tests/pm_agent/test_token_budget.py`