mirror of
https://github.com/SuperClaude-Org/SuperClaude_Framework.git
synced 2025-12-29 16:16:08 +00:00
* refactor: PM Agent complete independence from external MCP servers ## Summary Implement graceful degradation to ensure PM Agent operates fully without any MCP server dependencies. MCP servers now serve as optional enhancements rather than required components. ## Changes ### Responsibility Separation (NEW) - **PM Agent**: Development workflow orchestration (PDCA cycle, task management) - **mindbase**: Memory management (long-term, freshness, error learning) - **Built-in memory**: Session-internal context (volatile) ### 3-Layer Memory Architecture with Fallbacks 1. **Built-in Memory** [OPTIONAL]: Session context via MCP memory server 2. **mindbase** [OPTIONAL]: Long-term semantic search via airis-mcp-gateway 3. **Local Files** [ALWAYS]: Core functionality in docs/memory/ ### Graceful Degradation Implementation - All MCP operations marked with [ALWAYS] or [OPTIONAL] - Explicit IF/ELSE fallback logic for every MCP call - Dual storage: Always write to local files + optionally to mindbase - Smart lookup: Semantic search (if available) → Text search (always works) ### Key Fallback Strategies **Session Start**: - mindbase available: search_conversations() for semantic context - mindbase unavailable: Grep docs/memory/*.jsonl for text-based lookup **Error Detection**: - mindbase available: Semantic search for similar past errors - mindbase unavailable: Grep docs/mistakes/ + solutions_learned.jsonl **Knowledge Capture**: - Always: echo >> docs/memory/patterns_learned.jsonl (persistent) - Optional: mindbase.store() for semantic search enhancement ## Benefits - ✅ Zero external dependencies (100% functionality without MCP) - ✅ Enhanced capabilities when MCPs available (semantic search, freshness) - ✅ No functionality loss, only reduced search intelligence - ✅ Transparent degradation (no error messages, automatic fallback) ## Related Research - Serena MCP investigation: Exposes tools (not resources), memory = markdown files - mindbase superiority: PostgreSQL + pgvector > Serena memory features - Best practices alignment: /Users/kazuki/github/airis-mcp-gateway/docs/mcp-best-practices.md 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * chore: add PR template and pre-commit config - Add structured PR template with Git workflow checklist - Add pre-commit hooks for secret detection and Conventional Commits - Enforce code quality gates (YAML/JSON/Markdown lint, shellcheck) NOTE: Execute pre-commit inside Docker container to avoid host pollution: docker compose exec workspace uv tool install pre-commit docker compose exec workspace pre-commit run --all-files 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * docs: update PM Agent context with token efficiency architecture - Add Layer 0 Bootstrap (150 tokens, 95% reduction) - Document Intent Classification System (5 complexity levels) - Add Progressive Loading strategy (5-layer) - Document mindbase integration incentive (38% savings) - Update with 2025-10-17 redesign details * refactor: PM Agent command with progressive loading - Replace auto-loading with User Request First philosophy - Add 5-layer progressive context loading - Implement intent classification system - Add workflow metrics collection (.jsonl) - Document graceful degradation strategy * fix: installer improvements Update installer logic for better reliability * docs: add comprehensive development documentation - Add architecture overview - Add PM Agent improvements analysis - Add parallel execution architecture - Add CLI install improvements - Add code style guide - Add project overview - Add install process analysis * docs: add research documentation Add LLM agent token efficiency research and analysis * docs: add suggested commands reference * docs: add session logs and testing documentation - Add session analysis logs - Add testing documentation * feat: migrate CLI to typer + rich for modern UX ## What Changed ### New CLI Architecture (typer + rich) - Created `superclaude/cli/` module with modern typer-based CLI - Replaced custom UI utilities with rich native features - Added type-safe command structure with automatic validation ### Commands Implemented - **install**: Interactive installation with rich UI (progress, panels) - **doctor**: System diagnostics with rich table output - **config**: API key management with format validation ### Technical Improvements - Dependencies: Added typer>=0.9.0, rich>=13.0.0, click>=8.0.0 - Entry Point: Updated pyproject.toml to use `superclaude.cli.app:cli_main` - Tests: Added comprehensive smoke tests (11 passed) ### User Experience Enhancements - Rich formatted help messages with panels and tables - Automatic input validation with retry loops - Clear error messages with actionable suggestions - Non-interactive mode support for CI/CD ## Testing ```bash uv run superclaude --help # ✓ Works uv run superclaude doctor # ✓ Rich table output uv run superclaude config show # ✓ API key management pytest tests/test_cli_smoke.py # ✓ 11 passed, 1 skipped ``` ## Migration Path - ✅ P0: Foundation complete (typer + rich + smoke tests) - 🔜 P1: Pydantic validation models (next sprint) - 🔜 P2: Enhanced error messages (next sprint) - 🔜 P3: API key retry loops (next sprint) ## Performance Impact - **Code Reduction**: Prepared for -300 lines (custom UI → rich) - **Type Safety**: Automatic validation from type hints - **Maintainability**: Framework primitives vs custom code 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * refactor: consolidate documentation directories Merged claudedocs/ into docs/research/ for consistent documentation structure. Changes: - Moved all claudedocs/*.md files to docs/research/ - Updated all path references in documentation (EN/KR) - Updated RULES.md and research.md command templates - Removed claudedocs/ directory - Removed ClaudeDocs/ from .gitignore Benefits: - Single source of truth for all research reports - PEP8-compliant lowercase directory naming - Clearer documentation organization - Prevents future claudedocs/ directory creation 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * perf: reduce /sc:pm command output from 1652 to 15 lines - Remove 1637 lines of documentation from command file - Keep only minimal bootstrap message - 99% token reduction on command execution - Detailed specs remain in superclaude/agents/pm-agent.md 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * perf: split PM Agent into execution workflows and guide - Reduce pm-agent.md from 735 to 429 lines (42% reduction) - Move philosophy/examples to docs/agents/pm-agent-guide.md - Execution workflows (PDCA, file ops) stay in pm-agent.md - Guide (examples, quality standards) read once when needed Token savings: - Agent loading: ~6K → ~3.5K tokens (42% reduction) - Total with pm.md: 71% overall reduction 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * refactor: consolidate PM Agent optimization and pending changes PM Agent optimization (already committed separately): - superclaude/commands/pm.md: 1652→14 lines - superclaude/agents/pm-agent.md: 735→429 lines - docs/agents/pm-agent-guide.md: new guide file Other pending changes: - setup: framework_docs, mcp, logger, remove ui.py - superclaude: __main__, cli/app, cli/commands/install - tests: test_ui updates - scripts: workflow metrics analysis tools - docs/memory: session state updates 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * refactor: simplify MCP installer to unified gateway with legacy mode ## Changes ### MCP Component (setup/components/mcp.py) - Simplified to single airis-mcp-gateway by default - Added legacy mode for individual official servers (sequential-thinking, context7, magic, playwright) - Dynamic prerequisites based on mode: - Default: uv + claude CLI only - Legacy: node (18+) + npm + claude CLI - Removed redundant server definitions ### CLI Integration - Added --legacy flag to setup/cli/commands/install.py - Added --legacy flag to superclaude/cli/commands/install.py - Config passes legacy_mode to component installer ## Benefits - ✅ Simpler: 1 gateway vs 9+ individual servers - ✅ Lighter: No Node.js/npm required (default mode) - ✅ Unified: All tools in one gateway (sequential-thinking, context7, magic, playwright, serena, morphllm, tavily, chrome-devtools, git, puppeteer) - ✅ Flexible: --legacy flag for official servers if needed ## Usage ```bash superclaude install # Default: airis-mcp-gateway (推奨) superclaude install --legacy # Legacy: individual official servers ``` 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * refactor: rename CoreComponent to FrameworkDocsComponent and add PM token tracking ## Changes ### Component Renaming (setup/components/) - Renamed CoreComponent → FrameworkDocsComponent for clarity - Updated all imports in __init__.py, agents.py, commands.py, mcp_docs.py, modes.py - Better reflects the actual purpose (framework documentation files) ### PM Agent Enhancement (superclaude/commands/pm.md) - Added token usage tracking instructions - PM Agent now reports: 1. Current token usage from system warnings 2. Percentage used (e.g., "27% used" for 54K/200K) 3. Status zone: 🟢 <75% | 🟡 75-85% | 🔴 >85% - Helps prevent token exhaustion during long sessions ### UI Utilities (setup/utils/ui.py) - Added new UI utility module for installer - Provides consistent user interface components ## Benefits - ✅ Clearer component naming (FrameworkDocs vs Core) - ✅ PM Agent token awareness for efficiency - ✅ Better visual feedback with status zones 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * refactor(pm-agent): minimize output verbosity (471→284 lines, 40% reduction) **Problem**: PM Agent generated excessive output with redundant explanations - "System Status Report" with decorative formatting - Repeated "Common Tasks" lists user already knows - Verbose session start/end protocols - Duplicate file operations documentation **Solution**: Compress without losing functionality - Session Start: Reduced to symbol-only status (🟢 branch | nM nD | token%) - Session End: Compressed to essential actions only - File Operations: Consolidated from 2 sections to 1 line reference - Self-Improvement: 5 phases → 1 unified workflow - Output Rules: Explicit constraints to prevent Claude over-explanation **Quality Preservation**: - ✅ All core functions retained (PDCA, memory, patterns, mistakes) - ✅ PARALLEL Read/Write preserved (performance critical) - ✅ Workflow unchanged (session lifecycle intact) - ✅ Added output constraints (prevents verbose generation) **Reduction Method**: - Deleted: Explanatory text, examples, redundant sections - Retained: Action definitions, file paths, core workflows - Added: Explicit output constraints to enforce minimalism **Token Impact**: 40% reduction in agent documentation size **Before**: Verbose multi-section report with task lists **After**: Single line status: 🟢 integration | 15M 17D | 36% 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * refactor: consolidate MCP integration to unified gateway **Changes**: - Remove individual MCP server docs (superclaude/mcp/*.md) - Remove MCP server configs (superclaude/mcp/configs/*.json) - Delete MCP docs component (setup/components/mcp_docs.py) - Simplify installer (setup/core/installer.py) - Update components for unified gateway approach **Rationale**: - Unified gateway (airis-mcp-gateway) provides all MCP servers - Individual docs/configs no longer needed (managed centrally) - Reduces maintenance burden and file count - Simplifies installation process **Files Removed**: 17 MCP files (docs + configs) **Installer Changes**: Removed legacy MCP installation logic 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> * chore: update version and component metadata - Bump version (pyproject.toml, setup/__init__.py) - Update CLAUDE.md import service references - Reflect component structure changes 🤖 Generated with [Claude Code](https://claude.com/claude-code) Co-Authored-By: Claude <noreply@anthropic.com> --------- Co-authored-by: kazuki <kazuki@kazukinoMacBook-Air.local> Co-authored-by: Claude <noreply@anthropic.com>
402 lines
10 KiB
Markdown
402 lines
10 KiB
Markdown
# Workflow Metrics Schema
|
|
|
|
**Purpose**: Token efficiency tracking for continuous optimization and A/B testing
|
|
|
|
**File**: `docs/memory/workflow_metrics.jsonl` (append-only log)
|
|
|
|
## Data Structure (JSONL Format)
|
|
|
|
Each line is a complete JSON object representing one workflow execution.
|
|
|
|
```jsonl
|
|
{
|
|
"timestamp": "2025-10-17T01:54:21+09:00",
|
|
"session_id": "abc123def456",
|
|
"task_type": "typo_fix",
|
|
"complexity": "light",
|
|
"workflow_id": "progressive_v3_layer2",
|
|
"layers_used": [0, 1, 2],
|
|
"tokens_used": 650,
|
|
"time_ms": 1800,
|
|
"files_read": 1,
|
|
"mindbase_used": false,
|
|
"sub_agents": [],
|
|
"success": true,
|
|
"user_feedback": "satisfied",
|
|
"notes": "Optional implementation notes"
|
|
}
|
|
```
|
|
|
|
## Field Definitions
|
|
|
|
### Required Fields
|
|
|
|
| Field | Type | Description | Example |
|
|
|-------|------|-------------|---------|
|
|
| `timestamp` | ISO 8601 | Execution timestamp in JST | `"2025-10-17T01:54:21+09:00"` |
|
|
| `session_id` | string | Unique session identifier | `"abc123def456"` |
|
|
| `task_type` | string | Task classification | `"typo_fix"`, `"bug_fix"`, `"feature_impl"` |
|
|
| `complexity` | string | Intent classification level | `"ultra-light"`, `"light"`, `"medium"`, `"heavy"`, `"ultra-heavy"` |
|
|
| `workflow_id` | string | Workflow variant identifier | `"progressive_v3_layer2"` |
|
|
| `layers_used` | array | Progressive loading layers executed | `[0, 1, 2]` |
|
|
| `tokens_used` | integer | Total tokens consumed | `650` |
|
|
| `time_ms` | integer | Execution time in milliseconds | `1800` |
|
|
| `success` | boolean | Task completion status | `true`, `false` |
|
|
|
|
### Optional Fields
|
|
|
|
| Field | Type | Description | Example |
|
|
|-------|------|-------------|---------|
|
|
| `files_read` | integer | Number of files read | `1` |
|
|
| `mindbase_used` | boolean | Whether mindbase MCP was used | `false` |
|
|
| `sub_agents` | array | Delegated sub-agents | `["backend-architect", "quality-engineer"]` |
|
|
| `user_feedback` | string | Inferred user satisfaction | `"satisfied"`, `"neutral"`, `"unsatisfied"` |
|
|
| `notes` | string | Implementation notes | `"Used cached solution"` |
|
|
| `confidence_score` | float | Pre-implementation confidence | `0.85` |
|
|
| `hallucination_detected` | boolean | Self-check red flags found | `false` |
|
|
| `error_recurrence` | boolean | Same error encountered before | `false` |
|
|
|
|
## Task Type Taxonomy
|
|
|
|
### Ultra-Light Tasks
|
|
- `progress_query`: "進捗教えて"
|
|
- `status_check`: "現状確認"
|
|
- `next_action_query`: "次のタスクは?"
|
|
|
|
### Light Tasks
|
|
- `typo_fix`: README誤字修正
|
|
- `comment_addition`: コメント追加
|
|
- `variable_rename`: 変数名変更
|
|
- `documentation_update`: ドキュメント更新
|
|
|
|
### Medium Tasks
|
|
- `bug_fix`: バグ修正
|
|
- `small_feature`: 小機能追加
|
|
- `refactoring`: リファクタリング
|
|
- `test_addition`: テスト追加
|
|
|
|
### Heavy Tasks
|
|
- `feature_impl`: 新機能実装
|
|
- `architecture_change`: アーキテクチャ変更
|
|
- `security_audit`: セキュリティ監査
|
|
- `integration`: 外部システム統合
|
|
|
|
### Ultra-Heavy Tasks
|
|
- `system_redesign`: システム全面再設計
|
|
- `framework_migration`: フレームワーク移行
|
|
- `comprehensive_research`: 包括的調査
|
|
|
|
## Workflow Variant Identifiers
|
|
|
|
### Progressive Loading Variants
|
|
- `progressive_v3_layer1`: Ultra-light (memory files only)
|
|
- `progressive_v3_layer2`: Light (target file only)
|
|
- `progressive_v3_layer3`: Medium (related files 3-5)
|
|
- `progressive_v3_layer4`: Heavy (subsystem)
|
|
- `progressive_v3_layer5`: Ultra-heavy (full + external research)
|
|
|
|
### Experimental Variants (A/B Testing)
|
|
- `experimental_eager_layer3`: Always load Layer 3 for medium tasks
|
|
- `experimental_lazy_layer2`: Minimal Layer 2 loading
|
|
- `experimental_parallel_layer3`: Parallel file loading in Layer 3
|
|
|
|
## Complexity Classification Rules
|
|
|
|
```yaml
|
|
ultra_light:
|
|
keywords: ["進捗", "状況", "進み", "where", "status", "progress"]
|
|
token_budget: "100-500"
|
|
layers: [0, 1]
|
|
|
|
light:
|
|
keywords: ["誤字", "typo", "fix typo", "correct", "comment"]
|
|
token_budget: "500-2K"
|
|
layers: [0, 1, 2]
|
|
|
|
medium:
|
|
keywords: ["バグ", "bug", "fix", "修正", "error", "issue"]
|
|
token_budget: "2-5K"
|
|
layers: [0, 1, 2, 3]
|
|
|
|
heavy:
|
|
keywords: ["新機能", "new feature", "implement", "実装"]
|
|
token_budget: "5-20K"
|
|
layers: [0, 1, 2, 3, 4]
|
|
|
|
ultra_heavy:
|
|
keywords: ["再設計", "redesign", "overhaul", "migration"]
|
|
token_budget: "20K+"
|
|
layers: [0, 1, 2, 3, 4, 5]
|
|
```
|
|
|
|
## Recording Points
|
|
|
|
### Session Start (Layer 0)
|
|
```python
|
|
session_id = generate_session_id()
|
|
workflow_metrics = {
|
|
"timestamp": get_current_time(),
|
|
"session_id": session_id,
|
|
"workflow_id": "progressive_v3_layer0"
|
|
}
|
|
# Bootstrap: 150 tokens
|
|
```
|
|
|
|
### After Intent Classification (Layer 1)
|
|
```python
|
|
workflow_metrics.update({
|
|
"task_type": classify_task_type(user_request),
|
|
"complexity": classify_complexity(user_request),
|
|
"estimated_token_budget": get_budget(complexity)
|
|
})
|
|
```
|
|
|
|
### After Progressive Loading
|
|
```python
|
|
workflow_metrics.update({
|
|
"layers_used": [0, 1, 2], # Actual layers executed
|
|
"tokens_used": calculate_tokens(),
|
|
"files_read": len(files_loaded)
|
|
})
|
|
```
|
|
|
|
### After Task Completion
|
|
```python
|
|
workflow_metrics.update({
|
|
"success": task_completed_successfully,
|
|
"time_ms": execution_time_ms,
|
|
"user_feedback": infer_user_satisfaction()
|
|
})
|
|
```
|
|
|
|
### Session End
|
|
```python
|
|
# Append to workflow_metrics.jsonl
|
|
with open("docs/memory/workflow_metrics.jsonl", "a") as f:
|
|
f.write(json.dumps(workflow_metrics) + "\n")
|
|
```
|
|
|
|
## Analysis Scripts
|
|
|
|
### Weekly Analysis
|
|
```bash
|
|
# Group by task type and calculate averages
|
|
python scripts/analyze_workflow_metrics.py --period week
|
|
|
|
# Output:
|
|
# Task Type: typo_fix
|
|
# Count: 12
|
|
# Avg Tokens: 680
|
|
# Avg Time: 1,850ms
|
|
# Success Rate: 100%
|
|
```
|
|
|
|
### A/B Testing Analysis
|
|
```bash
|
|
# Compare workflow variants
|
|
python scripts/ab_test_workflows.py \
|
|
--variant-a progressive_v3_layer2 \
|
|
--variant-b experimental_eager_layer3 \
|
|
--metric tokens_used
|
|
|
|
# Output:
|
|
# Variant A (progressive_v3_layer2):
|
|
# Avg Tokens: 1,250
|
|
# Success Rate: 95%
|
|
#
|
|
# Variant B (experimental_eager_layer3):
|
|
# Avg Tokens: 2,100
|
|
# Success Rate: 98%
|
|
#
|
|
# Statistical Significance: p = 0.03 (significant)
|
|
# Recommendation: Keep Variant A (better efficiency)
|
|
```
|
|
|
|
## Usage (Continuous Optimization)
|
|
|
|
### Weekly Review Process
|
|
```yaml
|
|
every_monday_morning:
|
|
1. Run analysis: python scripts/analyze_workflow_metrics.py --period week
|
|
2. Identify patterns:
|
|
- Best-performing workflows per task type
|
|
- Inefficient patterns (high tokens, low success)
|
|
- User satisfaction trends
|
|
3. Update recommendations:
|
|
- Promote efficient workflows to standard
|
|
- Deprecate inefficient workflows
|
|
- Design new experimental variants
|
|
```
|
|
|
|
### A/B Testing Framework
|
|
```yaml
|
|
allocation_strategy:
|
|
current_best: 80% # Use best-known workflow
|
|
experimental: 20% # Test new variant
|
|
|
|
evaluation_criteria:
|
|
minimum_trials: 20 # Per variant
|
|
confidence_level: 0.95 # p < 0.05
|
|
metrics:
|
|
- tokens_used (primary)
|
|
- success_rate (gate: must be ≥95%)
|
|
- user_feedback (qualitative)
|
|
|
|
promotion_rules:
|
|
if experimental_better:
|
|
- Statistical significance confirmed
|
|
- Success rate ≥ current_best
|
|
- User feedback ≥ neutral
|
|
→ Promote to standard (80% allocation)
|
|
|
|
if experimental_worse:
|
|
→ Deprecate variant
|
|
→ Document learning in docs/patterns/
|
|
```
|
|
|
|
### Auto-Optimization Cycle
|
|
```yaml
|
|
monthly_cleanup:
|
|
1. Identify stale workflows:
|
|
- No usage in last 90 days
|
|
- Success rate <80%
|
|
- User feedback consistently negative
|
|
|
|
2. Archive deprecated workflows:
|
|
- Move to docs/patterns/deprecated/
|
|
- Document why deprecated
|
|
|
|
3. Promote new standards:
|
|
- Experimental → Standard (if proven better)
|
|
- Update pm.md with new best practices
|
|
|
|
4. Generate monthly report:
|
|
- Token efficiency trends
|
|
- Success rate improvements
|
|
- User satisfaction evolution
|
|
```
|
|
|
|
## Visualization
|
|
|
|
### Token Usage Over Time
|
|
```python
|
|
import pandas as pd
|
|
import matplotlib.pyplot as plt
|
|
|
|
df = pd.read_json("docs/memory/workflow_metrics.jsonl", lines=True)
|
|
df['date'] = pd.to_datetime(df['timestamp']).dt.date
|
|
|
|
daily_avg = df.groupby('date')['tokens_used'].mean()
|
|
plt.plot(daily_avg)
|
|
plt.title("Average Token Usage Over Time")
|
|
plt.ylabel("Tokens")
|
|
plt.xlabel("Date")
|
|
plt.show()
|
|
```
|
|
|
|
### Task Type Distribution
|
|
```python
|
|
task_counts = df['task_type'].value_counts()
|
|
plt.pie(task_counts, labels=task_counts.index, autopct='%1.1f%%')
|
|
plt.title("Task Type Distribution")
|
|
plt.show()
|
|
```
|
|
|
|
### Workflow Efficiency Comparison
|
|
```python
|
|
workflow_efficiency = df.groupby('workflow_id').agg({
|
|
'tokens_used': 'mean',
|
|
'success': 'mean',
|
|
'time_ms': 'mean'
|
|
})
|
|
print(workflow_efficiency.sort_values('tokens_used'))
|
|
```
|
|
|
|
## Expected Patterns
|
|
|
|
### Healthy Metrics (After 1 Month)
|
|
```yaml
|
|
token_efficiency:
|
|
ultra_light: 750-1,050 tokens (63% reduction)
|
|
light: 1,250 tokens (46% reduction)
|
|
medium: 3,850 tokens (47% reduction)
|
|
heavy: 10,350 tokens (40% reduction)
|
|
|
|
success_rates:
|
|
all_tasks: ≥95%
|
|
ultra_light: 100% (simple tasks)
|
|
light: 98%
|
|
medium: 95%
|
|
heavy: 92%
|
|
|
|
user_satisfaction:
|
|
satisfied: ≥70%
|
|
neutral: ≤25%
|
|
unsatisfied: ≤5%
|
|
```
|
|
|
|
### Red Flags (Require Investigation)
|
|
```yaml
|
|
warning_signs:
|
|
- success_rate < 85% for any task type
|
|
- tokens_used > estimated_budget by >30%
|
|
- time_ms > 10 seconds for light tasks
|
|
- user_feedback "unsatisfied" > 10%
|
|
- error_recurrence > 15%
|
|
```
|
|
|
|
## Integration with PM Agent
|
|
|
|
### Automatic Recording
|
|
PM Agent automatically records metrics at each execution point:
|
|
- Session start (Layer 0)
|
|
- Intent classification (Layer 1)
|
|
- Progressive loading (Layers 2-5)
|
|
- Task completion
|
|
- Session end
|
|
|
|
### No Manual Intervention
|
|
- All recording is automatic
|
|
- No user action required
|
|
- Transparent operation
|
|
- Privacy-preserving (local files only)
|
|
|
|
## Privacy and Security
|
|
|
|
### Data Retention
|
|
- Local storage only (`docs/memory/`)
|
|
- No external transmission
|
|
- Git-manageable (optional)
|
|
- User controls retention period
|
|
|
|
### Sensitive Data Handling
|
|
- No code snippets logged
|
|
- No user input content
|
|
- Only metadata (tokens, timing, success)
|
|
- Task types are generic classifications
|
|
|
|
## Maintenance
|
|
|
|
### File Rotation
|
|
```bash
|
|
# Archive old metrics (monthly)
|
|
mv docs/memory/workflow_metrics.jsonl \
|
|
docs/memory/archive/workflow_metrics_2025-10.jsonl
|
|
|
|
# Start fresh
|
|
touch docs/memory/workflow_metrics.jsonl
|
|
```
|
|
|
|
### Cleanup
|
|
```bash
|
|
# Remove metrics older than 6 months
|
|
find docs/memory/archive/ -name "workflow_metrics_*.jsonl" \
|
|
-mtime +180 -delete
|
|
```
|
|
|
|
## References
|
|
|
|
- Specification: `superclaude/commands/pm.md` (Line 291-355)
|
|
- Research: `docs/research/llm-agent-token-efficiency-2025.md`
|
|
- Tests: `tests/pm_agent/test_token_budget.py`
|