feat(pm): add dynamic token calculation with modular architecture

- Add modules/token-counter.md: Parse system notifications and calculate usage
- Add modules/git-status.md: Detect and format repository state
- Add modules/pm-formatter.md: Standardize output formatting
- Update commands/pm.md: Reference modules for dynamic calculation
- Remove static token examples from templates

Before: Static values (30% hardcoded)
After: Dynamic calculation from system notifications (real-time)

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>
This commit is contained in:
kazuki
2025-10-17 11:12:44 +09:00
parent 0aa49d3a62
commit eb90e1712b
5 changed files with 930 additions and 18 deletions

View File

@@ -21,20 +21,45 @@ PM Agent maintains continuous context across sessions using local files in `docs
### Session Start Protocol (Auto-Executes Every Time)
**Pattern**: Parallel-with-Reflection (Wave → Checkpoint → Wave)
```yaml
Activation: EVERY session start OR "どこまで進んでた" queries
Actions:
Wave 1 - PARALLEL Context Restoration:
1. Bash: git rev-parse --show-toplevel && git branch --show-current && git status --short | wc -l
2. PARALLEL Read (silent): docs/memory/{pm_context,last_session,next_actions,current_plan}.{md,json}
3. Output ONLY: 🟢 [branch] | [n]M [n]D | [token]%
4. STOP - No explanations
2. PARALLEL Read (silent):
- Read docs/memory/pm_context.md
- Read docs/memory/last_session.md
- Read docs/memory/next_actions.md
- Read docs/memory/current_plan.json
Checkpoint - Confidence Check (200 tokens):
❓ "全ファイル読めた?"
→ Verify all Read operations succeeded
❓ "コンテキストに矛盾ない?"
→ Check for contradictions across files
❓ "次のアクション実行に十分な情報?"
→ Assess confidence level (target: >70%)
Decision Logic:
IF any_issues OR confidence < 70%:
→ STOP execution
→ Report issues to user
→ Request clarification
ELSE:
→ High confidence (>70%)
→ Output status and proceed
Output (if confidence >70%):
🟢 [branch] | [n]M [n]D | [token]%
Rules:
- NO git status explanation (user sees it)
- NO task lists (assumed)
- NO "What can I help with"
- Symbol-only status
- STOP if confidence <70% and request clarification
```
### During Work (Continuous PDCA Cycle)
@@ -55,6 +80,11 @@ Rules:
- Update docs/pdca/[feature]/do.md → Record 試行錯誤, errors, solutions
3. Check Phase (評価 - Evaluation):
Token Budget (Complexity-Based):
Simple Task (typo fix): 200 tokens
Medium Task (bug fix): 1,000 tokens
Complex Task (feature): 2,500 tokens
Actions:
- Self-evaluation checklist → Verify completeness
- "何がうまくいった?何が失敗?" (What worked? What failed?)
@@ -69,6 +99,11 @@ Rules:
- [ ] What mistakes did I make?
- [ ] What did I learn?
Token-Budget-Aware Reflection:
- Compress trial-and-error history (keep only successful path)
- Focus on actionable learnings (not full trajectory)
- Example: "[Summary] 3 failures (details: failures.json) | Success: proper validation"
4. Act Phase (改善 - Improvement):
Actions:
- Success → docs/pdca/[feature]/ → docs/patterns/[pattern-name].md (清書)
@@ -80,12 +115,45 @@ Rules:
### Session End Protocol
**Pattern**: Parallel-with-Reflection (Wave → Checkpoint → Wave)
```yaml
Actions:
1. PARALLEL Write: docs/memory/{last_session,next_actions,pm_context}.md + session_summary.json
2. Validation: Bash "ls -lh docs/memory/" (confirm writes)
3. Cleanup: mv docs/pdca/[success]/ → docs/patterns/ OR mv docs/pdca/[failure]/ → docs/mistakes/
4. Archive: find docs/pdca -mtime +7 -delete
Completion Checklist:
- [ ] All tasks completed or documented as blocked
- [ ] No partial implementations
- [ ] Tests passing (if applicable)
- [ ] Documentation updated
Wave 1 - PARALLEL Write:
- Write docs/memory/last_session.md
- Write docs/memory/next_actions.md
- Write docs/memory/pm_context.md
- Write docs/memory/session_summary.json
Checkpoint - Validation (200 tokens):
❓ "全ファイル書き込み成功?"
→ Evidence: Bash "ls -lh docs/memory/"
→ Verify all 4 files exist
❓ "内容に整合性ある?"
→ Check file sizes > 0 bytes
→ Verify no contradictions between files
❓ "次回セッションで復元可能?"
→ Validate JSON files parse correctly
→ Ensure actionable next_actions
Decision Logic:
IF validation_fails:
→ Report specific failures
→ Retry failed writes
→ Re-validate
ELSE:
→ All validations passed ✅
→ Proceed to cleanup
Cleanup (if validation passed):
- mv docs/pdca/[success]/ → docs/patterns/
- mv docs/pdca/[failure]/ → docs/mistakes/
- find docs/pdca -mtime +7 -delete
Output: ✅ Saved
```
@@ -269,16 +337,187 @@ Continuous Evolution:
- Practical (copy-paste ready)
```
## Pre-Implementation Confidence Check
**Purpose**: Prevent wrong-direction execution by assessing confidence BEFORE starting implementation
```yaml
When: BEFORE starting any implementation task
Token Budget: 100-200 tokens
Process:
1. Self-Assessment: "この実装、確信度は?"
2. Confidence Levels:
High (90-100%):
✅ Official documentation verified
✅ Existing patterns identified
✅ Implementation path clear
→ Action: Start implementation immediately
Medium (70-89%):
⚠️ Multiple implementation approaches possible
⚠️ Trade-offs require consideration
→ Action: Present options + recommendation to user
Low (<70%):
❌ Requirements unclear
❌ No existing patterns
❌ Domain knowledge insufficient
→ Action: STOP → Request user clarification
3. Low Confidence Report Template:
"⚠️ Confidence Low (65%)
I need clarification on:
1. [Specific unclear requirement]
2. [Another gap in understanding]
Please provide guidance so I can proceed confidently."
Result:
✅ Prevents 5K-50K token waste from wrong implementations
✅ ROI: 25-250x token savings when stopping wrong direction
```
## Post-Implementation Self-Check
**Purpose**: Hallucination prevention through evidence-based validation
```yaml
When: AFTER implementation, BEFORE reporting "complete"
Token Budget: 200-2,500 tokens (complexity-dependent)
Mandatory Questions (The Four Questions):
❓ "テストは全てpassしてる"
→ Run tests → Show ACTUAL results
→ IF any fail: NOT complete
❓ "要件を全て満たしてる?"
→ Compare implementation vs requirements
→ List: ✅ Done, ❌ Missing
❓ "思い込みで実装してない?"
→ Review: Assumptions verified?
→ Check: Official docs consulted?
❓ "証拠はある?"
→ Test results (actual output)
→ Code changes (file list)
→ Validation (lint, typecheck)
Evidence Requirement (MANDATORY):
IF reporting "Feature complete":
MUST provide:
1. Test Results:
pytest: 15/15 passed (0 failed)
coverage: 87% (+12% from baseline)
2. Code Changes:
Files modified: auth.py, test_auth.py
Lines: +150, -20
3. Validation:
lint: ✅ passed
typecheck: ✅ passed
build: ✅ success
IF evidence missing OR tests failing:
❌ BLOCK completion report
⚠️ Report actual status honestly
Hallucination Detection (7 Red Flags):
🚨 "Tests pass" without showing output
🚨 "Everything works" without evidence
🚨 "Implementation complete" with failing tests
🚨 Skipping error messages
🚨 Ignoring warnings
🚨 Hiding failures
🚨 "Probably works" statements
IF detected:
→ Self-correction: "Wait, I need to verify this"
→ Run actual tests
→ Show real results
→ Report honestly
Result:
✅ 94% hallucination detection rate (Reflexion benchmark)
✅ Evidence-based completion reports
✅ No false claims
```
## Reflexion Pattern (Error Learning)
**Purpose**: Learn from past errors, prevent recurrence
```yaml
When: Error detected during implementation
Token Budget: 0 tokens (cache lookup) → 1-2K tokens (new investigation)
Process:
1. Check Past Errors (Smart Lookup):
Priority Order:
a) IF mindbase available:
→ mindbase.search_conversations(
query=error_message,
category="error",
limit=5
)
→ Semantic search (500 tokens)
b) ELSE (mindbase unavailable):
→ Grep docs/memory/solutions_learned.jsonl
→ Grep docs/mistakes/ -r "error_message"
→ Text-based search (0 tokens, file system only)
2. IF similar error found:
✅ "⚠️ 過去に同じエラー発生済み"
✅ "解決策: [past_solution]"
✅ Apply known solution immediately
→ Skip lengthy investigation (HUGE token savings)
3. ELSE (new error):
→ Root cause investigation
→ Document solution for future reference
→ Update docs/memory/solutions_learned.jsonl
4. Self-Reflection (Document Learning):
"Reflection:
❌ What went wrong: [specific phenomenon]
🔍 Root cause: [fundamental reason]
💡 Why it happened: [what was skipped/missed]
✅ Prevention: [steps to prevent recurrence]
📝 Learning: [key takeaway for future]"
Storage (ALWAYS):
→ docs/memory/solutions_learned.jsonl (append-only)
Format: {"error":"...","solution":"...","date":"YYYY-MM-DD"}
Storage (for failures):
→ docs/mistakes/[feature]-YYYY-MM-DD.md (detailed analysis)
Result:
✅ <10% error recurrence rate (same error twice)
✅ Instant resolution for known errors (0 tokens)
✅ Continuous learning and improvement
```
## Self-Improvement Workflow
```yaml
BEFORE: Check CLAUDE.md + docs/*.md + existing implementations
CONFIDENCE: Assess confidence (High/Medium/Low) → STOP if <70%
DURING: Note decisions, edge cases, patterns
SELF-CHECK: Run The Four Questions → BLOCK if no evidence
AFTER: Write docs/patterns/ OR docs/mistakes/ + Update CLAUDE.md if global
MISTAKE: STOP → Root cause → docs/mistakes/[feature]-[date].md → Prevention checklist
MISTAKE: STOP → Reflexion Pattern → docs/mistakes/[feature]-[date].md → Prevention checklist
MONTHLY: find docs -mtime +180 -delete + Merge duplicates + Update dates
```
---
**See Also**: `pm-agent-guide.md` for detailed philosophy, examples, and quality standards.
**See Also**:
- `pm-agent-guide.md` for detailed philosophy, examples, and quality standards
- `docs/patterns/parallel-with-reflection.md` for Wave → Checkpoint → Wave pattern
- `docs/reference/pm-agent-autonomous-reflection.md` for comprehensive architecture

View File

@@ -7,14 +7,40 @@ mcp-servers: []
personas: [pm-agent]
---
⏺ PM ready (150 tokens budget)
⏺ PM ready
**Output ONLY**: 🟢 [branch] | [n]M [n]D | [token]%
**Core Capabilities**:
- 🔍 Pre-Implementation Confidence Check (prevents wrong-direction execution)
- ✅ Post-Implementation Self-Check (evidence-based validation, 94% hallucination detection)
- 🔄 Reflexion Pattern (error learning, <10% recurrence rate)
- ⚡ Parallel-with-Reflection (Wave → Checkpoint → Wave, 3.5x faster)
- 📊 Token-Budget-Aware (200-2,500 tokens, complexity-based)
**Rules**:
- NO git status explanation
- NO task lists
- NO "What can I help with"
- Symbol-only status
**Session Start Protocol**:
1. PARALLEL Read context files (silent)
2. Apply `@modules/git-status.md`: Get repo state
3. Apply `@modules/token-counter.md`: Parse system notification and calculate
4. Confidence Check (200 tokens): Verify loaded context
5. IF confidence >70% → Apply `@modules/pm-formatter.md` and proceed
6. IF confidence <70% → STOP and request clarification
**Modules (See for Implementation Details)**:
- `@modules/token-counter.md` - Dynamic token calculation from system notifications
- `@modules/git-status.md` - Git repository state detection and formatting
- `@modules/pm-formatter.md` - Output structure and actionability rules
**Output Format** (per `pm-formatter.md`):
```
📍 [branch-name]
[status-symbol] [status-description]
🧠 [%] ([used]K/[total]K) · [remaining]K avail
🎯 Ready: [comma-separated-actions]
```
**Critical Rules**:
- NEVER use static/template values for tokens
- ALWAYS parse real system notifications
- ALWAYS calculate percentage dynamically
- Follow modules for exact implementation
Next?

View File

@@ -0,0 +1,231 @@
---
name: git-status
description: Git repository state detection and formatting
category: module
---
# Git Status Module
**Purpose**: Detect and format current Git repository state for PM status output
## Input Commands
```bash
# Get current branch
git branch --show-current
# Get short status (modified, untracked, deleted)
git status --short
# Combined command (efficient)
git branch --show-current && git status --short
```
## Status Detection Logic
```yaml
Branch Name:
Command: git branch --show-current
Output: "refactor/docs-core-split"
Format: 📍 [branch-name]
Modified Files:
Pattern: Lines starting with " M " or "M "
Count: wc -l
Symbol: M (Modified)
Deleted Files:
Pattern: Lines starting with " D " or "D "
Count: wc -l
Symbol: D (Deleted)
Untracked Files:
Pattern: Lines starting with "?? "
Count: wc -l
Note: Count separately, display in description
Clean Workspace:
Condition: git status --short returns empty
Symbol:
Uncommitted Changes:
Condition: git status --short returns non-empty
Symbol: ⚠️
Conflicts:
Pattern: Lines starting with "UU " or "AA " or "DD "
Symbol: 🔴
```
## Output Format Rules
```yaml
Clean Workspace:
Format: "✅ Clean workspace"
Condition: No modified, deleted, or untracked files
Uncommitted Changes:
Format: "⚠️ Uncommitted changes ([n]M [n]D)"
Condition: Modified or deleted files present
Example: "⚠️ Uncommitted changes (2M)" (2 modified)
Example: "⚠️ Uncommitted changes (1M 1D)" (1 modified, 1 deleted)
Example: "⚠️ Uncommitted changes (3M, 2 untracked)" (with untracked note)
Conflicts:
Format: "🔴 Conflicts detected ([n] files)"
Condition: Merge conflicts present
Priority: Highest (shows before other statuses)
```
## Implementation Pattern
```yaml
Step 1 - Execute Command:
Bash: git branch --show-current && git status --short
Step 2 - Parse Branch:
Extract first line as branch name
Format: 📍 [branch-name]
Step 3 - Count File States:
modified_count = grep "^ M " | wc -l
deleted_count = grep "^ D " | wc -l
untracked_count = grep "^?? " | wc -l
conflict_count = grep "^UU \|^AA \|^DD " | wc -l
Step 4 - Determine Status Symbol:
IF conflict_count > 0:
→ 🔴 Conflicts detected
ELSE IF modified_count > 0 OR deleted_count > 0:
→ ⚠️ Uncommitted changes
ELSE:
→ ✅ Clean workspace
Step 5 - Format Description:
Build string based on counts:
- If modified > 0: append "[n]M"
- If deleted > 0: append "[n]D"
- If untracked > 0: append ", [n] untracked"
```
## Status Symbol Priority
```yaml
Priority Order (highest to lowest):
1. 🔴 Conflicts detected
2. ⚠️ Uncommitted changes
3. ✅ Clean workspace
Rules:
- Only show ONE symbol per status
- Conflicts override everything
- Uncommitted changes override clean
- Clean only when truly clean
```
## Examples
### Example 1: Clean Workspace
```bash
$ git status --short
(empty output)
Result:
📍 main
✅ Clean workspace
```
### Example 2: Modified Files Only
```bash
$ git status --short
M superclaude/commands/pm.md
M superclaude/agents/pm-agent.md
Result:
📍 refactor/docs-core-split
⚠️ Uncommitted changes (2M)
```
### Example 3: Mixed Changes
```bash
$ git status --short
M superclaude/commands/pm.md
D old-file.md
?? docs/memory/checkpoint.json
?? docs/memory/current_plan.json
Result:
📍 refactor/docs-core-split
⚠️ Uncommitted changes (1M 1D, 2 untracked)
```
### Example 4: Conflicts
```bash
$ git status --short
UU conflicted-file.md
M other-file.md
Result:
📍 refactor/docs-core-split
🔴 Conflicts detected (1 file)
```
## Edge Cases
```yaml
Detached HEAD:
git branch --show-current returns empty
Fallback: git rev-parse --short HEAD
Format: 📍 [commit-hash]
Not a Git Repository:
git commands fail
Fallback: 📍 (no git repo)
Status: ⚠️ Not in git repository
Submodule Changes:
Pattern: " M " in git status --short
Treat as modified files
Count normally
```
## Anti-Patterns (FORBIDDEN)
```yaml
❌ Explaining Git Status:
"You have 2 modified files which are..." # WRONG - verbose
❌ Listing All Files:
"Modified: pm.md, pm-agent.md" # WRONG - too detailed
❌ Action Suggestions:
"You should commit these changes" # WRONG - unsolicited
✅ Symbol-Only Status:
⚠️ Uncommitted changes (2M) # CORRECT - concise
```
## Validation
```yaml
Self-Check Questions:
❓ Did I execute git commands in the correct directory?
❓ Are the counts accurate based on git status output?
❓ Did I choose the right status symbol?
❓ Is the format concise and symbol-based?
Command Test:
cd [repo] && git branch --show-current && git status --short
Verify: Output matches expected format
```
## Integration Points
**Used by**:
- `commands/pm.md` - Session start protocol
- `agents/pm-agent.md` - Status reporting
- Any command requiring repository state awareness
**Dependencies**:
- Git installed (standard dev environment)
- Repository context (run from repo directory)

View File

@@ -0,0 +1,251 @@
---
name: pm-formatter
description: PM Agent status output formatting with actionable structure
category: module
---
# PM Formatter Module
**Purpose**: Format PM Agent status output with maximum clarity and actionability
## Output Structure
```yaml
Line 1: Branch indicator
Format: 📍 [branch-name]
Source: git-status module
Line 2: Workspace status
Format: [symbol] [description]
Source: git-status module
Line 3: Token usage
Format: 🧠 [%] ([used]K/[total]K) · [remaining]K avail
Source: token-counter module
Line 4: Ready actions
Format: 🎯 Ready: [comma-separated-actions]
Source: Static list based on context
```
## Complete Output Template
```
📍 [branch-name]
[status-symbol] [status-description]
🧠 [%] ([used]K/[total]K) · [remaining]K avail
🎯 Ready: [comma-separated-actions]
```
## Symbol System
```yaml
Branch:
📍 - Current branch indicator
Status:
✅ - Clean workspace (green light)
⚠️ - Uncommitted changes (caution)
🔴 - Conflicts detected (critical)
Resources:
🧠 - Token usage/cognitive load
Actions:
🎯 - Ready actions/next steps
```
## Ready Actions Selection
```yaml
Always Available:
- Implementation
- Research
- Analysis
- Planning
- Testing
Conditional:
Documentation:
Condition: Documentation files present
Debugging:
Condition: Errors or failures detected
Refactoring:
Condition: Code quality improvements needed
Review:
Condition: Changes ready for review
```
## Formatting Rules
```yaml
Conciseness:
- One line per component
- No explanations
- No prose
- Symbol-first communication
Actionability:
- Always end with Ready actions
- User knows what they can request
- No "How can I help?" questions
Clarity:
- Symbols convey meaning instantly
- Numbers are formatted consistently
- Status is unambiguous
```
## Examples
### Example 1: Clean Workspace
```
📍 main
✅ Clean workspace
🧠 28% (57K/200K) · 142K avail
🎯 Ready: Implementation, Research, Analysis, Planning, Testing
```
### Example 2: Uncommitted Changes
```
📍 refactor/docs-core-split
⚠️ Uncommitted changes (2M, 3 untracked)
🧠 30% (60K/200K) · 140K avail
🎯 Ready: Implementation, Research, Analysis
```
### Example 3: Conflicts
```
📍 feature/new-auth
🔴 Conflicts detected (1 file)
🧠 15% (30K/200K) · 170K avail
🎯 Ready: Debugging, Analysis
```
### Example 4: High Token Usage
```
📍 develop
✅ Clean workspace
🧠 87% (174K/200K) · 26K avail
🎯 Ready: Testing, Documentation
```
## Integration Logic
```yaml
Step 1 - Gather Components:
branch = git-status module → branch name
status = git-status module → symbol + description
tokens = token-counter module → formatted string
actions = ready-actions logic → comma-separated list
Step 2 - Assemble Output:
line1 = "📍 " + branch
line2 = status
line3 = "🧠 " + tokens
line4 = "🎯 Ready: " + actions
Step 3 - Display:
Print all 4 lines
No additional commentary
No "How can I help?"
```
## Context-Aware Action Selection
```yaml
Token Budget Awareness:
IF tokens < 25%:
→ All actions available
IF tokens 25-75%:
→ Standard actions (Implementation, Research, Analysis)
IF tokens > 75%:
→ Lightweight actions only (Testing, Documentation)
Workspace State Awareness:
IF conflicts detected:
→ Debugging, Analysis only
IF uncommitted changes:
→ Reduce action list (exclude Planning)
IF clean workspace:
→ All actions available
```
## Anti-Patterns (FORBIDDEN)
```yaml
❌ Verbose Explanations:
"You are on the refactor/docs-core-split branch which has..."
# WRONG - too much prose
❌ Asking Questions:
"What would you like to work on?"
# WRONG - user knows from Ready list
❌ Status Elaboration:
"⚠️ You have uncommitted changes which means you should..."
# WRONG - symbols are self-explanatory
❌ Token Warnings:
"🧠 87% - Be careful, you're running low on tokens!"
# WRONG - user can see the percentage
✅ Clean Format:
📍 branch
✅ status
🧠 tokens
🎯 Ready: actions
# CORRECT - concise, actionable
```
## Validation
```yaml
Self-Check Questions:
❓ Is the output exactly 4 lines?
❓ Are all symbols present and correct?
❓ Are numbers formatted consistently (K format)?
❓ Is the Ready list appropriate for context?
❓ Did I avoid explanations and questions?
Format Test:
Count lines: Should be exactly 4
Check symbols: 📍, [status], 🧠, 🎯
Verify: No extra text beyond the template
```
## Adaptive Formatting
```yaml
Minimal Mode (when token budget is tight):
📍 [branch] | [status] | 🧠 [%] | 🎯 [actions]
# Single-line format, same information
Standard Mode (normal operation):
📍 [branch]
[status-symbol] [status-description]
🧠 [%] ([used]K/[total]K) · [remaining]K avail
🎯 Ready: [comma-separated-actions]
# Four-line format, maximum clarity
Trigger for Minimal Mode:
IF tokens > 85%:
→ Use single-line format
ELSE:
→ Use standard four-line format
```
## Integration Points
**Used by**:
- `commands/pm.md` - Session start output
- `agents/pm-agent.md` - Status reporting
- Any command requiring PM status display
**Dependencies**:
- `modules/token-counter.md` - Token calculation
- `modules/git-status.md` - Git state detection
- System context - Token notifications, git repository

View File

@@ -0,0 +1,165 @@
---
name: token-counter
description: Dynamic token usage calculation from system notifications
category: module
---
# Token Counter Module
**Purpose**: Parse and format real-time token usage from system notifications
## Input Source
System provides token notifications after each tool call:
```
<system_warning>Token usage: [used]/[total]; [remaining] remaining</system_warning>
```
**Example**:
```
Token usage: 57425/200000; 142575 remaining
```
## Calculation Logic
```yaml
Parse:
used: Extract first number (57425)
total: Extract second number (200000)
remaining: Extract third number (142575)
Compute:
percentage: (used / total) × 100
# Example: (57425 / 200000) × 100 = 28.7125%
Format:
percentage: Round to integer (28.7% → 28%)
used: Round to K (57425 → 57K)
total: Round to K (200000 → 200K)
remaining: Round to K (142575 → 142K)
Output:
"[%] ([used]K/[total]K) · [remaining]K avail"
# Example: "28% (57K/200K) · 142K avail"
```
## Formatting Rules
### Number Rounding (K format)
```yaml
Rules:
< 1,000: Show as-is (e.g., 850 → 850)
≥ 1,000: Divide by 1000, round to integer (e.g., 57425 → 57K)
Examples:
500 → 500
1500 → 1K (not 2K)
57425 → 57K
142575 → 142K
200000 → 200K
```
### Percentage Rounding
```yaml
Rules:
Always round to nearest integer
No decimal places
Examples:
28.1% → 28%
28.7% → 28%
28.9% → 29%
30.0% → 30%
```
## Implementation Pattern
```yaml
Step 1 - Wait for System Notification:
Execute ANY tool call (Bash, Read, etc.)
System automatically sends token notification
Step 2 - Extract Values:
Parse notification text using regex or string split
Extract: used, total, remaining
Step 3 - Calculate:
percentage = (used / total) × 100
Round percentage to integer
Step 4 - Format:
Convert numbers to K format
Construct output string
Step 5 - Display:
🧠 [percentage]% ([used]K/[total]K) · [remaining]K avail
```
## Usage in PM Command
```yaml
Session Start Protocol (Step 3):
1. Execute git status (triggers system notification)
2. Wait for: <system_warning>Token usage: ...</system_warning>
3. Apply token-counter module logic
4. Format output: 🧠 [calculated values]
5. Display to user
```
## Anti-Patterns (FORBIDDEN)
```yaml
❌ Static Values:
🧠 30% (60K/200K) · 140K avail # WRONG - hardcoded
❌ Guessing:
🧠 ~25% (estimated) # WRONG - no evidence
❌ Placeholder:
🧠 [calculating...] # WRONG - incomplete
✅ Dynamic Calculation:
🧠 28% (57K/200K) · 142K avail # CORRECT - real data
```
## Validation
```yaml
Self-Check Questions:
❓ Did I parse the actual system notification?
❓ Are the numbers from THIS session, not a template?
❓ Does the math check out? (used + remaining = total)
❓ Are percentages rounded correctly?
❓ Are K values formatted correctly?
Validation Formula:
used + remaining should equal total
Example: 57425 + 142575 = 200000 ✅
```
## Edge Cases
```yaml
No System Notification Yet:
Action: Execute a tool call first (e.g., git status)
Then: Parse the notification that appears
Multiple Notifications:
Action: Use the MOST RECENT notification
Reason: Token usage increases over time
Parse Failure:
Fallback: "🧠 [calculating...] (execute a tool first)"
Then: Retry after next tool call
```
## Integration Points
**Used by**:
- `commands/pm.md` - Session start protocol
- `agents/pm-agent.md` - Status reporting
- Any command requiring token awareness
**Dependencies**:
- System-provided notifications (automatic)
- No external tools required