docs: Replace Mindbase References with ReflexionMemory (#464)

* docs: fix mindbase syntax and document as optional MCP enhancement

Fix incorrect method call syntax and clarify mindbase as optional
enhancement that coexists with built-in ReflexionMemory.

Changes:
- Fix syntax: mindbase.search_conversations() → natural language
  instructions that allow Claude to autonomously select tools
- Clarify mindbase requires airis-mcp-gateway "recommended" profile
- Document ReflexionMemory as built-in fallback (always available)
- Show coexistence model: both systems work together

Architecture:
- ReflexionMemory (built-in): Keyword-based search, local JSONL
- Mindbase (optional MCP): Semantic search, PostgreSQL + pgvector
- Claude autonomously selects best available tool when needed

This approach allows users to enhance error learning with mindbase
when installed, while maintaining full functionality with
ReflexionMemory alone.

Related: #452

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* docs: add comprehensive ReflexionMemory user documentation

Add user-facing documentation for the ReflexionMemory error learning
system to address documentation gap identified during mindbase cleanup.

New Documentation:
- docs/user-guide/memory-system.md (283 lines)
  * Complete user guide for ReflexionMemory
  * How it works, storage format, usage examples
  * Performance benefits and troubleshooting
  * Manual inspection and management commands

- docs/memory/reflexion.jsonl.example (15 entries)
  * 15 realistic example reflexion entries
  * Covers common scenarios: auth, DB, CORS, uploads, etc.
  * Reference for understanding the data format

- docs/memory/README.md (277 lines)
  * Overview of memory directory structure
  * Explanation of all files (reflexion, metrics, patterns)
  * File management, backup, and git guidelines
  * Quick command reference

Context:
Previous mindbase cleanup removed references to non-existent external
MCP server, but didn't add sufficient user-facing documentation for
the actual ReflexionMemory implementation.

Related: #452

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

* docs: translate Japanese text to English in documentation

Address PR feedback to remove Japanese text from English documentation files.

Changes:
- docs/mcp/mcp-integration-policy.md: Translate headers and descriptions
- docs/reference/pm-agent-autonomous-reflection.md: Translate error messages
- docs/research/reflexion-integration-2025.md: Translate error messages
- docs/memory/pm_context.md: Translate example keywords

All Japanese text in English documentation files has been translated to English.

🤖 Generated with [Claude Code](https://claude.com/claude-code)

Co-Authored-By: Claude <noreply@anthropic.com>

---------

Co-authored-by: Claude <noreply@anthropic.com>
This commit is contained in:
Cedric Hurst
2025-10-30 22:14:35 -05:00
committed by GitHub
parent c733413d3c
commit bea4bfe289
11 changed files with 680 additions and 82 deletions

View File

@@ -32,7 +32,7 @@ SuperClaude is a comprehensive framework for Claude Code that provides:
### Intent Classification System
```yaml
Ultra-Light (100-500 tokens): "進捗", "progress", "status" → Layer 1 only
Ultra-Light (100-500 tokens): "progress", "status", "update" → Layer 1 only
Light (500-2K tokens): "typo", "rename", "comment" → Layer 2 (target file)
Medium (2-5K tokens): "bug", "fix", "refactor" → Layer 3 (related files)
Heavy (5-20K tokens): "feature", "architecture" → Layer 4 (subsystem)
@@ -52,10 +52,11 @@ Ultra-Heavy (20K+ tokens): "redesign", "migration" → Layer 5 (full + rese
- **Data**: task_type, complexity, workflow_id, tokens_used, time_ms, success
- **Strategy**: ε-greedy (80% best workflow, 20% experimental)
### mindbase Integration Incentive
- **Layer 1**: 500 tokens (mindbase) vs 800 tokens (fallback) = **38% savings**
- **Layer 3**: 3-4K tokens (mindbase) vs 4.5K tokens (fallback) = **20% savings**
- **Total Potential**: Up to **90% token reduction** with semantic search (industry benchmark)
### Error Learning & Memory Integration
- **ReflexionMemory (built-in)**: Layer 1: 650 tokens | Layer 3: 3.5-4K tokens
- **mindbase (optional)**: Layer 1: 500 tokens | Layer 3: 3-3.5K tokens (semantic search)
- **Profile**: Requires airis-mcp-gateway "recommended" profile for mindbase
- **Savings**: 20-35% with ReflexionMemory, additional 10-15% with mindbase enhancement
## Active Patterns