mirror of
https://github.com/SuperClaude-Org/SuperClaude_Framework.git
synced 2025-12-29 16:16:08 +00:00
- Core configuration files (CLAUDE.md, RULES.md, PERSONAS.md, MCP.md) - 17 slash commands for specialized workflows - 25 shared YAML resources for advanced configurations - Installation script for global deployment - 9 cognitive personas for specialized thinking modes - MCP integration patterns for intelligent tool usage - Token economy and ultracompressed mode support 🤖 Generated with [Claude Code](https://claude.ai/code) Co-Authored-By: Claude <noreply@anthropic.com>
5.5 KiB
5.5 KiB
Legend
| Symbol | Meaning | Abbrev | Meaning | |
|---|---|---|---|---|
| → | leads to | perf | performance | |
| & | and/with | ops | operations | |
| w/ | with | impl | implementation |
Execute immediately. Add --plan flag if user wants to see plan first.
Improve code, perf, or quality of system specified in $ARGUMENTS.
Improvement focus w/ flags:
--quality flag:
- Improve code structure & maintainability
- w/ --solid: Apply SOLID principles systematically | w/ --refactor: Clean code refactoring
- w/ --metrics: Generate quality metrics report
--perf flag:
- Optimize system perf | Analyze bottlenecks & resource usage
- Impl caching & async ops | Improve algorithm complexity
--iterate flag:
- Iteratively improve until threshold reached | w/ --threshold: Set target % (default 85%)
- Measure progress after each iteration | Stop at diminishing returns
--watch flag:
- Continuous improvement monitoring | Auto-apply safe optimizations
- Real-time perf tracking | Automated quality maintenance
When --interactive flag is present:
- Guided improvement process
- User choice on optimization strategies
- Step-by-step quality enhancement
- Interactive threshold adjustment
Code Quality Mode (--quality)
SOLID Principles application:
- Single Responsibility: One class, one purpose
- Open/Closed: Extensible but not modifiable
- Liskov Substitution: Subtypes must be substitutable
- Interface Segregation: Specific over general interfaces
- Dependency Inversion: Depend on abstractions
Refactoring techniques:
- Extract method/class for complex logic
- Inline unnecessary abstractions
- Rename for clarity and consistency
- Move code to appropriate modules
- Remove duplication (DRY principle)
- Simplify conditionals and loops
- Reduce coupling, increase cohesion
Quality metrics to track:
- Cyclomatic complexity (target < 5)
- Method length (target < 20 lines)
- Class cohesion and coupling
- Code duplication percentage
- Test coverage (target > 80%)
- Documentation completeness
Clean code principles:
- Meaningful, self-documenting names
- Functions do one thing well
- Consistent coding style
- Proper error handling
- No magic numbers/strings
- Comments explain why, not what
Performance Mode (--performance)
Performance optimization areas:
Code optimization:
- Algorithm complexity reduction (O(n²) → O(n log n))
- Efficient data structures
- Caching frequently accessed data
- Lazy loading and pagination
- Async/parallel processing
- Memory usage optimization
Database optimization:
- Query optimization and indexing
- N+1 query elimination
- Connection pooling
- Batch operations
- Denormalization where appropriate
- Query result caching
Frontend optimization:
- Bundle size reduction
- Code splitting and lazy loading
- Image and asset optimization
- Render performance improvements
- Service worker caching
- Reducing re-renders
System optimization:
- Load balancing strategies
- CDN implementation
- Compression (gzip/brotli)
- HTTP/2 and caching headers
- Resource pooling
- Microservice optimization
Iterative Mode (--iterate)
Iteration process:
-
Baseline Measurement
- Current performance metrics
- Quality scores
- Coverage percentage
- User satisfaction
-
Targeted Improvements
- Focus on highest impact areas
- Apply 80/20 rule
- Make incremental changes
- Maintain working state
-
Progress Tracking
- Measure after each change
- Document improvements
- Calculate ROI of changes
- Adjust strategy as needed
-
Completion Criteria
- Reach target threshold
- Diminishing returns detected
- Time/budget constraints
- "Good enough" achieved
Focus areas by iteration type:
- Quality: Complexity, duplication, coverage
- Performance: Response time, throughput, resources
- User Experience: Load time, responsiveness, errors
- Maintainability: Documentation, tests, structure
Best Practices
General improvement approach:
- Measure before changing
- Focus on bottlenecks first
- Make one change at a time
- Verify improvements
- Document changes made
- Consider trade-offs
Avoid common pitfalls:
- Premature optimization
- Over-engineering
- Breaking changes
- Ignoring tests
- Gold-plating
Balance considerations:
- Performance vs readability
- Flexibility vs simplicity
- Speed vs correctness
- Present vs future needs
Research Requirements
All optimization patterns must be verified:
- Performance optimizations → Research benchmarks and best practices via WebSearch
- Framework-specific improvements → C7 documentation lookup required
- Algorithm changes → Verify complexity analysis with authoritative sources
- Caching strategies → Check official recommendations for the platform
- Never apply "common" optimizations without documentation backing
- All improvements must cite sources: // Source: [optimization guide reference]
Report Output:
- Quality metrics:
.claudedocs/metrics/quality-<timestamp>.md - Performance benchmarks:
.claudedocs/metrics/performance-<timestamp>.md - Iteration logs:
.claudedocs/summaries/iteration-log-<timestamp>.md - Ensure directory exists:
mkdir -p .claudedocs/metrics/ .claudedocs/summaries/ - Include report location in output: "📄 Report saved to: [path]"
Deliverables:
- For quality: Refactored code, quality metrics report, improvement documentation
- For performance: Optimized system, performance benchmarks, bottleneck analysis
- For iterate: Final metrics, iteration log, recommendations for future improvements