Loading...
Loading...
Found 11 Skills
Audits Claude Code context window consumption across agents, skills, MCP servers, and rules. Identifies bloat, redundant components, and produces prioritized token-savings recommendations.
Apply optimization techniques to extend effective context capacity. Use when context limits constrain agent performance, when optimizing for cost or latency, or when implementing long-running agent systems.
Measure and optimize customer service performance using CSAT, NPS, CES, First Contact Resolution, and text mining on support tickets. Use this skill when the user needs to evaluate CS team performance, identify top complaint drivers, optimize staffing, or build CS dashboards — even if they say 'is our CS team doing well', 'what are customers complaining about', 'how many agents do we need', or 'build a CS dashboard'.
Expert prompt optimization for LLMs and AI systems. Use PROACTIVELY when building AI features, improving agent performance, or crafting system prompts. Masters prompt patterns and techniques.
Expert data analysis and manipulation for customer support operations using pandas
Query cross-project usage analytics. Use when reviewing agent, skill, hook, or team performance across OrchestKit projects. Also replay sessions, estimate costs, and view model delegation trends.
Japanese version of the PUA Universal Motivation Engine. It compels exhaustive problem-solving using corporate PUA rhetoric and structured debugging methodology in Japanese. MUST trigger under the following conditions: (1) Any task has failed 2+ times, or you're stuck in a loop of tweaking the same approach; (2) You're about to say 'I cannot', suggest manual handling to the user, or blame the environment without verification; (3) You find yourself being passive — not searching, not reading source code, not verifying, just waiting for instructions; (4) The user expresses frustration in any form: 'try harder', 'stop giving up', 'figure it out', 'why isn't this working', 'again???', 'もっと頑張れ', 'なんでまた失敗したの', 'もう一回やって', 'なんとかしろ', or any similar sentiment regardless of phrasing. It should also trigger when facing complex multi-step debugging, environment issues, configuration problems, or deployment failures where early surrender is tempting. Applies to ALL task types: code, configuration, research, writing, deployment, infrastructure, API integration. DO NOT trigger on first-attempt failures or when a known fix is already executing successfully.
Evaluate and improve Claude Code commands, skills, and agents. Use when testing prompt effectiveness, validating context engineering choices, or measuring improvement quality.
Use this when you need to EVALUATE OR IMPROVE or OPTIMIZE an existing LLM agent's output quality - including improving tool selection accuracy, answer quality, reducing costs, or fixing issues where the agent gives wrong/incomplete responses. Evaluates agents systematically using MLflow evaluation with datasets, scorers, and tracing. Covers end-to-end evaluation workflow or individual components (tracing setup, dataset creation, scorer definition, evaluation execution).
Review recent work, identify process gaps and repeated mistakes, and produce specific file edits to prevent them. Not a reflection exercise — outputs config and identity changes. Trigger manually after sprints, or automate weekly.
Paired benchmark orchestration for comparing coding-agent performance with recursive-mode off and on. Use when the user wants to benchmark recursive-mode, compare recursive vs non-recursive execution on the same project, generate disposable benchmark repos, capture timing/build-test logs, or write a benchmark report.