Loading...
Loading...
Found 21 Skills
Agent skill for performance-monitor - invoke with $agent-performance-monitor
Agent skill for performance-benchmarker - invoke with $agent-performance-benchmarker
Agent skill for performance-analyzer - invoke with $agent-performance-analyzer
Agent skill for performance-optimizer - invoke with $agent-performance-optimizer
Use this skill when the user's Copilot Studio agent evaluations have come back and they need to interpret scores, diagnose root causes of underperforming test cases, find remediation steps, or analyze patterns to improve their agent. Always use this skill when the user mentions: "eval failed", "why did this fail", "triage", "diagnose failure", "low pass rate", "fix evaluation results", "not passing", "failing test cases", "evaluation results", "improve my eval scores", or any situation where eval scores need interpretation and action.
Audits Claude Code context window consumption across agents, skills, MCP servers, and rules. Identifies bloat, redundant components, and produces prioritized token-savings recommendations.
Agent skill for v3-performance-engineer - invoke with $agent-v3-performance-engineer
Expert in observing, benchmarking, and optimizing AI agents. Specializes in token usage tracking, latency analysis, and quality evaluation metrics. Use when optimizing agent costs, measuring performance, or implementing evals. Triggers include "agent performance", "token usage", "latency optimization", "eval", "agent metrics", "cost optimization", "agent benchmarking".
Agent harness performance system for Claude Code and other AI coding agents — skills, instincts, memory, hooks, commands, and security scanning
Apply optimization techniques to extend effective context capacity. Use when context limits constrain agent performance, when optimizing for cost or latency, or when implementing long-running agent systems.
Japanese version of the PUA Universal Motivation Engine. It compels exhaustive problem-solving using corporate PUA rhetoric and structured debugging methodology in Japanese. MUST trigger under the following conditions: (1) Any task has failed 2+ times, or you're stuck in a loop of tweaking the same approach; (2) You're about to say 'I cannot', suggest manual handling to the user, or blame the environment without verification; (3) You find yourself being passive — not searching, not reading source code, not verifying, just waiting for instructions; (4) The user expresses frustration in any form: 'try harder', 'stop giving up', 'figure it out', 'why isn't this working', 'again???', 'もっと頑張れ', 'なんでまた失敗したの', 'もう一回やって', 'なんとかしろ', or any similar sentiment regardless of phrasing. It should also trigger when facing complex multi-step debugging, environment issues, configuration problems, or deployment failures where early surrender is tempting. Applies to ALL task types: code, configuration, research, writing, deployment, infrastructure, API integration. DO NOT trigger on first-attempt failures or when a known fix is already executing successfully.
Measure and optimize customer service performance using CSAT, NPS, CES, First Contact Resolution, and text mining on support tickets. Use this skill when the user needs to evaluate CS team performance, identify top complaint drivers, optimize staffing, or build CS dashboards — even if they say 'is our CS team doing well', 'what are customers complaining about', 'how many agents do we need', or 'build a CS dashboard'.