Loading...
Loading...
Found 1,668 Skills
Evaluate and improve skills through measured testing. Run trigger evaluations to test whether skill descriptions cause correct activation, optimize descriptions via automated train/test loops, benchmark skill output quality with A/B comparisons, and validate skill structure. Use when user says "improve skill", "test skill triggers", "optimize description", "benchmark skill", "eval skill", or "skill quality". Do NOT use for creating new skills (use skill-creator-engineer).
Generate blog post topic ideas through problem mining, gap analysis, and technology expansion. Use when user needs content ideas, wants to brainstorm articles, asks "what should I write about", or needs to fill content gaps. Use for "brainstorm", "topic ideas", "what to write", "content gaps", or "blog ideas". Do NOT use for writing posts, creating outlines, or SEO optimization without a specific ideation need.
Maintain /do routing tables and command references when skills or agents are added, modified, or removed. Use when skill/agent metadata changes, after skill-creator-engineer or agent-creator-engineer runs, or when routing tables need synchronization. Use for "update routes", "sync routing", "routing table", or "refresh /do". Do NOT use for creating new skills/agents, modifying skill logic, or manual /do table edits.
Spawn 10 independent parallel agents to analyze source material from distinct perspectives, synthesize findings, and apply improvements to a target agent or skill. Use when source material is complex and multi-angle extraction justifies 3-5x token cost over inline analysis. Use for "parallel analysis", "multi-perspective", or "deep extraction". Do NOT use for routine improvements, simple source material, or when token budget is limited.
Deterministic plan lifecycle management via scripts/plan-manager.py: create, track, check, complete, and abandon task plans. Use when user says "/plans", needs to create a multi-phase plan, track progress on active plans, or manage plan lifecycle (complete, abandon, audit). Do NOT use for one-off tasks that need no tracking, feature implementation, or debugging workflows.
Fetch PR feedback from all GitHub sources and validate every comment before acting. Use when user needs to process PR review comments, address reviewer feedback, or triage pull request discussions. Use for "review comments", "PR feedback", "address reviews", "address feedback", or "what did reviewers say". Do NOT use for creating new PRs, code review of local changes, or general code analysis without an existing PR.
Interactive debugging via DAP-MCP for multiple languages with natural language commands
Automatically suggests optimal files to preload based on the user's task description and historical context patterns. Activates when the user starts a new task, mentions reading files, or when session context is being set up.
Manage context templates for common task types
Show git-aware context suggestions for current working directory
Show context usage heatmap for current session
Configure token budget limits, auto-compact settings, and view current budget status