Loading...
Loading...
Found 240 Skills
Detects common LLM coding agent artifacts by spawning 4 parallel subagents
Debugs errors and traces failures in AI agents and their tools. Use this skill when the user says: "the agent is failing", "tool call not working", "error in the pipeline", "debug this", "why is the agent doing X instead of Y", "trace the execution", "agent is stuck", "infinite loop", "model response won't parse", "context overflow". Identifies context errors, infinite loops, malformed tool calls, response parsing issues and subagent conflicts.
Use when creating specialized subagents for Claude Code plugins or the Task tool - covers description writing for auto-delegation, tool selection, prompt structure, and testing agents
Manage context-mode GitHub issues, PRs, releases, and marketing with parallel subagent army. Orchestrates 10-20 dynamic agents per task. Use when triaging issues, reviewing PRs, releasing versions, writing LinkedIn posts, announcing releases, fixing bugs, merging contributions, validating ENV vars, testing adapters, or syncing branches.
Use when the user wants Claude Code to generate images from prompts, use /gi or /gi-setup, configure an OpenAI Images API-compatible endpoint, create placeholder images, delegate generation to a background subagent named painter, pass model parameters, choose output paths, or maintain an image index. If generation is requested before setup, guide the user through /gi-setup instead of failing.
Multi-agent collaboration plugin that spawns N parallel subagents competing on the same task via git worktree isolation. Agents work independently, results are evaluated by metric or LLM judge, and the best branch is merged. Use when: user wants multiple approaches tried in parallel — code optimization, content variation, research exploration, or any task that benefits from parallel competition. Requires: a git repo.
Use Browser Use cloud API to spin up cloud browsers for Clawdbot and run autonomous browser tasks. Primary use is creating browser sessions with profiles (persisted logins/cookies) that Clawdbot can control. Secondary use is running task subagents for fast autonomous browser automation. Docs at docs.browser-use.com and docs.cloud.browser-use.com.
Comprehensive technical research by combining multiple intelligence sources — Grok (X/Twitter developer discussions via Playwright), DeepWiki (AI-powered GitHub repository analysis), and WebSearch. Dispatches parallel subagents for each source and synthesizes findings into a unified report. This skill should be used when evaluating technologies, comparing libraries/frameworks, researching GitHub repos, gauging developer sentiment, or investigating technical architecture decisions. Trigger phrases include "tech research", "research this technology", "技术调研", "调研一下", "compare libraries", "evaluate framework", "investigate repo".
10-parallel code/design review using reviewer subagents. Use when: - Running code reviews on PRs, commits, or branches - Running design reviews on issues or documents - Need multi-perspective review (security, architecture, code, QA, historian)
Run configurable BMAD pipeline for story delivery using subagent
Full lifecycle orchestrator - spec/impl/test. Spawn-wait-close pipeline with inline discuss subagent, shared explore cache, fast-advance, and consensus severity routing.
Use when the user explicitly asks to invoke another coding agent CLI as a subagent. Triggers include phrases like 'get a second opinion from Codex', 'have Gemini review this', 'run this through Claude Code', 'ask another agent', or 'use a different model for this'. Supports Claude Code, Codex CLI, and Gemini CLI. Never invoke autonomously.