Loading...
Loading...
Found 10,051 Skills
Skill Evolver (Taotie) — Strengthen the target skill by "devouring" and analyzing the advantages of other skills. This skill must be triggered when users intend to: integrate two skills, optimize one skill with another, compare and analyze the pros and cons of two skills, extract the strengths of one skill into another, or express intentions like "feed X to Y", "use X to optimize Y", "integrate these two skills", "devour this skill", "skill evolution", "skill upgrade", "merge skills", etc. Even if users don't explicitly mention "Taotie", this skill should be used as long as it involves capability transfer, comparative analysis, or advantage extraction between two skills.
Use when a user has finished using one installed skill and wants to preserve actionable feedback about that skill while the session context is still fresh
Open a named, traced browser session into an RVF cognitive container with a ruvector trajectory recording every action
Printing Press CLI for Docker Hub. Docker Hub public API. Search container images, browse tags, check sizes, inspect Dockerfiles, and explore the...
Run a session retrospective against the CLAUDE.md, skills, and hooks — identify guidance violations, stale rules, and gaps. Use when user says 'reflect on this session', 'what did we learn', 'post-mortem this work', 'what should we update in CLAUDE.md', or 'are our skills still right'. Do NOT use for code review (use /review-diff), PR prep (use /create-pr), or creating new skills from scratch (use /create-or-audit-skill).
Drive an authentication flow once, sanitize cookies through AIDefence, and vault a reusable cookie handle in browser-cookies for future sessions
Read every docs/benchmarks/runs/*.json and surface drift in win rate, latency, escalation rate, and LLM-baseline cost over time
Publish or fetch learned patterns across projects via IPFS (Pinata) -- the cross-project pattern transfer that hooks_transfer enables
Per-conversation cost view — list every session in cost-tracking with started-at, message count, top model, and total cost
Wrap getTokenOptimizer().getCompactContext() to retrieve compacted ReasoningBank context for cost-analysis queries; report bridge-reported tokensSaved
Run the corpus benchmark — booster locally, optional Gemini/Sonnet/Opus baselines — and persist a verifiable measured-vs-claimed table
Use when conversation context is too long, hitting token limits, or responses are degrading. Compresses history while preserving critical information using anchored summarization and probe-based validation.