Loading...
Loading...
Found 7,605 Skills
Cisco Meraki integration. Manage data, records, and automate workflows. Use when the user wants to interact with Cisco Meraki data.
Two-way sync between a local paper directory and an Overleaf project via the Overleaf Git bridge (Premium feature). Lets you keep ARIS audit/edit workflows on the local copy while collaborators edit in the Overleaf web UI. Token never touches the agent — user does the one-time auth via macOS Keychain. Use when user says "同步 overleaf", "overleaf sync", "推送到 overleaf", "connect overleaf", "Overleaf 桥接", "pull overleaf", "push overleaf", or wants to bridge their ARIS paper directory with an Overleaf project.
Generate deterministic publication-quality architecture, workflow, and pipeline diagrams from structured JSON (FigureSpec) into editable SVG. Use when user says "架构图", "workflow 图", "pipeline 图", "确定性矢量图", "figure spec", "draw architecture", or needs precise, editable, publication-ready vector diagrams. Preferred over AI illustration for formal architecture/workflow figures.
Inspect, operate, and troubleshoot ERDA runtimes through erda-cli-backed workflows. Use when users need help with runtime status, deployment behavior, logs, scaling, restarts, or environment-oriented ERDA troubleshooting.
Ultra-lightweight channel for feature workflows: No need to write design docs, checklists, or conduct phased reviews. Let AI write code directly as it normally would, but before it starts, tell it where the CodeStable knowledge base in the project is and how to search it. This way, the code it writes will have fewer pitfalls and be more consistent with project conventions. Trigger scenarios: Users say "fast mode", "fastforward", "skip all those steps", "just start coding", "help me make xxx" and the requirement is too small to go through the design process.
Issue Workflow Stage 1 — Convert the user's problem into a reproducible, traceable {slug}-report.md through conversation. The AI only asks "what you saw, how to reproduce it, what should happen" here, and does not guess the root cause for the user (that's Stage 2's responsibility). Meanwhile, this stage is the only official decision point for choosing between the fast track and standard path: Based on the user's description, first review the relevant code; if the root cause can be identified at a glance and the required changes are minor, directly inform the user to take the fast track. Trigger scenarios: The user says "file an issue", "record this bug", "I found a problem". This is the starting point of the issue workflow with no pre-dependencies.
Ultra-lightweight channel for refactor processes - used when changes are clearly too small to go through the full scan → design → apply three-stage workflow. AI directly identifies 1-3 low-risk optimization points, confirms with the user once, modifies in-place using classic methods, and validates itself by running tests. No scan checklist, no design documentation, no multi-step human verification required. Trigger scenarios: User says "quick refactor", "small refactor", "simply optimize XX function", "modify directly", "skip the extra steps", and the scope of changes is clearly localized to a single function / single component with test coverage for self-validation.
Phase 3 of the issue workflow —— Fix code precisely according to confirmed root causes and solutions, verify the results, and document it in {slug}-fix-note.md. This is the final stage of the issue workflow —— no verification closure means the workflow is incomplete. Two entry points: the standard path is triggered from cs-issue-analyze (with existing {slug}-analysis.md), and the fast track is triggered directly from cs-issue-report (without {slug}-analysis.md, as the root cause was identified by AI through code reading during the report phase). Trigger scenarios: User says "Start fixing the bug", "Fix according to the analysis", "Start modifying the code". During the fix, only modify the files specified in the solution; do not make incidental optimizations or introduce new abstractions —— these actions will cause the scope to expand to an untraceable extent.
Document the pitfalls encountered or good practices discovered during this work into searchable learning documents, which can be accessed by both AI and humans when similar tasks arise in the future. Two tracks: The pitfall track records experiences where "things should have worked but didn't" — including bugs, configuration traps, environment issues, and integration failures; The knowledge track records findings that "should be the default approach going forward" — including best practices, workflow improvements, and reusable patterns. Trigger scenarios: Proactively prompt at the end of feature-acceptance or issue-fix workflows, or when the user mentions phrases like "document knowledge", "learning", "document learnings", or "record this experience". Spec documents record what was done, while learning documents record what pitfalls were encountered / what was learned — they complement each other and are not interchangeable.
Use Hive Intelligence MCP through UXC for broad crypto market, onchain, portfolio, and risk workflows with help-first discovery and convenience-layer guardrails.
The root entry of the CodeStable workflow family — introduces the overall system to users and routes users' specific requests to the correct cs-* sub-skills. Trigger scenarios: users only input `cs` / `/cs`, say "introduce codestable", "do something with codestable", "I want to do X, which skill should I use", "don't know which one to use", or users' described requests are open-ended (e.g., "start working") and haven't converged to a specific sub-skill. This skill itself **does not perform actual tasks** — it doesn't write specs, write code, or read/write content products in the codestable/ directory — it only performs scanning, routing, prompting, and then transfers control to the target sub-skill.
Neverbounce integration. Manage data, records, and automate workflows. Use when the user wants to interact with Neverbounce data.