Loading...
Loading...
Found 34 Skills
Execute, validate, and debug Midscene YAML automation files. Handles dry-run, execution, report analysis, and iterative debugging.
Coordinate smart-home actions across existing integrations with clear dry-run and safety confirmation.
Use when syncing or updating previously installed skills to their latest version. Always dry-run updates before applying, and check for breaking changes.
Use when regenerating README.md and WORK_AREAS.md in a managed library workspace. Always dry-run first to preview changes.
Use this skill whenever designing, building, or reviewing a command-line tool that AI agents or automation will invoke — covers non-interactive flags, layered --help with examples, stdin/pipeline composition, actionable errors, idempotency, dry-run, destructive-action safety, and predictable command structure. Trigger even if the user doesn't explicitly say "agent-friendly" — apply whenever they are writing `--help` text, adding a new subcommand, designing error messages, or reviewing a CLI's UX.
Comprehensive toolkit for validating, linting, testing, and automating Terragrunt configurations, HCL files, and Stacks. Use this skill when working with Terragrunt files (.hcl, terragrunt.hcl, terragrunt.stack.hcl), validating infrastructure-as-code, debugging Terragrunt configurations, performing dry-run testing with terragrunt plan, working with Terragrunt Stacks, or working with custom providers and modules.
Comprehensive toolkit for validating, linting, testing, and automating Ansible playbooks, roles, and collections. Use this skill when working with Ansible files (.yml, .yaml playbooks, roles, inventories), validating automation code, debugging playbook execution, performing dry-run testing with check mode, or working with custom modules and collections.
Fleet orchestration for distributed coding agents across Azure VMs. Invoked as `/fleet <command>`. Covers all fleet operations: status, scout, advance, adopt, watch, snapshot, dry-run, start, add-task, queue, auth, dashboard, tui, and more. Use when: user mentions fleet, agents, VMs, sessions, or asks "what are my agents doing".
Generate headless Claude Code cron jobs from a task description and schedule. Creates a wrapper script with safety mechanisms (lockfile, budget cap, dry-run default, logging) and installs crontab entries via deterministic Python script. Use when user says "schedule", "run every", "cron job", "run twice daily", "run hourly", "run daily", "run weekly", or "schedule task".
Generate EliteForge frontend projects with the same logic as cisdigital-generator-app. Reuse the exact project type to template mapping for frontend_app/frontend_ui/frontend_sdk, naming rules for company/product/service, and onebase-cli command assembly. Use when users ask to scaffold EliteForge frontend app projects, Vue3 component monorepo projects, JS SDK/lib projects, or request a dry-run command preview aligned with this generator. Always require user-provided required parameters and never infer missing required fields.
Investigate LLM analytics evaluations of both types — `hog` (deterministic code-based) and `llm_judge` (LLM-prompt-based). Find existing evaluations, inspect their configuration, run them against specific generations, query individual pass/fail results, and generate AI-powered summaries of patterns across many runs. Use when the user asks to debug why an evaluation is failing, surface common failure modes, compare results across filters, dry-run a Hog evaluator, prototype a new LLM-judge prompt, or manage the evaluation lifecycle (create, update, enable/disable, delete).
Interactively prune stale non-terminal workflows from the pipeline. Use when the user says 'prune workflows', 'clean stale workflows', 'pipeline cleanup', or runs /prune. Runs a dry-run preview, displays candidates with staleness and safeguard skips, prompts the user to proceed/abort/force, then bulk-cancels approved workflows with a workflow.pruned audit event. Safeguards skip workflows with open PRs or recent commits unless force is set.