Loading...
Loading...
Found 621 Skills
Use this agent when you need to verify that a UI implementation matches its Figma design specifications. This agent should be called after code has been written to implement a design, particularly after HTML/CSS/React components have been created or modified. The agent will visually compare the live implementation against the Figma design and provide detailed feedback on discrepancies.\n\nExamples:\n- <example>\n Context: The user has just implemented a new component based on a Figma design.\n user: "I've finished implementing the hero section based on the Figma design"\n assistant: "I'll review how well your implementation matches the Figma design."\n <commentary>\n Since UI implementation has been completed, use the design-implementation-reviewer agent to compare the live version with Figma.\n </commentary>\n </example>\n- <example>\n Context: After the general code agent has implemented design changes.\n user: "Update the button styles to match the new design system"\n assistant: "I've updated the butto...
Local Microsoft Open Specifications corpus navigator for Windows protocols. Use this skill when the user asks protocol-level questions, needs message/structure details, wants section-by-section summaries, or needs cross-references across related specifications.
Use this skill when crafting, reviewing, or improving prompts for LLM pipelines — including task prompts, system prompts, and LLM-as-Judge prompts. Triggers include: requests to write or refine a prompt, diagnose why an LLM produces inconsistent or incorrect outputs, bridge the gap between intent and model behavior, reduce ambiguity in instructions, add few-shot examples, structure complex prompts, or improve output formatting. Also use when the user needs help distinguishing specification failures (unclear instructions) from generalization failures (model limitations), or when iterating on prompts based on observed failure modes. Do NOT use for general coding tasks, document creation, or non-LLM writing.
Create semantic git commits following best practices and Conventional Commits specification.
Create a git commit following the Conventional Commits specification.
Gate 2 sub-skill - validates uncertain mappings from Gate 1 and confirms all field specifications through testing.
Co-author a Pantagruel (.pant) specification through guided conversation. Use when the user wants to write, design, or iterate on a Pantagruel spec file. Helps users think rigorously about domains, rules, invariants, actions, and initial state by asking clarifying questions rather than guessing. Trigger when user mentions writing a spec, modeling a system, or working with .pant files.
Verifies implementation against specifications by checking requirement fulfillment, task completion, and contract implementation. Generates a fulfillment report with coverage metrics. Always run after /speckit.implement completes.
Create Test Specifications (TSPEC) - Layer 10 artifact for unit, integration, smoke, and functional test cases
Automated fix skill that reads review reports and applies fixes to SPEC (Specification) documents - handles broken links, YAML structure issues, missing files, and iterative improvement
Automated fix skill that reads review reports and applies fixes to TSPEC (Test Specification) documents - handles broken links, element IDs, test case structure issues, and iterative improvement
Automated TSPEC generation from SPEC - generates test specifications for UTEST, ITEST, STEST, FTEST with TASKS-Ready scoring