Loading...
Loading...
Found 65 Skills
Use when verifying code works — after feature work, before committing, before deploy, or any request to "verify", "check", or "make sure it works"
Run quality gates on implemented feature: tests, lint, type checks, and custom validation. Use after /feature-implement completes. Use for "validate feature", "run quality gates", "check feature", or "/feature-validate". Do NOT use for ad-hoc linting or debugging.
Collaborative coding with enforced micro-steps: announce, show diff, wait for confirmation, apply, verify. User controls pace with commands. Works with any domain agent as the executor. Use when: "pair program", "pair with me", "let's code together", "step by step coding", "walk me through implementing", "code with me"
Run the full spec-driven workflow automatically. Proposes, implements, verifies, reviews, and archives a change with one mandatory proposal checkpoint plus any extra confirmations required by blocking conditions.
프로젝트의 모든 verify 스킬을 순차 실행하여 통합 검증 보고서를 생성합니다. 기능 구현 후, PR 전, 코드 리뷰 시 사용.
Database schema and migration workflow. Use when modifying database schemas in tools/atlas/ or SQLC queries in tools/sqlc/.
Phase-level implementation workflow for builder agents. Handles reading phase files, finding references, invoking domain skills, implementing all steps, and running verification (tests + typecheck). Preloaded into builder agents via skills: field — not user-invocable.
Test-driven development workflow with philosophy guide - plan → write tests → implement → validate
Verify completed implementation against the plan. Checks that all tasks were fully implemented, nothing was forgotten, code compiles, tests pass, and quality standards are met. Use after "/aif-implement" completes, or when user says "verify", "check work", "did we miss anything".
Use when you have a written implementation plan to execute with review checkpoints
Code quality gatekeeper and auditor. Enforces strict quality gates, resolves the AI verification gap, and evaluates codebases across 12 critical dimensions with evidence-based scoring. Use when auditing code quality, reviewing AI-generated code, scoring codebases against industry standards, or enforcing pre-commit quality gates. Use for quality audit, code review, codebase evaluation, security assessment, technical debt analysis.
TDD Skill for Team Workflows (Phase 3). Before writing any implementation code, you must first write a failing test and verify the failure, then write the minimal implementation until it passes, and finally refactor and re-test.