Loading...
Loading...
Found 80 Skills
Analyze raw prompts, identify intent and gaps, match ECC components (skills/commands/agents/hooks), and output a ready-to-paste optimized prompt. Advisory role only — never executes the task itself. TRIGGER when: user says "optimize prompt", "improve my prompt", "how to write a prompt for", "help me prompt", "rewrite this prompt", or explicitly asks to enhance prompt quality. Also triggers on Chinese equivalents: "优化prompt", "改进prompt", "怎么写prompt", "帮我优化这个指令". DO NOT TRIGGER when: user wants the task executed directly, or says "just do it" / "直接做". DO NOT TRIGGER when user says "优化代码", "优化性能", "optimize performance", "optimize this code" — those are refactoring/performance tasks, not prompt optimization.
Transform vague prompts into precise, well-structured specifications using EARS (Easy Approach to Requirements Syntax) methodology. This skill should be used when users provide loose requirements, ambiguous feature descriptions, or need to enhance prompts for AI-generated code, products, or documents. Triggers include requests to "optimize my prompt", "improve this requirement", "make this more specific", or when raw requirements lack detail and structure.
AI video pipeline validator for Veo 3 feasibility, 8-second scene chunking, and shot continuity. USE WHEN: Validating screenplays for AI video generation, chunking scenes into 8-second segments, generating continuation prompts, scoring feasibility risk, or adding editing metadata. PIPELINE POSITION: screenwriter → **production-validator** → imagine/arch-v INPUT: XML from screenwriter skill (scene tags with duration, action, key_visuals) OUTPUT: Enhanced XML with validation, chunks, continuity tags, and Veo 3 prompts KEY FUNCTIONS: - Veo 3 feasibility validation with risk scoring (LOW/MEDIUM/HIGH/CRITICAL) - 8-second scene chunking with continuation prompts - Shot continuity tagging for editors - Technical optimization for AI-friendly alternatives
Gas Town × DOK Framework - A two-dimensional model for analyzing AI collaboration maturity and cognitive complexity to reveal growth opportunities.
Master Anthropic's prompt engineering techniques to generate new prompts or improve existing ones using best practices for Claude AI models.
Analyze and optimize system prompts using a structured prompting guidelines framework — AI-powered analysis and rewriting. Use when a prompt needs improvement, experiment results show quality gaps, or you want a structured review of an existing system prompt. Do NOT use when production traces show failures (use analyze-trace-failures first to identify patterns). Do NOT use to build evaluators (use build-evaluator).
The meta-skill that powers all other AI tools. Prompt engineering for creative applications is the art and science of communicating with AI models to produce exactly what you envision—in images, video, audio, and text. This isn't just "write better prompts." It's understanding how different models interpret language, how to structure requests for different modalities, how to iterate systematically, and how to build prompt libraries that encode your creative vision. The best prompt engineers have developed intuition for what words trigger what responses in each model. This skill is foundational—it amplifies the effectiveness of every other AI creative skill. Master this, and you master the interface to all AI creation. Use when "prompt, prompting, prompt engineering, better prompts, prompt optimization, how to prompt, prompt strategy, prompt library, prompt template, make AI understand, prompt-engineering, prompting, meta-skill, ai-creative, foundational, optimization, iteration" mentioned.
Automatically intercepts and optimizes prompts using the prompt-learning MCP server. Learns from performance over time via embedding-indexed history. Uses APE, OPRO, DSPy patterns. Activate on "optimize prompt", "improve this prompt", "prompt engineering", or ANY complex task request. Requires prompt-learning MCP server. NOT for simple questions (just answer them), NOT for direct commands (just execute them), NOT for conversational responses (no optimization needed).
Transforms vague or simple user prompts into high-quality, structured, and high-performance AI instructions using systematic optimization techniques like XML tagging, few-shot examples, and Chain-of-Thought. Use this skill when you need to improve the reliability, accuracy, or formatting of an AI's output.
Generate synthetic training data when you don't have enough real examples. Use when you're starting from scratch with no data, need a proof of concept fast, have too few examples for optimization, can't use real customer data for privacy or compliance, need to fill gaps in edge cases, have unbalanced categories, added new categories, or changed your schema. Covers DSPy synthetic data generation, quality filtering, and bootstrapping from zero.
Reduce your AI API bill. Use when AI costs are too high, API calls are too expensive, you want to use cheaper models, optimize token usage, reduce LLM spending, route easy questions to cheap models, or make your AI feature more cost-effective. Covers DSPy cost optimization — cheaper models, smart routing, per-module LMs, fine-tuning, caching, and prompt reduction.
Reduces LLM costs and improves response times through caching, model selection, batching, and prompt optimization. Provides cost breakdowns, latency hotspots, and configuration recommendations. Use for "cost reduction", "performance optimization", "latency improvement", or "efficiency".