Loading...
Loading...
Found 80 Skills
Review a prompt to identify ambiguities, missing constraints, and hallucination risks, and provide an optimized version.
Forensic root cause analyzer for Antigravity sessions. Classifies scope deltas, rework patterns, root causes, hotspots, and auto-improves prompts/health.
Use when the user wants to turn a feature idea, change request, or rough requirement into a precise feature-development prompt for one or more codebase projects.
KERNEL-based prompt engineering — transforms vague requests into structured, high-performance prompts optimized for first-try success.
Ultra-compressed response mode. Cuts token usage by dropping articles, filler, pleasantries, and hedging. Uses symbols for relationships. Technical terms and code blocks remain exact and uncompressed. Use when user says "save tokens", "RTU mode", "compress", or "be brief".
Expert prompt engineering for LLM applications including prompt design, optimization, RAG systems, agent architectures, and AI product development.
Fine-tune models on your data to maximize quality and cut costs. Use when prompt optimization hit a ceiling, you need domain specialization, you want cheaper models to match expensive ones, you heard "fine-tuning will make us AI-native", you have 500+ training examples, or you need to train on proprietary data. Covers DSPy BootstrapFinetune, BetterTogether, model distillation, and when to fine-tune vs optimize prompts.
This skill should be used when the user asks to "refine a prompt", "optimize a prompt", "improve my prompt", "rewrite prompt for LLM", "craft a better prompt", or mentions prompt engineering, prompt optimization, or appending to PROMPT.md.
Use when prompts produce inconsistent or unreliable outputs, need explicit structure and constraints, require safety guardrails or quality checks, involve multi-step reasoning that needs decomposition, need domain expertise encoding, or when user mentions improving prompts, prompt templates, structured prompts, prompt optimization, reliable AI outputs, or prompt patterns.
Prompt engineering expert that helps users craft optimized prompts using 57 proven frameworks. Use when users want to optimize prompts, improve AI instructions, create better prompts for specific tasks, or need help selecting the best prompt framework for their use case.
Optimize, rewrite, and evaluate prompts using the Anthropic 1P interactive prompt-engineering tutorial patterns (clear/direct instructions, role prompting, XML-tag separation, output formatting + prefilling, step-by-step “precognition”, few-shot examples, hallucination reduction, complex prompt templates, prompt chaining, and tool-use XML formats). Use for 提示词优化/Prompt优化/Prompt engineering, rewriting system+user prompts, enforcing structured outputs (XML/JSON), reducing hallucinations, building multi-step prompt templates, adding few-shot examples, or designing prompt-chaining/tool-calling scaffolds.
프롬프트를 실증 기반 기법으로 분석하고 개선합니다. Few-shot, CoT, XML 구조화, Context Engineering 등 검증된 기법을 적용하여 프롬프트 품질을 높입니다. 프롬프트 개선, prompt 리뷰, 프롬프트 최적화, 프롬프팅 개선 요청 시 사용.