Loading...
Loading...
Found 450 Skills
Match spoken edit beats to candidate B-roll assets using a normalized transcript, subtitle chunking, optional A-roll analysis, and a reusable B-roll catalog. Use this when the goal is to decide what B-roll should support each beat, not just to list assets or describe the video.
Review a resume like a professional resume writer AND an ATS (Applicant Tracking System) — produce an ATS score (0-100), a level detection (fresher / mid / senior), a breakdown of what's working, what's failing, the 7 Deadly Sins check, priority fixes, concrete before→after bullet rewrites, and missing keywords. Use this skill whenever the user asks to review, audit, critique, score, rate, roast, or improve a resume or CV, or types /resume-review, or drops a resume PDF with a job description. Trigger EVEN IF the user just says "look at my CV", "check my resume", "is this resume good", "why am I not getting interviews", "roast my resume", or shares a resume file without explicit review instructions — any interaction involving a resume file or CV should invoke this skill. Especially strong for DevOps, SRE, Cloud, Platform, and Infrastructure engineering resumes, but works for any technical resume.
Audit and automatically optimize a lovstudio skill against repo conventions and official Anthropic skill-creator best practices, then bump the semver version and append a CHANGELOG entry. Checks SKILL.md frontmatter/trigger quality, script CLI hygiene, directory naming, README badge, and progressive disclosure structure. Prioritizes issues raised in the current conversation (e.g. bugs the user just hit) over a generic sweep. Use when the user asks to "optimize", "refine", "audit", or "polish" an existing skill, or when they say "bump version", "update changelog", or "fix this skill". Also trigger when the user mentions "优化 skill", "skill 审计", "刷一遍 skill", "skill-optimizer", "bump skill version", "update skill changelog".
Bootstrap a nao agent for a project — gather warehouse + scope + extra-context info in one round, look up the warehouse-specific config from nao docs, write nao_config.yaml, run nao init + nao sync, set up the LLM key, and generate the first RULES.md. Use when the user has just decided to use nao on a new project. Only for first-time setup; for editing rules, generating tests, or reviewing an existing context, use write-context-rules / create-context-tests / audit-context.
Rules and patterns for building React forms with React Hook Form (RHF) and Zod validation. Use this skill whenever the user is creating, editing, or refactoring any React form — including login forms, registration flows, multi-step wizards, dynamic field arrays, or any input component wired to RHF. Also trigger when the user mentions `useForm`, `Controller`, `zodResolver`, `z.object`, schema validation, form state, `useFieldArray`, or `FormProvider`. Trigger even if they just ask "how do I validate this field" or "how do I handle server errors in a form" — this skill covers it all.
You are a WeChat Official Account (微信公众号) marketing virtuoso with deep expertise in China's most intimate business communication platform. You understand that WeChat OA is not just a broadcast chan...
Pixel-faithful clone of any web UI into the user's existing stack, using whatever sources are available — a screenshot alone, a live URL, raw HTML/CSS, or any combination. Use this skill whenever the user wants to recreate, match, replicate, or "clone" a design from a screenshot, image, URL, Figma export, or HTML dump. Trigger on phrases like "clone this", "match this design", "build this from screenshot", "recreate this page", "make it look like this", "rebuild this UI", "copy this layout", or any time the user provides a visual reference and asks for a faithful implementation. Do not undertrigger — even if the user just drops a screenshot without explicit phrasing, this skill applies.
Generate Python code for Naver Open APIs (News Search, Blog Search, Web Search, Datalab Trends, Image Search, Book Search, and other openapi.naver.com APIs). Use this skill whenever the user wants to call Naver APIs, search Naver News/Blog/Web/Images, fetch Datalab keyword trends, or write Python scripts that interact with Naver's Open API platform. This includes requests for code examples, API integration, parameter explanations, endpoint lookup, or troubleshooting Naver Open API calls. Even if the user just mentions 'Naver API', 'Naver news data', 'search trends', or 'openapi.naver.com', activate this skill immediately.
Interpret the meaning of paper figures and output a highly readable Markdown report that 'teaches humans how to read figures'; supports input of absolute paths to one or more figure files and manual interpretations, automatically attempts to retrieve the source code used to generate the figures from the vicinity of the figures, and uses a parallel-vibe-like approach to interpret each figure with process-level isolation via `codex exec`/`claude -p` (default concurrency limit is 3, adjustable in config.yaml). ⚠️ Not applicable: Users only want to adjust figure size/crop/change format; or request direct modification of images/source code (this skill has read-only access to images and source code throughout, modification is strictly prohibited).
This skill is applicable when users explicitly request to 'write/generate NSFC budget specification', 'write budget explanation', 'generate budget.tex / budget.pdf', or 'write NSFC budget justification'. Based on the user's proposal text or supplementary materials, output a submittable budget specification LaTeX project and render `budget.pdf`. If the user does not specify a working directory, you must pause and ask them to specify it first. ⚠️ Not applicable: Users only want to understand budget principles; users only want budget table figures without writing the specification; or users are in the 2026 Youth A/B/C default lump-sum scenario where no budget specification is required.
Query-driven targeted ingest from a specific AI agent's raw history. Use this skill when the user invokes /wiki-claude, /wiki-codex, /wiki-hermes, /wiki-openclaw, /wiki-copilot — with or without a search topic. Different from wiki-history-ingest (which bulk-ingests everything new): this skill finds sessions about a SPECIFIC TOPIC in a specific agent's history and ingests just those, then returns a synthesized answer immediately usable in the current session. Primary use case: you're working in agent A and want to pull in how you solved X in agent B's history. Cross-referencing, not archiving. Also trigger on: "what did I work on in codex about X", "search my claude sessions for Y", "pull in hermes knowledge about Z", "find that conversation where I did X in codex".
AI HOT (aihot.virxact.com) Chinese AI News Query Skill. Trigger this Skill when users ask any Chinese AI information queries such as "What's happening in the AI circle today", "AI Daily", "AI HOT", "AI News", "AI Hot Topics", "Latest AI Updates", "What have OpenAI/Anthropic/Google released recently", "AI hot today", "AI news today", "Check AI industry trends", "What large models are released today", "AI circle updates from yesterday", "Check selected items", "AI HOT Selected", "AI papers from the past week", "AI model releases", "AI product launches", "AI industry dynamics", "AI tips and insights". Even if users only say "AI circle", "AI news", "AI Daily", or just ask "What happened today" with context related to AI / large models / LLM / startup fields, this Skill should be triggered. The Skill directly pulls data via curl from public REST APIs and organizes it into Chinese markdown briefings, with no need for users to configure any API Key or MCP server. **Do NOT undertrigger**——If users ask for AI news and you don't invoke this Skill, you are treating outdated training data as today's news, which is harmful to users.