Loading...
Loading...
Found 406 Skills
Experimental NotebookLM harness for listing notebooks, managing sources, asking questions, generating artifacts, and downloading outputs through an installed notebooklm CLI.
CLI harness for Krita digital painting — manage projects, layers, filters, and export via command line. Use when automating Krita workflows, batch processing images, or operating Krita without a GUI.
CLI harness for Zotero.
Agent-native CLI for Exa web search and content retrieval workflows.
Choose before `admin` when the user needs **Shopify CLI** to run or fix something now: validate app or extension config on disk (`shopify.app.toml`, `shopify.app.<name>.toml`, `shopify.extension.toml`) with **`shopify app config validate --json`** (not Admin GraphQL; MCP has no TOML validator); run or troubleshoot store workflows (`shopify store auth`, `shopify store execute`); inventory or product changes by handle, SKU, or location name; or CLI setup, auth, upgrade issues. Emphasize **commands and operational steps**, not only authoring GraphQL. Skip for API-only understanding or codegen with no CLI execution. Examples: validate before deploy; run an existing query via CLI; list products; missing `shopify store execute`.
For answering, using, and debugging @wot-ui/cli. Keywords: wot, @wot-ui/cli, CLI, MCP, doctor, usage, lint, list, info, doc, demo, token, changelog, extract, wot mcp. Applicable for command queries, parameter explanations, MCP integration, local debugging, data extraction, and open-wot repository maintenance.
Image generation and presentations. Use when: - User asks for images: logos, icons, app assets, diagrams, flowcharts, architecture diagrams, patterns, textures, photo edits, restorations - User needs a presentation or slide deck Covers nanobanana CLI for image generation and Slidev for presentations.
Quick reference card for all genshijin modes, skills, and commands. Displayed only once, no mode changes, no state persistence. Activate with "/genshijin-help", "原始人ヘルプ", or "原始人の使い方".
Use this skill when the user asks to call an authenticated HTTP API (for example "call the GitHub/OpenAI/Slack API", "hit an endpoint that needs a bearer token") and the `sesame` CLI is already installed on this device. The agent invokes `sesame request`, which forwards the HTTP call through the user's own broker and attaches the auth header server-side. The skill does not install software, does not read credentials from the environment, and runs shell only within the fixed `sesame` subcommand surface (`request`, `status`, `hostnames`, `login`, `refresh`). Skip for unauthenticated public endpoints, localhost services, or when the user has already exported a token in the environment for direct use.
Video Transcript Extraction Expert (based on Doubao Video Understanding Model). Supports links from Bilibili, Douyin, Xiaohongshu, YouTube, or local video files. Runs entirely in the background (headless) on the user's computer, no pop-ups, no requirement to log in to video platforms. Outputs strict verbatim transcripts with "semantic segmentation + paragraph-level timestamps" (retains colloquial words, internet memes, pauses). Long videos are automatically segmented to avoid being summarized by the model. Trigger scenarios: - User says "generate transcript", "extract transcript", "convert to text", "video to text" - User says "dictate video", "extract video copy", "video subtitles" - User uses the /video-transcript command - User pastes a video link (Bilibili/Douyin/Xiaohongshu/YouTube) with the intention of getting a text version
Defines command mode policy. Non-interactive is default; interactive is opt-in.
Use skill if you are building or repairing Raycast Script Commands in Python or Bash and need correct metadata, output modes, arguments, discovery, or runtime behavior.