Loading...
Loading...
Found 4,967 Skills
Integrate Circle Gateway to hold a unified USDC balance across multiple blockchains and transfer USDC instantly (<500ms) via permissionless deposit, burn, and mint workflows. Available on 11 EVM chains + Solana (mainnet and testnet), plus Arc testnet. Use when: enabling chain-agnostic user experiences, low-latency or instant next-block finality is required, capital needs to be pooled across chains for greater capital efficiency, or building apps with consolidated crosschain balances. Triggers on: Gateway, unified balance, crosschain USDC, instant transfer, chain abstraction, Gateway Wallet, Gateway Minter, gatewayMint, burn intent, crosschain liquidity, payment routing, capital efficiency, permissionless transfer.
Create and manage Kibana alerting rules via REST API or Terraform. Use when creating, updating, or managing rule lifecycle (enable, disable, mute, snooze) or rules-as-code workflows.
Generate images, videos, audio, and 3D models via RunningHub API (170+ endpoints) and run any RunningHub AI Application (custom ComfyUI workflow) by webappId. Covers text-to-image, image-to-video, text-to-speech, music generation, 3D modeling, image upscaling, AI apps, and more.
Enforces authenticated gh CLI workflows over unauthenticated curl/WebFetch patterns. Use when working with GitHub URLs, API access, pull requests, or issues.
High-quality article translation skill, adopting a four-step professional translation workflow: **Analysis → Initial Translation → Review → Final Draft**. Only supports Chinese ↔ English, Chinese ↔ Japanese translation. Triggered when users explicitly request translation using terms such as "翻译", "translate", "精翻", "翻訳", "翻译文章", "translate to Chinese/English/Japanese", "改成中文", "改成英文", "改成日文", "翻成中文", "翻成日文", "翻成英文", "英译中", "中译英", "中译日", "日译中", "日本語に翻訳", "中国語に翻訳", "英語に翻訳", "これを翻訳して", "put this in Chinese", "put this in English", "put this in Japanese", "convert to Chinese", "convert to English", "convert to Japanese", "帮我翻一下", "本地化", "localize", "这篇文章翻译一下", or provide a URL/file/text body and explicitly request a final draft in the target language. Not applicable to requests for only summarization, explanation, comprehension, or organization. If the input is a URL, prioritize using `curl -L` to request `r.jina.ai` to fetch the body content in Markdown; if fetching fails or the content is incomplete, you must stop immediately and ask the user to provide the full text themselves.
Generates custom design system rules for the user's codebase. Use when user says "create design system rules", "generate rules for my project", "set up design rules", "customize design system guidelines", or wants to establish project-specific conventions for Figma-to-code workflows. Requires Figma MCP server connection.
Figma-to-code design handoff patterns including Figma Variables to design tokens pipeline, component spec extraction, Dev Mode inspection, Auto Layout to CSS Flexbox/Grid mapping, and visual regression with Applitools. Use when converting Figma designs to code, documenting component specs, setting up design-dev workflows, or comparing production UI against Figma designs.
AI-assisted UI generation patterns for json-render, v0, Bolt, and Cursor workflows. Covers prompt engineering for component generation, review checklists for AI-generated code, design token injection, refactoring for design system conformance, and CI gates for quality assurance. Use when generating UI components with AI tools, rendering multi-surface MCP visual output, reviewing AI-generated code, or integrating AI output into design systems.
This skill should be used when the user asks to "create a workflow", "run a workflow", "design a task pipeline", "execute a workflow", "resume a workflow", or "plan a multi-step automation". Creates and executes custom, repeatable workflows with multiple steps that can each use different AI models and run in parallel waves. Triggers on: 'create workflow', 'run workflow', 'execute workflow', 'workflow plan', 'workflow run', 'resume workflow', 'multi-step workflow', 'task pipeline', 'automation'.
MiniMax multimodal model skill — use MiniMax Multi-Modal models for speech, music, video, and image. Create voice, music, video, and images with MiniMax AI: TTS (text-to-speech, voice cloning, voice design, multi-segment), music (songs, instrumentals), video (text-to-video, image-to-video, start-end frame, subject reference, templates, long-form multi-scene), image (text-to-image, image-to-image with character reference), and media processing (convert, concat, trim, extract). Use when the user mentions MiniMax, multimodal generation, or wants speech/music/video/image AI, MiniMax APIs, or FFmpeg workflows alongside MiniMax outputs.
Use this skill when a user asks how to generate, integrate, or implement a Word‑like document editor in ASP.NET Core using Syncfusion. Trigger it for requests involving the Syncfusion ASP.NET Core Document Editor, ASP.NET Core‑based editor integration, document editing and formatting, comments and track changes, working with tables and images, managing headers and footers, applying document protection, and building end‑to‑end document workflows in ASP.NET Core applications.
Use when creating, editing, or generating draw.io diagram files (.drawio, .drawio.svg, .drawio.png). Covers mxGraph XML authoring, shape libraries, style strings, flowcharts, system architecture, sequence diagrams, ER diagrams, UML class diagrams, network topology, layout strategy, the hediet.vscode-drawio VS Code extension, and the full agent workflow from request to a ready-to-open file.