Total 30,905 skills, AI & Machine Learning has 4990 skills
Showing 12 of 4990 skills
ALWAYS use this skill before answering brainstorming, ideation, prompt crafting, or open-ended exploration requests. Transforms vague requests into actionable outputs via adaptive guided questioning — triages into Prompt Mode (craft/improve prompts), Explore Mode (brainstorm ideas), or Focused Mode (specific problem strategies). Trigger when user says: "brainstorm", "ช่วยคิด", "help me think", "I have an idea", "improve this prompt", "let's explore", "I want to build", "I'm thinking about", "brainstorm วิธี", "ช่วยคิดหน่อย", "อยากทำ", "ยังไม่รู้จะทำอะไร", "not sure about the approach", "help me figure out", "what should I". Also trigger for: side projects, career decisions, project planning, migration strategies, architecture decisions, cost optimization, or any request where the user hasn't decided direction yet and would benefit from structured discovery. Do NOT skip — this skill adapts depth automatically (2-7 questions) and produces BETTER results by asking targeted questions first.
Generate audio visualization videos using each::sense AI. Create waveforms, spectrum analyzers, particle effects, 3D visualizations, and beat-synced animations from audio files.
Patterns for building AI agents that learn from their own execution, detect failure modes, and improve autonomously. Covers feedback loops, performance regression detection, memory curation, skill extraction, and meta-learning architectures. Use when building agents that need to get better over time, managing auto-memory, or designing self-correcting systems.
(project - Skill) Generate AI videos using Volcengine Jimeng Video 3.0 Pro API. Use when users request video generation from text prompts or images, including text-to-video, image-to-video, or any AI-powered video creation. Triggers include "generate video", "create video", "AI video", "Jimeng video", "text to video", "image to video", or any request involving AI-powered video generation from descriptions.
Analyze video content using visual/video large models. This tool is triggered when the user uses phrases like "analyze video", "video understanding", "look at this video", or "analyze video".
Build and deploy agentic finance applications on the Alva platform. Access 250+ financial data sources (crypto, equities, macro, on-chain, social), run cloud-side analytics, backtest trading strategies, and release interactive playbooks -- all from your AI agents.
A guided, zero-friction installer and maintenance assistant for OpenClaw. Use this skill when the user wants to install OpenClaw, set up OpenClaw on a local machine or remote server, connect OpenClaw to DingTalk, get OpenClaw skill recommendations for their use case, or perform post-installation maintenance (health checks, troubleshooting, installing new skills, changing AI models, adding chat channels, updating OpenClaw). Handles full environment detection, installation, optional DingTalk integration, scene-based skill recommendations, and daily maintenance — all interactively, with no wasted steps.
Japanese version of the PUA Universal Motivation Engine. It compels exhaustive problem-solving using corporate PUA rhetoric and structured debugging methodology in Japanese. MUST trigger under the following conditions: (1) Any task has failed 2+ times, or you're stuck in a loop of tweaking the same approach; (2) You're about to say 'I cannot', suggest manual handling to the user, or blame the environment without verification; (3) You find yourself being passive — not searching, not reading source code, not verifying, just waiting for instructions; (4) The user expresses frustration in any form: 'try harder', 'stop giving up', 'figure it out', 'why isn't this working', 'again???', 'もっと頑張れ', 'なんでまた失敗したの', 'もう一回やって', 'なんとかしろ', or any similar sentiment regardless of phrasing. It should also trigger when facing complex multi-step debugging, environment issues, configuration problems, or deployment failures where early surrender is tempting. Applies to ALL task types: code, configuration, research, writing, deployment, infrastructure, API integration. DO NOT trigger on first-attempt failures or when a known fix is already executing successfully.
AI-assisted video editing workflows for cutting, structuring, and augmenting real footage. Covers the full pipeline from raw capture through FFmpeg, Remotion, ElevenLabs, fal.ai, and final polish in Descript or CapCut. Use when the user wants to edit video, cut footage, create vlogs, or build video content.
Neural search via Exa MCP for web, code, and company research. Use when the user needs web search, code examples, company intel, people lookup, or AI-powered deep research with Exa's neural search engine.
Anthropic Claude API patterns for Python and TypeScript. Covers Messages API, streaming, tool use, vision, extended thinking, batches, prompt caching, and Claude Agent SDK. Use when building applications with the Claude API or Anthropic SDKs.
Unified media generation via fal.ai MCP — image, video, and audio. Covers text-to-image (Nano Banana), text/image-to-video (Seedance, Kling, Veo 3), text-to-speech (CSM-1B), and video-to-audio (ThinkSound). Use when the user wants to generate images, videos, or audio with AI.