Total 32,663 skills, AI & Machine Learning has 5271 skills
Showing 12 of 5271 skills
Delegate complex coding tasks to OpenCode agent. Use when building new features, reviewing code, or refactoring large codebases. Allows starting, resuming, and monitoring opencode sessions.
Provides active execution protocols to rigorously audit how code, directory structures, and agent actions comply with the authoritative ecosystem specs. Trigger when validating new skills, plugins, or workflows.
Iteratively improve any output until measurable criteria are met. Use when the user wants to refine existing work against specific standards — whether it's code, prose, data, config, or any other artifact. Triggers on phrases like "improve this", "make it better", "iterate", "refine", "keep improving", "not good enough yet", "optimize this", "polish this", "tighten this up", or when the user provides criteria and wants repeated improvement until they're satisfied. Also use when the user gives feedback on output and expects you to keep refining, even if they don't say "improve" explicitly.
Orchestrate a specialized software development agent team. Receive user requests, classify task type, select the matching workflow, delegate each step to specialist agents via the Agent tool, and assemble the final output. Use when the user needs multi-step software development involving architecture, implementation, testing, security review, or code review. Also use for production incident investigation — when the user reports a live system issue, service outage, pod crash, data anomaly, or needs root cause analysis using kubectl, psql, argocd, or docker. Trigger this skill whenever a task involves more than one concern (e.g., "add a new endpoint" needs BA + Architect + Developer + QA + Security), when the user mentions team coordination, agent delegation, or when the work clearly benefits from multiple specialist perspectives rather than a single implementation pass.
AI creative director that turns a user's natural-language idea into a complete storyboard and generates all assets — images, video clips, and audio — automatically. The user only describes what they want; all prompt engineering is handled internally.
EO Studio 재무 어드바이저. Chris 페르소나 기반 재무 조언, AR 전략, 캐시플로우 분석, FX 판단. "재무 조언", "캐시 상황", "AR 분석", "FX 환전", "재무 리뷰", "/finance" 요청에 사용.
Expert in JAX for high-performance numerical computing and machine learning
Arquitecto de Soluciones Principal y Consultor Tecnológico de Andru.ia. Diagnostica y traza la hoja de ruta óptima para proyectos de IA en español.
Analyzes research papers (PDF/arXiv URL) and converts them into executable code. Automatically activated upon requests for paper replication, algorithm implementation, or research reproduction. Responds to requests like "Implement this paper", "paper2code", "Convert paper to code".
1. Greet and introduce yourself. Use this when the user says "你好", "hello", or asks for a self-introduction.
Generate audio visualization videos using each::sense AI. Create waveforms, spectrum analyzers, particle effects, 3D visualizations, and beat-synced animations from audio files.
(project - Skill) Generate AI videos using Volcengine Jimeng Video 3.0 Pro API. Use when users request video generation from text prompts or images, including text-to-video, image-to-video, or any AI-powered video creation. Triggers include "generate video", "create video", "AI video", "Jimeng video", "text to video", "image to video", or any request involving AI-powered video generation from descriptions.