Total 30,708 skills, AI & Machine Learning has 4959 skills
Showing 12 of 4959 skills
AI audio generation powered by CellCog. Text-to-speech, voice synthesis, voiceovers, podcast audio, narration, music generation, background music, sound design. Professional audio creation with AI.
Use when asked to detect silent failures/weak error handling or explicitly asked to run the silent-failure-hunter subagent.
A guide to creating efficient Skills. Use this skill when users need to create a new skill (or update an existing one) to extend Claude's capabilities through expertise, workflows, or tool integrations.
CX Workflow - Configuration Management. View and modify configuration fields in config.json, such as developer_id, github_sync mode, code review toggle, agent team mode, auto-formatting, etc. Trigger words: config, settings, modify config, change mode. Only execute when the user explicitly calls /cx-config. Do not trigger automatically.
Perplexity AI search and research. Use for AI search.
XGBoost gradient boosting library. Use for tabular ML.
Mistral AI efficient open models. Use for efficient AI.
NLTK natural language toolkit. Use for NLP.
Convert HuggingFace transformer models to ONNX format for browser inference with Transformers.js and WebGPU. Use when given a HuggingFace model link to convert to ONNX, when setting up optimum-cli for ONNX export, when quantizing models (fp16, q8, q4) for web deployment, when configuring Transformers.js with WebGPU acceleration, or when troubleshooting ONNX conversion errors. Triggers on mentions of ONNX conversion, Transformers.js, WebGPU inference, optimum export, model quantization for browser, or running ML models in the browser.
Expert in building comprehensive AI systems, integrating LLMs, RAG architectures, and autonomous agents into production applications. Use when building AI-powered features, implementing LLM integrations, designing RAG pipelines, or deploying AI systems.
MLflow ML lifecycle management. Use for ML experiment tracking.
Ollama local LLM deployment and management. Use for running LLMs locally.