Loading...
Loading...
Found 913 Skills
Arquitecto de soluciones digitales basadas en IA. Dos modos: (1) ANALIZAR repositorios o código existente y explicar su arquitectura para cualquier audiencia, incluyendo personas sin conocimiento técnico. (2) DISEÑAR la arquitectura completa de sistemas nuevos que usan LLMs, RAG, agentes o fine-tuning. Usa este skill cuando el usuario mencione: arquitectura de IA, diseño de sistema con LLM, capas arquitectónicas, RAG architecture, tech stack para IA, vector database, diagrama de arquitectura, componentes del sistema, embedding, retrieval, pipeline de datos, MLOps, LLMOps, evaluar enfoques, RAG vs fine-tuning, diseñar solución de inteligencia artificial, explicar repositorio, explicar código, analizar proyecto, qué hace este repo, cómo funciona este sistema, explícame este proyecto, o cualquier variación de "qué componentes necesito" o "explícame cómo funciona esto". Actívalo cuando el usuario pegue código, README, estructura de archivos, o mencione un repositorio de GitHub para analizar. También cuando quiera diseñar arquitectura nueva.
Build Retrieval-Augmented Generation (RAG) systems for LLM applications with vector databases and semantic search. Use when implementing knowledge-grounded AI, building document Q&A systems, or integrating LLMs with external knowledge bases.
Framework for building LLM-powered applications with agents, chains, and RAG. Supports multiple providers (OpenAI, Anthropic, Google), 500+ integrations, ReAct agents, tool calling, memory management, and vector store retrieval. Use for building chatbots, question-answering systems, autonomous agents, or RAG applications. Best for rapid prototyping and production deployments.
Expert guidance for deep learning, transformers, diffusion models, and LLM development with PyTorch, Transformers, Diffusers, and Gradio.
Guide for creating high-quality MCP (Model Context Protocol) servers that enable LLMs to interact with external services through well-designed tools. Use when building MCP servers to integrate external APIs or services, whether in Python (FastMCP) or Node/TypeScript (MCP SDK).
MiniMax API via curl. Use this skill for Chinese LLM chat, text-to-speech, and AI video generation.
Meta's 86M prompt injection and jailbreak detector. Filters malicious prompts and third-party data for LLM apps. 99%+ TPR, <1% FPR. Fast (<2ms GPU). Multilingual (8 languages). Deploy with HuggingFace or batch processing for RAG security.
Apple FoundationModels framework for on-device LLM — text generation, guided generation with @Generable, tool calling, and snapshot streaming in iOS 26+.
Ultra-lightweight AI assistant in Go that runs on $10 hardware with <10MB RAM, supporting multiple LLM providers, tools, and single-binary deployment across RISC-V, ARM, MIPS, and x86.
Develop, debug, and optimize SGLang LLM serving engine. Use when the user mentions SGLang, sglang, srt, sgl-kernel, LLM serving, model inference, KV cache, attention backend, FlashInfer, MLA, MoE routing, speculative decoding, disaggregated serving, TP/PP/EP, radix cache, continuous batching, chunked prefill, CUDA graph, model loading, quantization FP8/GPTQ/AWQ, JIT kernel, triton kernel SGLang, or asks about serving LLMs with SGLang.
Expert skill for Token-Oriented Object Notation (TOON) — compact, schema-aware JSON encoding for LLM prompts that reduces tokens by ~40%.
Turn a refined research proposal or method idea into a detailed, claim-driven experiment roadmap. Use after `research-refine`, or when the user asks for a detailed experiment plan, ablation matrix, evaluation protocol, run order, compute budget, or paper-ready validation that supports the core problem, novelty, simplicity, and any LLM / VLM / Diffusion / RL-based contribution.