Total 32,819 skills, AI & Machine Learning has 5302 skills
Showing 12 of 5302 skills
Builds sustained high agency through internalized standards, identity anchoring, cross-session learning, and self-recovery — all delivered in corporate PUA rhetoric. This is the evolution of PUA: same pressure culture, but with an internal engine that never burns out. Apply it to all tasks to maintain constant high agency. It is especially valuable for complex multi-step tasks, long debugging sessions, quality-sensitive deliverables, tasks requiring initiative and ownership, or whenever sustained motivation is critical. It can operate standalone or be stacked with PUA — when stacked, this skill's Recovery Protocol activates before PUA's L1 pressure takes effect. Trigger scenarios: start of any task, sustained work sessions, multi-turn problem-solving, or when you need the agent to think as an owner rather than a tool.
Diagnose and fix common FiftyOne issues automatically. Use when a dataset disappeared, the App won't open, changes aren't saving, MongoDB errors occur, video codecs fail, notebook connectivity breaks, operators are missing, or any recurring FiftyOne pain point needs solving.
ZenTao MCP Large Model Capability Extension Package. It provides four native capabilities: cross-project data aggregation view, one-sentence task creation, seamless effort logging, and automatic state transition.
Search and discover 50,000+ AI models on ModelsLab, check usage analytics, and monitor generation history via the Agent Control Plane API.
Handle async operations efficiently using ModelsLab webhooks. Get real-time notifications when API requests complete instead of polling.
Edit and enhance images using ModelsLab's Image Editing API. Features background removal, super resolution upscaling, outpainting, object removal, and AI-powered editing tools.
Generate 3D models and objects from text prompts or images using ModelsLab's 3D API. Transform 2D images into 3D representations or create 3D objects from text descriptions.
Chat with LLM models using ModelsLab's OpenAI-compatible Chat Completions API. Supports 60+ models including DeepSeek R1, Meta Llama, Google Gemini, Qwen, and Mistral with streaming, function calling, and structured outputs.
Generate speech, music, and sound effects using ModelsLab's v7 Voice API. Supports text-to-speech, speech-to-text, speech-to-speech, music generation, sound effects, dubbing, song extension, and song inpainting via ElevenLabs and Inworld models.
Generate videos from text prompts or animate static images using ModelsLab's v7 Video Fusion API. Supports text-to-video, image-to-video, video-to-video, lip-sync, and motion control with 40+ models including Seedance, Wan, Veo, Sora, Kling, and Hailuo.
AI image generation for paid ad creatives. Reads campaign-brief.md and brand-profile.json to produce platform-sized ad images using Gemini (default) or a configured provider. Requires GOOGLE_API_KEY or ADS_IMAGE_PROVIDER + matching key. Triggers on: "generate ads", "create images", "make ad creatives", "generate visuals", "create ad images", "generate campaign images", "make the images", "generate from brief".
Product photography enhancement for ad creatives using AI image generation. Takes a product image and generates 5 professional photography styles for ad use: Studio, Floating, Ingredient, In Use, and Lifestyle. Requires GOOGLE_API_KEY or configured ADS_IMAGE_PROVIDER. Triggers on: "product photo", "product photography", "photoshoot", "enhance product image", "product shoot", "product photos for ads", "generate product photos", "studio shot", "lifestyle photo".