Loading...
Loading...
Found 24 Skills
Prepare datasets and configure LoRA training for character consistency. Covers FLUX (AI-Toolkit, SimpleTuner, FluxGym) and SDXL (Kohya_ss) training with step-by-step guidance. Use when training custom character LoRAs.
Conducts a guided conversational interview with the user about the image or scene they want to create, asking pertinent questions to fully understand their vision, then synthesizes a perfect model-appropriate prompt with recommended settings. Use when the user says "I want to create...", "help me make an image of...", "I have an idea for...", or asks for help crafting a prompt.
ComfyUI custom node project structure - directory layout, __init__.py, registration, requirements.txt, publishing, WEB_DIRECTORY. Use when setting up a new custom node project, packaging nodes, or publishing to the registry.
Manage video production projects including character profiles, project manifests, workflow history, and asset tracking. Use when creating new projects, managing characters, or tracking production state.
Master AI-powered game asset pipelines using ComfyUI, Stable Diffusion, FLUX, ControlNet, and IP-Adapter. Creates production-ready sprites, textures, UI, and environments with consistency, proper licensing, and game engine integration. Use when "AI game art, generate game assets, ComfyUI game, stable diffusion sprites, AI texture generation, character consistency AI, procedural art generation, SDXL game assets, FLUX textures, train LoRA game, AI tileable texture, spritesheet generation, " mentioned.
Generate images, videos, audio, and 3D models via RunningHub API (170+ endpoints) and run any RunningHub AI Application (custom ComfyUI workflow) by webappId. Covers text-to-image, image-to-video, text-to-speech, music generation, 3D modeling, image upscaling, AI apps, and more.
Use when editing ComfyUI workflow JSON, adding nodes, wiring connections, modifying workflows, adding ControlNet/LoRA/upscaling to a workflow, or submitting workflows to ComfyUI.
Publish assembled videos to YouTube and other platforms. Orchestrates existing youtube-uploader, youtube-strategy, and youtube-plan-new-video skills. Use when ready to publish or plan distribution for completed videos.
Full production pipeline — story to scenes, Z-Image start frames, Qwen Edit end frames, WAN FLF video clips, ffmpeg concatenation
Xiaohongshu End-to-End Content Workflow: From data collection to AI analysis to content generation, using the Kimi K2 model. Supports web crawling, analysis, note generation, and AI image prompt generation.
Expert guidance for integrating ViewComfy API into web applications using Python and FastAPI
Configure MeiGen plugin provider and API keys. Use this when the user runs /meigen:setup, asks to "configure meigen", "set up image generation", "add API key", or needs help configuring the plugin.