Loading...
Loading...
Found 9 Skills
Generate and edit images with OpenAI GPT-Image-2 via inference.sh CLI. Models: GPT-Image-2. Capabilities: text-to-image, image editing, inpainting, mask-based editing, multi-image reference, batch generation. Use for: product mockups, marketing visuals, image editing, concept art, inpainting, photo manipulation. Triggers: gpt image, gpt-image-2, openai image, chatgpt image, dall-e, dalle, openai image generation, gpt image edit, gpt inpainting, openai dall-e, gpt 4o image
Standalone multi-agent image generation skill for Hermes. Includes an internal design compiler, GPT-Image-2 generation via apimart.ai, case library reuse, interactive reference selection, batch workflows, and style-consistent series generation.
Novel Cover Generation. Automatically analyze the genre style based on the book title and author's name, call GPT-Image-2 to directly generate a professional web novel cover with title and signature. Trigger methods: /story-cover, /封面, "Help me make a cover", "Generate cover image", "Make a novel cover", "Cover design"
Image Generation Skill: Use this skill when users need to generate images, visual infographics, create graphics, or edit/modify/adjust existing images. Based on the official formal version of the ChatGPT Image 2 model (gpt-image-2) from Apiyi Platform (https://api.apiyi.com/). This model supports precise size/quality control (including 4K) and is billed by token. Key differences from gpt-image-2-all (official reverse version): Uses /v1/images/generations and /v1/images/edits endpoints; Has explicit size parameter; Has quality parameter; Billed by token; Uses multipart/form-data to upload reference images; b64_json is pure base64 without prefix.
AI Image Generation Skill, using the latest ChatGPT image generation model gpt-image-2-all. This skill is applied when users need to generate images, visual infographics, create graphics, or edit/modify/adjust existing images. Based on the image generation service of the latest ChatGPT image generation model gpt-image-2-all from APIYI Platform (https://api.apiyi.com/), no external network access is required. The model is charged per image at $0.03 per piece, supporting text-to-image generation, single image editing, multi-image fusion, and natural language-based image modification, with high text restoration accuracy and friendly Chinese prompts. The size is controlled by prompt description (no explicit size parameter). Key differences from NanoBanana2: no size parameter, need to describe the size at the beginning of the prompt; unified $0.03 per image with no resolution tiering; the conversational endpoint /v1/chat/completions is the recommended one.
Single-image generation skill for posters, key art, and editorial illustrations. Defaults to gpt-image-2 but is provider-agnostic — the same workflow drives Flux, Imagen, or Midjourney via the active upstream tooling. Output is one or more PNG/JPEG files saved to the project folder.
Upgrade a coded website to award-tier, editorially-crafted design using fal.ai. Takes a local HTML file or a dev-server URL, screenshots it, has an opus-4.7 vision model write a gpt-image-2 edit prompt, uses fal-ai/gpt-image-2/edit to produce the redesigned reference image, then opus-4.7 vision writes a Markdown build-spec with a "Hard constraints" section + a tokens.json. Also supports iterate (screenshot implemented site → delta-spec vs reference) and greenfield generate (brief → mockup → single-file HTML). Invoke when the user says "improve the design", "make it world-class", "redesign this landing page", "upgrade this site", "design pass", or points at a local HTML / dev server for a visual review.
Generate brand-quality product images via mode-specific prompt enhancement on Higgsfield's gpt_image_2 model. The single entry point for any professional brand visual involving a product. Use when: "make a product photo", "studio shot", "lifestyle photo", "in use", "Pinterest pin", "hero banner", "website header", "carousel", "Meta ads", "ad creatives", "model wearing", "virtual try-on", "person holding product", "closeup with hands", "levitating product", "floating", "splash shot", "CGI style", "surreal product", "restyle", "Christmas version", "in [aesthetic] style", or any request involving a product, brand, or paid social creative. Modes: product_shot, lifestyle_scene, closeup_product_with_person, pinterest_pin, hero_banner, social_carousel, ad_creative_pack, virtual_model_tryout, conceptual_product, restyle. Backend assembles the final prompt — never write gpt_image_2 prompts freehand. Always go through this skill. NOT for: raw text-to-image with no brand/product (use higgsfield-generate), branded marketing video with avatars (use higgsfield-generate's Marketing Studio), Soul Character training (use higgsfield-soul-id).
Conversation-first, image-first PPT generation workflow skill using GPT Image 2 for full-page visual slides packaged into PPTX files.