giggle-generation-video
Original:🇺🇸 English
Translated
1 scriptsChecked / no sensitive code detected
Supports text-to-video and image-to-video (start/end frame). Use when the user needs to generate video, create short videos, or convert text to video. Use cases: (1) Generate video from text description, (2) Use reference images as start/end frame for image-to-video, (3) Customize model, aspect ratio, duration, resolution. Triggers: generate video, text-to-video, image-to-video, AI video.
3installs
Sourcegiggle-official/skills
Added on
NPX Install
npx skill4agent add giggle-official/skills giggle-generation-videoTags
Translated version includes tags in frontmatterSKILL.md Content
View Translation Comparison →Giggle Video Generation
Source: giggle-official/skills · API: giggle.pro
Generates AI videos via giggle.pro's Generation API. Supports text-to-video and image-to-video. Submit task → query when ready. No polling, no Cron, no file writes—all operations via exec.
Installation Requirements
| Requirement | Value |
|---|---|
| Binary | |
| Environment | |
| Pip | |
Set before use. The script will prompt if not configured.
GIGGLE_API_KEYNo Retry on Error: If script execution encounters an error, do not retry. Report the error to the user directly and stop.
Supported Models
| Model | Supported Durations (s) | Default | Description |
|---|---|---|---|
| grok | 6, 10 | 6 | Strong overall capability, recommended |
| grok-fast | 6, 10 | 6 | grok fast version |
| sora2 | 4, 8, 12 | 4 | OpenAI Sora 2 |
| sora2-pro | 4, 8, 12 | 4 | Sora 2 Pro |
| sora2-fast | 10, 15 | 10 | Sora 2 Fast |
| sora2-pro-fast | 10, 15 | 10 | Sora 2 Pro Fast |
| kling25 | 5, 10 | 5 | Kling video model |
| seedance15-pro | 4, 8, 12 | 4 | Seedance Pro (with audio) |
| seedance15-pro-no-audio | 4, 8, 12 | 4 | Seedance Pro (no audio) |
| veo31 | 4, 6, 8 | 4 | Google Veo 3.1 (with audio) |
| veo31-no-audio | 4, 6, 8 | 4 | Google Veo 3.1 (no audio) |
| minimax23 | 6 | 6 | MiniMax model |
| wan25 | 5, 10 | 0 | Wanxiang model |
Note: must be chosen from the model's supported durations, otherwise the API will error.
--durationFrame Reference (Image-to-Video)
For image-to-video, and support three mutually exclusive formats:
--start-frame--end-frame| Method | Format | Example |
|---|---|---|
| asset_id | | |
| url | | |
| base64 | | |
Each frame parameter can only use one of these methods.
Execution Flow: Submit and Query
Video generation is asynchronous (typically 60–300 seconds). Submit a task to get , then query when the user wants to check status. All commands run via ; API key from system env.
task_idexecStep 1: Submit Task
First send a message to the user: "Video generation submitted. Usually takes 1–5 minutes. You can ask me about the progress anytime."
bash
# Text-to-video (default grok-fast)
python3 scripts/generation_api.py \
--prompt "Camera slowly pushes forward, person smiling in frame" \
--model grok-fast --duration 6 \
--aspect-ratio 16:9 --resolution 720p
# Image-to-video - use asset_id as start frame
python3 scripts/generation_api.py \
--prompt "Person slowly turns around" \
--start-frame "asset_id:lkllv0yv81" \
--model grok-fast --duration 6 \
--aspect-ratio 16:9 --resolution 720p
# Image-to-video - use URL as start frame
python3 scripts/generation_api.py \
--prompt "Scenery from still to motion" \
--start-frame "url:https://example.com/img.jpg" \
--model grok-fast --duration 6
# Image-to-video - both start and end frame
python3 scripts/generation_api.py \
--prompt "Scene transition" \
--start-frame "asset_id:abc123" \
--end-frame "url:https://example.com/end.jpg" \
--model grok --duration 6Response example:
json
{"status": "started", "task_id": "55bf24ca-e92a-4d9b-a172-8f585a7c5969"}Store task_id in memory ():
addMemorygiggle-generation-video task_id: xxx (submitted: YYYY-MM-DD HH:mm)Step 2: Query When User Asks
When the user asks about video progress (e.g. "is my video ready?", "progress?"), run:
bash
python3 scripts/generation_api.py --query --task-id <task_id>Output handling:
| stdout pattern | Action |
|---|---|
| Plain text with video links (视频已就绪) | Forward to user as-is |
| Plain text with error | Forward to user as-is |
JSON | Tell user "Still in progress, please ask again in a moment" |
Link return rule: Video links in results must be full signed URLs (with Policy, Key-Pair-Id, Signature query params). Keep as-is when forwarding.
New Request vs Query Old Task
When the user initiates a new video generation request, must run Step 1 to submit a new task. Do not reuse old task_id from memory.
Only when the user explicitly asks about a previous task's progress should you query the old task_id from memory.
Parameter Reference
| Parameter | Default | Description |
|---|---|---|
| required | Video description prompt |
| grok | See "Supported Models" table |
| model default | Must choose from model's supported durations |
| 16:9 | 16:9, 9:16, 1:1, 3:4, 4:3 |
| 720p | 480p, 720p, 1080p |
| - | Image-to-video start frame: |
| - | Image-to-video end frame, same format as start |
Note: base64 parameter supports base64-encoded images. Pass the raw Base64 string directly, do not add the prefix.
data:image/xxx;base64,Interaction Guide
When the user request is vague, guide per the steps below. If the user has provided enough info, run the command directly.
Step 1: Model Selection (required)
Before generating, must introduce available models and let the user choose. Display the model list from "Supported Models" table. Wait for explicit user choice before continuing.
Step 2: Video Duration
For the chosen model, show supported duration options. Default to the model's default duration.
Step 3: Generation Mode
Question: "Do you need reference images as start/end frame?"
Options: No - text-to-video only / Yes - image-to-video (set start/end frame)Step 4: Aspect Ratio
Question: "What aspect ratio do you need?"
Options: 16:9 - Landscape (recommended) / 9:16 - Portrait / 1:1 - SquareStep 5: Execute and Display
Follow the flow: send message → Step 1 submit → user asks → Step 2 query. Forward exec stdout to the user as-is.