aliyun-pixverse-generation
Original:🇺🇸 English
Translated
1 scriptsChecked / no sensitive code detected
Use when generating videos with Alibaba Cloud Model Studio PixVerse models (`pixverse/pixverse-v5.6-t2v`, `pixverse/pixverse-v5.6-it2v`, `pixverse/pixverse-v5.6-kf2v`, `pixverse/pixverse-v5.6-r2v`). Use when building non-Wan text-to-video, first-frame image-to-video, keyframe-to-video, or multi-image reference-to-video workflows on Model Studio.
3installs
Sourcecinience/alicloud-skills
Added on
NPX Install
npx skill4agent add cinience/alicloud-skills aliyun-pixverse-generationTags
Translated version includes tags in frontmatterSKILL.md Content
View Translation Comparison →Category: provider
Model Studio Aishi Video Generation
Validation
bash
mkdir -p output/aliyun-pixverse-generation
python -m py_compile skills/ai/video/aliyun-pixverse-generation/scripts/prepare_aishi_request.py && echo "py_compile_ok" > output/aliyun-pixverse-generation/validate.txtPass criteria: command exits 0 and is generated.
output/aliyun-pixverse-generation/validate.txtOutput And Evidence
- Save normalized request payloads, chosen model variant, and task polling snapshots under .
output/aliyun-pixverse-generation/ - Record region, resolution/size, duration, and whether audio generation was enabled.
Use Aishi when the user explicitly wants the non-Wan PixVerse family for video generation.
Critical model names
Use one of these exact model strings:
pixverse/pixverse-v5.6-t2vpixverse/pixverse-v5.6-it2vpixverse/pixverse-v5.6-kf2vpixverse/pixverse-v5.6-r2v
Selection guidance:
- Use for text-only generation.
pixverse/pixverse-v5.6-t2v - Use for first-frame image-to-video.
pixverse/pixverse-v5.6-it2v - Use for first-frame + last-frame transitions.
pixverse/pixverse-v5.6-kf2v - Use for multi-image character/style consistency.
pixverse/pixverse-v5.6-r2v
Prerequisites
- This family currently only supports China mainland (Beijing).
- Install SDK or call HTTP directly:
bash
python3 -m venv .venv
. .venv/bin/activate
python -m pip install dashscope- Set in your environment, or add
DASHSCOPE_API_KEYtodashscope_api_key.~/.alibabacloud/credentials
Normalized interface (video.generate)
Request
- (string, required)
model - (string, optional for
prompt, required for other variants)it2v - (array<object>, optional)
media - (string, optional): direct pixel size such as
size, used by1280*720andt2vr2v - (string, optional):
resolution/360P/540P/720P, used by1080Pandit2vkf2v - (int, required):
duration/5/8, except 1080P only supports10/58 - (bool, optional)
audio - (bool, optional)
watermark - (int, optional)
seed
Response
- (string)
task_id - (string)
task_status - (string, when finished)
video_url
Endpoint and execution model
- Submit task:
POST https://dashscope.aliyuncs.com/api/v1/services/aigc/video-generation/video-synthesis - Poll task:
GET https://dashscope.aliyuncs.com/api/v1/tasks/{task_id} - HTTP calls are async only and must set header .
X-DashScope-Async: enable
Quick start
Text-to-video:
bash
python skills/ai/video/aliyun-pixverse-generation/scripts/prepare_aishi_request.py \
--model pixverse/pixverse-v5.6-t2v \
--prompt "A compact robot walks through a rainy neon alley." \
--size 1280*720 \
--duration 5Image-to-video:
bash
python skills/ai/video/aliyun-pixverse-generation/scripts/prepare_aishi_request.py \
--model pixverse/pixverse-v5.6-it2v \
--prompt "The turtle swims slowly as the camera rises." \
--media image_url=https://example.com/turtle.webp \
--resolution 720P \
--duration 5Operational guidance
- and
t2vuser2v;sizeandit2vusekf2v.resolution - For , provide exactly one
kf2vand onefirst_frame.last_frame - For , you can pass up to 7 reference images.
r2v - Aishi returns task IDs first; do not treat the initial response as the final video result.
Output location
- Default output:
output/aliyun-pixverse-generation/request.json - Override base dir with .
OUTPUT_DIR
References
references/sources.md