Loading...
Loading...
Perform image understanding and analysis using GLM Vision MCP. Trigger conditions: (1) Users request image analysis, image understanding, or description of image content (2) Need to identify objects, text, or scenes in images (3) Use GLM's visual understanding capabilities
npx skill4agent add thincher/awsome_skills glm-understand-imagenpx -y mcporter --versioncat ~/.openclaw/config/glm.json 2>/dev/null | python3 -c "import json,sys; d=json.load(sys.stdin); print(d.get('api_key', ''))"mkdir -p ~/.openclaw/config
cat > ~/.openclaw/config/glm.json << EOF
{
"api_key": "API_KEY"
}
EOFmcporter config add glm-vision \
--command "npx -y @z_ai/mcp-server" \
--env Z_AI_API_KEY="your-key" \
--env Z_AI_MODE="ZHIPU" \
--env HOME="$PWD"your-keyHOMEmcporter listglm-vision~/.openclaw/workspace/images/image-name.jpgmcporter call glm-vision.analyze_image prompt="<Question about the image>" image_source="<Image path or URL>"# Describe image content
mcporter call glm-vision.analyze_image prompt="Describe the content of this image in detail" image_source="~/image.jpg"
# Use URL
mcporter call glm-vision.analyze_image prompt="What does this image show?" image_source="https://example.com/image.jpg"
# Extract text from image
mcporter call glm-vision.extract_text_from_screenshot image_source="~/screenshot.png"
# Diagnose error screenshot
mcporter call glm-vision.diagnose_error_screenshot prompt="Analyze this error" image_source="~/error.png"| Parameter | Description | Type |
|---|---|---|
| image_source | Image path or URL | string (Required) |
| prompt | Question about the image | string (Required) |
ui_to_artifactextract_text_from_screenshotdiagnose_error_screenshotunderstand_technical_diagramanalyze_data_visualizationui_diff_checkanalyze_imagevideo_analysisglm-vision@z_ai/mcp-serverZ_AI_API_KEYZ_AI_MODEZHIPU