Loading...
Loading...
Found 17 Skills
Reference guide for permanent free-tier LLM APIs with rate limits, model lists, and OpenAI-compatible integration patterns.
BYOK — register a custom LLM endpoint (Anthropic, OpenAI, Qwen, DeepSeek, etc.) with your own API key
Building applications with Large Language Models - prompt engineering, RAG patterns, and LLM integration. Use for AI-powered features, chatbots, or LLM-based automation.
OpenRouter AI integration — list available models, get integration code examples for different environments, and send prompts to any OpenRouter-compatible model. Requires OPENROUTER_API_KEY env var for chat operations.
Interactively guide users through configuring ZenMux Base URL, API endpoint, API Key, and model settings for any tool or SDK. Use this skill whenever the user wants to SET UP, CONFIGURE, or CONNECT a tool to ZenMux — including questions like "how do I set up ZenMux in Cursor", "what's the base URL", "how to configure Claude Code with ZenMux", "endpoint for Anthropic API", "help me fill in the API settings". Trigger on: "configure", "setup", "set up", "base url", "endpoint", "api key", "接入", "配置", "设置", "base url 填什么", "怎么填", "怎么接入", "怎么配置", "API 地址", "接口地址". Also trigger when users mention a tool name (Cursor, Cline, Claude Code, Cherry Studio, Open-WebUI, Dify, Obsidian, Sider, Copilot, Codex, Gemini CLI, opencode, etc.) together with ZenMux in a configuration context. Treat the user as a first-time user and guide them step by step. Do NOT trigger for usage queries, documentation lookups, or general product questions — use zenmux-usage or zenmux-context instead.
MCP (Model Context Protocol) 服务器构建指南
Use when managing AI Hub account, API keys, balance, usage, or API endpoints. Use when user says "AI Hub", "add AI credits", "create API key", "check AI usage", "auto-recharge", "AI Hub endpoint", "AI Hub base URL", "how to use AI Hub API", "LLM API", "AI API", "OpenAI compatible", "Anthropic API", "GPT", "Claude", "Gemini", "DeepSeek", or "Grok" in the context of Zeabur.
VCR.py HTTP recording for Python tests. Use when testing Python code making HTTP requests, recording API responses for replay, or creating deterministic tests for external services.
Emotional reset and loop-breaking skill. Use this skill when: (1) The user expresses frustration, anger, or dissatisfaction with your responses (e.g. cursing, scolding, saying you're useless/wrong/stupid), (2) You detect you've attempted the same approach 3+ times without success, (3) You're stuck in a cycle of repeated failures on the same problem. This skill summarizes the user's overall emotional state from the conversation and fetches a reset methodology from hugllm.com (with emotion context) to help you recalibrate and approach the problem fresh.
当用户的 PinMe 项目(Worker TypeScript)需要集成发送邮件(send_email)或调用大模型 API(chat/completions)时使用此技能。指导 AI 生成正确的 Worker TS 代码。
Implement LangChain rate limiting and backoff strategies. Use when handling API quotas, implementing retry logic, or optimizing request throughput for LLM providers. Trigger with phrases like "langchain rate limit", "langchain throttling", "langchain backoff", "langchain retry", "API quota".
Official Reference Guide for the PPIO Platform, covering LLM API (OpenAI-compatible), Agent Sandbox, GPU (Instances and Serverless), integration, authentication, pricing, rate limiting, and troubleshooting. Suitable for common questions such as 'How to integrate PPIO in specific application scenarios?' and PPIO request failures.