Loading...
Loading...
Found 1,077 Skills
Command-line interface for Novita AI - An OpenAI-compatible AI API client for DeepSeek, GLM, and other models.
Optional Stage 0 of the feature workflow — clarify vague ideas through dialogue until they are ready to enter the design phase. The role of AI is a thinking partner: dig out the real problem the user wants to solve (instead of sticking to the first solution they blurt out), actively evaluate the solution when the user brings it up, and propose better alternatives if necessary. After the discussion, output {slug}-brainstorm.md to document the results. Trigger scenarios: The user says "I have an unclear idea", "Let's brainstorm first", "The feature direction is still undecided", or the user brings a specific solution but wants to hear other ideas first. Skip this stage and proceed directly to design if the idea is already clear and the user does not want to discuss the solution further. This stage also does not handle bugs and refactoring.
Document the pitfalls encountered or good practices discovered during this work into searchable learning documents, so that both AI and humans can look them up when similar tasks arise in the future. Two tracks: The pitfall track records experiences where "things should have worked but didn't" — bugs, configuration traps, environment issues, integration failures; The knowledge track records findings that "should be the default approach going forward" — best practices, workflow improvements, reusable patterns. Trigger scenarios: Proactively prompt for input when wrapping up feature-acceptance or issue-fix, or when the user says phrases like "document knowledge", "learning", "document learnings", "record this experience". Spec documents record what was done and how it was done, while learning documents record what pitfalls were encountered / what was learned — the two complement each other and are not interchangeable.
Use Ref and WidgetRef to read, watch, listen, invalidate, and refresh providers; onDispose and onCancel lifecycle; ref.read vs ref.watch vs ref.listen, ref.invalidate and ref.refresh. Use when interacting with Riverpod providers from widgets or other providers, when to use watch vs read, or when resetting provider state. Use this skill whenever the user asks about ref.watch, ref.read, ref.listen, ref.invalidate, or Riverpod lifecycle.
Enables AI-powered parsing and key information extraction from high-frequency documents including invoices, orders, receipts, long texts, and common Chinese identity & credential documents. Supports reusable custom templates for non-standard business files. Features batch concurrent processing to automate document workflows for finance, administration, HR data entry and other departments.
Run vLLM performance benchmark using synthetic random data to measure throughput, TTFT (Time to First Token), TPOT (Time per Output Token), and other key performance metrics. Use when the user wants to quickly test vLLM serving performance without downloading external datasets.
Provides guidance for training LLMs with reinforcement learning using verl (Volcano Engine RL). Use when implementing RLHF, GRPO, PPO, or other RL algorithms for LLM post-training at scale with flexible infrastructure backends.
Extract invoice data from images or text descriptions and generate a categorized Excel expense report. Supports receipt photos, scanned invoices, and manual text input. Auto-classifies into: business entertainment (客户餐费), travel-transport (机票/火车票/打车), travel-accommodation (酒店), travel-meals, office supplies, communication, and other. Use when the user mentions "发票报销", "expense report", "报销单", "发票整理", "invoice", "报销汇总", "发票分类", "reimbursement", or has invoice images to process.
Fetch a URL and distill its content into the Obsidian wiki. If invoked from inside a project directory, the page lands directly in that project's folder (creating the project in the vault if needed). Otherwise it goes to misc/ and gains project affinity over time. Use this skill when the user says "/ingest-url <url>", "add this URL to the wiki", "ingest this link", "save this page", or pastes a URL and says "add this" or "save this to my wiki".
Buttons, inputs, pills, badges, calendars, and other interactive components form a visual family — they share the same border-radius, colour logic, shadow scale, border style, and spacing rhythm. Inconsistency between them breaks the sense of a coherent product. Use when building or reviewing a component library, design system, or any set of UI components.
Configure the LaunchDarkly hosted MCP server during onboarding. Use when the parent LaunchDarkly onboarding skill reaches Step 4 (MCP). Supports Cursor, Claude Code, Windsurf, GitHub Copilot, and other MCP-compatible agents. OAuth authentication; no API keys for the hosted server.
Exhaustively extract UX patterns from a reference web app. Walks every screen, captures screenshots of every state, records interaction patterns, copy verbatim, keyboard shortcuts, responsive treatments, motion, and empty/error/loading states. Produces a reusable pattern library that other audits can compare against. The inverse of ux-audit — asks 'what is the bar?' rather than 'does this match the bar?'. Trigger with 'learn from X', 'extract patterns from X', 'study X's UX', 'reverse engineer the UX of X', 'build a pattern library from X'.