Loading...
Loading...
Found 315 Skills
Generate (TTS), Transcribe (STT), and Clone voices using Google's GenAI and Cloud Speech SDKs. Supports Gemini-TTS, Chirp 3, and Instant Custom Voice.
Senior Visualization Architect & Documentation Engineer for 2026. Specialized in Mermaid.js orchestration, Diagram-as-Code (DaC) workflows, and automated system behavior modeling. Expert in generating high-fidelity Sequence, ERD, Gitgraph, and State diagrams to visualize complex logic, data flows, and project timelines within the Gemini Elite Core.
Setup universal code quality standards in your project. Use when the user wants to generate coding standards files (CLAUDE.md, AGENTS.md, GEMINI.md, etc.) or mentions 'code standards', 'code review setup', or similar intent in any language.
Automatically sync Agents.md, claude.md and gemini.md files in the project to maintain content consistency. Supports automatic monitoring and manual triggering.
Generate new images from text prompts using EachLabs AI models. Supports text-to-image with multiple model families including Flux, GPT Image, Gemini, Imagen, Seedream, and more. Use when the user wants to create new images from text. For editing existing images, see eachlabs-image-edit.
Securely manages API credentials for multiple providers (Anthropic Claude, Google Gemini, GitHub). Use when skills need to access stored API keys for external service invocations.
Cheat sheet + workflow for launching interactive coding-agent CLIs (Claude Code, Gemini CLI, Codex CLI, Cursor CLI, and pi itself) via the interactive_shell overlay or headless dispatch. Use for TUI agents and long-running processes that need supervision, fire-and-forget delegation, or headless background execution. Regular bash commands should use the bash tool instead.
Analyzes and improves LLM prompts and agent instructions for token efficiency, determinism, and clarity. Use when (1) writing a new system prompt, skill, or CLAUDE.md file, (2) reviewing or improving an existing prompt for clarity and efficiency, (3) diagnosing why a prompt produces inconsistent or unexpected results, (4) converting natural language instructions into imperative LLM directives, or (5) evaluating prompt anti-patterns and suggesting fixes. Applies to all LLM platforms (Claude, GPT, Gemini, Llama).
CLI interface for Perplexity AI. Perform AI-powered searches, queries, and research directly from terminal. Use when user mentions Perplexity, AI search, web research, or needs to query AI models like GPT, Claude, Grok, Gemini. Commands: query.
Access and interact with Large Language Models from the command line using Simon Willison's llm CLI tool. Supports OpenAI, Anthropic, Gemini, Llama, and dozens of other models via plugins. Features include chat sessions, embeddings, structured data extraction with schemas, prompt templates, conversation logging, and tool use. This skill is triggered when the user says things like "run a prompt with llm", "use the llm command", "call an LLM from the command line", "set up llm API keys", "install llm plugins", "create embeddings", or "extract structured data from text".
Sync a commit from the gemini remote into sync-upstream for empower-site.
Actionable design system complementing official frontend-design plugin. The plugin provides philosophy; this skill provides executable patterns. Invoke when: - Building web components, pages, or applications - Designing UI that needs to stand out (not generic) - Implementing visual polish and micro-interactions - Choosing typography, color palettes, or spatial layouts CRITICAL: Avoid "AI slop" aesthetics. Make creative, unexpected choices. MANDATORY: Consult Gemini before any frontend work. See also: references/dna-codes.md, references/banned-patterns.md