Loading...
Loading...
Found 1,221 Skills
Cluely platform help — real-time AI meeting assistant with live coaching overlay, pre-call briefs, meeting notes, conversation analytics, and knowledge base RAG. Use when setting up Cluely for live AI prompts during sales calls, configuring the knowledge base with company docs for real-time RAG retrieval, connecting Cluely to HubSpot or Salesforce via Merge.dev, troubleshooting transcription accuracy or speaker attribution errors, comparing Cluely Pro vs Pro + Undetectability plans, or setting up team coaching scorecards and missed opportunity tracking. Do NOT use for choosing between AI note-takers across vendors (use /sales-note-taker) or reviewing a call for coaching (use /sales-call-review).
Meta-skill for understanding and customizing Mindfold Trellis — the all-in-one AI workflow system for 11 AI coding platforms (Claude Code, Cursor, OpenCode, iFlow, Codex, Kilo, Kiro, Gemini CLI, Antigravity, Qoder, CodeBuddy). Documents the original Trellis system design including architecture, commands, hooks, multi-agent pipelines, monorepo support, and task lifecycle hooks. Use when understanding Trellis architecture, customizing workflows, adding commands or agents, troubleshooting issues, or adapting Trellis to specific projects. Modifications should be recorded in a project-local trellis-local skill, not here.
Cleft Notes platform help — Apple-native AI voice-to-notes app with on-device transcription that turns spoken thoughts into organized markdown notes with auto-headings. Use when setting up Cleft Notes for capturing voice memos and converting rambling thoughts into structured notes, configuring Obsidian or Notion sync to route Cleft notes into an existing knowledge base, troubleshooting recordings that fail after a couple minutes or produce garbled transcription output, setting up Zapier automations to send Cleft notes to project management or CRM tools, choosing between Cleft free and Plus plans, deciding whether Cleft or Voicenotes or AudioPen fits your voice capture workflow, or evaluating Cleft for ADHD-friendly voice-first note-taking on Apple devices. Do NOT use for comparing AI meeting note-takers across platforms (use /sales-note-taker) or reviewing a sales call for coaching (use /sales-call-review).
GetMoreBacklinks platform help — managed directory submission service for startups. Use when deciding whether to pay for directory submissions vs doing it yourself, comparing GetMoreBacklinks plans (Starter $87 vs Business $187), setting expectations for DR improvement timeline, evaluating if a managed submission service is worth it for your budget, or troubleshooting why directory submissions didn't improve rankings. Do NOT use for DIY directory submission strategy (use /sales-launch-directory). Do NOT use for backlink analysis or SEO audits (use /sales-semrush).
Treasure Data platform help — enterprise AI-native CDP with profile unification, audience segmentation, journey orchestration, 400+ connectors, AI Marketing Cloud suites. Use when setting up Treasure Data CDP, unifying customer profiles across channels, building audience segments in Treasure Data, activating audiences to ad platforms or CRMs, configuring Treasure Data connectors or integrations, writing SQL queries in Treasure Data, troubleshooting Treasure Data workflows or jobs, or working with the TD API. Do NOT use for general CRM data quality without Treasure Data context — use /sales-data-hygiene instead.
This skill deploys frontend and full-stack projects to EdgeOne Pages (Tencent EdgeOne). It should be used when the user's primary intent is to deploy, publish, ship, host, launch, go live, or release a new version — e.g. "deploy my app", "publish this site", "push this live", "create a preview deployment", "deploy to EdgeOne", "ship to production", "上线", "发布", "发一版", "重新部署". Do NOT trigger when deployment is only mentioned as a secondary step (e.g. "write an API and deploy it" — primary intent is writing code, use edgeone-pages-dev). Do NOT trigger for post-deployment runtime errors (e.g. CORS issues, 500 errors after deploy — use edgeone-pages-dev for troubleshooting).
DataWorks Operations Center assistant for task and workflow operations, alert rule creation and management. Covers troubleshooting, failure recovery, baseline assurance, monitoring and alerting. Supports periodic, manual, and triggered tasks/workflows (excludes real-time/streaming tasks). Uses aliyun CLI to call dataworks-public OpenAPI (2024-05-18). Trigger keywords: query task, task instance, instance log, workflow, workflow instance, alert rule, operations center, task failure, instance status, upstream/downstream dependency, rerun, monitoring alert, custom monitoring, alert rule, task instance, workflow instance, operation log, baseline assurance, failure recovery, DataWorks operations. Do NOT trigger: data source management, compute resources, resource groups, data development, MaxCompute table management, ECS/RDS/OSS operations, workspace member management, data quality, data lineage, data preview.
Generate Python code for Naver Open APIs (News Search, Blog Search, Web Search, Datalab Trends, Image Search, Book Search, and other openapi.naver.com APIs). Use this skill whenever the user wants to call Naver APIs, search Naver News/Blog/Web/Images, fetch Datalab keyword trends, or write Python scripts that interact with Naver's Open API platform. This includes requests for code examples, API integration, parameter explanations, endpoint lookup, or troubleshooting Naver Open API calls. Even if the user just mentions 'Naver API', 'Naver news data', 'search trends', or 'openapi.naver.com', activate this skill immediately.
Import data into the AWS data lake from S3 files, local uploads, JDBC databases (Oracle, SQL Server, PostgreSQL, MySQL, RDS, Aurora), Amazon Redshift, Snowflake, BigQuery, DynamoDB, or existing Glue catalog tables (migration). Default target is S3 Tables; standard Iceberg on a general purpose bucket is supported where S3 Tables is not adopted. Handles one-time loads, recurring pipelines, migrations. Triggers on: import data, load data, ingest, sync database, migrate table, move data to AWS, set up pipeline, ETL, pull from Snowflake, query BigQuery into S3, export DynamoDB, CTAS, convert to Iceberg. Do NOT use for setting up or troubleshooting Glue connections (use connecting-to-data-source), creating empty tables (use creating-data-lake-table), running queries (use querying-data-lake), finding tables by fuzzy name (use finding-data-lake-assets), catalog audit (use exploring-data-catalog), or SaaS platforms like Salesforce, ServiceNow, SAP, MongoDB, Kafka.
Build integrations with Google Workspace APIs (Gmail, Calendar, Drive, Sheets, Docs, Chat, Meet, Forms, Tasks, Admin SDK). Covers OAuth 2.0, service accounts, rate limits, batch operations, and Cloudflare Workers patterns. Use when building MCP servers, automation tools, or integrations with any Google Workspace API, or troubleshooting OAuth errors, rate limit 429 errors, scope issues, or API-specific gotchas.
Automate GitHub repository setup with CI/CD workflows, issue templates, Dependabot, and CodeQL security scanning. Includes 12 production-tested workflows and prevents 18 errors: YAML syntax, action pinning, and configuration. Use when: setting up GitHub Actions CI/CD, creating issue/PR templates, enabling Dependabot or CodeQL scanning, deploying to Cloudflare Workers, implementing matrix testing, or troubleshooting YAML indentation, action version pinning, secrets syntax, runner versions, or CodeQL configuration. Keywords: github actions, github workflow, ci/cd, issue templates, pull request templates, dependabot, codeql, security scanning, yaml syntax, github automation, repository setup, workflow templates, github actions matrix, secrets management, branch protection, codeowners, github projects, continuous integration, continuous deployment, workflow syntax error, action version pinning, runner version, github context, yaml indentation error
Best practices for building integrations with NetBox REST and GraphQL APIs. Use when building NetBox API integrations, reviewing integration code, troubleshooting NetBox performance issues, planning automation architecture, writing scripts that interact with NetBox, using pynetbox, configuring Diode for data ingestion, or implementing NetBox webhooks.