Loading...
Loading...
Found 10,357 Skills
Use TinyFish web agent to extract/scrape websites, extract data, and automate browser actions using natural language. Use when you need to extract/scrape data from websites, handle bot-protected sites, or automate web tasks.
Conducts security investigations on SOC Compass. The AI agent reads workspace context, asks the user to run SIEM queries, analyzes results, and writes verdicts. Supports multiple alerts in parallel via subagent dispatch. Use when the user mentions SOC Compass, security investigations, alert triage, SIEM queries, threat analysis, Splunk, Elastic, Sentinel, IOC lookups, investigation workspaces, or multiple alerts. Do not use for general cybersecurity questions not involving the SOC Compass platform.
Deploy and configure the Havoc C2 framework with teamserver, HTTPS listeners, redirectors, and Demon agents for authorized red team operations.
Core consulting thinking frameworks and methodologies for structuring business problems, communicating findings, analyzing strategy, building financial models, and designing operations. Use when any agent or command needs to apply MECE decomposition, pyramid principle, hypothesis-driven analysis, issue trees, SCR communication, Porter's Five Forces, TAM/SAM/SOM market sizing, value chain analysis, NPV/IRR decision criteria, build/buy/partner evaluation, RACI matrices, or any standard consulting framework. This skill provides procedural guidance — not just framework names, but how to apply them correctly.
Gate Exchange affiliate program data query and management skill. Use this skill when users ask about their affiliate/partner commission, trading volume, net fees, customer count, trading users, or want to apply for the affiliate program. Supports queries for up to 180 days (API limited to 30 days per request, agent should split longer queries). IMPORTANT: user_id parameter in APIs refers to 'trader' not 'commission receiver' - avoid using unless explicitly specified. Aggregated data from API lists should be calculated using custom scripts, not simple summation. CRITICAL TIME CONSTRAINT: All query times are calculated based on user's system current date in UTC+8 timezone. For relative time descriptions (e.g., 'last 7 days', 'last 30 days', 'this week', 'last month'), calculate start date by subtracting days from current date, then convert both start and end dates to UTC+8 00:00:00 and 23:59:59 respectively, then convert to Unix timestamps. NEVER use future timestamps as query conditions. When timestamps are needed, obtain them via system functions, never generate manually. The 'to' parameter must always be less than or equal to the current Unix timestamp. Trigger phrases include 'my affiliate data', 'commission this week', 'partner earnings', 'team performance', 'customer trading volume', 'rebate income', 'apply for affiliate', 'can I apply', 'am I eligible', 'my application status', 'recent application', 'partner application status'.
Deterministic 4-phase documentation drift detector: Scan, Cross-Reference, Detect, Report. Use when skills/agents/commands are added, removed, or renamed, when README files seem outdated, or before committing documentation changes. Use for "check docs", "sync README", "documentation audit", or "stale entries". Do NOT use for writing documentation content, improving descriptions, or generating new README files.
Full-repo SAP Converged Cloud Go compliance audit. Reviews every package against established review standards — focusing on over-engineering, error message quality, dead code, interface contracts, copy-paste structs, and pattern consistency with keppel. Dispatches parallel agents by package group, each reading ALL sapcc rules. Produces code-level findings with actual before/after diffs. Invoked via "/sapcc-audit" or through /do.
Run comprehensive 3-wave review against all source files in the repo, producing a prioritized issue backlog. Use for "full repo review", "review entire repo", "codebase health check", "review all files", "full codebase review". Do NOT use for PR-scoped reviews (use comprehensive-review) or single-concern reviews (use individual agents).
Deterministic 3-phase GitHub PR review comment extraction: Authenticate, Mine, Validate. Use when mining tribal knowledge from PR reviews, extracting coding standards from review history, or building datasets for the Code Archaeologist agent. Use for "mine PRs", "extract review comments", "tribal knowledge", or "PR review history". Do NOT use for analyzing patterns, generating rules, or interpreting comments — that is the Code Archaeologist agent's responsibility.
Salesforce Data Cloud Retrieve phase. TRIGGER when: user runs Data Cloud SQL, describe, async queries, vector search, search-index workflows, or metadata introspection for Data Cloud objects. DO NOT TRIGGER when: the task is standard CRM SOQL (use sf-soql), segment creation or calculated insight design (use sf-datacloud-segment), or STDM/session tracing/parquet analysis (use sf-ai-agentforce-observability).
Salesforce Data Cloud Harmonize phase. TRIGGER when: user works with DMOs, mappings, relationships, identity resolution, unified profiles, data graphs, or universal IDs. DO NOT TRIGGER when: the task is only about streams/DLOs (use sf-datacloud-prepare), segments/insights (use sf-datacloud-segment), retrieval/search (use sf-datacloud-retrieve), or STDM/session tracing (use sf-ai-agentforce-observability).
Salesforce Data Cloud product orchestrator for connect→prepare→harmonize→segment→act workflows. TRIGGER when: user needs a multi-step Data Cloud pipeline, asks to set up or troubleshoot Data Cloud across phases, manages data spaces or data kits, or wants a cross-phase `sf data360` workflow. DO NOT TRIGGER when: work is isolated to a single phase (use the matching sf-datacloud-* skill), the task is STDM/session tracing/parquet telemetry (use sf-ai-agentforce-observability), standard CRM SOQL (use sf-soql), or Apex implementation (use sf-apex).