Loading...
Loading...
Found 47 Skills
Healthcare Enterprise Funding Monitoring System. Real-time monitoring of industrial and commercial changes of healthcare enterprises, identification of funding signals, and automatic alert pushing. Supports data collection from Tianyancha/Qichacha, AI funding judgment, and multi-channel pushing.
基于 Reddit API 进行深度调研并生成研究报告。支持搜索、获取帖子、评论分析、情感分析等功能
COO Social Media Operation Skill: Automated Operation of Xiaohongshu, Douyin, WeChat Official Account, and WeChat Channels using Playwright MCP
GitHub data collection patterns for workflow agents. Covers search query construction by intent, date range handling, repository scope narrowing, preferences.md integration, cross-repo intelligence, parallel stream collection model, and auto-recovery for empty results. Use when building agents that search GitHub for issues, PRs, discussions, releases, security alerts, or CI status.
Batch content grabber — bulk fetch bookmarks, user tweets, search results, author notes, wiki pages, and more from X/Twitter, Xiaohongshu, WeChat, YouTube, Feishu. Use when user wants to batch/bulk fetch, search keywords, or grab all posts from an account.
Web scraping via progressive escalation and social media platform scrapers. USE WHEN scraping, crawl, scrape URL, bot detection, CAPTCHA, spider.
Points to the BlockchainSpider open-source Python/Scrapy toolkit for collecting on-chain data—transfer subgraphs around an address or tx, EVM and Solana block/transaction ingestion, receipts/logs, and optional label plugins. Use when the user wants to build datasets, offline traces, or research pipelines alongside blockchain-analytics-operations and solana-tracing-specialist—not as a substitute for RPC provider ToS, rate limits, or legal review of sensitive crawls.
Collect app events via evalpkgs into sqlite, then filter/report capture_results to Feishu Bitable with retry-safe writeback. Use for collect-start/collect-stop/filter/report/retry-reset workflows.
Scrape social media profiles, posts, comments, followers, and search across 6 platforms via x402. USE FOR: - Getting TikTok, Instagram, X/Twitter, Facebook, Reddit, or LinkedIn profiles - Fetching a user's posts, stories, highlights, or videos - Getting comments, replies, and reactions on posts - Listing followers and following for any account - Searching posts, hashtags, profiles, jobs, and ads across platforms - Cross-platform social media research and monitoring TRIGGERS: - "tiktok", "instagram", "facebook", "linkedin profile", "linkedin posts" - "get followers", "who follows", "following list" - "scrape profile", "get posts from", "social media data" - "instagram stories", "tiktok videos", "facebook page" - "linkedin company", "linkedin jobs", "linkedin ads" - "cross-platform", "social media research" IMPORTANT: StableSocial uses an async two-step flow. Step 1: POST triggers data collection (paid, $0.06). Step 2: Poll GET /api/jobs?token=... until finished (free). All endpoints are $0.06 per call. Use `npx agentcash fetch` for paid POST triggers. Use `npx agentcash fetch` for free GET polling. IMPORTANT: Use exact endpoint paths from the Quick Reference tables below. All paths include a platform prefix (e.g. `https://stablesocial.dev/api/tiktok/...`).
Trigger this skill when the user wants to collect structured data, create forms, or set up submission endpoints. Covers any form type (contact, feedback, signup, waitlist, bug report, support, lead capture, surveys, applications) and any collection method (HTML forms, scripts, AI agents). Also trigger for managing existing forms or submissions, generating frontend form integration code, setting up webhooks or Discord/Slack notifications, configuring spam protection or AI features (translation, smart replies), connecting MCP for AI-assisted form management, or anything referencing Postbox or usepostbox.com. If the user needs data flowing in from anywhere, this skill applies.
Lobstr.io platform help — no-code web scraping platform with 50+ ready-made scrapers for Google Maps, LinkedIn Sales Navigator, Twitter, YouTube, and more. Features cookie-based login sync, scheduled automation, multi-threading, and a full API with Python SDK and MCP Server. Use when configuring a Lobstr scraper, exporting data to Google Sheets or S3, setting up scheduled scraping, working with the Lobstr API or Python SDK, or managing credits. Do NOT use for general prospect list strategy (use /sales-prospect-list), cross-platform enrichment strategy (use /sales-enrich), or integration strategy (use /sales-integration).
This skill helps users automatically extract complete Markdown content from any website via the BrowserAct Web Search Scraper API. The Agent should proactively apply this skill when users express needs like extract complete markdown from a specific website, scrape the content of an article link, get the text from a target url, convert a webpage to markdown format, fetch the main content of a blog post, extract data from a given web page, parse the html of a website into markdown, download the readable text from a news article, obtain the content of a tutorial page, extract all the markdown text from any http or https url, scrape documentation from a web link, or grab the text of a single webpage.