Loading...
Loading...
Found 84 Skills
Web page and website scraping with Firecrawl API. Use this skill when scraping web articles, blog posts, documentation pages, paywalled content, or JavaScript-heavy sites. Triggers on requests to scrape websites, extract article content, convert pages to markdown, or handle anti-bot protection.
Writing and authoring Quarto documents (.qmd), including code cell options, figure and table captions, cross-references, callout blocks (notes, warnings, tips), citations and bibliography, page layout and columns, Mermaid diagrams, YAML metadata configuration, and Quarto extensions. Also covers converting and migrating R Markdown (.Rmd), bookdown, blogdown, xaringan, and distill projects to Quarto, and creating Quarto websites, books, presentations, and reports.
Posts content to WeChat Official Account (微信公众号) via API or Chrome CDP. Supports article posting (文章) with HTML, markdown, or plain text input, and image-text posting (贴图, formerly 图文) with multiple images. Use when user mentions "发布公众号", "post to wechat", "微信公众号", or "贴图/图文/文章".
Use the official MinerU (mineru.net) parsing API to convert a URL (HTML pages like WeChat articles, or direct PDF/Office/image links) into clean Markdown + structured outputs. Use when web_fetch/browser can’t access or extracts messy content, and you want higher-fidelity parsing (layout/table/formula/OCR).
Fetch and parse WeChat Official Account articles. Extract titles, authors, official account names, main content, images and metadata from WeChat article links. It is used when users provide WeChat article links (mp.weixin.qq.com/s/...) and want to read, extract, download or convert article content. Applicable scenarios include obtaining/downloading WeChat articles, extracting text or metadata from WeChat articles, converting WeChat articles to Markdown, and saving WeChat articles along with images locally. Keywords: WeChat Official Account, article acquisition, article scraping, article download.
Batch-process web pages via headless Playwright browser, extract HTML, convert to markdown using Turndown, and save to timestamped scratchpad file. Use when user asks to "capture these pages as markdown", "save web content", "fetch and convert webpages", or needs clean markdown from HTML. All URLs from one prompt → single file at docs/web-captures/<timestamp>.md.
Read content from Feishu cloud documents or knowledge bases and analyze document structure. It is used when users request to "view", "read", "analyze", "fetch", "open" Feishu documents or knowledge bases. It supports reading via document ID, knowledge base Token or URL. Markdown is used as the intermediate format and stored in the /tmp directory.
Fetch and read web pages as Markdown via Jina Reader (https://r.jina.ai/). Use when you need the readable content of a URL (articles, docs, blog posts) without a browser/JS, or when you want a quick plain-text/markdown snapshot for summarization, quoting, or extraction.
Scrape any webpage as clean markdown via Bright Data Web Unlocker API. Bypasses bot detection and CAPTCHA. Requires BRIGHTDATA_API_KEY and BRIGHTDATA_UNLOCKER_ZONE environment variables.
通过 API 或 Chrome CDP 发布内容到微信公众号。支持文章发布(HTML、Markdown 或纯文本输入)和图文消息发布(多图)。当用户提到"发布公众号"、"微信公众号"、"贴图"、"图文"、"文章"时使用。
Publish Markdown articles to X (Twitter) Articles editor with proper formatting. Use when user wants to publish a Markdown file/URL to X Articles, or mentions "publish to X", "post article to Twitter", "X article", or wants help with X Premium article publishing. Handles cover image upload and converts Markdown to rich text automatically.
Fetch and extract web content as clean Markdown when provided with URLs. Use this skill whenever a user provides a URL (http/https link) that needs to be read, analyzed, summarized, or extracted. Converts web pages to Markdown with 80% fewer tokens than raw HTML. Handles all content types including JS-heavy sites, documentation, articles, and blog posts. Supports three conversion methods (auto, AI, browser rendering). Always use this instead of web_fetch when working with URLs - it's more efficient and provides cleaner output.