sales-lobstr
Compare original and translation side by side
🇺🇸
Original
English🇨🇳
Translation
ChineseLobstr.io Platform Help
Lobstr.io平台帮助
Help the user with Lobstr.io platform questions — from configuring ready-made scrapers and running data collection jobs through scheduling automation, exporting results, managing credits, and API integration. Lobstr.io is a no-code web scraping and data collection platform based in Saint-Mandé, France. It offers 50+ ready-made scrapers, multi-threading, cookie-based account sync, scheduled automation, and a full API with Python SDK, CLI, and MCP Server.
为用户解答Lobstr.io平台相关问题,涵盖配置现成爬虫、通过定时自动化运行数据采集任务、导出结果、管理点数、API集成等全场景。Lobstr.io是一款位于法国圣芒代的无代码网页抓取和数据采集平台,提供50+现成爬虫、多线程、基于Cookie的账户同步、定时自动化能力,以及配套Python SDK、CLI和MCP Server的完整API。
Step 1 — Gather context
步骤1 — 收集上下文信息
Ask the user:
-
What area of Lobstr.io do you need help with?
- A) Ready-made scrapers — choosing and configuring a pre-built scraper (Google Maps, LinkedIn, Twitter, YouTube, etc.)
- B) No-code web app — running scrapers, managing runs, viewing results in the UI
- C) Cookie-based account sync — scraping behind login walls (LinkedIn, etc.) using the cookie picking add-on
- D) Scheduled automation — setting up recurring scraping jobs with triggers
- E) Data export — exporting results to Google Sheets, Amazon S3, or SFTP
- F) Webhooks — configuring run status notifications (run.running, run.paused, run.done, run.error)
- G) API / SDK / CLI — programmatic access, Python SDK, CLI, MCP Server
- H) Make integration — connecting Lobstr to make.com automation workflows
- I) Enterprise custom scrapers — tailored data collectors for specific use cases
- J) Account / Billing / Credits — pricing tiers, credit usage, browsing limits
- K) Safety management — browsing limit protection to avoid account bans
- L) Something else — describe it
-
What's your role?
- A) Sales / business development
- B) Marketing / growth
- C) Developer / engineer
- D) Data analyst / researcher
- E) Agency / freelancer
- F) Other
-
What are you trying to accomplish? (describe your specific goal or question)
If the user's request already provides most of this context, skip directly to the relevant step. Lead with your best-effort answer using reasonable assumptions (stated explicitly), then ask only the most critical 1-2 clarifying questions at the end — don't gate your response behind gathering complete context.
Note: If the user needs a specialized skill, route them there with a brief explanation of why that skill is a better fit.
询问用户以下问题:
-
你需要Lobstr.io哪个模块的帮助?
- A) 现成爬虫 —— 选择和配置预构建爬虫(Google Maps、LinkedIn、Twitter、YouTube等)
- B) 无代码网页应用 —— 运行爬虫、管理运行任务、在UI中查看结果
- C) 基于Cookie的账户同步 —— 使用Cookie拾取插件抓取需要登录的平台内容(LinkedIn等)
- D) 定时自动化 —— 通过触发器设置周期性抓取任务
- E) 数据导出 —— 导出结果到Google Sheets、Amazon S3或SFTP
- F) Webhooks —— 配置运行状态通知(run.running、run.paused、run.done、run.error)
- G) API / SDK / CLI —— 程序化访问、Python SDK、CLI、MCP Server
- H) Make集成 —— 连接Lobstr到make.com自动化工作流
- I) 企业定制爬虫 —— 为特定场景定制的数据采集器
- J) 账户/账单/点数 —— 定价档位、点数使用、浏览限制
- K) 安全管理 —— 浏览限制保护,避免账户被封禁
- L) 其他问题 —— 请描述
-
你的角色是什么?
- A) 销售/业务拓展
- B) 营销/增长
- C) 开发/工程师
- D) 数据分析师/研究员
- E) 代理服务商/自由职业者
- F) 其他
-
你想要达成什么目标?(描述你的具体目标或问题)
如果用户的请求已经提供了大部分上下文信息,可以直接跳到相关步骤。 先基于合理假设(需要明确说明假设内容)给出你力所能及的解答,最后只补充询问1-2个最关键的澄清问题——不要等收集完所有信息才给出回复。
注意:如果用户需要使用专用技能,引导用户前往对应技能页面,并简要说明该技能更适配的原因。
Step 2 — Route or answer directly
步骤2 — 引导跳转或直接解答
If the request maps to a specialized skill, route:
- General prospect list strategy / building target lists ->
/sales-prospect-list - Cross-platform data enrichment strategy ->
/sales-enrich - Email deliverability / verification strategy ->
/sales-deliverability - Connecting Lobstr to CRM or other tools via middleware ->
/sales-integration - Outscraper-specific questions ->
/sales-outscraper
Otherwise, answer directly from platform knowledge using the reference below.
如果请求匹配专用技能,引导跳转:
- 通用潜客列表策略/构建目标列表 ->
/sales-prospect-list - 跨平台数据补全策略 ->
/sales-enrich - 邮件送达率/验证策略 ->
/sales-deliverability - 通过中间件连接Lobstr到CRM或其他工具 ->
/sales-integration - Outscraper相关问题 ->
/sales-outscraper
否则,使用下方的平台知识库直接解答。
Step 3 — Lobstr.io platform reference
步骤3 — Lobstr.io平台参考
Provide module-by-module guidance based on the user's area:
根据用户的问题场景提供分模块指导:
Key terminology
核心术语
| Term | Meaning |
|---|---|
| Squid | A scraper configuration — the template that defines what to scrape and how |
| Task | A URL or set of parameters to feed into a squid (the input) |
| Run | An execution of a squid with its tasks (the job) |
| Credit | 1 unique result = 1 credit (the billing unit) |
| 术语 | 含义 |
|---|---|
| Squid | 爬虫配置——定义抓取内容和方式的模板 |
| Task | 输入到Squid中的URL或参数集合(输入项) |
| Run | Squid关联其Task后的一次执行(任务实例) |
| Credit | 1条唯一结果=1个点数(计费单位) |
Ready-Made Scrapers (50+)
现成爬虫(50+)
- What it does: Pre-built scraper configurations for popular platforms — no coding required, just configure inputs and run
- Popular scrapers: Google Maps Leads, Google Maps Reviews, LinkedIn Sales Navigator Leads, Twitter User Tweets, YouTube Channel Scraper, Vinted Products, and many more
- How it works: Select a scraper from the catalog, configure its parameters (search terms, URLs, filters), add tasks (inputs), and start a run
- Premium scrapers: Some scrapers are marked Premium and are only available on paid plans
- 功能:面向主流平台的预构建爬虫配置,无需编码,仅需配置输入即可运行
- 热门爬虫:Google Maps线索、Google Maps评论、LinkedIn Sales Navigator线索、Twitter用户推文、YouTube频道爬虫、Vinted商品等
- 使用方式:从爬虫目录选择对应爬虫,配置参数(搜索词、URL、筛选条件),添加Task(输入项),启动Run即可
- 高级爬虫:部分标记为Premium的爬虫仅付费套餐可用
No-Code Web App
无代码网页应用
- What it does: Browser-based interface to configure, run, and manage all your scrapers without writing code
- Workflow: Select squid → configure parameters → add tasks → start run → monitor progress → download or export results
- Run management: View active, paused, completed, and failed runs with timestamps, credit usage, and result counts
- 功能:基于浏览器的操作界面,无需编写代码即可配置、运行和管理所有爬虫
- 工作流:选择Squid → 配置参数 → 添加Task → 启动Run → 监控进度 → 下载或导出结果
- 运行管理:查看活跃、暂停、完成、失败的Run,附带时间戳、点数消耗、结果数量等信息
Multi-Threading
多线程
- What it does: Run hundreds of concurrent scrapers simultaneously for high-throughput data collection
- Use case: Large-scale scraping projects where you need results from many sources quickly
- Benefit: Dramatically reduces total collection time compared to sequential execution
- 功能:同时运行数百个并发爬虫,实现高吞吐量数据采集
- 适用场景:需要快速从多个来源获取结果的大规模抓取项目
- 优势:相比顺序执行,大幅降低总采集时间
Cookie-Based Account Sync
基于Cookie的账户同步
- What it does: 1-click cookie picking browser add-on that lets you scrape behind login walls (LinkedIn, etc.) without sharing your credentials
- How it works: Install the browser extension, log into the target platform normally, click to sync cookies — Lobstr uses the session cookies to access authenticated pages
- Safety: Your password is never shared with Lobstr — only session cookies are transferred
- Use case: LinkedIn Sales Navigator scraping, any platform requiring authentication
- 功能:一键拾取Cookie的浏览器插件,无需共享凭证即可抓取需要登录的平台内容(LinkedIn等)
- 使用方式:安装浏览器扩展,正常登录目标平台,点击同步Cookie即可——Lobstr会使用会话Cookie访问需要认证的页面
- 安全性:你的密码绝不会共享给Lobstr,仅会传输会话Cookie
- 适用场景:LinkedIn Sales Navigator抓取、任何需要认证的平台抓取
Scheduled Automation
定时自动化
- What it does: Set up recurring scraping jobs that run on autopilot with configurable triggers
- Scheduling options: Define frequency (daily, weekly, custom intervals) and triggers for automatic execution
- Plan requirement: Scheduling is not available on the Free plan — requires a paid subscription
- Use case: Ongoing lead generation, regular competitive monitoring, periodic data refreshes
- 功能:通过可配置的触发器设置自动运行的周期性抓取任务
- 调度选项:定义执行频率(每日、每周、自定义间隔)和自动执行触发器
- 套餐要求:免费套餐不支持调度功能,需要付费订阅
- 适用场景:持续获客、定期竞品监控、周期性数据更新
Data Export
数据导出
- What it does: Export scraped results to external destinations
- Destinations: Google Sheets (direct integration), Amazon S3 (cloud storage), SFTP (server transfer)
- Export limits: Free plan limited to 30 rows per export; paid plans allow full exports
- Use case: Feeding scraped data into spreadsheets, data warehouses, or downstream tools
- 功能:将抓取结果导出到外部目的地
- 支持目的地:Google Sheets(直接集成)、Amazon S3(云存储)、SFTP(服务器传输)
- 导出限制:免费套餐每次最多导出30行,付费套餐支持全量导出
- 适用场景:将抓取数据导入表格、数据仓库或下游工具
Webhooks
Webhooks
- What it does: Receive HTTP POST notifications when run status changes
- Events: (run started),
run.running(run paused),run.paused(run completed),run.done(run failed)run.error - Configuration: to set up webhook endpoint for a squid
POST /v1/delivery?squid={id} - Payload: JSON with run ID, squid info, and timestamp
- Retry policy: 3 attempts with 15-minute delay between retries
- Response requirement: Your endpoint must respond with 200, 201, or 202 within 30 seconds
- 功能:当Run状态变更时接收HTTP POST通知
- 支持事件:(Run启动)、
run.running(Run暂停)、run.paused(Run完成)、run.done(Run失败)run.error - 配置方式:调用为Squid设置Webhook端点
POST /v1/delivery?squid={id} - Payload:包含Run ID、Squid信息和时间戳的JSON数据
- 重试策略:最多重试3次,每次重试间隔15分钟
- 响应要求:你的端点必须在30秒内返回200、201或202状态码
Gmail Notifications
Gmail通知
- What it does: Email alerts on run completion or failure
- Use case: Simple monitoring when you don't need programmatic webhook handling
- 功能:Run完成或失败时发送邮件提醒
- 适用场景:不需要程序化Webhook处理的简单监控场景
Make Integration
Make集成
- What it does: Connect Lobstr to make.com for no-code automation workflows
- Use case: Chain Lobstr scraping results into downstream actions — CRM updates, spreadsheet writes, Slack notifications, email sends, etc.
- 功能:将Lobstr连接到make.com实现无代码自动化工作流
- 适用场景:将Lobstr抓取结果串联到下游动作——CRM更新、表格写入、Slack通知、邮件发送等
API + Python SDK + CLI + MCP Server
API + Python SDK + CLI + MCP Server
- What it does: Full programmatic access to all Lobstr scrapers and platform features
- API: REST API for creating squids, managing tasks, starting runs, retrieving results
- Python SDK: Python library for scripted access
- CLI: Command-line interface for automation and scripting
- MCP Server: Model Context Protocol server for AI agent integration
- 功能:对所有Lobstr爬虫和平台功能的全量程序化访问能力
- API:REST API,可用于创建Squid、管理Task、启动Run、获取结果
- Python SDK:用于脚本化访问的Python库
- CLI:用于自动化和脚本编写的命令行界面
- MCP Server:用于AI Agent集成的模型上下文协议服务器
Enterprise Custom Scrapers
企业定制爬虫
- What it does: Tailored data collectors built by the Lobstr team for your specific use case
- SLA: 99.5% task completion rate, 24-hour failure resolution
- Pricing: Custom pricing based on requirements
- Use case: Complex scraping needs that aren't covered by ready-made scrapers
- 功能:Lobstr团队为你的特定场景定制的数据采集器
- SLA:99.5%的任务完成率,24小时故障修复
- 定价:根据需求定制定价
- 适用场景:现成爬虫无法覆盖的复杂抓取需求
Safety Management
安全管理
- What it does: Browsing limit protection to avoid account bans when scraping platforms like LinkedIn
- How it works: Configurable limits on browsing activity to stay within safe thresholds and avoid detection
- Use case: Protecting your LinkedIn or other platform accounts from being flagged or banned due to excessive automated activity
- 功能:浏览限制保护,避免抓取LinkedIn等平台时账户被封禁
- 工作方式:可配置的浏览活动限制,保持在安全阈值内避免被检测
- 适用场景:保护你的LinkedIn或其他平台账户不会因过度自动化活动被标记或封禁
Data model
数据模型
| Object | Description | Key fields |
|---|---|---|
| Squid | Scraper configuration template | id, name, description, parameters, status, created_at |
| Task | Input URL or parameters for a squid | id, squid_id, url, parameters, status |
| Run | Execution of a squid with its tasks | id, squid_id, status (running/paused/done/error), started_at, completed_at, credits_used, results_count |
| Result | Single scraped data record (1 credit) | id, run_id, data (varies by scraper), scraped_at |
| Delivery/Webhook | Webhook configuration for a squid | id, squid_id, url, events, retry_count, created_at |
| 对象 | 描述 | 核心字段 |
|---|---|---|
| Squid | 爬虫配置模板 | id、name、description、parameters、status、created_at |
| Task | Squid的输入URL或参数 | id、squid_id、url、parameters、status |
| Run | Squid关联Task后的执行实例 | id、squid_id、status(running/paused/done/error)、started_at、completed_at、credits_used、results_count |
| Result | 单条抓取数据记录(消耗1个Credit) | id、run_id、data(不同爬虫结构不同)、scraped_at |
| Delivery/Webhook | Squid的Webhook配置 | id、squid_id、url、events、retry_count、created_at |
API quick reference
API快速参考
- Base URL:
https://api.lobstr.io/v1/ - Authentication: header on all requests
Authorization: Token {your_api_key} - Rate limits:
- — 120 requests/min
/squids - — 90 requests/min
/tasks - — 90 requests/min
/runtasks - — 120 requests/min
/runs - — 2 requests/sec
/results
- Async workflow: Create squid → configure parameters → add tasks → start run → poll status → retrieve results as JSON
- Webhook setup: — configure endpoint URL and events
POST /v1/delivery?squid={id} - Key endpoints:
- — List your squids
GET /squids - — Create a new squid
POST /squids - — Get squid details
GET /squids/{id} - — Add tasks to a squid
POST /tasks - — List tasks for a squid
GET /tasks - — Start a run
POST /runs - — List runs
GET /runs - — Get run status
GET /runs/{id} - — Retrieve results for a run
GET /results - — Configure webhooks
POST /delivery?squid={id}
- SDKs: Python SDK, CLI, MCP Server
- Docs: lobstr.io
- 基础URL:
https://api.lobstr.io/v1/ - 认证方式:所有请求携带请求头
Authorization: Token {your_api_key} - 速率限制:
- — 120次请求/分钟
/squids - — 90次请求/分钟
/tasks - — 90次请求/分钟
/runtasks - — 120次请求/分钟
/runs - — 2次请求/秒
/results
- 异步工作流:创建Squid → 配置参数 → 添加Task → 启动Run → 轮询状态 → 获取JSON格式结果
- Webhook设置:— 配置端点URL和监听事件
POST /v1/delivery?squid={id} - 核心端点:
- — 列出你的所有Squid
GET /squids - — 创建新的Squid
POST /squids - — 获取Squid详情
GET /squids/{id} - — 为Squid添加Task
POST /tasks - — 列出Squid的所有Task
GET /tasks - — 启动Run
POST /runs - — 列出所有Run
GET /runs - — 获取Run状态
GET /runs/{id} - — 获取Run的结果
GET /results - — 配置Webhook
POST /delivery?squid={id}
- SDK:Python SDK、CLI、MCP Server
- 文档:lobstr.io
Integrations
集成能力
| Category | Tools | Status |
|---|---|---|
| Cloud storage | Amazon S3 | Live |
| Spreadsheets | Google Sheets | Live |
| File transfer | SFTP | Live |
| Webhooks | Custom HTTP endpoints | Live |
| Gmail notifications | Live | |
| Automation | Make (make.com) | Live |
| CRM | HubSpot | Coming soon |
| Messaging | Slack | Coming soon |
| Automation | Zapier | Coming soon |
| Database | Airtable | Coming soon |
| 类别 | 工具 | 状态 |
|---|---|---|
| 云存储 | Amazon S3 | 已上线 |
| 电子表格 | Google Sheets | 已上线 |
| 文件传输 | SFTP | 已上线 |
| Webhooks | 自定义HTTP端点 | 已上线 |
| 邮件 | Gmail通知 | 已上线 |
| 自动化 | Make (make.com) | 已上线 |
| CRM | HubSpot | 即将上线 |
| 即时通讯 | Slack | 即将上线 |
| 自动化 | Zapier | 即将上线 |
| 数据库 | Airtable | 即将上线 |
Pricing (as of March 2026 — verify current pricing at lobstr.io)
定价(2026年3月版本 —— 请以lobstr.io当前定价为准)
| Tier | Price | Credits | Export limit | Scheduling | Premium scrapers | Data retention |
|---|---|---|---|---|---|---|
| Free | €0/mo | 100 credits | 30 rows/export | No | No | 7 days |
| Paid | From €50/mo | More credits | Full exports | Yes | Yes | Up to 28 days |
| Enterprise | Custom | Custom | Full exports | Yes | Yes | Custom |
Key pricing notes:
- 1 credit = 1 unique result — credits are the universal billing unit
- Credits refresh monthly and do not roll over to the next month
- Premium scrapers are only available on paid plans
- Scheduling/automation requires a paid plan
- Free plan is limited to 30 rows per export and 7-day data retention
- Enterprise tier includes custom scrapers with 99.5% task completion SLA
| 档位 | 价格 | 点数 | 导出限制 | 调度能力 | 高级爬虫 | 数据留存 |
|---|---|---|---|---|---|---|
| 免费版 | €0/月 | 100点数 | 30行/次导出 | 不支持 | 不支持 | 7天 |
| 付费版 | €50/月起 | 更多点数 | 全量导出 | 支持 | 支持 | 最长28天 |
| 企业版 | 定制 | 定制 | 全量导出 | 支持 | 支持 | 定制 |
核心定价说明:
- 1点数=1条唯一结果——点数是通用计费单位
- 点数每月刷新,不会结转到下个月
- 高级爬虫仅付费套餐可用
- 调度/自动化功能需要付费套餐
- 免费版每次最多导出30行,数据仅留存7天
- 企业版包含定制爬虫,享受99.5%任务完成率SLA
Step 4 — Actionable guidance
步骤4 — 可落地指导
Based on the user's specific question:
-
Setting up a Google Maps lead scraping job:
- Go to the Lobstr scraper catalog and select "Google Maps Leads"
- Configure the scraper parameters — search query (e.g., "dentists in Austin, TX"), result limits, and any filters
- Add tasks — each task is a search query or URL to scrape
- Start the run — Lobstr executes the scraper with multi-threading for speed
- Monitor run status in the web app or via webhook notifications
- Export results to Google Sheets, S3, or download as file
- Each unique result costs 1 credit — monitor your credit usage in the dashboard
-
Scraping LinkedIn Sales Navigator with cookie sync:
- Install the Lobstr cookie picking browser extension
- Log into LinkedIn normally in your browser
- Click the extension to sync your session cookies with Lobstr — your password is never shared
- Select the "LinkedIn Sales Navigator Leads" scraper in Lobstr
- Configure search parameters (title, industry, location, company size, etc.)
- Enable safety management — set browsing limits to avoid LinkedIn detecting excessive automated activity
- Start the run and monitor progress — results include name, title, company, location, and profile URL
- Be conservative with volume — stay within safe browsing limits to protect your LinkedIn account
-
Setting up scheduled recurring scraping:
- Requires a paid plan (scheduling not available on Free tier)
- Configure your squid with the desired scraper and parameters
- Add your tasks (URLs or search queries)
- In the scheduling settings, define frequency (daily, weekly, or custom interval) and trigger conditions
- Set up notifications — configure webhooks for and
run.doneevents, or enable Gmail alertsrun.error - Optionally configure auto-export to Google Sheets or S3 so results are delivered automatically
- Monitor credit consumption — recurring jobs use credits each time they run
-
Using the API to automate scraping programmatically:
- Get your API key from the Lobstr dashboard
- Create a squid: with scraper configuration
POST /v1/squids - Add tasks: with URLs or parameters to scrape
POST /v1/tasks - Start a run: referencing the squid
POST /v1/runs - Poll run status: until status is
GET /v1/runs/{id}done - Retrieve results: for the completed run
GET /v1/results - Alternative to polling: configure a webhook via to receive a POST notification when the run completes
POST /v1/delivery?squid={id} - Mind rate limits: is limited to 2 requests/sec — implement backoff when fetching large result sets
/results
-
Exporting results to Google Sheets:
- Connect your Google account in Lobstr's integration settings
- After a run completes, select "Export to Google Sheets" from the results view
- Choose an existing spreadsheet or create a new one
- Free plan exports are capped at 30 rows — upgrade to a paid plan for full exports
- For automated exports, combine scheduling with the Google Sheets integration so each recurring run auto-exports
根据用户的具体问题提供对应指导:
-
设置Google Maps线索抓取任务:
- 进入Lobstr爬虫目录,选择「Google Maps Leads」
- 配置爬虫参数——搜索词(比如“德克萨斯州奥斯汀的牙医”)、结果限制、其他筛选条件
- 添加Task——每个Task对应一个搜索词或要抓取的URL
- 启动Run——Lobstr会通过多线程快速执行爬虫
- 在网页应用中监控Run状态,或通过Webhook通知获取进度
- 导出结果到Google Sheets、S3,或直接下载为文件
- 每条唯一结果消耗1点数——可在看板中监控点数使用情况
-
通过Cookie同步抓取LinkedIn Sales Navigator线索:
- 安装Lobstr Cookie拾取浏览器扩展
- 在浏览器中正常登录LinkedIn
- 点击扩展将会话Cookie同步到Lobstr——你的密码绝不会被共享
- 在Lobstr中选择「LinkedIn Sales Navigator Leads」爬虫
- 配置搜索参数(职位、行业、地点、公司规模等)
- 开启安全管理——设置浏览限制,避免LinkedIn检测到过度自动化活动
- 启动Run并监控进度——结果包含姓名、职位、公司、地点、个人主页URL
- 控制抓取量级——保持在安全浏览限制内,保护你的LinkedIn账户
-
设置周期性定时抓取:
- 需要付费套餐(免费版不支持调度功能)
- 配置好对应爬虫的Squid和参数
- 添加Task(URL或搜索词)
- 在调度设置中定义执行频率(每日、每周或自定义间隔)和触发条件
- 设置通知——为和
run.done事件配置Webhook,或开启Gmail提醒run.error - 可选配置自动导出到Google Sheets或S3,实现结果自动交付
- 监控点数消耗——周期性任务每次运行都会消耗点数
-
使用API实现程序化抓取自动化:
- 从Lobstr看板获取你的API密钥
- 创建Squid:携带爬虫配置调用
POST /v1/squids - 添加Task:携带要抓取的URL或参数调用
POST /v1/tasks - 启动Run:关联对应Squid调用
POST /v1/runs - 轮询Run状态:调用直到状态变为
GET /v1/runs/{id}done - 获取结果:为已完成的Run调用
GET /v1/results - 轮询替代方案:通过配置Webhook,Run完成时会收到POST通知
POST /v1/delivery?squid={id} - 注意速率限制:接口限制为2次请求/秒——拉取大量结果时需要实现退避逻辑
/results
-
导出结果到Google Sheets:
- 在Lobstr集成设置中绑定你的Google账户
- Run完成后,在结果页选择「导出到Google Sheets」
- 选择现有表格或创建新表格
- 免费版导出最多30行——升级到付费套餐可全量导出
- 如需自动导出,可结合调度功能和Google Sheets集成,实现每次周期性Run自动导出
Gotchas
注意事项
Best-effort from research — review these, especially items about plan-gated features and integration gotchas that may be outdated.
- Free plan is very limited — 100 credits, 30 rows/export, no scheduling, no Premium scrapers, 7-day data retention. The free tier is suitable only for testing. Any serious data collection requires a paid plan starting at €50/month. Credits do not roll over, and the 30-row export cap means you cannot download full result sets on the free plan.
- Credits are consumed per unique result and do not roll over month to month. Each unique result costs 1 credit regardless of which scraper produced it. Unused credits expire at the end of each billing cycle. Plan your scraping volumes to match your credit allowance, and monitor usage to avoid running out mid-month.
- Cookie-based account sync requires careful safety management to avoid platform bans. When scraping behind login walls (especially LinkedIn), always configure browsing limits in Lobstr's safety management settings. Excessive automated activity on your account can trigger detection and lead to restrictions or bans. Start with conservative limits and increase gradually.
- Webhook endpoints must respond within 30 seconds with a 2xx status code, or Lobstr will retry. If your endpoint is slow or returns an error, Lobstr retries up to 3 times with a 15-minute delay between attempts. After 3 failed attempts, the delivery is abandoned. Process webhook payloads asynchronously — accept the request quickly and handle the data in a background job.
- The /results endpoint has a strict 2 requests/second rate limit — much tighter than other endpoints. When fetching large result sets, implement proper rate limiting and backoff in your client code. Batch your result retrieval and avoid tight polling loops. The /squids and /runs endpoints are more generous at 120 requests/minute.
- HubSpot, Slack, Zapier, and Airtable integrations are listed as "coming soon" — they are not yet available. If you need to connect Lobstr to these tools today, use the Make integration as a bridge or build a custom integration using Lobstr's webhook and API capabilities.
- Data retention is 7 days on Free and up to 28 days on paid plans — export or retrieve results promptly. Results are automatically deleted after the retention period. Set up auto-export to Google Sheets or S3, or retrieve results via API as soon as runs complete. Do not rely on the Lobstr dashboard as a long-term data store.
基于调研的经验总结——请留意这些内容,尤其是套餐限制功能、集成相关的注意点可能会过时。
- 免费版限制非常多——仅100点数、每次导出30行、无调度功能、无高级爬虫、数据仅留存7天。 免费版仅适合测试使用,任何正式数据采集都需要每月50欧元起的付费套餐。点数不会结转,30行的导出上限意味着免费版无法下载完整结果集。
- 点数按唯一结果消耗,每月不会结转。 无论哪个爬虫产生的结果,每条唯一结果都消耗1点数。未使用的点数会在每个账单周期结束时过期。请根据你的点数配额规划抓取量级,监控使用情况避免月中点数耗尽。
- 基于Cookie的账户同步需要严格的安全管理,避免平台封禁。 抓取需要登录的平台(尤其是LinkedIn)时,一定要在Lobstr安全管理设置中配置浏览限制。账户过度自动化活动会触发检测,导致限制或封禁。请从保守的限制开始,逐步提高量级。
- Webhook端点必须在30秒内返回2xx状态码,否则Lobstr会重试。 如果你的端点响应缓慢或返回错误,Lobstr最多重试3次,每次间隔15分钟。3次重试失败后会放弃推送。请异步处理Webhook payload——快速接收请求,在后台任务中处理数据。
- 接口有严格的2次/秒速率限制——比其他接口严格很多。 拉取大量结果时,请在客户端代码中实现合理的速率限制和退避逻辑。批量拉取结果,避免紧凑的轮询循环。
/results和/squids接口限制更宽松,为120次请求/分钟。/runs - HubSpot、Slack、Zapier和Airtable集成标注为「即将上线」——目前暂不可用。 如果你现在需要连接Lobstr到这些工具,可以使用Make集成作为桥梁,或者通过Lobstr的Webhook和API能力构建自定义集成。
- 免费版数据留存7天,付费版最长28天——请及时导出或获取结果。 结果会在留存期结束后自动删除。请设置自动导出到Google Sheets或S3,或Run完成后立即通过API获取结果。不要将Lobstr看板作为长期数据存储。
Step 5 — Related skills
步骤5 — 相关技能
- — Build and refine prospect lists using data from Lobstr and other sources
/sales-prospect-list - — Cross-platform data enrichment strategy (combine Lobstr data with other enrichment tools)
/sales-enrich - — Connect Lobstr to CRM, automation, or other tools
/sales-integration - — Email deliverability and verification strategy for scraped contact data
/sales-deliverability - — Similar scraping and data extraction platform — compare capabilities and choose the right tool
/sales-outscraper - — Not sure which skill to use? The router matches any sales objective to the right skill. Install:
/sales-donpx skills add sales-skills/sales --skills sales-do
- —— 使用Lobstr和其他来源的数据构建和优化潜客列表
/sales-prospect-list - —— 跨平台数据补全策略(将Lobstr数据与其他补全工具结合)
/sales-enrich - —— 连接Lobstr到CRM、自动化工具或其他工具
/sales-integration - —— 抓取到的联系数据的邮件送达率和验证策略
/sales-deliverability - —— 同类抓取和数据提取平台——对比能力选择合适的工具
/sales-outscraper - —— 不确定该用哪个技能?路由工具会为任何销售目标匹配合适的技能。安装命令:
/sales-donpx skills add sales-skills/sales --skills sales-do
Examples
示例
Example 1: Building a local business lead list from Google Maps
示例1:从Google Maps构建本地商家线索列表
User says: "I need to scrape all coffee shops in Paris with their addresses, phone numbers, and ratings using Lobstr."
Skill does:
- Recommends the "Google Maps Leads" ready-made scraper from the Lobstr catalog
- Advises breaking Paris into arrondissements or neighborhoods to maximize coverage — e.g., "coffee shops, Paris 1er", "coffee shops, Paris 2ème", etc. — since broad queries may miss results
- Creates tasks for each geographic segment and starts a multi-threaded run
- Results include business name, address, phone, website, Google rating, review count, hours, and categories
- Exports full results to Google Sheets for the sales team (requires paid plan for more than 30 rows)
- Suggests setting up a monthly scheduled run to capture newly opened coffee shops Result: User has a comprehensive list of Paris coffee shops with contact details and ratings, exported to Google Sheets and set up for monthly refresh
用户提问:「我需要用Lobstr抓取巴黎所有咖啡店的地址、电话和评分。」
技能处理流程:
- 推荐Lobstr爬虫目录中的「Google Maps Leads」现成爬虫
- 建议将巴黎拆分为行政区或社区来最大化覆盖——比如「巴黎1区咖啡店」、「巴黎2区咖啡店」等——因为宽泛的查询可能会遗漏结果
- 为每个地理分区创建Task,启动多线程Run
- 结果包含商家名称、地址、电话、网站、Google评分、评论数、营业时间、分类
- 为销售团队导出全量结果到Google Sheets(超过30行需要付费套餐)
- 建议设置每月定时Run,捕获新开业的咖啡店 结果:用户获得完整的巴黎咖啡店列表,包含联系方式和评分,已导出到Google Sheets并设置每月自动刷新
Example 2: Scraping LinkedIn Sales Navigator leads with cookie sync
示例2:通过Cookie同步抓取LinkedIn Sales Navigator线索
User says: "I want to extract leads from a LinkedIn Sales Navigator search for CTOs at SaaS companies in Germany."
Skill does:
- Walks through installing the Lobstr cookie picking browser extension and syncing LinkedIn session cookies
- Selects the "LinkedIn Sales Navigator Leads" scraper (Premium — requires paid plan)
- Configures search parameters: title = "CTO", industry = "SaaS / Software", location = "Germany"
- Enables safety management with conservative browsing limits (e.g., 100 profiles/day) to protect the LinkedIn account
- Starts the run and monitors via webhook notifications for event
run.done - Results include name, title, company, location, LinkedIn profile URL, and other available fields
- Warns about LinkedIn's Terms of Service — scraping may violate their TOS, so user should assess their own risk tolerance Result: User has a targeted list of German SaaS CTOs from Sales Navigator, collected safely with browsing limits, ready for outreach campaign planning
用户提问:「我想从LinkedIn Sales Navigator搜索结果中提取德国SaaS公司CTO的线索。」
技能处理流程:
- 指导用户安装Lobstr Cookie拾取浏览器扩展,同步LinkedIn会话Cookie
- 选择「LinkedIn Sales Navigator Leads」爬虫(高级功能——需要付费套餐)
- 配置搜索参数:职位=「CTO」,行业=「SaaS / Software」,地点=「德国」
- 开启安全管理,设置保守的浏览限制(比如每天100个Profile)保护LinkedIn账户
- 启动Run,通过Webhook监听事件获取完成通知
run.done - 结果包含姓名、职位、公司、地点、LinkedIn个人主页URL和其他可用字段
- 提醒用户LinkedIn的服务条款——抓取可能违反其条款,用户需要自行评估风险承受能力 结果:用户获得来自Sales Navigator的德国SaaS公司CTO精准列表,通过浏览限制安全采集,可直接用于 outreach 活动规划
Example 3: Automating recurring data collection with API and webhooks
示例3:通过API和Webhook实现周期性数据采集自动化
User says: "I need to set up an automated pipeline that scrapes Google Maps reviews for our competitors every week and pushes results to our S3 bucket."
Skill does:
- Creates a squid via configured for the Google Maps Reviews scraper
POST /v1/squids - Adds tasks with competitor Google Maps URLs or place IDs via
POST /v1/tasks - Configures a webhook via pointing to a Lambda function URL, listening for
POST /v1/delivery?squid={id}eventsrun.done - Sets up scheduled automation to run the squid weekly (requires paid plan)
- In the Lambda function: on webhook, calls
run.done(respecting the 2 req/sec rate limit) and writes results to the S3 bucketGET /v1/results - Adds error handling: also listens for webhook events and sends alerts via SNS or Slack
run.error - Monitors credit usage — weekly runs consume credits each cycle, so verifies the plan has sufficient monthly credits Result: User has a fully automated weekly pipeline: Lobstr scrapes competitor reviews on schedule, webhook triggers Lambda, results land in S3 — all running hands-free with error alerting
用户提问:「我需要搭建一个自动化流水线,每周抓取竞品的Google Maps评论,推送到我们的S3存储桶。」
技能处理流程:
- 调用创建配置了Google Maps评论爬虫的Squid
POST /v1/squids - 调用添加竞品Google Maps URL或地点ID作为Task
POST /v1/tasks - 调用配置Webhook指向Lambda函数URL,监听
POST /v1/delivery?squid={id}事件run.done - 设置定时自动化,每周运行该Squid(需要付费套餐)
- Lambda函数逻辑:收到Webhook时,调用
run.done(遵守2次/秒的速率限制),将结果写入S3存储桶GET /v1/results - 添加错误处理:同时监听Webhook事件,通过SNS或Slack发送告警
run.error - 监控点数使用——每周运行每次都会消耗点数,确认套餐有足够的月度点数 结果:用户获得完全自动化的周度流水线:Lobstr按计划抓取竞品评论,Webhook触发Lambda,结果自动存入S3——全程无需人工操作,配备错误告警
Troubleshooting
故障排查
Run stuck in "running" status or taking too long
Run卡在「running」状态或耗时过长
Symptom: A run has been in "running" status for an unexpectedly long time with no results appearing.
Cause: Large task volumes, rate-limited target platforms, or network issues can cause runs to take longer than expected. Multi-threaded runs against rate-limited platforms (like LinkedIn) may be intentionally throttled by safety management settings.
Solution: Check the run details in the web app for progress indicators (tasks completed vs. total). If safety management is enabled, browsing limits may be throttling execution speed — this is expected behavior to protect your accounts. For very large runs, consider breaking them into smaller batches. If the run appears genuinely stuck (no progress for an extended period), pause and restart it. Configure webhooks so you are notified immediately if something fails rather than discovering it later.
run.error症状:Run长时间处于「running」状态,没有新结果产生。
原因:任务体量过大、目标平台速率限制、网络问题都可能导致Run耗时超出预期。针对有限速的平台(比如LinkedIn)的多线程Run可能会被安全管理设置主动限流。
解决方案:在网页应用中查看Run详情的进度指示器(已完成Task/总Task)。如果开启了安全管理,浏览限制可能会降低执行速度——这是保护账户的预期行为。如果是超大型Run,建议拆分为更小的批次。如果Run确实卡住了(长时间没有进度),暂停后重启即可。配置Webhook,故障发生时会立即收到通知,避免后续才发现问题。
run.errorCredits running out mid-month
月中点数耗尽
Symptom: You receive a notification that your credits are exhausted, or runs stop producing results before you expected.
Cause: Each unique result costs 1 credit. Large scraping jobs can consume credits faster than anticipated, especially with high-volume scrapers like Google Maps or LinkedIn. Credits do not roll over, so any unused credits from last month do not help.
Solution: Monitor credit usage in the dashboard before starting large runs. Estimate credit consumption: number of tasks multiplied by expected results per task gives approximate credit cost. For ongoing scheduled jobs, calculate monthly credit needs (weekly runs × results per run × 4 weeks) and ensure your plan covers the total. If you consistently run out, upgrade to a higher plan tier. Consider reducing result volumes by narrowing search parameters or filtering inputs more tightly.
症状:你收到点数耗尽的通知,或Run提前停止产生结果。
原因:每条唯一结果消耗1点数。大型抓取任务消耗点数的速度可能超出预期,尤其是Google Maps、LinkedIn这类高量级爬虫。点数不会结转,上个月未使用的点数无法用于本月。
解决方案:启动大型Run前在看板中监控点数剩余量。预估点数消耗:Task数量乘以每个Task的预期结果数可以得到大概的点数成本。对于持续性的定时任务,计算月度点数需求(每周运行×每次运行结果数×4周),确保你的套餐可以覆盖总量。如果经常出现点数耗尽的情况,升级到更高档位的套餐。可以通过缩小搜索参数范围、更严格的筛选输入来降低结果量级。
Webhook not receiving notifications
Webhook收不到通知
Symptom: You configured a webhook via but your endpoint is not receiving POST requests when runs complete.
Cause: Common issues include: endpoint URL not publicly accessible (localhost or firewalled), endpoint not responding with 200/201/202 within 30 seconds (Lobstr treats slow responses as failures), TLS/SSL certificate issues, or incorrect webhook configuration.
Solution: Verify your endpoint is publicly accessible — test with a tool like webhook.site first to confirm Lobstr is sending requests. Ensure your endpoint responds with a 2xx status code within 30 seconds — offload heavy processing to a background job. Check that your TLS certificate is valid and not self-signed. Review the webhook configuration: confirm you are listening for the correct events (, , etc.) and that the squid ID is correct. Remember that Lobstr retries failed deliveries up to 3 times with a 15-minute delay — check if retries are arriving. If all else fails, use Gmail notifications as a fallback while debugging the webhook setup.
POST /v1/delivery?squid={id}run.donerun.error症状:你通过配置了Webhook,但Run完成时你的端点没有收到POST请求。
原因:常见问题包括:端点URL不是公网可访问(localhost或被防火墙拦截)、端点没有在30秒内返回200/201/202(Lobstr会将慢响应视为失败)、TLS/SSL证书问题、Webhook配置错误。
解决方案:确认你的端点是公网可访问的——先用webhook.site这类工具测试,确认Lobstr会发送请求。确保你的端点在30秒内返回2xx状态码——把 heavy 处理逻辑放到后台任务中。检查你的TLS证书有效,不是自签名证书。核对Webhook配置:确认你监听了正确的事件(、等),Squid ID正确。注意Lobstr最多重试3次失败推送,每次间隔15分钟——检查是否有重试请求到达。如果所有方法都无效,调试Webhook配置期间可以先用Gmail通知作为替代方案。
POST /v1/delivery?squid={id}run.donerun.error