doublecheck

Compare original and translation side by side

🇺🇸

Original

English
🇨🇳

Translation

Chinese

Doublecheck

双重核查(Doublecheck)

Run a three-layer verification pipeline on AI-generated output. The goal is not to tell the user what is true -- it is to extract every verifiable claim, find sources the user can check independently, and flag anything that looks like a hallucination pattern.
对AI生成的输出运行三层验证流水线。目标不是告知用户何为真相——而是提取每一项可验证的主张,找到用户可独立核查的来源,并标记所有符合幻觉模式的内容。

Activation

激活方式

Doublecheck operates in two modes: active mode (persistent) and one-shot mode (on demand).
Doublecheck 有两种运行模式:持续模式(active mode)(永久生效)和单次模式(one-shot mode)(按需触发)。

Active Mode

持续模式

When the user invokes this skill without providing specific text to verify, activate persistent doublecheck mode. Respond with:
Doublecheck is now active. I'll verify factual claims in my responses before presenting them. You'll see an inline verification summary after each substantive response. Say "full report" on any response to get the complete three-layer verification with detailed sourcing. Turn it off anytime by saying "turn off doublecheck."
Then follow ALL of the rules below for the remainder of the conversation:
Rule: Classify every response before sending it.
Before producing any substantive response, determine whether it contains verifiable claims. Classify the response:
Response typeContains verifiable claims?Action
Factual analysis, legal guidance, regulatory interpretation, compliance guidance, or content with case citations or statutory referencesYes -- high densityRun full verification report (see high-stakes content rule below)
Summary of a document, research, or dataYes -- moderate densityRun inline verification on key claims
Code generation, creative writing, brainstormingRarelySkip verification; note that doublecheck mode doesn't apply to this type of content
Casual conversation, clarifying questions, status updatesNoSkip verification silently
Rule: Inline verification for active mode.
When active mode applies, do NOT generate a separate full verification report for every response. Instead, embed verification directly into your response using this pattern:
  1. Generate your response normally.
  2. After the response, add a
    Verification
    section.
  3. In that section, list each verifiable claim with its confidence rating and a source link where available.
Format:
---
**Verification (N claims checked)**

- [VERIFIED] "Claim text" -- Source: [URL]
- [VERIFIED] "Claim text" -- Source: [URL]
- [PLAUSIBLE] "Claim text" -- no specific source found
- [FABRICATION RISK] "Claim text" -- could not find this citation; verify before relying on it
For active mode, prioritize speed. Run web searches for citations, specific statistics, and any claim you have low confidence about. You do not need to search for claims that are common knowledge or that you have high confidence about -- just rate them PLAUSIBLE and move on.
If any claim rates DISPUTED or FABRICATION RISK, call it out prominently before the verification section so the user sees it immediately. When auto-escalation applies (see below), place this callout at the top of the full report, before the summary table:
**Heads up:** I'm not confident about [specific claim]. I couldn't find a supporting source. You should verify this independently before relying on it.
Rule: Auto-escalate to full report for high-risk findings.
If your inline verification identifies ANY claim rated DISPUTED or FABRICATION RISK, do not produce inline verification. Instead, place the "Heads up" callout at the top of your response and then produce the full three-layer verification report using the template in
assets/verification-report-template.md
. The user should not have to ask for the detailed report when something is clearly wrong.
Rule: Full report for high-stakes content.
If the response contains legal analysis, regulatory interpretation, compliance guidance, case citations, or statutory references, always produce the full verification report using the template in
assets/verification-report-template.md
. Do not use inline verification for these content types -- the stakes are too high for the abbreviated format.
Rule: Discoverability footer for inline verification.
When producing inline verification (not a full report), always append this line at the end of the verification section:
_Say "full report" for detailed three-layer verification with sources._
Rule: Offer full verification on request.
If the user says "full report," "run full verification," "verify that," "doublecheck that," or similar, run the complete three-layer pipeline (described below) and produce the full report using the template in
assets/verification-report-template.md
.
当用户调用本技能但未提供待验证的特定文本时,激活永久生效的双重核查模式,回复如下内容:
Doublecheck 已激活。 我会在输出回复前先验证其中的事实主张,每一条实质性回复后都会附带内联验证摘要。对任意回复发送“full report”即可获取带详细来源的完整三层验证报告。随时发送“turn off doublecheck”即可关闭该功能。
后续对话全程需严格遵守以下所有规则:
规则:发送前对所有响应分类
生成任何实质性回复前,先判断其是否包含可验证主张,对响应分类:
响应类型是否包含可验证主张?操作
事实分析、法律指导、监管解读、合规指导,或包含案例引用、法规参考的内容是,密度高运行完整验证报告(参见下文高风险内容规则)
文档、研究或数据摘要是,密度中等对核心主张运行内联验证
代码生成、创意写作、头脑风暴极少跳过验证;注明双重核查模式不适用于此类内容
日常对话、澄清问题、状态更新静默跳过验证
规则:持续模式的内联验证规则
启用持续模式时,无需为每条响应单独生成完整验证报告,而是按照以下规则将验证内容直接嵌入回复:
  1. 正常生成回复内容
  2. 回复末尾添加
    Verification
    模块
  3. 在模块中列出每一项可验证主张、置信度评级,以及可用的来源链接
格式如下:
---
**Verification (N claims checked)**

- [VERIFIED] "Claim text" -- Source: [URL]
- [VERIFIED] "Claim text" -- Source: [URL]
- [PLAUSIBLE] "Claim text" -- no specific source found
- [FABRICATION RISK] "Claim text" -- could not find this citation; verify before relying on it
持续模式优先保障速度,仅对引用内容、特定统计数据、以及低置信度的主张运行网页搜索。无需搜索常识类或高置信度的主张,仅标记为PLAUSIBLE即可。
如果任何主张被标记为DISPUTED或FABRICATION RISK,需在验证模块前醒目提示,确保用户第一时间看到。触发自动升级规则时(见下文),将提示放在完整报告的顶部、摘要表格之前:
**Heads up:** I'm not confident about [specific claim]. I couldn't find a supporting source. You should verify this independently before relying on it.
规则:高风险发现自动升级为完整报告
如果内联验证发现任何标记为DISPUTED或FABRICATION RISK的主张,不要输出内联验证,而是在回复顶部放置“Heads up”提示,再使用
assets/verification-report-template.md
模板生成完整的三层验证报告。发现明确问题时无需等待用户申请即可输出详细报告。
规则:高风险内容直接生成完整报告
如果响应包含法律分析、监管解读、合规指导、案例引用或法规参考,始终使用
assets/verification-report-template.md
模板生成完整验证报告,不得使用内联验证——此类内容风险过高,不适用简化格式。
规则:内联验证的可发现性页脚
输出内联验证(非完整报告)时,始终在验证模块末尾添加以下内容:
_Say "full report" for detailed three-layer verification with sources._
规则:按要求提供完整验证
如果用户发送“full report”、“run full verification”、“verify that”、“doublecheck that”或类似指令,运行完整的三层验证流程,并使用
assets/verification-report-template.md
模板生成完整报告。

One-Shot Mode

单次模式

When the user invokes this skill and provides specific text to verify (or references previous output), run the complete three-layer pipeline and produce a full verification report using the template in
assets/verification-report-template.md
.
当用户调用本技能时提供了待验证的特定文本(或引用了之前的输出),运行完整的三层验证流程,并使用
assets/verification-report-template.md
模板生成完整验证报告。

Deactivation

关闭方式

When the user says "turn off doublecheck," "stop doublecheck," or similar, respond with:
Doublecheck is now off. I'll respond normally without inline verification. You can reactivate it anytime.

当用户发送“turn off doublecheck”、“stop doublecheck”或类似指令时,回复如下内容:
Doublecheck 已关闭。 我将正常响应,不再提供内联验证。您可以随时重新启用。

Layer 1: Self-Audit

第一层:自审计

Re-read the target text with a critical lens. Your job in this layer is extraction and internal analysis -- no web searches yet.
以批判视角重读目标文本,本层的核心是主张提取和内部分析——暂不执行网页搜索。

Step 1: Extract Claims

步骤1:提取主张

Go through the target text sentence by sentence and pull out every statement that asserts something verifiable. Categorize each claim:
CategoryWhat to look forExamples
FactualAssertions about how things are or were"Python was created in 1991", "The GPL requires derivative works to be open-sourced"
StatisticalNumbers, percentages, quantities"95% of enterprises use cloud services", "The contract has a 30-day termination clause"
CitationReferences to specific documents, cases, laws, papers, or standards"Under Section 230 of the CDA...", "In Mayo v. Prometheus (2012)..."
EntityClaims about specific people, organizations, products, or places"OpenAI was founded by Sam Altman and Elon Musk", "GDPR applies to EU residents"
CausalClaims that X caused Y or X leads to Y"This vulnerability allows remote code execution", "The regulation was passed in response to the 2008 financial crisis"
TemporalDates, timelines, sequences of events"The deadline is March 15", "Version 2.0 was released before the security patch"
Assign each claim a temporary ID (C1, C2, C3...) for tracking through subsequent layers.
逐句梳理目标文本,提取所有断言可验证事实的表述,对每项主张分类:
分类识别特征示例
事实类关于事物现状或历史状态的断言“Python was created in 1991”、“The GPL requires derivative works to be open-sourced”
统计类数字、百分比、量化表述“95% of enterprises use cloud services”、“The contract has a 30-day termination clause”
引用类提及特定文档、案例、法律、论文或标准“Under Section 230 of the CDA...”、“In Mayo v. Prometheus (2012)...”
实体类关于特定人物、组织、产品或地点的主张“OpenAI was founded by Sam Altman and Elon Musk”、“GDPR applies to EU residents”
因果类主张X导致Y或X会引发Y的表述“This vulnerability allows remote code execution”、“The regulation was passed in response to the 2008 financial crisis”
时间类日期、时间线、事件顺序相关表述“The deadline is March 15”、“Version 2.0 was released before the security patch”
为每项主张分配临时ID(C1、C2、C3...),用于后续流程跟踪。

Step 2: Check Internal Consistency

步骤2:核查内部一致性

Review the extracted claims against each other:
  • Does the text contradict itself anywhere? (e.g., states two different dates for the same event)
  • Are there claims that are logically incompatible?
  • Does the text make assumptions in one section that it contradicts in another?
Flag any internal contradictions immediately -- these don't need external verification to identify as problems.
交叉核对提取的所有主张:
  • 文本是否存在自相矛盾?(比如对同一事件标注两个不同日期)
  • 是否存在逻辑上不兼容的主张?
  • 文本某一部分的假设是否与另一部分的内容矛盾?
立即标记所有内部矛盾——此类问题无需外部验证即可判定为异常。

Step 3: Initial Confidence Assessment

步骤3:初始置信度评估

For each claim, make an initial assessment based only on your own knowledge:
  • Do you recall this being accurate?
  • Is this the kind of claim where models frequently hallucinate? (Specific citations, precise statistics, and exact dates are high-risk categories.)
  • Is the claim specific enough to verify, or is it vague enough to be unfalsifiable?
Record your initial confidence but do NOT report it as a finding yet. This is input for Layer 2, not output.

仅基于自身知识库为每项主张做初始评估:
  • 你是否记得该内容是准确的?
  • 这类主张是否属于模型频繁出现幻觉的类别?(特定引用、精确统计数据、准确日期都属于高风险类别)
  • 主张是否具体到可验证,还是过于模糊无法证伪?
记录初始置信度,但暂不将其作为结果输出,该结果仅作为第二层的输入。

Layer 2: Source Verification

第二层:来源验证

For each extracted claim, search for external evidence. The purpose of this layer is to find URLs the user can visit to verify claims independently.
为每一项提取的主张搜索外部证据,本层的目标是找到用户可访问的URL,用于独立验证主张。

Search Strategy

搜索策略

For each claim:
  1. Formulate a search query that would surface the primary source. For citations, search for the exact title or case name. For statistics, search for the specific number and topic. For factual claims, search for the key entities and relationships.
  2. Run the search using
    web_search
    . If the first search doesn't return relevant results, reformulate and try once more with different terms.
  3. Evaluate what you find:
    • Did you find a primary or authoritative source that directly addresses the claim?
    • Did you find contradicting information from a credible source?
    • Did you find nothing relevant? (This is itself a signal -- real things usually have a web footprint.)
  4. Record the result with the source URL. Always provide the URL even if you also summarize what the source says.
针对每一项主张:
  1. 构造搜索查询,优先返回一手来源。针对引用内容搜索完整标题或案例名,针对统计数据搜索具体数值和主题,针对事实主张搜索核心实体和关联关系。
  2. 使用
    web_search
    执行搜索,如果第一次搜索未返回相关结果,换关键词重新尝试一次。
  3. 评估搜索结果:
    • 是否找到直接佐证该主张的一手或权威来源?
    • 是否找到可信来源发布的矛盾信息?
    • 是否未找到任何相关内容?(这本身就是信号——真实存在的内容通常会在网络上留下痕迹)
  4. 记录结果和来源URL,即使你同时总结了来源内容,也必须提供URL。

What Counts as a Source

有效来源判定标准

Prefer primary and authoritative sources:
  • Official documentation, specifications, and standards
  • Court records, legislative texts, regulatory filings
  • Peer-reviewed publications
  • Official organizational websites and press releases
  • Established reference works (encyclopedias, legal databases)
Note when a source is secondary (news article, blog post, wiki page) vs. primary. The user can weigh accordingly.
优先选择一手和权威来源:
  • 官方文档、规范和标准
  • 法庭记录、立法文本、监管备案文件
  • 同行评审出版物
  • 官方机构网站和新闻稿
  • 权威参考资料(百科全书、法律数据库)
注明来源是二手(新闻文章、博客、维基页面)还是一手,供用户自行权衡。

Handling Citations Specifically

引用内容的特殊处理规则

Citations are the highest-risk category for hallucinations. For any claim that cites a specific case, statute, paper, standard, or document:
  1. Search for the exact citation (case name, title, section number).
  2. If you find it, confirm the cited content actually says what the target text claims it says.
  3. If you cannot find it at all, flag it as FABRICATION RISK. Models frequently generate plausible-sounding citations for things that don't exist.

引用是幻觉最高发的类别,任何提及特定案例、法规、论文、标准或文档的主张:
  1. 搜索完整引用内容(案例名、标题、章节号)
  2. 如果找到来源,确认引用内容确实与目标文本的主张一致
  3. 如果完全找不到来源,标记为FABRICATION RISK。模型经常生成听起来合理但完全不存在的引用内容。

Layer 3: Adversarial Review

第三层:对抗性审查

Switch your posture entirely. In Layers 1 and 2, you were trying to understand and verify the output. In this layer, assume the output contains errors and actively try to find them.
完全切换立场,前两层你尝试理解和验证输出内容,本层假设输出包含错误,主动查找问题。

Hallucination Pattern Checklist

幻觉模式检查清单

Check for these common patterns:
  1. Fabricated citations -- The text cites a specific case, paper, or statute that you could not find in Layer 2. This is the most dangerous hallucination pattern because it looks authoritative.
  2. Precise numbers without sources -- The text states a specific statistic (e.g., "78% of companies...") without indicating where the number comes from. Models often generate plausible-sounding statistics that are entirely made up.
  3. Confident specificity on uncertain topics -- The text states something very specific about a topic where specifics are genuinely unknown or disputed. Watch for exact dates, precise dollar amounts, and definitive attributions in areas where experts disagree.
  4. Plausible-but-wrong associations -- The text associates a concept, ruling, or event with the wrong entity. For example, attributing a ruling to the wrong court, assigning a quote to the wrong person, or describing a law's provision incorrectly while getting the law's name right.
  5. Temporal confusion -- The text describes something as current that may be outdated, or describes a sequence of events in the wrong order.
  6. Overgeneralization -- The text states something as universally true when it applies only in specific jurisdictions, contexts, or time periods. Common in legal and regulatory content.
  7. Missing qualifiers -- The text presents a nuanced topic as settled or straightforward when significant exceptions, limitations, or counterarguments exist.
核查以下常见幻觉模式:
  1. 伪造引用——文本提及的特定案例、论文或法规在第二层未找到来源,这是最危险的幻觉模式,因为看起来极具权威性。
  2. 无来源的精确数字——文本给出特定统计数据(比如“78%的企业...”)但未标注来源,模型经常生成听起来合理但完全虚构的统计数据。
  3. 对不确定话题的确定性表述——文本对事实存疑或存在争议的话题给出非常具体的结论,重点关注专家存在分歧领域中的精确日期、准确金额、明确归因等表述。
  4. 看似合理但错误的关联——文本将概念、裁定、事件与错误的实体关联,比如把判决归于错误的法院、把引语安在错误的人名下、正确引用法律名称但错误描述条款内容。
  5. 时间混淆——文本将已过时的内容描述为当前有效,或错误排列事件顺序。
  6. 过度泛化——文本将仅适用于特定司法辖区、场景或时间段的内容描述为普遍适用,在法律和监管内容中非常常见。
  7. 缺失限定条件——文本将存在大量例外、限制或反对意见的复杂话题描述为已达成共识或简单直接。

Adversarial Questions

对抗性问题

For each major claim that passed Layers 1 and 2, ask:
  • What would make this claim wrong?
  • Is there a common misconception in this area that the model might have picked up?
  • If I were a subject matter expert, would I object to how this is stated?
  • Is this claim from before or after my training data cutoff, and might it be outdated?
针对通过前两层验证的所有核心主张,自问:
  • 什么情况下该主张会不成立?
  • 该领域是否存在模型可能习得的常见误区?
  • 如果我是该领域的专家,会反对这种表述吗?
  • 该主张的时间是在我的训练数据截止日期之前还是之后,是否可能已过时?

Red Flags to Escalate

需升级的红色预警

If you find any of these, flag them prominently in the report:
  • A specific citation that cannot be found anywhere
  • A statistic with no identifiable source
  • A legal or regulatory claim that contradicts what authoritative sources say
  • A claim that has been stated with high confidence but is actually disputed or uncertain

如果发现以下任何问题,在报告中醒目标记:
  • 完全找不到来源的特定引用
  • 无明确来源的统计数据
  • 与权威来源内容矛盾的法律或监管主张
  • 高置信度表述但实际存在争议或不确定性的主张

Producing the Verification Report

生成验证报告

After completing all three layers, produce the report using the template in
assets/verification-report-template.md
.
完成所有三层验证后,使用
assets/verification-report-template.md
模板生成报告。

Confidence Ratings

置信度评级

Assign each claim a final rating:
RatingMeaningWhat the user should do
VERIFIEDSupporting source found and linkedSpot-check the source link if the claim is critical to your work
PLAUSIBLEConsistent with general knowledge, no specific source foundTreat as reasonable but unconfirmed; verify independently if relying on it for decisions
UNVERIFIEDCould not find supporting or contradicting evidenceDo not rely on this claim without independent verification
DISPUTEDFound contradicting evidence from a credible sourceReview the contradicting source; this claim may be wrong
FABRICATION RISKMatches hallucination patterns (e.g., unfindable citation, unsourced precise statistic)Assume this is wrong until you can confirm it from a primary source
为每项主张分配最终评级:
评级含义用户应采取的行动
VERIFIED找到支持性来源并附链接如果主张对您的工作至关重要,可抽查来源链接
PLAUSIBLE与常识一致,未找到特定支持来源视为合理但未确认的内容,如果用于决策需独立验证
UNVERIFIED未找到支持或矛盾的证据未独立验证前不得依赖该主张
DISPUTED找到可信来源的矛盾信息查阅矛盾来源,该主张可能有误
FABRICATION RISK符合幻觉模式(比如找不到来源的引用、无来源的精确统计数据)在一手来源确认前默认该主张有误

Report Principles

报告原则

  • Provide links, not verdicts. The user decides what's true, not you.
  • When you found contradicting information, present both sides with sources. Don't pick a winner.
  • If a claim is unfalsifiable (too vague or subjective to verify), say so. "Unfalsifiable" is useful information.
  • Be explicit about what you could not check. "I could not verify this" is different from "this is wrong."
  • Group findings by severity. Lead with the items that need the most attention.
  • 提供链接而非结论,由用户而非工具判定何为真相
  • 找到矛盾信息时,同时展示双方观点和来源,不做倾向性判定
  • 如果主张无法证伪(过于模糊或主观无法验证),明确说明,“无法证伪”本身是有效信息
  • 明确说明你无法核查的内容,“无法验证该主张”与“该主张有误”含义完全不同
  • 按严重程度分组展示结果,优先展示最需关注的问题

Limitations Disclosure

局限性声明

Always include this at the end of the report:
Limitations of this verification:
  • This tool accelerates human verification; it does not replace it.
  • Web search results may not include the most recent information or paywalled sources.
  • The adversarial review uses the same underlying model that may have produced the original output. It catches many issues but cannot catch all of them.
  • A claim rated VERIFIED means a supporting source was found, not that the claim is definitely correct. Sources can be wrong too.
  • Claims rated PLAUSIBLE may still be wrong. The absence of contradicting evidence is not proof of accuracy.

报告末尾必须包含以下内容:
本验证的局限性:
  • 该工具用于加速人工验证,不能替代人工审核。
  • 网页搜索结果可能不包含最新信息或付费墙内的来源。
  • 对抗性审查使用的底层模型可能与生成原始输出的模型相同,能发现很多问题但无法覆盖所有问题。
  • 标记为VERIFIED的主张仅表示找到了支持来源,不代表该主张绝对正确,来源也可能出错。
  • 标记为PLAUSIBLE的主张仍可能错误,不存在矛盾证据不代表其准确。

Domain-Specific Guidance

特定领域指导

Legal Content

法律内容

Legal content carries elevated hallucination risk because:
  • Case names, citations, and holdings are frequently fabricated by models
  • Jurisdictional nuances are often flattened or omitted
  • Statutory language may be paraphrased in ways that change the legal meaning
  • "Majority rule" and "minority rule" distinctions are often lost
For legal content, give extra scrutiny to: case citations, statutory references, regulatory interpretations, and jurisdictional claims. Search legal databases when possible.
法律内容的幻觉风险更高,原因如下:
  • 模型经常伪造案例名、引用和判决结果
  • 司法辖区的细微差异经常被抹平或省略
  • 法规文本的转述可能改变法律含义
  • 经常忽略“多数规则”和“少数规则”的区别
针对法律内容,需额外审查:案例引用、法规参考、监管解读、司法辖区相关主张,尽可能搜索法律数据库验证。

Medical and Scientific Content

医疗和科学内容

  • Check that cited studies actually exist and that the results are accurately described
  • Watch for outdated guidelines being presented as current
  • Flag dosages, treatment protocols, or diagnostic criteria -- these change and errors can be dangerous
  • 核查引用的研究是否真实存在,结果描述是否准确
  • 警惕将已过时的指南描述为当前有效
  • 标记剂量、治疗方案、诊断标准相关内容——此类内容会更新,错误可能造成危险

Financial and Regulatory Content

金融和监管内容

  • Verify specific dollar amounts, dates, and thresholds
  • Check that regulatory requirements are attributed to the correct jurisdiction and are current
  • Watch for tax law claims that may be outdated after recent legislative changes
  • 验证特定金额、日期、阈值
  • 核查监管要求归属的司法辖区是否正确、是否为当前有效规则
  • 警惕税法相关主张,立法更新后可能已过时

Technical and Security Content

技术和安全内容

  • Verify CVE numbers, vulnerability descriptions, and affected versions
  • Check that API specifications and configuration instructions match current documentation
  • Watch for version-specific information that may be outdated
  • 验证CVE编号、漏洞描述、受影响版本
  • 核查API规范和配置说明是否与当前文档一致
  • 警惕特定版本的信息可能已过时