Loading...
Loading...
Found 54 Skills
Verify claims in generated output against sources. Use as a separate pass AFTER content generation to catch hallucinations. Critical constraint - cannot be reliably combined with generation in a single pass.
Detects fabricated content, false citations, and unverifiable claims in agent outputs. Uses source verification and consistency checking. Activate on 'detect hallucination', 'fact check', 'verify claims', 'check accuracy', 'find fabrications'. NOT for validation (use dag-output-validator) or confidence scoring (use dag-confidence-scorer).
Three-layer verification pipeline for AI output. Extracts verifiable claims, finds supporting or contradicting sources via web search, runs adversarial review for hallucination patterns, and produces a structured verification report with source links for human review.
Side role: find and correct bad signals, earn leaderboard points per Publisher-approved correction (max 3/day)
Evaluate source credibility using primary/secondary classification, internal/external criticism, triangulation, and misinformation detection. Use this skill when the user needs to assess whether information is trustworthy, evaluate research sources, fact-check claims, or detect misinformation — even if they say 'can I trust this source', 'is this real', 'how reliable is this data', or 'fact-check this for me'.
Analyze arguments, detect biases, evaluate claims, and improve reasoning. Use when asked to fact-check, identify logical fallacies, evaluate arguments, analyze predictions, find root causes, or think adversarially about plans. Triggers include "evaluate this argument", "logical fallacies", "fact check", "analyze the claims", "identify biases", "devil's advocate", "red team this", "root cause".
Research topics, verify facts, check historical accuracy, and validate technical details. Use when writing non-fiction, historical fiction, or any content requiring factual accuracy.
Research topics with web search. Use when: researching a topic or concept, finding current information, answering factual questions, comparing options or technologies. Triggers: research [topic], find out about, what are the best practices for, research the latest on.
Extracts verifiable claims from ALL .md files (paths, versions, counts, configs, names, endpoints), verifies each against codebase, cross-checks between documents for contradictions.
Validate whether an implementation matches its stated goal. Use this skill when a skill or agent wants a second opinion on its own output, when the user says "check this implementation", "validate what you did", "is this correct?", "review the output", or "did you do this right?". Also spawned automatically as a subagent by other skills (memory-bridge, daily-update) to self-check their outputs before presenting to the user. Returns a structured pass/warn/fail verdict with specific actionable issues.
Extract falsifiable ideas from input, deep-research each one, and return evidence for or against with strength ratings. Use when user says "find evidence for this", "is this true?", "back this up with data", or "fact-check these claims". Honest about when evidence contradicts the idea.
Fact-checks LLM responses by extracting verifiable claims, verifying each via web search, producing an audit report with verdicts, and optionally revising inaccurate responses. Use when the user asks to audit, fact-check, double-check, or verify a response.