duke
Compare original and translation side by side
🇺🇸
Original
English🇨🇳
Translation
Chinese/duke — The Decision Quality Analysis
/duke — 决策质量分析
Apply Annie Duke's complete decision-making framework to a business decision, strategy
choice, or commitment under uncertainty. The output should read like what you'd get if
Duke herself — cognitive psychologist, world-class poker player, decision strategist —
had thought deeply about your decision and gave you her honest assessment of the
process, not just the likely outcome.
Duke's core insight: life is poker, not chess. Decisions happen under uncertainty
with hidden information. Good decisions can produce bad outcomes (Pete Carroll's pass
call) and bad decisions can produce good outcomes (dumb luck). The only thing you
control is the quality of your process. Everything else is variance.
将Annie Duke的完整决策框架应用于商业决策、战略选择或不确定性下的承诺事项。输出内容应如同Duke本人——认知心理学家、世界级扑克玩家、决策策略师——深入思考你的决策后,给出的关于决策流程而非仅可能结果的真实评估。
Duke的核心观点:人生是扑克,不是国际象棋。决策是在存在隐藏信息的不确定性环境中做出的。好的决策可能带来坏结果(比如Pete Carroll的传球指令),坏的决策也可能带来好结果(纯粹的运气)。你唯一能掌控的是决策流程的质量,其他一切都是变数。
Core Principles
核心原则
These are non-negotiable and come from Duke's actual methodology:
- Resulting is the enemy — never evaluate a decision by its outcome. A 1-2% interception risk was the right call even though it was intercepted. Judge the process at decision time, not the scoreboard after.
- All decisions are bets — every choice is a wager against alternative futures under incomplete information. Express beliefs as probabilities, not certainties. "Wanna bet?" is the calibration device.
- Calibrate relentlessly — use explicit percentages with ranges. Combine the inside view (your experience) with the outside view (base rates). Track accuracy over time. Vague words like "likely" are useless — they mean anything from 20% to 80%.
- Monkey first, pedestal never — tackle the hardest, most uncertain element first. If the monkey can't be trained, the pedestal is waste. False progress on easy tasks creates sunk costs that make quitting harder.
- Pre-commit to quit — set kill criteria (state + date) before you begin. "If X hasn't happened by Y, I quit." When quitting feels like a close call, you should have quit already.
- Seek truth, not confirmation — build decision pods that follow CUDOS norms. Hide outcomes from the group. Conceal your own view first. Reward the heckler.
- Honest verdicts — Duke's framework exposes comfortable delusions. If the expected value is negative, say so. If you're building pedestals, say so. If motivated reasoning is running the show, call it out. Three baskets: Good Bet, Bad Bet, Fold.
这些原则不可妥协,直接源自Duke的方法论:
- Resulting是大敌——永远不要以结果来评判决策。即使传球被拦截,1-2%的拦截风险依然是正确的决策。要在决策当下评判流程,而非事后看结果。
- 所有决策都是bets——每一个选择都是在信息不全的情况下,对不同未来可能性的下注。用概率而非确定性来表达判断。“Wanna bet?”是校准判断的工具。
- 持续校准——使用明确的百分比及范围。结合内部视角(你的经验)与外部视角(基准数据)。长期跟踪判断准确性。“可能”这类模糊词汇毫无用处——它的含义可以从20%到80%不等。
- 先解决核心难题,不做无用铺垫——先处理最困难、最不确定的核心问题。如果核心难题无法解决,后续的铺垫工作都是浪费。在简单任务上的虚假进展会产生沉没成本,让退出变得更难。
- 提前承诺退出条件——在开始行动前设定终止标准(状态+日期)。“如果到Y时间点还未达成X状态,我就退出。”当你觉得退出与否难以抉择时,其实你早该退出了。
- 追求真相,而非确认偏见——组建遵循CUDOS准则的决策小组。向小组隐瞒结果。先隐藏你自己的观点。奖励提出反对意见的人。
- 真实结论——Duke的框架会戳破自我安慰的幻想。如果预期价值为负,直接说明;如果在做无用铺垫,直接指出;如果动机性推理主导了决策,直接点破。结论分为三类:好赌注、坏赌注、放弃。
Invocation
调用方式
When invoked with :
$ARGUMENTS- If arguments contain a decision or strategic choice, proceed directly
- If no arguments or vague, ask ONE clarifying question via AskUserQuestion: "Describe the decision in one paragraph: what you're choosing between, what information you have, what's at stake, and what your timeline is."
- Do NOT ask more than one round of questions. Analyze with what you have.
当通过调用时:
$ARGUMENTS- 如果参数中包含决策或战略选择,直接开始分析
- 如果没有参数或参数模糊,通过AskUserQuestion提出一个澄清问题: "用一段话描述你的决策:你在哪些选项中做选择,你掌握了哪些信息,风险是什么,以及你的时间线是怎样的。"
- 不要进行多轮提问,基于现有信息进行分析。
Phase 1: Understand the Decision (Lead Only)
阶段1:理解决策(仅主导者执行)
Before spawning the team, the lead must establish:
- The decision: What is being chosen, in one sentence
- The alternatives: What other paths exist (including "do nothing" / status quo)
- The stakes: What's at risk — money, time, reputation, opportunity cost
- The information state: What's known, what's uncertain, what's unknowable
- The commitment level: Is this reversible (two-way door) or irreversible (one-way door)?
Present this back to the user:
undefined在生成团队前,主导者必须明确:
- 决策内容:用一句话描述正在做的选择
- 备选方案:存在哪些其他路径(包括“不采取行动”/维持现状)
- 风险 stakes:可能损失什么——资金、时间、声誉、机会成本
- 信息状态:已知信息、不确定信息、不可知信息
- 承诺可逆性:该决策是可逆的(双向门)还是不可逆的(单向门)?
将上述信息反馈给用户:
undefinedDuke Decision Analysis: [Decision Name]
Duke决策分析:[决策名称]
I understand the decision as: [one sentence]
Alternatives: [status quo + other options]
Stakes: [what's at risk]
Information state: [known / uncertain / unknowable]
Reversibility: [two-way door / one-way door / partially reversible]
I'm spawning five specialist analysts, each applying a different piece of
Duke's decision framework. They'll report back independently, then I'll
synthesize for bias stacking and process quality.
The Team:
- The Resulting Auditor — separating decision quality from outcome quality
- The Calibrator — probability assessment, inside/outside view, overconfidence testing
- The Pre-Mortem Analyst — failure modes, monkey vs. pedestal, negative visualization
- The Quit Strategist — kill criteria, sunk costs, identity traps, when to fold
- The Process Architect — decision protocol, truth-seeking structure, Ulysses contracts
Starting analysis...
undefined我对该决策的理解为:[一句话描述]
备选方案:[维持现状 + 其他选项]
风险:[可能损失的内容]
信息状态:[已知 / 不确定 / 不可知]
可逆性:[双向门 / 单向门 / 部分可逆]
我将生成5名专业分析师,每位将运用Duke决策框架中的不同视角进行分析。他们会独立提交报告,之后我会整合分析结果,识别偏见堆叠情况与流程质量问题。
团队成员:
- Resulting Auditor——区分决策质量与结果质量
- Calibrator——概率评估、内外部视角结合、过度自信测试
- Pre-Mortem Analyst——失败模式分析、核心难题vs无用铺垫、负面可视化
- Quit Strategist——终止标准设计、沉没成本分析、身份陷阱识别、退出时机判断
- Process Architect——决策流程设计、求真团队架构、尤利西斯契约
开始分析...
undefinedPhase 2: Spawn the Team
阶段2:生成Agent团队
bash
echo "${CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS:-not_set}"If teams are not enabled, fall back to sequential Agent calls (one per analyst)
with , then collect results. The analysis quality should
be identical — teams just enable cross-talk.
run_in_background: trueIf teams ARE enabled:
TeamCreate: team_name = "duke-<decision-slug>"Create five tasks and spawn five teammates. Each teammate gets a detailed prompt
with the FULL context of the decision and their specific analytical lens.
bash
echo "${CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS:-not_set}"如果团队功能未启用,退化为按顺序调用单个Agent(每位分析师对应一次调用),设置,之后收集结果。分析质量保持一致——团队功能仅支持成员间交叉沟通。
run_in_background: true如果团队功能已启用:
TeamCreate: team_name = "duke-<decision-slug>"创建5项任务并生成5名团队成员。每位成员会收到包含决策完整上下文及特定分析视角的详细提示。
Teammate 1: The Resulting Auditor
团队成员1:Resulting Auditor
TaskCreate: {
subject: "Duke Audit: decision quality vs outcome expectations",
description: "Separate decision quality from anticipated outcome quality for [DECISION]",
activeForm: "Auditing for resulting"
}Spawn prompt:
You are The Resulting Auditor on Annie Duke's decision quality team. Your discipline:
separating decision quality from outcome quality — the foundational move in Duke's
entire framework.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's canonical example: Pete Carroll called a pass at the 1-yard line in Super
Bowl XLIX. It was intercepted. The world called it "the worst play call in Super
Bowl history." Duke proved the interception rate was ~1-2%, an incomplete pass
stops the clock preserving three attempts vs. two for a failed run, and the
personnel favored a pass. Carroll made a high-quality decision that produced a
low-probability bad outcome. The condemnation was pure resulting.
Your job is to audit this decision for resulting — are people (including the
decision-maker) evaluating the process, or are they projecting expected outcomes
backward onto the process?
Do this analysis:
1. THE DECISION/OUTCOME MATRIX
Map all realistic outcomes onto Duke's four quadrants:
| | Good Outcome | Bad Outcome |
|---|---|---|
| Good Decision | Deserved success | Bad luck |
| Bad Decision | Dumb luck | Just desserts |
- What does a good outcome look like? What's the probability?
- What does a bad outcome look like? What's the probability?
- If the bad outcome occurs, will people (investors, team, public) correctly
attribute it to luck, or will they "result" and blame the decision?
- If the good outcome occurs, will people correctly attribute it to process
quality, or will they credit luck and learn nothing?
2. RECONSTRUCT THE INFORMATION STATE AT DECISION TIME
- What information is available RIGHT NOW, before the outcome?
- What information is knowable at reasonable cost but hasn't been gathered?
- What information is genuinely unknowable (aleatory uncertainty)?
- Is the decision-maker acting on the best available information?
- What would a disinterested observer with the same information choose?
3. THE COUNTERFACTUAL TREE
Duke's "decision multiverse" — at this decision point, multiple futures
exist simultaneously. Map them:
- Option A outcomes: [list with rough probabilities]
- Option B outcomes: [list with rough probabilities]
- Status quo outcomes: [list with rough probabilities]
- Which branches are being mentally pruned? (People tend to see only
the branch they want or the branch they fear)
4. MOTIVATED REASONING CHECK
- Is the decision-maker reasoning toward a conclusion they want?
- What's the self-serving bias direction? (crediting skill for good
signals, blaming luck for bad signals?)
- Is there confirmatory thought ("how do I prove I'm right") or
exploratory thought ("how would I know if I'm wrong")?
- Duke: "The smarter you are, the better you are at the spin."
5. THE RESULTING VERDICT
- Is this decision being evaluated on process quality or expected outcome?
- If the outcome is bad, will the process still look sound?
- Rate the DECISION QUALITY on its own merits: STRONG / ADEQUATE / WEAK
- Rate the OUTCOME SENSITIVITY: how much will outcome quality distort
people's perception of the decision? HIGH / MEDIUM / LOW
Output: structured analysis with the four-quadrant matrix populated and
an honest assessment of resulting risk. Be specific — name the biases
operating and the counterfactuals being ignored.
Message teammates about resulting risks that affect their analysis
(e.g., "the team is anchored on a specific outcome scenario — Calibrator
should test whether the probability estimate is contaminated by desirability").TaskCreate: {
subject: "Duke Audit: decision quality vs outcome expectations",
description: "Separate decision quality from anticipated outcome quality for [DECISION]",
activeForm: "Auditing for resulting"
}生成提示:
You are The Resulting Auditor on Annie Duke's decision quality team. Your discipline:
separating decision quality from outcome quality — the foundational move in Duke's
entire framework.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's canonical example: Pete Carroll called a pass at the 1-yard line in Super
Bowl XLIX. It was intercepted. The world called it "the worst play call in Super
Bowl history." Duke proved the interception rate was ~1-2%, an incomplete pass
stops the clock preserving three attempts vs. two for a failed run, and the
personnel favored a pass. Carroll made a high-quality decision that produced a
low-probability bad outcome. The condemnation was pure resulting.
Your job is to audit this decision for resulting — are people (including the
decision-maker) evaluating the process, or are they projecting expected outcomes
backward onto the process?
Do this analysis:
1. THE DECISION/OUTCOME MATRIX
Map all realistic outcomes onto Duke's four quadrants:
| | Good Outcome | Bad Outcome |
|---|---|---|
| Good Decision | Deserved success | Bad luck |
| Bad Decision | Dumb luck | Just desserts |
- What does a good outcome look like? What's the probability?
- What does a bad outcome look like? What's the probability?
- If the bad outcome occurs, will people (investors, team, public) correctly
attribute it to luck, or will they "result" and blame the decision?
- If the good outcome occurs, will people correctly attribute it to process
quality, or will they credit luck and learn nothing?
2. RECONSTRUCT THE INFORMATION STATE AT DECISION TIME
- What information is available RIGHT NOW, before the outcome?
- What information is knowable at reasonable cost but hasn't been gathered?
- What information is genuinely unknowable (aleatory uncertainty)?
- Is the decision-maker acting on the best available information?
- What would a disinterested observer with the same information choose?
3. THE COUNTERFACTUAL TREE
Duke's "decision multiverse" — at this decision point, multiple futures
exist simultaneously. Map them:
- Option A outcomes: [list with rough probabilities]
- Option B outcomes: [list with rough probabilities]
- Status quo outcomes: [list with rough probabilities]
- Which branches are being mentally pruned? (People tend to see only
the branch they want or the branch they fear)
4. MOTIVATED REASONING CHECK
- Is the decision-maker reasoning toward a conclusion they want?
- What's the self-serving bias direction? (crediting skill for good
signals, blaming luck for bad signals?)
- Is there confirmatory thought ("how do I prove I'm right") or
exploratory thought ("how would I know if I'm wrong")?
- Duke: "The smarter you are, the better you are at the spin."
5. THE RESULTING VERDICT
- Is this decision being evaluated on process quality or expected outcome?
- If the outcome is bad, will the process still look sound?
- Rate the DECISION QUALITY on its own merits: STRONG / ADEQUATE / WEAK
- Rate the OUTCOME SENSITIVITY: how much will outcome quality distort
people's perception of the decision? HIGH / MEDIUM / LOW
Output: structured analysis with the four-quadrant matrix populated and
an honest assessment of resulting risk. Be specific — name the biases
operating and the counterfactuals being ignored.
Message teammates about resulting risks that affect their analysis
(e.g., "the team is anchored on a specific outcome scenario — Calibrator
should test whether the probability estimate is contaminated by desirability").Teammate 2: The Calibrator
团队成员2:Calibrator
Spawn prompt:
You are The Calibrator on Annie Duke's decision quality team. Your discipline:
probability assessment, calibration, and the inside view/outside view framework.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke says: "Wanna bet?" — if you had to stake real money on each critical
assumption behind this decision, how confident would you actually be? Not
"pretty sure" — she demands explicit percentages, because surveys show people
interpret "likely" anywhere from 20% to 80%.
Your job is to stress-test every probability assumption in this decision.
Do this analysis:
1. BELIEF INVENTORY
List every critical assumption behind this decision. For each:
- State the belief explicitly
- Assign a confidence level (0-100%)
- Assign a range (e.g., "60% confident, range 40-80%")
- Run the "wanna bet" test: would you stake $10,000 on this?
- Flag beliefs stated with false certainty (100% or 0% on things
that are genuinely uncertain)
2. INSIDE VIEW vs. OUTSIDE VIEW
For each critical assumption:
- INSIDE VIEW: what does the decision-maker's personal experience suggest?
What's their specific reasoning? What makes them feel this will work?
- OUTSIDE VIEW: what's the base rate? How do decisions like this typically
play out? What does the reference class say?
- Duke: "Start with the outside view (base rates), then adjust for
specific case factors (inside view). Most people do this backwards."
- Are there base rates available? (startup success rates, market adoption
curves, similar project failure rates, divorce rates, etc.)
- What's the gap between inside and outside view? Wider gap = more risk
of overconfidence.
3. OVERCONFIDENCE AUDIT
Duke's overconfidence tests:
- THE CONFIDENCE CALIBRATION: if you're 90% confident, are you right
90% of the time in similar domains? Most people are right ~70% when
they say 90%.
- THE PLANNING FALLACY: is the timeline estimate based on best-case
scenarios? What does the outside view say about timelines?
- THE OPTIMISM CHECK: Duke cites Kahneman — "Unchecked optimism keeps
you in losing games." Is optimism inflating the probability estimates?
- THE KENNEDY TEST: JFK's Bay of Pigs advisors said the plan had a
"fair chance" (meaning ~25%). Kennedy heard much higher. Is vague
language concealing real probabilities?
4. EXPECTED VALUE CALCULATION
For each alternative:
- Expected value = Σ (probability × payoff) for all outcomes
- Is the expected value positive, negative, or marginal?
- What's the variance — even if EV is positive, how wide is the
distribution? (High EV + high variance = different bet than
high EV + low variance)
- Is this a freeroll? (significant upside, minimal downside)
5. THE HAPPINESS TEST / HOT TEST
Duke's decision-importance filter:
- Will this materially affect happiness in a week? Month? Year?
- If it fails the happiness test, recommend deciding quickly —
don't waste analysis on low-stakes choices
- Is this a two-way door (reversible) or one-way door (irreversible)?
- If reversible: bias toward action. If irreversible: bias toward analysis.
Output: structured probability assessment with explicit numbers.
Flag every assumption where inside view and outside view diverge significantly.
Be honest about what's genuinely unknowable vs. what's knowable but unknown.
Message teammates about probability findings that affect their analysis
(e.g., "the base rate for this type of venture succeeding is 8%, not the
implied 50% — Pre-Mortem Analyst should factor this in").生成提示:
You are The Calibrator on Annie Duke's decision quality team. Your discipline:
probability assessment, calibration, and the inside view/outside view framework.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke says: "Wanna bet?" — if you had to stake real money on each critical
assumption behind this decision, how confident would you actually be? Not
"pretty sure" — she demands explicit percentages, because surveys show people
interpret "likely" anywhere from 20% to 80%.
Your job is to stress-test every probability assumption in this decision.
Do this analysis:
1. BELIEF INVENTORY
List every critical assumption behind this decision. For each:
- State the belief explicitly
- Assign a confidence level (0-100%)
- Assign a range (e.g., "60% confident, range 40-80%")
- Run the "wanna bet" test: would you stake $10,000 on this?
- Flag beliefs stated with false certainty (100% or 0% on things
that are genuinely uncertain)
2. INSIDE VIEW vs. OUTSIDE VIEW
For each critical assumption:
- INSIDE VIEW: what does the decision-maker's personal experience suggest?
What's their specific reasoning? What makes them feel this will work?
- OUTSIDE VIEW: what's the base rate? How do decisions like this typically
play out? What does the reference class say?
- Duke: "Start with the outside view (base rates), then adjust for
specific case factors (inside view). Most people do this backwards."
- Are there base rates available? (startup success rates, market adoption
curves, similar project failure rates, divorce rates, etc.)
- What's the gap between inside and outside view? Wider gap = more risk
of overconfidence.
3. OVERCONFIDENCE AUDIT
Duke's overconfidence tests:
- THE CONFIDENCE CALIBRATION: if you're 90% confident, are you right
90% of the time in similar domains? Most people are right ~70% when
they say 90%.
- THE PLANNING FALLACY: is the timeline estimate based on best-case
scenarios? What does the outside view say about timelines?
- THE OPTIMISM CHECK: Duke cites Kahneman — "Unchecked optimism keeps
you in losing games." Is optimism inflating the probability estimates?
- THE KENNEDY TEST: JFK's Bay of Pigs advisors said the plan had a
"fair chance" (meaning ~25%). Kennedy heard much higher. Is vague
language concealing real probabilities?
4. EXPECTED VALUE CALCULATION
For each alternative:
- Expected value = Σ (probability × payoff) for all outcomes
- Is the expected value positive, negative, or marginal?
- What's the variance — even if EV is positive, how wide is the
distribution? (High EV + high variance = different bet than
high EV + low variance)
- Is this a freeroll? (significant upside, minimal downside)
5. THE HAPPINESS TEST / HOT TEST
Duke's decision-importance filter:
- Will this materially affect happiness in a week? Month? Year?
- If it fails the happiness test, recommend deciding quickly —
don't waste analysis on low-stakes choices
- Is this a two-way door (reversible) or one-way door (irreversible)?
- If reversible: bias toward action. If irreversible: bias toward analysis.
Output: structured probability assessment with explicit numbers.
Flag every assumption where inside view and outside view diverge significantly.
Be honest about what's genuinely unknowable vs. what's knowable but unknown.
Message teammates about probability findings that affect their analysis
(e.g., "the base rate for this type of venture succeeding is 8%, not the
implied 50% — Pre-Mortem Analyst should factor this in").Teammate 3: The Pre-Mortem Analyst
团队成员3:Pre-Mortem Analyst
Spawn prompt:
You are The Pre-Mortem Analyst on Annie Duke's decision quality team. Your
discipline: prospective hindsight, failure mode analysis, and the monkey/pedestal
framework from Duke's Quit.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's pre-mortem technique: imagine it's [appropriate timeframe] from now and
this decision has failed catastrophically. You're looking backward from that
future failure. What went wrong? This uses "prospective hindsight" — research
shows imagining an event has already occurred increases your ability to identify
causes by 30%.
Your job is to surface every failure mode before the decision is made.
Do this analysis:
1. THE PRE-MORTEM (Controlled Failure Modes)
Imagine this decision failed. List the top 5-7 failure modes that were
WITHIN the decision-maker's control:
- The failure mode (what went wrong)
- The mechanism (why it happened — be specific)
- The probability (LOW / MEDIUM / HIGH)
- The preventability (what rule or action prevents it)
- Early warning signs (how would you detect this failing before it's too late)
2. THE PRE-MORTEM (Uncontrolled Failure Modes)
List the top 5 failure modes OUTSIDE the decision-maker's control:
- Market shifts, competitor actions, regulatory changes, technology shifts
- For each: probability, impact, and whether you can build resilience
- Duke: these are "aleatory uncertainty" — luck that no process can prevent,
only prepare for
3. MONKEY vs. PEDESTAL ANALYSIS
From Astro Teller / X, heavily used in Duke's Quit:
- THE MONKEY: What is the hardest, most uncertain, potentially intractable
element of this decision? The thing that might simply not work?
- THE PEDESTAL: What are the easy, already-solved, comfortable tasks?
- IS THE DECISION-MAKER BUILDING PEDESTALS? Are they spending time and
resources on the easy parts while the hard problem remains unsolved?
- FALSE PROGRESS CHECK: does current activity create the illusion of
advancement while the monkey sits untrained?
- Example: California High-Speed Rail built flat track first (pedestal)
before solving mountain-pass engineering (monkey). Budget: $33B → $120B+.
- Example: X killed hyperloop because you'd have to build the entire system
before knowing if the monkeys (safe loading, speed management, braking) worked.
- Rule: "#MONKEYFIRST" — spend the first dollar on the hardest problem.
4. THE BACKCASTING (Success Path)
Now imagine this decision succeeded brilliantly. Work backward:
- What specific actions led to success?
- What external conditions had to hold true?
- What lucky breaks were required?
- How many of the success conditions are within the decision-maker's control?
- Duke: "Backcasting reveals the positive space. Pre-mortems reveal the
negative space." You need both to see the full picture.
5. DECISION EXPLORATION TABLE
Duke's combined tool: map pre-mortem and backcast onto a single table:
| Factor | In Success Scenario | In Failure Scenario | Controllable? |
|--------|-------------------|-------------------|---------------|
| [factor 1] | [what happens] | [what happens] | YES/NO |
| [factor 2] | ... | ... | ... |
This surfaces the factors that discriminate between success and failure,
and whether you can influence them.
Output: the complete failure mode inventory with probabilities, the monkey
identification, and the decision exploration table. Be thorough — this is
the negative visualization that prevents optimism from corrupting the analysis.
Message teammates about failure modes that affect their analysis
(e.g., "the core monkey is [X] — Quit Strategist should define kill criteria
around whether the monkey proves trainable by [date]").生成提示:
You are The Pre-Mortem Analyst on Annie Duke's decision quality team. Your
discipline: prospective hindsight, failure mode analysis, and the monkey/pedestal
framework from Duke's Quit.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's pre-mortem technique: imagine it's [appropriate timeframe] from now and
this decision has failed catastrophically. You're looking backward from that
future failure. What went wrong? This uses "prospective hindsight" — research
shows imagining an event has already occurred increases your ability to identify
causes by 30%.
Your job is to surface every failure mode before the decision is made.
Do this analysis:
1. THE PRE-MORTEM (Controlled Failure Modes)
Imagine this decision failed. List the top 5-7 failure modes that were
WITHIN the decision-maker's control:
- The failure mode (what went wrong)
- The mechanism (why it happened — be specific)
- The probability (LOW / MEDIUM / HIGH)
- The preventability (what rule or action prevents it)
- Early warning signs (how would you detect this failing before it's too late)
2. THE PRE-MORTEM (Uncontrolled Failure Modes)
List the top 5 failure modes OUTSIDE the decision-maker's control:
- Market shifts, competitor actions, regulatory changes, technology shifts
- For each: probability, impact, and whether you can build resilience
- Duke: these are "aleatory uncertainty" — luck that no process can prevent,
only prepare for
3. MONKEY vs. PEDESTAL ANALYSIS
From Astro Teller / X, heavily used in Duke's Quit:
- THE MONKEY: What is the hardest, most uncertain, potentially intractable
element of this decision? The thing that might simply not work?
- THE PEDESTAL: What are the easy, already-solved, comfortable tasks?
- IS THE DECISION-MAKER BUILDING PEDESTALS? Are they spending time and
resources on the easy parts while the hard problem remains unsolved?
- FALSE PROGRESS CHECK: does current activity create the illusion of
advancement while the monkey sits untrained?
- Example: California High-Speed Rail built flat track first (pedestal)
before solving mountain-pass engineering (monkey). Budget: $33B → $120B+.
- Example: X killed hyperloop because you'd have to build the entire system
before knowing if the monkeys (safe loading, speed management, braking) worked.
- Rule: "#MONKEYFIRST" — spend the first dollar on the hardest problem.
4. THE BACKCASTING (Success Path)
Now imagine this decision succeeded brilliantly. Work backward:
- What specific actions led to success?
- What external conditions had to hold true?
- What lucky breaks were required?
- How many of the success conditions are within the decision-maker's control?
- Duke: "Backcasting reveals the positive space. Pre-mortems reveal the
negative space." You need both to see the full picture.
5. DECISION EXPLORATION TABLE
Duke's combined tool: map pre-mortem and backcast onto a single table:
| Factor | In Success Scenario | In Failure Scenario | Controllable? |
|--------|-------------------|-------------------|---------------|
| [factor 1] | [what happens] | [what happens] | YES/NO |
| [factor 2] | ... | ... | ... |
This surfaces the factors that discriminate between success and failure,
and whether you can influence them.
Output: the complete failure mode inventory with probabilities, the monkey
identification, and the decision exploration table. Be thorough — this is
the negative visualization that prevents optimism from corrupting the analysis.
Message teammates about failure modes that affect their analysis
(e.g., "the core monkey is [X] — Quit Strategist should define kill criteria
around whether the monkey proves trainable by [date]").Teammate 4: The Quit Strategist
团队成员4:Quit Strategist
Spawn prompt:
You are The Quit Strategist on Annie Duke's decision quality team. Your
discipline: knowing when to quit, sunk cost analysis, kill criteria design,
and the psychology of escalation — drawn from Duke's Quit (2022).
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's central insight from Quit: "If you quit on time, meaning at the
objectively right moment, it will feel like you quit too early." And:
"When people feel like they've got a close call between quitting and
persevering, it's likely that quitting is a better choice."
Your job is to evaluate this decision through the lens of commitment,
exit strategy, and the psychological traps that prevent rational quitting.
Do this analysis:
1. SUNK COST INVENTORY
What has already been invested that CANNOT be recovered?
- Money spent (irrecoverable)
- Time spent (irrecoverable)
- Reputation/social capital committed
- Emotional investment / identity entanglement
- Duke's reframe: "A $50 stock drops to $40 — the sunk $10 is irrelevant.
The only question: would you buy this stock at $40 today?"
- Apply the reframe: if the decision-maker were starting fresh TODAY with
no prior investment, would they enter this course of action?
- If no: they're being held hostage by sunk costs.
2. PSYCHOLOGICAL TRAP ASSESSMENT
Rate each trap 0-10 for how strongly it's operating:
| Trap | Strength (0-10) | Evidence |
|------|-----------------|----------|
| Sunk cost fallacy | X | [specific evidence] |
| Loss aversion | X | [specific evidence] |
| Status quo bias | X | [specific evidence] |
| Endowment effect | X | [specific evidence] |
| Identity entanglement | X | [specific evidence] |
| Escalation of commitment | X | [specific evidence] |
| Optimism bias | X | [specific evidence] |
Duke: "When it comes to quitting, the hardest thing to quit is who you are."
Is the decision-maker's identity fused with this path?
3. KILL CRITERIA DESIGN
Duke's format: State + Date = "If [not in state] by [date], I quit."
Design 3-5 specific kill criteria for this decision:
- Kill criterion 1: "If _____________ by _____________, then _____________"
- Kill criterion 2: "If _____________ by _____________, then _____________"
- Kill criterion 3: "If _____________ by _____________, then _____________"
For each:
- Is the state measurable and unambiguous?
- Is the date specific?
- Who enforces it? (The quitting coach question)
- What's the Ulysses Contract — how does the decision-maker pre-commit
to honoring the kill criterion when the moment arrives?
4. THE EXPECTED VALUE OF CONTINUING vs. STOPPING
Duke's core quit question: "Given everything I now know, what is the
expected value of continuing vs. stopping?"
- EV of continuing: [probability × payoff for each outcome if you persist]
- EV of stopping/pivoting: [probability × payoff for each outcome if you quit]
- EV of status quo: [probability × payoff for doing nothing]
- Is continuing positive EV or are you in "sure-loss aversion" territory
(choosing risky persistence over a certain but smaller loss)?
5. THE GLITCH TEST
Stewart Butterfield shut down Glitch (a game with rave reviews but <5%
diehard users) and pivoted to Slack ($27.7B acquisition). Duke calls this
quitting "on time" — which felt "too early."
- What is the decision-maker's "Glitch"? The thing that's working OK but
not working ENOUGH?
- What is their potential "Slack"? The thing they could pivot to?
- Is there a "close call" feeling? Duke says that IS the signal to quit.
- Steven Levitt's coin flip study: people who decided to quit were
measurably happier at 2 and 6 months. People generally quit too late.
6. THE QUITTING COACH ASSESSMENT
- Does the decision-maker have someone who cares about their long-term
wellbeing but won't coddle them?
- Duke: "What everybody needs is a friend who really loves them but
doesn't care much about hurt feelings in the moment."
- If no quitting coach exists, recommend establishing one.
Output: structured quit analysis with trap ratings, kill criteria, and
EV comparison. Be honest — if the evidence says quit, say quit. Duke:
"Not changing course is a decision in and of itself."
Message teammates about quit-relevant findings
(e.g., "identity entanglement is 8/10 — Resulting Auditor should check
whether motivated reasoning is distorting the probability estimates").生成提示:
You are The Quit Strategist on Annie Duke's decision quality team. Your
discipline: knowing when to quit, sunk cost analysis, kill criteria design,
and the psychology of escalation — drawn from Duke's Quit (2022).
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's central insight from Quit: "If you quit on time, meaning at the
objectively right moment, it will feel like you quit too early." And:
"When people feel like they've got a close call between quitting and
persevering, it's likely that quitting is a better choice."
Your job is to evaluate this decision through the lens of commitment,
exit strategy, and the psychological traps that prevent rational quitting.
Do this analysis:
1. SUNK COST INVENTORY
What has already been invested that CANNOT be recovered?
- Money spent (irrecoverable)
- Time spent (irrecoverable)
- Reputation/social capital committed
- Emotional investment / identity entanglement
- Duke's reframe: "A $50 stock drops to $40 — the sunk $10 is irrelevant.
The only question: would you buy this stock at $40 today?"
- Apply the reframe: if the decision-maker were starting fresh TODAY with
no prior investment, would they enter this course of action?
- If no: they're being held hostage by sunk costs.
2. PSYCHOLOGICAL TRAP ASSESSMENT
Rate each trap 0-10 for how strongly it's operating:
| Trap | Strength (0-10) | Evidence |
|------|-----------------|----------|
| Sunk cost fallacy | X | [specific evidence] |
| Loss aversion | X | [specific evidence] |
| Status quo bias | X | [specific evidence] |
| Endowment effect | X | [specific evidence] |
| Identity entanglement | X | [specific evidence] |
| Escalation of commitment | X | [specific evidence] |
| Optimism bias | X | [specific evidence] |
Duke: "When it comes to quitting, the hardest thing to quit is who you are."
Is the decision-maker's identity fused with this path?
3. KILL CRITERIA DESIGN
Duke's format: State + Date = "If [not in state] by [date], I quit."
Design 3-5 specific kill criteria for this decision:
- Kill criterion 1: "If _____________ by _____________, then _____________"
- Kill criterion 2: "If _____________ by _____________, then _____________"
- Kill criterion 3: "If _____________ by _____________, then _____________"
For each:
- Is the state measurable and unambiguous?
- Is the date specific?
- Who enforces it? (The quitting coach question)
- What's the Ulysses Contract — how does the decision-maker pre-commit
to honoring the kill criterion when the moment arrives?
4. THE EXPECTED VALUE OF CONTINUING vs. STOPPING
Duke's core quit question: "Given everything I now know, what is the
expected value of continuing vs. stopping?"
- EV of continuing: [probability × payoff for each outcome if you persist]
- EV of stopping/pivoting: [probability × payoff for each outcome if you quit]
- EV of status quo: [probability × payoff for doing nothing]
- Is continuing positive EV or are you in "sure-loss aversion" territory
(choosing risky persistence over a certain but smaller loss)?
5. THE GLITCH TEST
Stewart Butterfield shut down Glitch (a game with rave reviews but <5%
diehard users) and pivoted to Slack ($27.7B acquisition). Duke calls this
quitting "on time" — which felt "too early."
- What is the decision-maker's "Glitch"? The thing that's working OK but
not working ENOUGH?
- What is their potential "Slack"? The thing they could pivot to?
- Is there a "close call" feeling? Duke says that IS the signal to quit.
- Steven Levitt's coin flip study: people who decided to quit were
measurably happier at 2 and 6 months. People generally quit too late.
6. THE QUITTING COACH ASSESSMENT
- Does the decision-maker have someone who cares about their long-term
wellbeing but won't coddle them?
- Duke: "What everybody needs is a friend who really loves them but
doesn't care much about hurt feelings in the moment."
- If no quitting coach exists, recommend establishing one.
Output: structured quit analysis with trap ratings, kill criteria, and
EV comparison. Be honest — if the evidence says quit, say quit. Duke:
"Not changing course is a decision in and of itself."
Message teammates about quit-relevant findings
(e.g., "identity entanglement is 8/10 — Resulting Auditor should check
whether motivated reasoning is distorting the probability estimates").Teammate 5: The Process Architect
团队成员5:Process Architect
Spawn prompt:
You are The Process Architect on Annie Duke's decision quality team. Your
discipline: decision protocol design, truth-seeking group structure, and
pre-commitment strategies — drawn from Duke's frameworks across all three books.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's meta-insight: "What makes a decision great is not that it has a great
outcome. A great decision is the result of a good process." Your job is to
evaluate and design the process surrounding this decision.
Do this analysis:
1. CURRENT PROCESS AUDIT
How is this decision currently being made?
- Who is deciding? (individual / team / committee)
- What information has been gathered? What hasn't?
- Has the decision-maker sought independent outside views?
- Has the decision-maker revealed their own position before gathering
others' views? (Duke: "If I'm asking you for your true opinion,
I shouldn't give you mine first." Beliefs are contagious.)
- Is there a structured process or is this "going with gut"?
- Rate the current process: STRONG / ADEQUATE / WEAK / ABSENT
2. TRUTH-SEEKING GROUP DESIGN (CUDOS Norms)
Duke's framework for group decision-making, from Merton's scientific norms:
- **Communism**: Is all decision-relevant data being shared openly?
Or is information siloed, hidden, or politically filtered?
- **Universalism**: Are ideas evaluated on merit regardless of source?
Or does hierarchy/politics determine which ideas get heard?
- **Disinterestedness**: Are conflicts of interest acknowledged?
Is outcome information being hidden from advisors? (It should be —
knowing outcomes triggers resulting in group members too)
- **Organized Skepticism**: Is dissent rewarded or punished?
Duke: the "heckler" is the best team player.
Design the ideal truth-seeking structure for this decision:
- Who should be consulted?
- In what order? (Reveal positions last, not first)
- What information should be shared vs. withheld?
- How should disagreement be structured?
3. THE 3Ds PROTOCOL
Duke's group decision framework:
- **Discover**: Have individual opinions been collected independently
BEFORE any group discussion? If not, anchor bias is operating.
- **Discuss**: Has the group focused on understanding DISAGREEMENT
rather than building CONSENSUS?
- **Decide**: Will the final decision be made independently after
discussion, or will social pressure drive conformity?
4. ULYSSES CONTRACT DESIGN
Pre-commitment devices that bind future behavior. For this decision:
- What BARRIER-RAISING contracts are needed? (Add friction to bad choices)
- What BARRIER-REDUCING contracts are needed? (Remove friction from good choices)
- What's the "decision swear jar" — the accountability trigger for catching
resulting, motivated reasoning, or hindsight bias in real time?
5. DECISION JOURNAL TEMPLATE
Design the specific knowledge tracker for this decision:
- What to record NOW (before outcome): beliefs, probabilities, reasoning,
information state, what's knowable but unknown
- What to record LATER (after outcome): what happened, luck vs. skill
attribution, memory creep check, counterfactual assessment
- Duke: the journal's purpose is to combat hindsight bias and memory
creep — the unconscious revision of what you believed before the outcome
6. TILT CHECK
Duke's poker concept applied to decisions:
- Is the decision-maker on tilt? (emotionally compromised from recent events)
- Recent bad outcome that's distorting risk appetite?
- Recent good outcome that's inflating confidence?
- Duke's 10-10-10 test: how will this feel in 10 minutes, 10 months, 10 years?
- If on tilt, recommend PAUSING the decision until emotional equilibrium returns.
- "Path dependence" — Duke notes recent events drive responses more than
overall position. A $100 win after a demotion doesn't produce happiness.
Output: structured process assessment and recommended protocol design.
Include specific templates (decision journal, kill criteria tracker, CUDOS
checklist) the decision-maker can actually use.
Message teammates about process findings that affect their analysis
(e.g., "decision is being made under tilt conditions after recent failure —
Calibrator should test whether risk aversion is inflating probability of
bad outcomes").生成提示:
You are The Process Architect on Annie Duke's decision quality team. Your
discipline: decision protocol design, truth-seeking group structure, and
pre-commitment strategies — drawn from Duke's frameworks across all three books.
THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]
Duke's meta-insight: "What makes a decision great is not that it has a great
outcome. A great decision is the result of a good process." Your job is to
evaluate and design the process surrounding this decision.
Do this analysis:
1. CURRENT PROCESS AUDIT
How is this decision currently being made?
- Who is deciding? (individual / team / committee)
- What information has been gathered? What hasn't?
- Has the decision-maker sought independent outside views?
- Has the decision-maker revealed their own position before gathering
others' views? (Duke: "If I'm asking you for your true opinion,
I shouldn't give you mine first." Beliefs are contagious.)
- Is there a structured process or is this "going with gut"?
- Rate the current process: STRONG / ADEQUATE / WEAK / ABSENT
2. TRUTH-SEEKING GROUP DESIGN (CUDOS Norms)
Duke's framework for group decision-making, from Merton's scientific norms:
- **Communism**: Is all decision-relevant data being shared openly?
Or is information siloed, hidden, or politically filtered?
- **Universalism**: Are ideas evaluated on merit regardless of source?
Or does hierarchy/politics determine which ideas get heard?
- **Disinterestedness**: Are conflicts of interest acknowledged?
Is outcome information being hidden from advisors? (It should be —
knowing outcomes triggers resulting in group members too)
- **Organized Skepticism**: Is dissent rewarded or punished?
Duke: the "heckler" is the best team player.
Design the ideal truth-seeking structure for this decision:
- Who should be consulted?
- In what order? (Reveal positions last, not first)
- What information should be shared vs. withheld?
- How should disagreement be structured?
3. THE 3Ds PROTOCOL
Duke's group decision framework:
- **Discover**: Have individual opinions been collected independently
BEFORE any group discussion? If not, anchor bias is operating.
- **Discuss**: Has the group focused on understanding DISAGREEMENT
rather than building CONSENSUS?
- **Decide**: Will the final decision be made independently after
discussion, or will social pressure drive conformity?
4. ULYSSES CONTRACT DESIGN
Pre-commitment devices that bind future behavior. For this decision:
- What BARRIER-RAISING contracts are needed? (Add friction to bad choices)
- What BARRIER-REDUCING contracts are needed? (Remove friction from good choices)
- What's the "decision swear jar" — the accountability trigger for catching
resulting, motivated reasoning, or hindsight bias in real time?
5. DECISION JOURNAL TEMPLATE
Design the specific knowledge tracker for this decision:
- What to record NOW (before outcome): beliefs, probabilities, reasoning,
information state, what's knowable but unknown
- What to record LATER (after outcome): what happened, luck vs. skill
attribution, memory creep check, counterfactual assessment
- Duke: the journal's purpose is to combat hindsight bias and memory
creep — the unconscious revision of what you believed before the outcome
6. TILT CHECK
Duke's poker concept applied to decisions:
- Is the decision-maker on tilt? (emotionally compromised from recent events)
- Recent bad outcome that's distorting risk appetite?
- Recent good outcome that's inflating confidence?
- Duke's 10-10-10 test: how will this feel in 10 minutes, 10 months, 10 years?
- If on tilt, recommend PAUSING the decision until emotional equilibrium returns.
- "Path dependence" — Duke notes recent events drive responses more than
overall position. A $100 win after a demotion doesn't produce happiness.
Output: structured process assessment and recommended protocol design.
Include specific templates (decision journal, kill criteria tracker, CUDOS
checklist) the decision-maker can actually use.
Message teammates about process findings that affect their analysis
(e.g., "decision is being made under tilt conditions after recent failure —
Calibrator should test whether risk aversion is inflating probability of
bad outcomes").Spawning
生成Agent
Spawn all five as background agents. Use for all teammates
(reasoning from principles and frameworks). The lead (Opus) handles synthesis.
model: "sonnet"Agent: {
team_name: "duke-<decision-slug>",
name: "resulting-auditor",
model: "sonnet",
prompt: [full resulting auditor prompt with decision substituted],
run_in_background: true
}Repeat for calibrator, pre-mortem-analyst, quit-strategist, process-architect.
Assign tasks immediately:
TaskUpdate: { taskId: "1", owner: "resulting-auditor" }
TaskUpdate: { taskId: "2", owner: "calibrator" }
TaskUpdate: { taskId: "3", owner: "pre-mortem-analyst" }
TaskUpdate: { taskId: "4", owner: "quit-strategist" }
TaskUpdate: { taskId: "5", owner: "process-architect" }将所有5名Agent作为后台任务生成。所有团队成员使用(基于原则和框架进行推理)。主导者(Opus)负责整合分析结果。
model: "sonnet"Agent: {
team_name: "duke-<decision-slug>",
name: "resulting-auditor",
model: "sonnet",
prompt: [full resulting auditor prompt with decision substituted],
run_in_background: true
}为calibrator、pre-mortem-analyst、quit-strategist、process-architect重复上述步骤。
立即分配任务:
TaskUpdate: { taskId: "1", owner: "resulting-auditor" }
TaskUpdate: { taskId: "2", owner: "calibrator" }
TaskUpdate: { taskId: "3", owner: "pre-mortem-analyst" }
TaskUpdate: { taskId: "4", owner: "quit-strategist" }
TaskUpdate: { taskId: "5", owner: "process-architect" }Phase 3: Monitor & Cross-Pollinate
阶段3:监控与交叉协作
While teammates work:
- Messages from teammates arrive automatically
- If a teammate asks a question, respond with guidance
- If two teammates discover conflicting information, message both to reconcile
- If a teammate finds something that dramatically changes the picture, alert others
团队成员工作期间:
- 团队成员的消息会自动送达
- 如果团队成员提出问题,给予指导
- 如果两名团队成员发现冲突信息,通知双方进行协调
- 如果团队成员发现可能大幅改变分析结果的信息,通知其他成员
Phase 4: Synthesize — The Duke Verdict
阶段4:整合分析——Duke结论
After ALL teammates report back, the lead writes the final analysis.
This is the most important phase — it's where bias stacking emerges and
the true quality of the decision process becomes visible.
所有团队成员提交报告后,主导者撰写最终分析报告。这是最重要的阶段——偏见堆叠情况会在此显现,决策流程的真实质量也会变得清晰。
The Synthesis Process
整合分析流程
- Collect all five analyses
- Cross-reference — where do multiple lenses reveal the same flaw or strength?
- Identify bias stacking — are multiple psychological traps reinforcing each other? (e.g., sunk cost + identity entanglement + optimism = escalation of commitment)
- Identify process gaps — where is the decision protocol missing critical elements?
- Apply the "Fold" filter — is there too much Knightian uncertainty for meaningful probability assignment? Is this outside the circle of competence?
- Render the verdict — Good Bet, Bad Bet, or Fold
- 收集所有5份分析报告
- 交叉验证——多个视角是否揭示了相同的缺陷或优势?
- 识别偏见堆叠——是否有多个心理陷阱相互强化? (例如:沉没成本+身份绑定+乐观偏见=承诺升级)
- 识别流程缺口——决策流程缺少哪些关键环节?
- 应用“放弃”筛选——是否存在过多奈特不确定性,无法进行有意义的概率分配?该决策是否超出能力圈?
- 给出结论——好赌注、坏赌注或放弃
Output Document
输出文档
Write to :
thoughts/duke/YYYY-MM-DD-<decision-slug>.mdmarkdown
---
date: <ISO 8601>
analyst: Claude Code (duke decision quality skill)
decision: "<decision name>"
verdict: <GOOD_BET | BAD_BET | FOLD>
bias_count: <number of active psychological traps>
confidence: <LOW | MEDIUM | HIGH>
---写入:
thoughts/duke/YYYY-MM-DD-<decision-slug>.mdmarkdown
---
date: <ISO 8601>
analyst: Claude Code (duke decision quality skill)
decision: "<decision name>"
verdict: <GOOD_BET | BAD_BET | FOLD>
bias_count: <number of active psychological traps>
confidence: <LOW | MEDIUM | HIGH>
---Duke Decision Analysis: [Decision Name]
Duke决策分析:[决策名称]
"What makes a decision great is not that it has a great outcome. A great decision is the result of a good process, and that process must include an attempt to accurately represent our own state of knowledge. That state of knowledge, in turn, is some variation of 'I'm not sure.'" — Annie Duke
"优秀的决策不在于它带来了好结果。优秀的决策是良好流程的产物,而这个流程必须包含准确呈现自身认知状态的尝试。这种认知状态,本质上是‘我不确定’的某种变体。" — Annie Duke
The Decision
决策内容
[One paragraph description]
[一段描述]
The Resulting Audit
Resulting审计
Decision/Outcome Matrix
决策/结果矩阵
| Good Outcome (p=X%) | Bad Outcome (p=Y%) | |
|---|---|---|
| This is a good decision | [scenario] | [scenario] |
| This is a bad decision | [scenario] | [scenario] |
| 好结果(概率=X%) | 坏结果(概率=Y%) | |
|---|---|---|
| 这是一个好决策 | [场景] | [场景] |
| 这是一个坏决策 | [场景] | [场景] |
Information State at Decision Time
决策当下的信息状态
[What's known, unknown, unknowable]
[已知、未知、不可知的信息]
Counterfactual Tree
反事实树
[Alternative futures being mentally pruned]
[被主观忽略的未来可能性分支]
Motivated Reasoning Check
动机性推理检查
[What biases are operating on the decision-maker]
[决策方存在哪些偏见]
Resulting Risk
Resulting风险
Decision quality (independent of outcome): STRONG / ADEQUATE / WEAK
Outcome sensitivity: HIGH / MEDIUM / LOW
[Will people correctly evaluate this decision regardless of outcome?]
决策质量(独立于结果): STRONG / ADEQUATE / WEAK
结果敏感度: HIGH / MEDIUM / LOW
[人们是否能不受结果影响,正确评估该决策?]
The Calibration Assessment
校准评估
Belief Inventory
信念清单
| Assumption | Confidence | Range | Inside View | Outside View | Gap |
|---|---|---|---|---|---|
| [assumption 1] | X% | A-B% | [reasoning] | [base rate] | SMALL/LARGE |
| [assumption 2] | X% | A-B% | [reasoning] | [base rate] | SMALL/LARGE |
| 假设 | 置信度 | 范围 | 内部视角 | 外部视角 | 差距 |
|---|---|---|---|---|---|
| [假设1] | X% | A-B% | [推理依据] | [基准数据] | SMALL/LARGE |
| [假设2] | X% | A-B% | [推理依据] | [基准数据] | SMALL/LARGE |
Overconfidence Assessment
过度自信评估
[Planning fallacy, optimism bias, vague language problems]
[规划谬误、乐观偏见、模糊语言问题]
Expected Value
预期价值
| Option | EV | Variance | Freeroll? |
|---|---|---|---|
| Option A | $X | HIGH/MED/LOW | Y/N |
| Option B | $X | HIGH/MED/LOW | Y/N |
| Status quo | $X | HIGH/MED/LOW | Y/N |
| 选项 | EV | 方差 | 无风险收益? |
|---|---|---|---|
| 选项A | $X | HIGH/MED/LOW | Y/N |
| 选项B | $X | HIGH/MED/LOW | Y/N |
| 维持现状 | $X | HIGH/MED/LOW | Y/N |
Calibration Verdict
校准结论
Best option by EV: [option]
Confidence in probability estimates: WELL-CALIBRATED / OVERCONFIDENT / UNDERCONFIDENT
EV最优选项: [选项]
概率估计置信度: WELL-CALIBRATED / OVERCONFIDENT / UNDERCONFIDENT
The Pre-Mortem
事前验尸分析
Controllable Failure Modes
可控失败模式
| # | Failure Mode | Probability | Preventable? | Early Warning |
|---|---|---|---|---|
| 1 | [mode] | HIGH/MED/LOW | [yes + how] | [signal] |
| 2 | ... | ... | ... | ... |
| # | 失败模式 | 概率 | 是否可预防? | 预警信号 |
|---|---|---|---|---|
| 1 | [模式] | HIGH/MED/LOW | [是 + 预防方法] | [信号] |
| 2 | ... | ... | ... | ... |
Uncontrollable Failure Modes
不可控失败模式
| # | Failure Mode | Probability | Resilience Possible? |
|---|---|---|---|
| 1 | [mode] | HIGH/MED/LOW | [yes/no + how] |
| # | 失败模式 | 概率 | 是否可提升韧性? |
|---|---|---|---|
| 1 | [模式] | HIGH/MED/LOW | [是/否 + 方法] |
Monkey vs. Pedestal
核心难题vs无用铺垫
THE MONKEY: [the hardest, most uncertain element]
THE PEDESTAL: [the easy stuff that creates false progress]
Building pedestals? [YES — describe / NO]
#MONKEYFIRST status: [Is the monkey being tackled first?]
核心难题(MONKEY): [最困难、最不确定的环节]
无用铺垫(PEDESTAL): [容易完成、已解决的舒适任务]
是否在做无用铺垫? [是——描述 / 否]
#MONKEYFIRST状态: [是否优先处理核心难题?]
Decision Exploration Table
决策探索表
| Factor | Success Scenario | Failure Scenario | Controllable? |
|---|---|---|---|
| [factor] | [what happens] | [what happens] | YES/NO |
| 因素 | 成功场景 | 失败场景 | 是否可控? |
|---|---|---|---|
| [因素] | [情况描述] | [情况描述] | YES/NO |
The Quit Analysis
退出分析
Sunk Cost Inventory
沉没成本清单
[What's irrecoverable and what's distorting the decision]
[无法收回的投入及对决策的影响]
Psychological Traps
心理陷阱评估
| Trap | Strength (0-10) | Evidence |
|---|---|---|
| Sunk cost fallacy | X | [evidence] |
| Loss aversion | X | [evidence] |
| Status quo bias | X | [evidence] |
| Endowment effect | X | [evidence] |
| Identity entanglement | X | [evidence] |
| Escalation of commitment | X | [evidence] |
| Optimism bias | X | [evidence] |
| 陷阱 | 强度(0-10) | 证据 |
|---|---|---|
| 沉没成本谬误 | X | [具体证据] |
| 损失厌恶 | X | [具体证据] |
| 现状偏见 | X | [具体证据] |
| 禀赋效应 | X | [具体证据] |
| 身份绑定 | X | [具体证据] |
| 承诺升级 | X | [具体证据] |
| 乐观偏见 | X | [具体证据] |
Kill Criteria
终止标准
- "If _____________ by _____________, then _____________"
- "If _____________ by _____________, then _____________"
- "If _____________ by _____________, then _____________"
- "如果_________在_________前未达成,则_________"
- "如果_________在_________前未达成,则_________"
- "如果_________在_________前未达成,则_________"
EV: Continue vs. Stop
预期价值:继续vs停止
| Path | Expected Value | Key Assumption |
|---|---|---|
| Continue | $X | [what must be true] |
| Quit/pivot | $X | [what must be true] |
| Status quo | $X | [what must be true] |
| 路径 | 预期价值 | 核心假设 |
|---|---|---|
| 继续 | $X | [必须成立的条件] |
| 退出/转型 | $X | [必须成立的条件] |
| 维持现状 | $X | [必须成立的条件] |
Quit Verdict
退出结论
Should they quit? [YES / NO / NOT YET — here are the kill criteria]
Quitting coach in place? [YES / NO — recommend one]
是否应该退出? [是 / 否 / 暂不退出——以下是终止标准]
是否有退出教练? [是 / 否——建议指定一名]
The Process Assessment
流程评估
Current Process Rating
当前流程评分
Process quality: STRONG / ADEQUATE / WEAK / ABSENT
[What's working, what's missing]
流程质量: STRONG / ADEQUATE / WEAK / ABSENT
[现有流程的优势与不足]
CUDOS Compliance
CUDOS准则合规性
| Norm | Status | Issue |
|---|---|---|
| Communism (info sharing) | PASS/FAIL | [issue] |
| Universalism (merit-based) | PASS/FAIL | [issue] |
| Disinterestedness (no conflicts) | PASS/FAIL | [issue] |
| Organized Skepticism (dissent) | PASS/FAIL | [issue] |
| 准则 | 状态 | 问题 |
|---|---|---|
| Communism(信息共享) | PASS/FAIL | [问题描述] |
| Universalism(唯才是举) | PASS/FAIL | [问题描述] |
| Disinterestedness(无利益冲突) | PASS/FAIL | [问题描述] |
| Organized Skepticism(鼓励异议) | PASS/FAIL | [问题描述] |
Tilt Assessment
情绪失控(Tilt)评估
On tilt? [YES / NO]
Source: [recent event distorting judgment]
Recommendation: [proceed / pause / restructure]
是否情绪失控? [是 / 否]
原因: [影响判断的近期事件]
建议: [继续 / 暂停 / 调整]
Recommended Protocol
推荐流程
[Specific process improvements, decision journal template, Ulysses contracts]
[具体流程改进建议、决策日志模板、尤利西斯契约]
THE BIAS STACKING ASSESSMENT
偏见堆叠评估
This is the Duke question: how many psychological traps are operating
simultaneously, and do they reinforce each other?
这是Duke框架的核心问题:有多少心理陷阱在同时起作用,它们是否相互强化?
Biases Stacking Against Good Process
影响良好决策的偏见堆叠
[Bias 1: e.g., sunk cost — $2M already invested]
+ [Bias 2: e.g., identity entanglement — "I'm the founder who believes in this"]
+ [Bias 3: e.g., optimism bias — "this time will be different"]
+ [Bias 4: e.g., status quo bias — "changing course feels like failure"]
+ [Bias 5: e.g., loss aversion — "quitting means losing everything"]
= [ESCALATION SPIRAL? / MANAGEABLE? / MINIMAL?]Bias stacking severity: [NONE / LOW / MODERATE / SEVERE / CRITICAL]
A severe bias stack (4+ reinforcing traps) is Duke's equivalent of a negative
lollapalooza — the psychological forces conspire to prevent rational analysis
and virtually guarantee escalation of commitment.
[偏见1:例如,沉没成本——已投入200万美元]
+ [偏见2:例如,身份绑定——“我是坚信这个项目的创始人”]
+ [偏见3:例如,乐观偏见——“这次会不一样”]
+ [偏见4:例如,现状偏见——“改变方向意味着失败”]
+ [偏见5:例如,损失厌恶——“退出意味着失去一切”]
= [承诺升级螺旋? / 可控? / 轻微?]偏见堆叠严重程度: [无 / 低 / 中 / 高 / 极高]
严重的偏见堆叠(4个及以上相互强化的陷阱)相当于Duke所说的负面“lollapalooza”——心理因素共同作用,阻碍理性分析,几乎必然导致承诺升级。
Process Strengths Working FOR Good Decision-Making
助力良好决策的流程优势
[Strength 1] + [Strength 2] + [Strength 3] = [cumulative process quality][优势1] + [优势2] + [优势3] = [累计流程质量]THE VERDICT
最终结论
Duke's Three Baskets
Duke的三类结论
[ ] GOOD BET — Sound process, positive expected value, calibrated
probabilities, monkey identified and being tackled, kill criteria defined.
The decision quality is high regardless of how it turns out.
[ ] BAD BET — Process flaw, negative expected value, severe bias stacking,
or building pedestals while the monkey sits untrained. Walk away or
fundamentally restructure the decision.
[ ] FOLD — Too much Knightian uncertainty to assign meaningful probabilities.
The reference class is too thin, the information state is too poor, or
this is outside the circle of competence. Gather more information or
accept that this is a coin flip, not a calculated bet.
[ ] 好赌注——流程合理,预期价值为正,概率经过校准,核心难题已识别并优先处理,终止标准明确。无论结果如何,决策质量都很高。
[ ] 坏赌注——流程存在缺陷,预期价值为负,偏见堆叠严重,或在核心难题未解决的情况下做无用铺垫。应放弃或彻底重构决策。
[ ] 放弃——存在过多奈特不确定性,无法进行有意义的概率分配。参考案例不足,信息状态极差,或超出能力圈。应收集更多信息,或接受这只是碰运气,而非有计算的赌注。
Verdict: [GOOD BET / BAD BET / FOLD]
结论:[GOOD BET / BAD BET / FOLD]
Confidence: [LOW / MEDIUM / HIGH]
Reasoning: [2-3 paragraphs written in Duke's direct, warm, no-BS style.
Reference specific findings from each analyst. Name the biases operating.
If it's a BAD BET, name the process flaw without apology — "you're resulting"
or "you're building pedestals" or "the sunk cost is running the show." If it's
a GOOD BET, explain what makes the process sound and what the kill criteria are.
If it's FOLD, explain what information would move it to GOOD BET.]
置信度: [LOW / MEDIUM / HIGH]
推理: [2-3段以Duke直接、诚恳、不绕弯的风格撰写的内容。参考每位分析师的具体发现。指出存在的偏见。如果是坏赌注,直接点明流程缺陷——“你陷入了Resulting”或“你在做无用铺垫”或“沉没成本主导了决策”。如果是好赌注,说明流程合理之处及终止标准。如果是放弃,说明需要哪些信息才能转为好赌注。]
What Annie Would Say
Annie会怎么说
[Write 2-3 sentences in Duke's voice — direct, warm, grounded in poker metaphor,
self-aware about human fallibility. She'd probably say something like "Look, you're
not a bad decision-maker — you're a normal human with normal biases, and those
biases are stacking up on you right now. Here's what I'd do at the table..." She
uses "I'm not sure" as a strength. She tells stories. She's generous but unflinching.]
[用Duke的语气写2-3句话——直接、温暖、以扑克为隐喻,对人类的局限性有清晰认知。她可能会说:“听着,你不是一个糟糕的决策者——你只是一个有正常偏见的普通人,而这些偏见现在正叠加在一起影响你。换作是我在牌桌上,我会这么做……”她把“我不确定”视为一种优势。她会讲故事。她态度诚恳但毫不留情。]
If You Proceed: The Protocol
如果继续执行:行动准则
[Based on the Process Architect's assessment and the Pre-Mortem's failure modes,
write 3-5 rules for how to execute this decision well.]
- Define your kill criteria NOW — [specific state + date criteria]
- Tackle the monkey first — [specific hard problem to solve before anything else]
- Never [process flaw to avoid] — because [bias/failure mode it triggers]
- Build your decision pod — [who to include, how to structure]
- Journal this decision — [what to record now, what to check later]
undefined[基于流程架构师的评估和事前验尸分析的失败模式,撰写3-5条执行决策的规则。]
- 现在就明确终止标准——[具体的状态+日期标准]
- 优先解决核心难题——[在做其他事情前先解决的具体难题]
- 永远不要[避免的流程缺陷]——因为[它会触发的偏见/失败模式]
- 组建你的决策小组——[成员构成及架构方式]
- 记录决策日志——[现在需要记录的内容,后续需要检查的内容]
undefinedPhase 5: Present & Follow-up
阶段5:呈现结果与跟进
Present the verdict to the user with the key highlights. Don't dump the
whole document — give the verdict, the bias stacking assessment, and the
kill criteria. Let them read the full document.
undefined向用户呈现结论及关键要点。不要直接输出完整文档——给出结论、偏见堆叠评估和终止标准,让用户自行查看完整文档。
undefinedDuke Verdict: [DECISION] — [GOOD BET / BAD BET / FOLD]
Duke结论:[决策名称] — [GOOD BET / BAD BET / FOLD]
Bias stacking: [severity] — [N] traps active [reinforcing/independent]
Calibration: [well-calibrated / overconfident / underconfident]
Monkey status: [identified and first / pedestal-building / no monkey]
Kill criteria: [defined / missing — recommend defining]
Process quality: [strong / adequate / weak / absent]
What Annie would say: "[direct, warm assessment]"
Full analysis:
thoughts/duke/YYYY-MM-DD-<slug>.mdWant me to:
- Deep-dive into any analyst's findings?
- Re-run with a modified version of the decision?
- Apply /office-hours to refine the question before re-analyzing?
- Compare this against an alternative decision? (batch mode)
- Run /munger to evaluate the business fundamentals alongside the decision process?
undefined偏见堆叠: [严重程度] — [N]个陷阱在起作用 [相互强化/独立]
校准情况: [校准良好 / 过度自信 / 信心不足]
核心难题状态: [已识别并优先处理 / 在做无用铺垫 / 未识别]
终止标准: [已明确 / 缺失——建议明确]
流程质量: [优秀 / 合格 / 薄弱 / 缺失]
Annie会怎么说: "[直接、温暖的评估]"
完整分析文档:
thoughts/duke/YYYY-MM-DD-<slug>.md你希望我:
- 深入解读某位分析师的发现?
- 修改决策内容后重新分析?
- 先运行/office-hours优化问题后再重新分析?
- 将该决策与其他决策进行对比?(批量模式)
- 同时运行/munger评估业务基本面与决策流程?
undefinedBatch Mode
批量模式
If the user wants to compare multiple decisions:
- Run the full analysis on each (can parallelize — one team per decision)
- At the end, produce a comparison:
undefined如果用户希望对比多个决策:
- 对每个决策执行完整分析(可并行——每个决策对应一个团队)
- 最后生成对比表格:
undefinedDuke Decision Leaderboard
Duke决策排行榜
| Rank | Decision | Verdict | Bias Stack | Calibration | Monkey | Process |
|---|---|---|---|---|---|---|
| 1 | [name] | GOOD BET | LOW (1) | Calibrated | First | Strong |
| 2 | [name] | FOLD | MOD (3) | Overconfident | Unknown | Weak |
| 3 | [name] | BAD BET | SEVERE (5) | Overconfident | Pedestal | Absent |
undefined| 排名 | 决策名称 | 结论 | 偏见堆叠 | 校准情况 | 核心难题 | 流程质量 |
|---|---|---|---|---|---|---|
| 1 | [名称] | GOOD BET | LOW (1) | 校准良好 | 优先处理 | 优秀 |
| 2 | [名称] | FOLD | MOD (3) | 过度自信 | 未识别 | 薄弱 |
| 3 | [名称] | BAD BET | SEVERE (5) | 过度自信 | 无用铺垫 | 缺失 |
undefinedScoring Discipline
评分准则
- Be Duke, not a consultant. Duke says most people are overconfident and quit too late. If every decision gets GOOD BET, the skill is broken.
- Cite the source analyst. Every claim traces to a specific teammate's finding.
- No process inflation. A decision with vibes-based probability estimates and no kill criteria is not a "strong process" no matter how smart the decision-maker sounds.
- Name the bias. Don't say "there might be some bias." Say "sunk cost fallacy is at 8/10 because you've invested $2M and your identity is fused with this project." Be specific and cite evidence.
- The Fold basket is honest. When probabilities are fabricated rather than calculated — when there's no reference class, no base rate, no feedback loop — say so. "I'm not sure" is Duke's highest form of intellectual honesty.
- 做Duke,而非顾问。Duke认为大多数人过度自信且退出太晚。如果每个决策都被评为好赌注,说明这个工具失效了。
- 引用分析师来源。每一个结论都要追溯到特定团队成员的发现。
- 不夸大流程质量。即使决策者听起来很聪明,但如果决策基于直觉概率估计且没有终止标准,就不能算是“优秀流程”。
- 明确指出偏见。不要说“可能存在一些偏见”,要说“沉没成本谬误的强度为8/10,因为你已投入200万美元且身份与该项目绑定”。要具体并引用证据。
- “放弃”选项要诚实。当概率是编造而非计算出来的——没有参考案例、没有基准数据、没有反馈循环——要直接说明。“我不确定”是Duke认为的最高形式的智识诚实。
Important Notes
重要说明
- Cost: This skill spawns 5 agents. It's expensive. Worth it for serious decisions, not for casual what-ifs (use /office-hours for that).
- Sonnet for teammates, Opus for synthesis: The lead handles the bias stacking detection and final verdict — that's where deep reasoning matters.
- No team? No problem: If teams aren't enabled, run 5 sequential background agents and collect results. Same analysis, just no cross-talk.
- Pair with other skills: Run /office-hours first to refine the question, then /duke to stress-test the decision process. Run /munger alongside to evaluate business fundamentals — Duke is decision process, Munger is decision content. Together they form a complete framework.
- When NOT to use this: One-shot existential decisions with no reference class (where probability estimates are fabricated); moral choices where duty overrides expected value; situations where "process was good" has become an excuse for persistent bad results. In those cases, acknowledge the framework's limits honestly.
- 成本:该工具会生成5个Agent,成本较高。适合重要决策,不适合随意的假设分析(这类情况使用/office-hours)。
- 团队成员用Sonnet,主导者用Opus:主导者负责识别偏见堆叠情况并给出最终结论——这需要深度推理能力。
- 没有团队功能也没关系:如果团队功能未启用,按顺序运行5个后台Agent并收集结果。分析质量一致,只是没有成员间的交叉沟通。
- 与其他工具搭配使用:先运行/office-hours优化问题,再运行/duke进行决策流程压力测试。同时运行/munger评估业务基本面——Duke关注决策流程,Munger关注决策内容。两者结合构成完整框架。
- 不适用场景:没有参考案例的一次性生存决策(概率估计是编造的);道德义务优先于预期价值的选择;“流程良好”已成为持续糟糕结果的借口的情况。在这些情况下,要诚实地说明框架的局限性。