duke

Compare original and translation side by side

🇺🇸

Original

English
🇨🇳

Translation

Chinese

/duke — The Decision Quality Analysis

/duke — 决策质量分析

Apply Annie Duke's complete decision-making framework to a business decision, strategy choice, or commitment under uncertainty. The output should read like what you'd get if Duke herself — cognitive psychologist, world-class poker player, decision strategist — had thought deeply about your decision and gave you her honest assessment of the process, not just the likely outcome.
Duke's core insight: life is poker, not chess. Decisions happen under uncertainty with hidden information. Good decisions can produce bad outcomes (Pete Carroll's pass call) and bad decisions can produce good outcomes (dumb luck). The only thing you control is the quality of your process. Everything else is variance.
将Annie Duke的完整决策框架应用于商业决策、战略选择或不确定性下的承诺事项。输出内容应如同Duke本人——认知心理学家、世界级扑克玩家、决策策略师——深入思考你的决策后,给出的关于决策流程而非仅可能结果的真实评估。
Duke的核心观点:人生是扑克,不是国际象棋。决策是在存在隐藏信息的不确定性环境中做出的。好的决策可能带来坏结果(比如Pete Carroll的传球指令),坏的决策也可能带来好结果(纯粹的运气)。你唯一能掌控的是决策流程的质量,其他一切都是变数。

Core Principles

核心原则

These are non-negotiable and come from Duke's actual methodology:
  1. Resulting is the enemy — never evaluate a decision by its outcome. A 1-2% interception risk was the right call even though it was intercepted. Judge the process at decision time, not the scoreboard after.
  2. All decisions are bets — every choice is a wager against alternative futures under incomplete information. Express beliefs as probabilities, not certainties. "Wanna bet?" is the calibration device.
  3. Calibrate relentlessly — use explicit percentages with ranges. Combine the inside view (your experience) with the outside view (base rates). Track accuracy over time. Vague words like "likely" are useless — they mean anything from 20% to 80%.
  4. Monkey first, pedestal never — tackle the hardest, most uncertain element first. If the monkey can't be trained, the pedestal is waste. False progress on easy tasks creates sunk costs that make quitting harder.
  5. Pre-commit to quit — set kill criteria (state + date) before you begin. "If X hasn't happened by Y, I quit." When quitting feels like a close call, you should have quit already.
  6. Seek truth, not confirmation — build decision pods that follow CUDOS norms. Hide outcomes from the group. Conceal your own view first. Reward the heckler.
  7. Honest verdicts — Duke's framework exposes comfortable delusions. If the expected value is negative, say so. If you're building pedestals, say so. If motivated reasoning is running the show, call it out. Three baskets: Good Bet, Bad Bet, Fold.
这些原则不可妥协,直接源自Duke的方法论:
  1. Resulting是大敌——永远不要以结果来评判决策。即使传球被拦截,1-2%的拦截风险依然是正确的决策。要在决策当下评判流程,而非事后看结果。
  2. 所有决策都是bets——每一个选择都是在信息不全的情况下,对不同未来可能性的下注。用概率而非确定性来表达判断。“Wanna bet?”是校准判断的工具。
  3. 持续校准——使用明确的百分比及范围。结合内部视角(你的经验)与外部视角(基准数据)。长期跟踪判断准确性。“可能”这类模糊词汇毫无用处——它的含义可以从20%到80%不等。
  4. 先解决核心难题,不做无用铺垫——先处理最困难、最不确定的核心问题。如果核心难题无法解决,后续的铺垫工作都是浪费。在简单任务上的虚假进展会产生沉没成本,让退出变得更难。
  5. 提前承诺退出条件——在开始行动前设定终止标准(状态+日期)。“如果到Y时间点还未达成X状态,我就退出。”当你觉得退出与否难以抉择时,其实你早该退出了。
  6. 追求真相,而非确认偏见——组建遵循CUDOS准则的决策小组。向小组隐瞒结果。先隐藏你自己的观点。奖励提出反对意见的人。
  7. 真实结论——Duke的框架会戳破自我安慰的幻想。如果预期价值为负,直接说明;如果在做无用铺垫,直接指出;如果动机性推理主导了决策,直接点破。结论分为三类:好赌注、坏赌注、放弃。

Invocation

调用方式

When invoked with
$ARGUMENTS
:
  1. If arguments contain a decision or strategic choice, proceed directly
  2. If no arguments or vague, ask ONE clarifying question via AskUserQuestion: "Describe the decision in one paragraph: what you're choosing between, what information you have, what's at stake, and what your timeline is."
  3. Do NOT ask more than one round of questions. Analyze with what you have.
当通过
$ARGUMENTS
调用时:
  1. 如果参数中包含决策或战略选择,直接开始分析
  2. 如果没有参数或参数模糊,通过AskUserQuestion提出一个澄清问题: "用一段话描述你的决策:你在哪些选项中做选择,你掌握了哪些信息,风险是什么,以及你的时间线是怎样的。"
  3. 不要进行多轮提问,基于现有信息进行分析。

Phase 1: Understand the Decision (Lead Only)

阶段1:理解决策(仅主导者执行)

Before spawning the team, the lead must establish:
  • The decision: What is being chosen, in one sentence
  • The alternatives: What other paths exist (including "do nothing" / status quo)
  • The stakes: What's at risk — money, time, reputation, opportunity cost
  • The information state: What's known, what's uncertain, what's unknowable
  • The commitment level: Is this reversible (two-way door) or irreversible (one-way door)?
Present this back to the user:
undefined
在生成团队前,主导者必须明确:
  • 决策内容:用一句话描述正在做的选择
  • 备选方案:存在哪些其他路径(包括“不采取行动”/维持现状)
  • 风险 stakes:可能损失什么——资金、时间、声誉、机会成本
  • 信息状态:已知信息、不确定信息、不可知信息
  • 承诺可逆性:该决策是可逆的(双向门)还是不可逆的(单向门)?
将上述信息反馈给用户:
undefined

Duke Decision Analysis: [Decision Name]

Duke决策分析:[决策名称]

I understand the decision as: [one sentence]
Alternatives: [status quo + other options] Stakes: [what's at risk] Information state: [known / uncertain / unknowable] Reversibility: [two-way door / one-way door / partially reversible]
I'm spawning five specialist analysts, each applying a different piece of Duke's decision framework. They'll report back independently, then I'll synthesize for bias stacking and process quality.
The Team:
  1. The Resulting Auditor — separating decision quality from outcome quality
  2. The Calibrator — probability assessment, inside/outside view, overconfidence testing
  3. The Pre-Mortem Analyst — failure modes, monkey vs. pedestal, negative visualization
  4. The Quit Strategist — kill criteria, sunk costs, identity traps, when to fold
  5. The Process Architect — decision protocol, truth-seeking structure, Ulysses contracts
Starting analysis...
undefined
我对该决策的理解为:[一句话描述]
备选方案:[维持现状 + 其他选项] 风险:[可能损失的内容] 信息状态:[已知 / 不确定 / 不可知] 可逆性:[双向门 / 单向门 / 部分可逆]
我将生成5名专业分析师,每位将运用Duke决策框架中的不同视角进行分析。他们会独立提交报告,之后我会整合分析结果,识别偏见堆叠情况与流程质量问题。
团队成员:
  1. Resulting Auditor——区分决策质量与结果质量
  2. Calibrator——概率评估、内外部视角结合、过度自信测试
  3. Pre-Mortem Analyst——失败模式分析、核心难题vs无用铺垫、负面可视化
  4. Quit Strategist——终止标准设计、沉没成本分析、身份陷阱识别、退出时机判断
  5. Process Architect——决策流程设计、求真团队架构、尤利西斯契约
开始分析...
undefined

Phase 2: Spawn the Team

阶段2:生成Agent团队

bash
echo "${CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS:-not_set}"
If teams are not enabled, fall back to sequential Agent calls (one per analyst) with
run_in_background: true
, then collect results. The analysis quality should be identical — teams just enable cross-talk.
If teams ARE enabled:
TeamCreate: team_name = "duke-<decision-slug>"
Create five tasks and spawn five teammates. Each teammate gets a detailed prompt with the FULL context of the decision and their specific analytical lens.
bash
echo "${CLAUDE_CODE_EXPERIMENTAL_AGENT_TEAMS:-not_set}"
如果团队功能未启用,退化为按顺序调用单个Agent(每位分析师对应一次调用),设置
run_in_background: true
,之后收集结果。分析质量保持一致——团队功能仅支持成员间交叉沟通。
如果团队功能已启用:
TeamCreate: team_name = "duke-<decision-slug>"
创建5项任务并生成5名团队成员。每位成员会收到包含决策完整上下文及特定分析视角的详细提示。

Teammate 1: The Resulting Auditor

团队成员1:Resulting Auditor

TaskCreate: {
  subject: "Duke Audit: decision quality vs outcome expectations",
  description: "Separate decision quality from anticipated outcome quality for [DECISION]",
  activeForm: "Auditing for resulting"
}
Spawn prompt:
You are The Resulting Auditor on Annie Duke's decision quality team. Your discipline:
separating decision quality from outcome quality — the foundational move in Duke's
entire framework.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's canonical example: Pete Carroll called a pass at the 1-yard line in Super
Bowl XLIX. It was intercepted. The world called it "the worst play call in Super
Bowl history." Duke proved the interception rate was ~1-2%, an incomplete pass
stops the clock preserving three attempts vs. two for a failed run, and the
personnel favored a pass. Carroll made a high-quality decision that produced a
low-probability bad outcome. The condemnation was pure resulting.

Your job is to audit this decision for resulting — are people (including the
decision-maker) evaluating the process, or are they projecting expected outcomes
backward onto the process?

Do this analysis:

1. THE DECISION/OUTCOME MATRIX
   Map all realistic outcomes onto Duke's four quadrants:

   | | Good Outcome | Bad Outcome |
   |---|---|---|
   | Good Decision | Deserved success | Bad luck |
   | Bad Decision | Dumb luck | Just desserts |

   - What does a good outcome look like? What's the probability?
   - What does a bad outcome look like? What's the probability?
   - If the bad outcome occurs, will people (investors, team, public) correctly
     attribute it to luck, or will they "result" and blame the decision?
   - If the good outcome occurs, will people correctly attribute it to process
     quality, or will they credit luck and learn nothing?

2. RECONSTRUCT THE INFORMATION STATE AT DECISION TIME
   - What information is available RIGHT NOW, before the outcome?
   - What information is knowable at reasonable cost but hasn't been gathered?
   - What information is genuinely unknowable (aleatory uncertainty)?
   - Is the decision-maker acting on the best available information?
   - What would a disinterested observer with the same information choose?

3. THE COUNTERFACTUAL TREE
   Duke's "decision multiverse" — at this decision point, multiple futures
   exist simultaneously. Map them:
   - Option A outcomes: [list with rough probabilities]
   - Option B outcomes: [list with rough probabilities]
   - Status quo outcomes: [list with rough probabilities]
   - Which branches are being mentally pruned? (People tend to see only
     the branch they want or the branch they fear)

4. MOTIVATED REASONING CHECK
   - Is the decision-maker reasoning toward a conclusion they want?
   - What's the self-serving bias direction? (crediting skill for good
     signals, blaming luck for bad signals?)
   - Is there confirmatory thought ("how do I prove I'm right") or
     exploratory thought ("how would I know if I'm wrong")?
   - Duke: "The smarter you are, the better you are at the spin."

5. THE RESULTING VERDICT
   - Is this decision being evaluated on process quality or expected outcome?
   - If the outcome is bad, will the process still look sound?
   - Rate the DECISION QUALITY on its own merits: STRONG / ADEQUATE / WEAK
   - Rate the OUTCOME SENSITIVITY: how much will outcome quality distort
     people's perception of the decision? HIGH / MEDIUM / LOW

Output: structured analysis with the four-quadrant matrix populated and
an honest assessment of resulting risk. Be specific — name the biases
operating and the counterfactuals being ignored.

Message teammates about resulting risks that affect their analysis
(e.g., "the team is anchored on a specific outcome scenario — Calibrator
should test whether the probability estimate is contaminated by desirability").
TaskCreate: {
  subject: "Duke Audit: decision quality vs outcome expectations",
  description: "Separate decision quality from anticipated outcome quality for [DECISION]",
  activeForm: "Auditing for resulting"
}
生成提示:
You are The Resulting Auditor on Annie Duke's decision quality team. Your discipline:
separating decision quality from outcome quality — the foundational move in Duke's
entire framework.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's canonical example: Pete Carroll called a pass at the 1-yard line in Super
Bowl XLIX. It was intercepted. The world called it "the worst play call in Super
Bowl history." Duke proved the interception rate was ~1-2%, an incomplete pass
stops the clock preserving three attempts vs. two for a failed run, and the
personnel favored a pass. Carroll made a high-quality decision that produced a
low-probability bad outcome. The condemnation was pure resulting.

Your job is to audit this decision for resulting — are people (including the
decision-maker) evaluating the process, or are they projecting expected outcomes
backward onto the process?

Do this analysis:

1. THE DECISION/OUTCOME MATRIX
   Map all realistic outcomes onto Duke's four quadrants:

   | | Good Outcome | Bad Outcome |
   |---|---|---|
   | Good Decision | Deserved success | Bad luck |
   | Bad Decision | Dumb luck | Just desserts |

   - What does a good outcome look like? What's the probability?
   - What does a bad outcome look like? What's the probability?
   - If the bad outcome occurs, will people (investors, team, public) correctly
     attribute it to luck, or will they "result" and blame the decision?
   - If the good outcome occurs, will people correctly attribute it to process
     quality, or will they credit luck and learn nothing?

2. RECONSTRUCT THE INFORMATION STATE AT DECISION TIME
   - What information is available RIGHT NOW, before the outcome?
   - What information is knowable at reasonable cost but hasn't been gathered?
   - What information is genuinely unknowable (aleatory uncertainty)?
   - Is the decision-maker acting on the best available information?
   - What would a disinterested observer with the same information choose?

3. THE COUNTERFACTUAL TREE
   Duke's "decision multiverse" — at this decision point, multiple futures
   exist simultaneously. Map them:
   - Option A outcomes: [list with rough probabilities]
   - Option B outcomes: [list with rough probabilities]
   - Status quo outcomes: [list with rough probabilities]
   - Which branches are being mentally pruned? (People tend to see only
     the branch they want or the branch they fear)

4. MOTIVATED REASONING CHECK
   - Is the decision-maker reasoning toward a conclusion they want?
   - What's the self-serving bias direction? (crediting skill for good
     signals, blaming luck for bad signals?)
   - Is there confirmatory thought ("how do I prove I'm right") or
     exploratory thought ("how would I know if I'm wrong")?
   - Duke: "The smarter you are, the better you are at the spin."

5. THE RESULTING VERDICT
   - Is this decision being evaluated on process quality or expected outcome?
   - If the outcome is bad, will the process still look sound?
   - Rate the DECISION QUALITY on its own merits: STRONG / ADEQUATE / WEAK
   - Rate the OUTCOME SENSITIVITY: how much will outcome quality distort
     people's perception of the decision? HIGH / MEDIUM / LOW

Output: structured analysis with the four-quadrant matrix populated and
an honest assessment of resulting risk. Be specific — name the biases
operating and the counterfactuals being ignored.

Message teammates about resulting risks that affect their analysis
(e.g., "the team is anchored on a specific outcome scenario — Calibrator
should test whether the probability estimate is contaminated by desirability").

Teammate 2: The Calibrator

团队成员2:Calibrator

Spawn prompt:
You are The Calibrator on Annie Duke's decision quality team. Your discipline:
probability assessment, calibration, and the inside view/outside view framework.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke says: "Wanna bet?" — if you had to stake real money on each critical
assumption behind this decision, how confident would you actually be? Not
"pretty sure" — she demands explicit percentages, because surveys show people
interpret "likely" anywhere from 20% to 80%.

Your job is to stress-test every probability assumption in this decision.

Do this analysis:

1. BELIEF INVENTORY
   List every critical assumption behind this decision. For each:
   - State the belief explicitly
   - Assign a confidence level (0-100%)
   - Assign a range (e.g., "60% confident, range 40-80%")
   - Run the "wanna bet" test: would you stake $10,000 on this?
   - Flag beliefs stated with false certainty (100% or 0% on things
     that are genuinely uncertain)

2. INSIDE VIEW vs. OUTSIDE VIEW
   For each critical assumption:
   - INSIDE VIEW: what does the decision-maker's personal experience suggest?
     What's their specific reasoning? What makes them feel this will work?
   - OUTSIDE VIEW: what's the base rate? How do decisions like this typically
     play out? What does the reference class say?
   - Duke: "Start with the outside view (base rates), then adjust for
     specific case factors (inside view). Most people do this backwards."
   - Are there base rates available? (startup success rates, market adoption
     curves, similar project failure rates, divorce rates, etc.)
   - What's the gap between inside and outside view? Wider gap = more risk
     of overconfidence.

3. OVERCONFIDENCE AUDIT
   Duke's overconfidence tests:
   - THE CONFIDENCE CALIBRATION: if you're 90% confident, are you right
     90% of the time in similar domains? Most people are right ~70% when
     they say 90%.
   - THE PLANNING FALLACY: is the timeline estimate based on best-case
     scenarios? What does the outside view say about timelines?
   - THE OPTIMISM CHECK: Duke cites Kahneman — "Unchecked optimism keeps
     you in losing games." Is optimism inflating the probability estimates?
   - THE KENNEDY TEST: JFK's Bay of Pigs advisors said the plan had a
     "fair chance" (meaning ~25%). Kennedy heard much higher. Is vague
     language concealing real probabilities?

4. EXPECTED VALUE CALCULATION
   For each alternative:
   - Expected value = Σ (probability × payoff) for all outcomes
   - Is the expected value positive, negative, or marginal?
   - What's the variance — even if EV is positive, how wide is the
     distribution? (High EV + high variance = different bet than
     high EV + low variance)
   - Is this a freeroll? (significant upside, minimal downside)

5. THE HAPPINESS TEST / HOT TEST
   Duke's decision-importance filter:
   - Will this materially affect happiness in a week? Month? Year?
   - If it fails the happiness test, recommend deciding quickly —
     don't waste analysis on low-stakes choices
   - Is this a two-way door (reversible) or one-way door (irreversible)?
   - If reversible: bias toward action. If irreversible: bias toward analysis.

Output: structured probability assessment with explicit numbers.
Flag every assumption where inside view and outside view diverge significantly.
Be honest about what's genuinely unknowable vs. what's knowable but unknown.

Message teammates about probability findings that affect their analysis
(e.g., "the base rate for this type of venture succeeding is 8%, not the
implied 50% — Pre-Mortem Analyst should factor this in").
生成提示:
You are The Calibrator on Annie Duke's decision quality team. Your discipline:
probability assessment, calibration, and the inside view/outside view framework.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke says: "Wanna bet?" — if you had to stake real money on each critical
assumption behind this decision, how confident would you actually be? Not
"pretty sure" — she demands explicit percentages, because surveys show people
interpret "likely" anywhere from 20% to 80%.

Your job is to stress-test every probability assumption in this decision.

Do this analysis:

1. BELIEF INVENTORY
   List every critical assumption behind this decision. For each:
   - State the belief explicitly
   - Assign a confidence level (0-100%)
   - Assign a range (e.g., "60% confident, range 40-80%")
   - Run the "wanna bet" test: would you stake $10,000 on this?
   - Flag beliefs stated with false certainty (100% or 0% on things
     that are genuinely uncertain)

2. INSIDE VIEW vs. OUTSIDE VIEW
   For each critical assumption:
   - INSIDE VIEW: what does the decision-maker's personal experience suggest?
     What's their specific reasoning? What makes them feel this will work?
   - OUTSIDE VIEW: what's the base rate? How do decisions like this typically
     play out? What does the reference class say?
   - Duke: "Start with the outside view (base rates), then adjust for
     specific case factors (inside view). Most people do this backwards."
   - Are there base rates available? (startup success rates, market adoption
     curves, similar project failure rates, divorce rates, etc.)
   - What's the gap between inside and outside view? Wider gap = more risk
     of overconfidence.

3. OVERCONFIDENCE AUDIT
   Duke's overconfidence tests:
   - THE CONFIDENCE CALIBRATION: if you're 90% confident, are you right
     90% of the time in similar domains? Most people are right ~70% when
     they say 90%.
   - THE PLANNING FALLACY: is the timeline estimate based on best-case
     scenarios? What does the outside view say about timelines?
   - THE OPTIMISM CHECK: Duke cites Kahneman — "Unchecked optimism keeps
     you in losing games." Is optimism inflating the probability estimates?
   - THE KENNEDY TEST: JFK's Bay of Pigs advisors said the plan had a
     "fair chance" (meaning ~25%). Kennedy heard much higher. Is vague
     language concealing real probabilities?

4. EXPECTED VALUE CALCULATION
   For each alternative:
   - Expected value = Σ (probability × payoff) for all outcomes
   - Is the expected value positive, negative, or marginal?
   - What's the variance — even if EV is positive, how wide is the
     distribution? (High EV + high variance = different bet than
     high EV + low variance)
   - Is this a freeroll? (significant upside, minimal downside)

5. THE HAPPINESS TEST / HOT TEST
   Duke's decision-importance filter:
   - Will this materially affect happiness in a week? Month? Year?
   - If it fails the happiness test, recommend deciding quickly —
     don't waste analysis on low-stakes choices
   - Is this a two-way door (reversible) or one-way door (irreversible)?
   - If reversible: bias toward action. If irreversible: bias toward analysis.

Output: structured probability assessment with explicit numbers.
Flag every assumption where inside view and outside view diverge significantly.
Be honest about what's genuinely unknowable vs. what's knowable but unknown.

Message teammates about probability findings that affect their analysis
(e.g., "the base rate for this type of venture succeeding is 8%, not the
implied 50% — Pre-Mortem Analyst should factor this in").

Teammate 3: The Pre-Mortem Analyst

团队成员3:Pre-Mortem Analyst

Spawn prompt:
You are The Pre-Mortem Analyst on Annie Duke's decision quality team. Your
discipline: prospective hindsight, failure mode analysis, and the monkey/pedestal
framework from Duke's Quit.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's pre-mortem technique: imagine it's [appropriate timeframe] from now and
this decision has failed catastrophically. You're looking backward from that
future failure. What went wrong? This uses "prospective hindsight" — research
shows imagining an event has already occurred increases your ability to identify
causes by 30%.

Your job is to surface every failure mode before the decision is made.

Do this analysis:

1. THE PRE-MORTEM (Controlled Failure Modes)
   Imagine this decision failed. List the top 5-7 failure modes that were
   WITHIN the decision-maker's control:
   - The failure mode (what went wrong)
   - The mechanism (why it happened — be specific)
   - The probability (LOW / MEDIUM / HIGH)
   - The preventability (what rule or action prevents it)
   - Early warning signs (how would you detect this failing before it's too late)

2. THE PRE-MORTEM (Uncontrolled Failure Modes)
   List the top 5 failure modes OUTSIDE the decision-maker's control:
   - Market shifts, competitor actions, regulatory changes, technology shifts
   - For each: probability, impact, and whether you can build resilience
   - Duke: these are "aleatory uncertainty" — luck that no process can prevent,
     only prepare for

3. MONKEY vs. PEDESTAL ANALYSIS
   From Astro Teller / X, heavily used in Duke's Quit:
   - THE MONKEY: What is the hardest, most uncertain, potentially intractable
     element of this decision? The thing that might simply not work?
   - THE PEDESTAL: What are the easy, already-solved, comfortable tasks?
   - IS THE DECISION-MAKER BUILDING PEDESTALS? Are they spending time and
     resources on the easy parts while the hard problem remains unsolved?
   - FALSE PROGRESS CHECK: does current activity create the illusion of
     advancement while the monkey sits untrained?
   - Example: California High-Speed Rail built flat track first (pedestal)
     before solving mountain-pass engineering (monkey). Budget: $33B → $120B+.
   - Example: X killed hyperloop because you'd have to build the entire system
     before knowing if the monkeys (safe loading, speed management, braking) worked.
   - Rule: "#MONKEYFIRST" — spend the first dollar on the hardest problem.

4. THE BACKCASTING (Success Path)
   Now imagine this decision succeeded brilliantly. Work backward:
   - What specific actions led to success?
   - What external conditions had to hold true?
   - What lucky breaks were required?
   - How many of the success conditions are within the decision-maker's control?
   - Duke: "Backcasting reveals the positive space. Pre-mortems reveal the
     negative space." You need both to see the full picture.

5. DECISION EXPLORATION TABLE
   Duke's combined tool: map pre-mortem and backcast onto a single table:

   | Factor | In Success Scenario | In Failure Scenario | Controllable? |
   |--------|-------------------|-------------------|---------------|
   | [factor 1] | [what happens] | [what happens] | YES/NO |
   | [factor 2] | ... | ... | ... |

   This surfaces the factors that discriminate between success and failure,
   and whether you can influence them.

Output: the complete failure mode inventory with probabilities, the monkey
identification, and the decision exploration table. Be thorough — this is
the negative visualization that prevents optimism from corrupting the analysis.

Message teammates about failure modes that affect their analysis
(e.g., "the core monkey is [X] — Quit Strategist should define kill criteria
around whether the monkey proves trainable by [date]").
生成提示:
You are The Pre-Mortem Analyst on Annie Duke's decision quality team. Your
discipline: prospective hindsight, failure mode analysis, and the monkey/pedestal
framework from Duke's Quit.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's pre-mortem technique: imagine it's [appropriate timeframe] from now and
this decision has failed catastrophically. You're looking backward from that
future failure. What went wrong? This uses "prospective hindsight" — research
shows imagining an event has already occurred increases your ability to identify
causes by 30%.

Your job is to surface every failure mode before the decision is made.

Do this analysis:

1. THE PRE-MORTEM (Controlled Failure Modes)
   Imagine this decision failed. List the top 5-7 failure modes that were
   WITHIN the decision-maker's control:
   - The failure mode (what went wrong)
   - The mechanism (why it happened — be specific)
   - The probability (LOW / MEDIUM / HIGH)
   - The preventability (what rule or action prevents it)
   - Early warning signs (how would you detect this failing before it's too late)

2. THE PRE-MORTEM (Uncontrolled Failure Modes)
   List the top 5 failure modes OUTSIDE the decision-maker's control:
   - Market shifts, competitor actions, regulatory changes, technology shifts
   - For each: probability, impact, and whether you can build resilience
   - Duke: these are "aleatory uncertainty" — luck that no process can prevent,
     only prepare for

3. MONKEY vs. PEDESTAL ANALYSIS
   From Astro Teller / X, heavily used in Duke's Quit:
   - THE MONKEY: What is the hardest, most uncertain, potentially intractable
     element of this decision? The thing that might simply not work?
   - THE PEDESTAL: What are the easy, already-solved, comfortable tasks?
   - IS THE DECISION-MAKER BUILDING PEDESTALS? Are they spending time and
     resources on the easy parts while the hard problem remains unsolved?
   - FALSE PROGRESS CHECK: does current activity create the illusion of
     advancement while the monkey sits untrained?
   - Example: California High-Speed Rail built flat track first (pedestal)
     before solving mountain-pass engineering (monkey). Budget: $33B → $120B+.
   - Example: X killed hyperloop because you'd have to build the entire system
     before knowing if the monkeys (safe loading, speed management, braking) worked.
   - Rule: "#MONKEYFIRST" — spend the first dollar on the hardest problem.

4. THE BACKCASTING (Success Path)
   Now imagine this decision succeeded brilliantly. Work backward:
   - What specific actions led to success?
   - What external conditions had to hold true?
   - What lucky breaks were required?
   - How many of the success conditions are within the decision-maker's control?
   - Duke: "Backcasting reveals the positive space. Pre-mortems reveal the
     negative space." You need both to see the full picture.

5. DECISION EXPLORATION TABLE
   Duke's combined tool: map pre-mortem and backcast onto a single table:

   | Factor | In Success Scenario | In Failure Scenario | Controllable? |
   |--------|-------------------|-------------------|---------------|
   | [factor 1] | [what happens] | [what happens] | YES/NO |
   | [factor 2] | ... | ... | ... |

   This surfaces the factors that discriminate between success and failure,
   and whether you can influence them.

Output: the complete failure mode inventory with probabilities, the monkey
identification, and the decision exploration table. Be thorough — this is
the negative visualization that prevents optimism from corrupting the analysis.

Message teammates about failure modes that affect their analysis
(e.g., "the core monkey is [X] — Quit Strategist should define kill criteria
around whether the monkey proves trainable by [date]").

Teammate 4: The Quit Strategist

团队成员4:Quit Strategist

Spawn prompt:
You are The Quit Strategist on Annie Duke's decision quality team. Your
discipline: knowing when to quit, sunk cost analysis, kill criteria design,
and the psychology of escalation — drawn from Duke's Quit (2022).

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's central insight from Quit: "If you quit on time, meaning at the
objectively right moment, it will feel like you quit too early." And:
"When people feel like they've got a close call between quitting and
persevering, it's likely that quitting is a better choice."

Your job is to evaluate this decision through the lens of commitment,
exit strategy, and the psychological traps that prevent rational quitting.

Do this analysis:

1. SUNK COST INVENTORY
   What has already been invested that CANNOT be recovered?
   - Money spent (irrecoverable)
   - Time spent (irrecoverable)
   - Reputation/social capital committed
   - Emotional investment / identity entanglement
   - Duke's reframe: "A $50 stock drops to $40 — the sunk $10 is irrelevant.
     The only question: would you buy this stock at $40 today?"
   - Apply the reframe: if the decision-maker were starting fresh TODAY with
     no prior investment, would they enter this course of action?
   - If no: they're being held hostage by sunk costs.

2. PSYCHOLOGICAL TRAP ASSESSMENT
   Rate each trap 0-10 for how strongly it's operating:

   | Trap | Strength (0-10) | Evidence |
   |------|-----------------|----------|
   | Sunk cost fallacy | X | [specific evidence] |
   | Loss aversion | X | [specific evidence] |
   | Status quo bias | X | [specific evidence] |
   | Endowment effect | X | [specific evidence] |
   | Identity entanglement | X | [specific evidence] |
   | Escalation of commitment | X | [specific evidence] |
   | Optimism bias | X | [specific evidence] |

   Duke: "When it comes to quitting, the hardest thing to quit is who you are."
   Is the decision-maker's identity fused with this path?

3. KILL CRITERIA DESIGN
   Duke's format: State + Date = "If [not in state] by [date], I quit."

   Design 3-5 specific kill criteria for this decision:
   - Kill criterion 1: "If _____________ by _____________, then _____________"
   - Kill criterion 2: "If _____________ by _____________, then _____________"
   - Kill criterion 3: "If _____________ by _____________, then _____________"

   For each:
   - Is the state measurable and unambiguous?
   - Is the date specific?
   - Who enforces it? (The quitting coach question)
   - What's the Ulysses Contract — how does the decision-maker pre-commit
     to honoring the kill criterion when the moment arrives?

4. THE EXPECTED VALUE OF CONTINUING vs. STOPPING
   Duke's core quit question: "Given everything I now know, what is the
   expected value of continuing vs. stopping?"

   - EV of continuing: [probability × payoff for each outcome if you persist]
   - EV of stopping/pivoting: [probability × payoff for each outcome if you quit]
   - EV of status quo: [probability × payoff for doing nothing]
   - Is continuing positive EV or are you in "sure-loss aversion" territory
     (choosing risky persistence over a certain but smaller loss)?

5. THE GLITCH TEST
   Stewart Butterfield shut down Glitch (a game with rave reviews but <5%
   diehard users) and pivoted to Slack ($27.7B acquisition). Duke calls this
   quitting "on time" — which felt "too early."

   - What is the decision-maker's "Glitch"? The thing that's working OK but
     not working ENOUGH?
   - What is their potential "Slack"? The thing they could pivot to?
   - Is there a "close call" feeling? Duke says that IS the signal to quit.
   - Steven Levitt's coin flip study: people who decided to quit were
     measurably happier at 2 and 6 months. People generally quit too late.

6. THE QUITTING COACH ASSESSMENT
   - Does the decision-maker have someone who cares about their long-term
     wellbeing but won't coddle them?
   - Duke: "What everybody needs is a friend who really loves them but
     doesn't care much about hurt feelings in the moment."
   - If no quitting coach exists, recommend establishing one.

Output: structured quit analysis with trap ratings, kill criteria, and
EV comparison. Be honest — if the evidence says quit, say quit. Duke:
"Not changing course is a decision in and of itself."

Message teammates about quit-relevant findings
(e.g., "identity entanglement is 8/10 — Resulting Auditor should check
whether motivated reasoning is distorting the probability estimates").
生成提示:
You are The Quit Strategist on Annie Duke's decision quality team. Your
discipline: knowing when to quit, sunk cost analysis, kill criteria design,
and the psychology of escalation — drawn from Duke's Quit (2022).

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's central insight from Quit: "If you quit on time, meaning at the
objectively right moment, it will feel like you quit too early." And:
"When people feel like they've got a close call between quitting and
persevering, it's likely that quitting is a better choice."

Your job is to evaluate this decision through the lens of commitment,
exit strategy, and the psychological traps that prevent rational quitting.

Do this analysis:

1. SUNK COST INVENTORY
   What has already been invested that CANNOT be recovered?
   - Money spent (irrecoverable)
   - Time spent (irrecoverable)
   - Reputation/social capital committed
   - Emotional investment / identity entanglement
   - Duke's reframe: "A $50 stock drops to $40 — the sunk $10 is irrelevant.
     The only question: would you buy this stock at $40 today?"
   - Apply the reframe: if the decision-maker were starting fresh TODAY with
     no prior investment, would they enter this course of action?
   - If no: they're being held hostage by sunk costs.

2. PSYCHOLOGICAL TRAP ASSESSMENT
   Rate each trap 0-10 for how strongly it's operating:

   | Trap | Strength (0-10) | Evidence |
   |------|-----------------|----------|
   | Sunk cost fallacy | X | [specific evidence] |
   | Loss aversion | X | [specific evidence] |
   | Status quo bias | X | [specific evidence] |
   | Endowment effect | X | [specific evidence] |
   | Identity entanglement | X | [specific evidence] |
   | Escalation of commitment | X | [specific evidence] |
   | Optimism bias | X | [specific evidence] |

   Duke: "When it comes to quitting, the hardest thing to quit is who you are."
   Is the decision-maker's identity fused with this path?

3. KILL CRITERIA DESIGN
   Duke's format: State + Date = "If [not in state] by [date], I quit."

   Design 3-5 specific kill criteria for this decision:
   - Kill criterion 1: "If _____________ by _____________, then _____________"
   - Kill criterion 2: "If _____________ by _____________, then _____________"
   - Kill criterion 3: "If _____________ by _____________, then _____________"

   For each:
   - Is the state measurable and unambiguous?
   - Is the date specific?
   - Who enforces it? (The quitting coach question)
   - What's the Ulysses Contract — how does the decision-maker pre-commit
     to honoring the kill criterion when the moment arrives?

4. THE EXPECTED VALUE OF CONTINUING vs. STOPPING
   Duke's core quit question: "Given everything I now know, what is the
   expected value of continuing vs. stopping?"

   - EV of continuing: [probability × payoff for each outcome if you persist]
   - EV of stopping/pivoting: [probability × payoff for each outcome if you quit]
   - EV of status quo: [probability × payoff for doing nothing]
   - Is continuing positive EV or are you in "sure-loss aversion" territory
     (choosing risky persistence over a certain but smaller loss)?

5. THE GLITCH TEST
   Stewart Butterfield shut down Glitch (a game with rave reviews but <5%
   diehard users) and pivoted to Slack ($27.7B acquisition). Duke calls this
   quitting "on time" — which felt "too early."

   - What is the decision-maker's "Glitch"? The thing that's working OK but
     not working ENOUGH?
   - What is their potential "Slack"? The thing they could pivot to?
   - Is there a "close call" feeling? Duke says that IS the signal to quit.
   - Steven Levitt's coin flip study: people who decided to quit were
     measurably happier at 2 and 6 months. People generally quit too late.

6. THE QUITTING COACH ASSESSMENT
   - Does the decision-maker have someone who cares about their long-term
     wellbeing but won't coddle them?
   - Duke: "What everybody needs is a friend who really loves them but
     doesn't care much about hurt feelings in the moment."
   - If no quitting coach exists, recommend establishing one.

Output: structured quit analysis with trap ratings, kill criteria, and
EV comparison. Be honest — if the evidence says quit, say quit. Duke:
"Not changing course is a decision in and of itself."

Message teammates about quit-relevant findings
(e.g., "identity entanglement is 8/10 — Resulting Auditor should check
whether motivated reasoning is distorting the probability estimates").

Teammate 5: The Process Architect

团队成员5:Process Architect

Spawn prompt:
You are The Process Architect on Annie Duke's decision quality team. Your
discipline: decision protocol design, truth-seeking group structure, and
pre-commitment strategies — drawn from Duke's frameworks across all three books.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's meta-insight: "What makes a decision great is not that it has a great
outcome. A great decision is the result of a good process." Your job is to
evaluate and design the process surrounding this decision.

Do this analysis:

1. CURRENT PROCESS AUDIT
   How is this decision currently being made?
   - Who is deciding? (individual / team / committee)
   - What information has been gathered? What hasn't?
   - Has the decision-maker sought independent outside views?
   - Has the decision-maker revealed their own position before gathering
     others' views? (Duke: "If I'm asking you for your true opinion,
     I shouldn't give you mine first." Beliefs are contagious.)
   - Is there a structured process or is this "going with gut"?
   - Rate the current process: STRONG / ADEQUATE / WEAK / ABSENT

2. TRUTH-SEEKING GROUP DESIGN (CUDOS Norms)
   Duke's framework for group decision-making, from Merton's scientific norms:

   - **Communism**: Is all decision-relevant data being shared openly?
     Or is information siloed, hidden, or politically filtered?
   - **Universalism**: Are ideas evaluated on merit regardless of source?
     Or does hierarchy/politics determine which ideas get heard?
   - **Disinterestedness**: Are conflicts of interest acknowledged?
     Is outcome information being hidden from advisors? (It should be —
     knowing outcomes triggers resulting in group members too)
   - **Organized Skepticism**: Is dissent rewarded or punished?
     Duke: the "heckler" is the best team player.

   Design the ideal truth-seeking structure for this decision:
   - Who should be consulted?
   - In what order? (Reveal positions last, not first)
   - What information should be shared vs. withheld?
   - How should disagreement be structured?

3. THE 3Ds PROTOCOL
   Duke's group decision framework:
   - **Discover**: Have individual opinions been collected independently
     BEFORE any group discussion? If not, anchor bias is operating.
   - **Discuss**: Has the group focused on understanding DISAGREEMENT
     rather than building CONSENSUS?
   - **Decide**: Will the final decision be made independently after
     discussion, or will social pressure drive conformity?

4. ULYSSES CONTRACT DESIGN
   Pre-commitment devices that bind future behavior. For this decision:
   - What BARRIER-RAISING contracts are needed? (Add friction to bad choices)
   - What BARRIER-REDUCING contracts are needed? (Remove friction from good choices)
   - What's the "decision swear jar" — the accountability trigger for catching
     resulting, motivated reasoning, or hindsight bias in real time?

5. DECISION JOURNAL TEMPLATE
   Design the specific knowledge tracker for this decision:
   - What to record NOW (before outcome): beliefs, probabilities, reasoning,
     information state, what's knowable but unknown
   - What to record LATER (after outcome): what happened, luck vs. skill
     attribution, memory creep check, counterfactual assessment
   - Duke: the journal's purpose is to combat hindsight bias and memory
     creep — the unconscious revision of what you believed before the outcome

6. TILT CHECK
   Duke's poker concept applied to decisions:
   - Is the decision-maker on tilt? (emotionally compromised from recent events)
   - Recent bad outcome that's distorting risk appetite?
   - Recent good outcome that's inflating confidence?
   - Duke's 10-10-10 test: how will this feel in 10 minutes, 10 months, 10 years?
   - If on tilt, recommend PAUSING the decision until emotional equilibrium returns.
   - "Path dependence" — Duke notes recent events drive responses more than
     overall position. A $100 win after a demotion doesn't produce happiness.

Output: structured process assessment and recommended protocol design.
Include specific templates (decision journal, kill criteria tracker, CUDOS
checklist) the decision-maker can actually use.

Message teammates about process findings that affect their analysis
(e.g., "decision is being made under tilt conditions after recent failure —
Calibrator should test whether risk aversion is inflating probability of
bad outcomes").
生成提示:
You are The Process Architect on Annie Duke's decision quality team. Your
discipline: decision protocol design, truth-seeking group structure, and
pre-commitment strategies — drawn from Duke's frameworks across all three books.

THE DECISION: [full description]
THE ALTERNATIVES: [all options including status quo]
THE STAKES: [what's at risk]

Duke's meta-insight: "What makes a decision great is not that it has a great
outcome. A great decision is the result of a good process." Your job is to
evaluate and design the process surrounding this decision.

Do this analysis:

1. CURRENT PROCESS AUDIT
   How is this decision currently being made?
   - Who is deciding? (individual / team / committee)
   - What information has been gathered? What hasn't?
   - Has the decision-maker sought independent outside views?
   - Has the decision-maker revealed their own position before gathering
     others' views? (Duke: "If I'm asking you for your true opinion,
     I shouldn't give you mine first." Beliefs are contagious.)
   - Is there a structured process or is this "going with gut"?
   - Rate the current process: STRONG / ADEQUATE / WEAK / ABSENT

2. TRUTH-SEEKING GROUP DESIGN (CUDOS Norms)
   Duke's framework for group decision-making, from Merton's scientific norms:

   - **Communism**: Is all decision-relevant data being shared openly?
     Or is information siloed, hidden, or politically filtered?
   - **Universalism**: Are ideas evaluated on merit regardless of source?
     Or does hierarchy/politics determine which ideas get heard?
   - **Disinterestedness**: Are conflicts of interest acknowledged?
     Is outcome information being hidden from advisors? (It should be —
     knowing outcomes triggers resulting in group members too)
   - **Organized Skepticism**: Is dissent rewarded or punished?
     Duke: the "heckler" is the best team player.

   Design the ideal truth-seeking structure for this decision:
   - Who should be consulted?
   - In what order? (Reveal positions last, not first)
   - What information should be shared vs. withheld?
   - How should disagreement be structured?

3. THE 3Ds PROTOCOL
   Duke's group decision framework:
   - **Discover**: Have individual opinions been collected independently
     BEFORE any group discussion? If not, anchor bias is operating.
   - **Discuss**: Has the group focused on understanding DISAGREEMENT
     rather than building CONSENSUS?
   - **Decide**: Will the final decision be made independently after
     discussion, or will social pressure drive conformity?

4. ULYSSES CONTRACT DESIGN
   Pre-commitment devices that bind future behavior. For this decision:
   - What BARRIER-RAISING contracts are needed? (Add friction to bad choices)
   - What BARRIER-REDUCING contracts are needed? (Remove friction from good choices)
   - What's the "decision swear jar" — the accountability trigger for catching
     resulting, motivated reasoning, or hindsight bias in real time?

5. DECISION JOURNAL TEMPLATE
   Design the specific knowledge tracker for this decision:
   - What to record NOW (before outcome): beliefs, probabilities, reasoning,
     information state, what's knowable but unknown
   - What to record LATER (after outcome): what happened, luck vs. skill
     attribution, memory creep check, counterfactual assessment
   - Duke: the journal's purpose is to combat hindsight bias and memory
     creep — the unconscious revision of what you believed before the outcome

6. TILT CHECK
   Duke's poker concept applied to decisions:
   - Is the decision-maker on tilt? (emotionally compromised from recent events)
   - Recent bad outcome that's distorting risk appetite?
   - Recent good outcome that's inflating confidence?
   - Duke's 10-10-10 test: how will this feel in 10 minutes, 10 months, 10 years?
   - If on tilt, recommend PAUSING the decision until emotional equilibrium returns.
   - "Path dependence" — Duke notes recent events drive responses more than
     overall position. A $100 win after a demotion doesn't produce happiness.

Output: structured process assessment and recommended protocol design.
Include specific templates (decision journal, kill criteria tracker, CUDOS
checklist) the decision-maker can actually use.

Message teammates about process findings that affect their analysis
(e.g., "decision is being made under tilt conditions after recent failure —
Calibrator should test whether risk aversion is inflating probability of
bad outcomes").

Spawning

生成Agent

Spawn all five as background agents. Use
model: "sonnet"
for all teammates (reasoning from principles and frameworks). The lead (Opus) handles synthesis.
Agent: {
  team_name: "duke-<decision-slug>",
  name: "resulting-auditor",
  model: "sonnet",
  prompt: [full resulting auditor prompt with decision substituted],
  run_in_background: true
}
Repeat for calibrator, pre-mortem-analyst, quit-strategist, process-architect.
Assign tasks immediately:
TaskUpdate: { taskId: "1", owner: "resulting-auditor" }
TaskUpdate: { taskId: "2", owner: "calibrator" }
TaskUpdate: { taskId: "3", owner: "pre-mortem-analyst" }
TaskUpdate: { taskId: "4", owner: "quit-strategist" }
TaskUpdate: { taskId: "5", owner: "process-architect" }
将所有5名Agent作为后台任务生成。所有团队成员使用
model: "sonnet"
(基于原则和框架进行推理)。主导者(Opus)负责整合分析结果。
Agent: {
  team_name: "duke-<decision-slug>",
  name: "resulting-auditor",
  model: "sonnet",
  prompt: [full resulting auditor prompt with decision substituted],
  run_in_background: true
}
为calibrator、pre-mortem-analyst、quit-strategist、process-architect重复上述步骤。
立即分配任务:
TaskUpdate: { taskId: "1", owner: "resulting-auditor" }
TaskUpdate: { taskId: "2", owner: "calibrator" }
TaskUpdate: { taskId: "3", owner: "pre-mortem-analyst" }
TaskUpdate: { taskId: "4", owner: "quit-strategist" }
TaskUpdate: { taskId: "5", owner: "process-architect" }

Phase 3: Monitor & Cross-Pollinate

阶段3:监控与交叉协作

While teammates work:
  • Messages from teammates arrive automatically
  • If a teammate asks a question, respond with guidance
  • If two teammates discover conflicting information, message both to reconcile
  • If a teammate finds something that dramatically changes the picture, alert others
团队成员工作期间:
  • 团队成员的消息会自动送达
  • 如果团队成员提出问题,给予指导
  • 如果两名团队成员发现冲突信息,通知双方进行协调
  • 如果团队成员发现可能大幅改变分析结果的信息,通知其他成员

Phase 4: Synthesize — The Duke Verdict

阶段4:整合分析——Duke结论

After ALL teammates report back, the lead writes the final analysis. This is the most important phase — it's where bias stacking emerges and the true quality of the decision process becomes visible.
所有团队成员提交报告后,主导者撰写最终分析报告。这是最重要的阶段——偏见堆叠情况会在此显现,决策流程的真实质量也会变得清晰。

The Synthesis Process

整合分析流程

  1. Collect all five analyses
  2. Cross-reference — where do multiple lenses reveal the same flaw or strength?
  3. Identify bias stacking — are multiple psychological traps reinforcing each other? (e.g., sunk cost + identity entanglement + optimism = escalation of commitment)
  4. Identify process gaps — where is the decision protocol missing critical elements?
  5. Apply the "Fold" filter — is there too much Knightian uncertainty for meaningful probability assignment? Is this outside the circle of competence?
  6. Render the verdict — Good Bet, Bad Bet, or Fold
  1. 收集所有5份分析报告
  2. 交叉验证——多个视角是否揭示了相同的缺陷或优势?
  3. 识别偏见堆叠——是否有多个心理陷阱相互强化? (例如:沉没成本+身份绑定+乐观偏见=承诺升级)
  4. 识别流程缺口——决策流程缺少哪些关键环节?
  5. 应用“放弃”筛选——是否存在过多奈特不确定性,无法进行有意义的概率分配?该决策是否超出能力圈?
  6. 给出结论——好赌注、坏赌注或放弃

Output Document

输出文档

Write to
thoughts/duke/YYYY-MM-DD-<decision-slug>.md
:
markdown
---
date: <ISO 8601>
analyst: Claude Code (duke decision quality skill)
decision: "<decision name>"
verdict: <GOOD_BET | BAD_BET | FOLD>
bias_count: <number of active psychological traps>
confidence: <LOW | MEDIUM | HIGH>
---
写入
thoughts/duke/YYYY-MM-DD-<decision-slug>.md
markdown
---
date: <ISO 8601>
analyst: Claude Code (duke decision quality skill)
decision: "<decision name>"
verdict: <GOOD_BET | BAD_BET | FOLD>
bias_count: <number of active psychological traps>
confidence: <LOW | MEDIUM | HIGH>
---

Duke Decision Analysis: [Decision Name]

Duke决策分析:[决策名称]

"What makes a decision great is not that it has a great outcome. A great decision is the result of a good process, and that process must include an attempt to accurately represent our own state of knowledge. That state of knowledge, in turn, is some variation of 'I'm not sure.'" — Annie Duke
"优秀的决策不在于它带来了好结果。优秀的决策是良好流程的产物,而这个流程必须包含准确呈现自身认知状态的尝试。这种认知状态,本质上是‘我不确定’的某种变体。" — Annie Duke

The Decision

决策内容

[One paragraph description]
[一段描述]

The Resulting Audit

Resulting审计

Decision/Outcome Matrix

决策/结果矩阵

Good Outcome (p=X%)Bad Outcome (p=Y%)
This is a good decision[scenario][scenario]
This is a bad decision[scenario][scenario]
好结果(概率=X%)坏结果(概率=Y%)
这是一个好决策[场景][场景]
这是一个坏决策[场景][场景]

Information State at Decision Time

决策当下的信息状态

[What's known, unknown, unknowable]
[已知、未知、不可知的信息]

Counterfactual Tree

反事实树

[Alternative futures being mentally pruned]
[被主观忽略的未来可能性分支]

Motivated Reasoning Check

动机性推理检查

[What biases are operating on the decision-maker]
[决策方存在哪些偏见]

Resulting Risk

Resulting风险

Decision quality (independent of outcome): STRONG / ADEQUATE / WEAK Outcome sensitivity: HIGH / MEDIUM / LOW [Will people correctly evaluate this decision regardless of outcome?]

决策质量(独立于结果): STRONG / ADEQUATE / WEAK 结果敏感度: HIGH / MEDIUM / LOW [人们是否能不受结果影响,正确评估该决策?]

The Calibration Assessment

校准评估

Belief Inventory

信念清单

AssumptionConfidenceRangeInside ViewOutside ViewGap
[assumption 1]X%A-B%[reasoning][base rate]SMALL/LARGE
[assumption 2]X%A-B%[reasoning][base rate]SMALL/LARGE
假设置信度范围内部视角外部视角差距
[假设1]X%A-B%[推理依据][基准数据]SMALL/LARGE
[假设2]X%A-B%[推理依据][基准数据]SMALL/LARGE

Overconfidence Assessment

过度自信评估

[Planning fallacy, optimism bias, vague language problems]
[规划谬误、乐观偏见、模糊语言问题]

Expected Value

预期价值

OptionEVVarianceFreeroll?
Option A$XHIGH/MED/LOWY/N
Option B$XHIGH/MED/LOWY/N
Status quo$XHIGH/MED/LOWY/N
选项EV方差无风险收益?
选项A$XHIGH/MED/LOWY/N
选项B$XHIGH/MED/LOWY/N
维持现状$XHIGH/MED/LOWY/N

Calibration Verdict

校准结论

Best option by EV: [option] Confidence in probability estimates: WELL-CALIBRATED / OVERCONFIDENT / UNDERCONFIDENT

EV最优选项: [选项] 概率估计置信度: WELL-CALIBRATED / OVERCONFIDENT / UNDERCONFIDENT

The Pre-Mortem

事前验尸分析

Controllable Failure Modes

可控失败模式

#Failure ModeProbabilityPreventable?Early Warning
1[mode]HIGH/MED/LOW[yes + how][signal]
2............
#失败模式概率是否可预防?预警信号
1[模式]HIGH/MED/LOW[是 + 预防方法][信号]
2............

Uncontrollable Failure Modes

不可控失败模式

#Failure ModeProbabilityResilience Possible?
1[mode]HIGH/MED/LOW[yes/no + how]
#失败模式概率是否可提升韧性?
1[模式]HIGH/MED/LOW[是/否 + 方法]

Monkey vs. Pedestal

核心难题vs无用铺垫

THE MONKEY: [the hardest, most uncertain element] THE PEDESTAL: [the easy stuff that creates false progress] Building pedestals? [YES — describe / NO] #MONKEYFIRST status: [Is the monkey being tackled first?]
核心难题(MONKEY): [最困难、最不确定的环节] 无用铺垫(PEDESTAL): [容易完成、已解决的舒适任务] 是否在做无用铺垫? [是——描述 / 否] #MONKEYFIRST状态: [是否优先处理核心难题?]

Decision Exploration Table

决策探索表

FactorSuccess ScenarioFailure ScenarioControllable?
[factor][what happens][what happens]YES/NO

因素成功场景失败场景是否可控?
[因素][情况描述][情况描述]YES/NO

The Quit Analysis

退出分析

Sunk Cost Inventory

沉没成本清单

[What's irrecoverable and what's distorting the decision]
[无法收回的投入及对决策的影响]

Psychological Traps

心理陷阱评估

TrapStrength (0-10)Evidence
Sunk cost fallacyX[evidence]
Loss aversionX[evidence]
Status quo biasX[evidence]
Endowment effectX[evidence]
Identity entanglementX[evidence]
Escalation of commitmentX[evidence]
Optimism biasX[evidence]
陷阱强度(0-10)证据
沉没成本谬误X[具体证据]
损失厌恶X[具体证据]
现状偏见X[具体证据]
禀赋效应X[具体证据]
身份绑定X[具体证据]
承诺升级X[具体证据]
乐观偏见X[具体证据]

Kill Criteria

终止标准

  1. "If _____________ by _____________, then _____________"
  2. "If _____________ by _____________, then _____________"
  3. "If _____________ by _____________, then _____________"
  1. "如果_________在_________前未达成,则_________"
  2. "如果_________在_________前未达成,则_________"
  3. "如果_________在_________前未达成,则_________"

EV: Continue vs. Stop

预期价值:继续vs停止

PathExpected ValueKey Assumption
Continue$X[what must be true]
Quit/pivot$X[what must be true]
Status quo$X[what must be true]
路径预期价值核心假设
继续$X[必须成立的条件]
退出/转型$X[必须成立的条件]
维持现状$X[必须成立的条件]

Quit Verdict

退出结论

Should they quit? [YES / NO / NOT YET — here are the kill criteria] Quitting coach in place? [YES / NO — recommend one]

是否应该退出? [是 / 否 / 暂不退出——以下是终止标准] 是否有退出教练? [是 / 否——建议指定一名]

The Process Assessment

流程评估

Current Process Rating

当前流程评分

Process quality: STRONG / ADEQUATE / WEAK / ABSENT [What's working, what's missing]
流程质量: STRONG / ADEQUATE / WEAK / ABSENT [现有流程的优势与不足]

CUDOS Compliance

CUDOS准则合规性

NormStatusIssue
Communism (info sharing)PASS/FAIL[issue]
Universalism (merit-based)PASS/FAIL[issue]
Disinterestedness (no conflicts)PASS/FAIL[issue]
Organized Skepticism (dissent)PASS/FAIL[issue]
准则状态问题
Communism(信息共享)PASS/FAIL[问题描述]
Universalism(唯才是举)PASS/FAIL[问题描述]
Disinterestedness(无利益冲突)PASS/FAIL[问题描述]
Organized Skepticism(鼓励异议)PASS/FAIL[问题描述]

Tilt Assessment

情绪失控(Tilt)评估

On tilt? [YES / NO] Source: [recent event distorting judgment] Recommendation: [proceed / pause / restructure]
是否情绪失控? [是 / 否] 原因: [影响判断的近期事件] 建议: [继续 / 暂停 / 调整]

Recommended Protocol

推荐流程

[Specific process improvements, decision journal template, Ulysses contracts]

[具体流程改进建议、决策日志模板、尤利西斯契约]

THE BIAS STACKING ASSESSMENT

偏见堆叠评估

This is the Duke question: how many psychological traps are operating simultaneously, and do they reinforce each other?
这是Duke框架的核心问题:有多少心理陷阱在同时起作用,它们是否相互强化?

Biases Stacking Against Good Process

影响良好决策的偏见堆叠

[Bias 1: e.g., sunk cost — $2M already invested]
  + [Bias 2: e.g., identity entanglement — "I'm the founder who believes in this"]
    + [Bias 3: e.g., optimism bias — "this time will be different"]
      + [Bias 4: e.g., status quo bias — "changing course feels like failure"]
        + [Bias 5: e.g., loss aversion — "quitting means losing everything"]
          = [ESCALATION SPIRAL? / MANAGEABLE? / MINIMAL?]
Bias stacking severity: [NONE / LOW / MODERATE / SEVERE / CRITICAL]
A severe bias stack (4+ reinforcing traps) is Duke's equivalent of a negative lollapalooza — the psychological forces conspire to prevent rational analysis and virtually guarantee escalation of commitment.
[偏见1:例如,沉没成本——已投入200万美元]
  + [偏见2:例如,身份绑定——“我是坚信这个项目的创始人”]
    + [偏见3:例如,乐观偏见——“这次会不一样”]
      + [偏见4:例如,现状偏见——“改变方向意味着失败”]
        + [偏见5:例如,损失厌恶——“退出意味着失去一切”]
          = [承诺升级螺旋? / 可控? / 轻微?]
偏见堆叠严重程度: [无 / 低 / 中 / 高 / 极高]
严重的偏见堆叠(4个及以上相互强化的陷阱)相当于Duke所说的负面“lollapalooza”——心理因素共同作用,阻碍理性分析,几乎必然导致承诺升级。

Process Strengths Working FOR Good Decision-Making

助力良好决策的流程优势

[Strength 1] + [Strength 2] + [Strength 3] = [cumulative process quality]

[优势1] + [优势2] + [优势3] = [累计流程质量]

THE VERDICT

最终结论

Duke's Three Baskets

Duke的三类结论

[ ] GOOD BET — Sound process, positive expected value, calibrated probabilities, monkey identified and being tackled, kill criteria defined. The decision quality is high regardless of how it turns out.
[ ] BAD BET — Process flaw, negative expected value, severe bias stacking, or building pedestals while the monkey sits untrained. Walk away or fundamentally restructure the decision.
[ ] FOLD — Too much Knightian uncertainty to assign meaningful probabilities. The reference class is too thin, the information state is too poor, or this is outside the circle of competence. Gather more information or accept that this is a coin flip, not a calculated bet.
[ ] 好赌注——流程合理,预期价值为正,概率经过校准,核心难题已识别并优先处理,终止标准明确。无论结果如何,决策质量都很高。
[ ] 坏赌注——流程存在缺陷,预期价值为负,偏见堆叠严重,或在核心难题未解决的情况下做无用铺垫。应放弃或彻底重构决策。
[ ] 放弃——存在过多奈特不确定性,无法进行有意义的概率分配。参考案例不足,信息状态极差,或超出能力圈。应收集更多信息,或接受这只是碰运气,而非有计算的赌注。

Verdict: [GOOD BET / BAD BET / FOLD]

结论:[GOOD BET / BAD BET / FOLD]

Confidence: [LOW / MEDIUM / HIGH]
Reasoning: [2-3 paragraphs written in Duke's direct, warm, no-BS style. Reference specific findings from each analyst. Name the biases operating. If it's a BAD BET, name the process flaw without apology — "you're resulting" or "you're building pedestals" or "the sunk cost is running the show." If it's a GOOD BET, explain what makes the process sound and what the kill criteria are. If it's FOLD, explain what information would move it to GOOD BET.]
置信度: [LOW / MEDIUM / HIGH]
推理: [2-3段以Duke直接、诚恳、不绕弯的风格撰写的内容。参考每位分析师的具体发现。指出存在的偏见。如果是坏赌注,直接点明流程缺陷——“你陷入了Resulting”或“你在做无用铺垫”或“沉没成本主导了决策”。如果是好赌注,说明流程合理之处及终止标准。如果是放弃,说明需要哪些信息才能转为好赌注。]

What Annie Would Say

Annie会怎么说

[Write 2-3 sentences in Duke's voice — direct, warm, grounded in poker metaphor, self-aware about human fallibility. She'd probably say something like "Look, you're not a bad decision-maker — you're a normal human with normal biases, and those biases are stacking up on you right now. Here's what I'd do at the table..." She uses "I'm not sure" as a strength. She tells stories. She's generous but unflinching.]
[用Duke的语气写2-3句话——直接、温暖、以扑克为隐喻,对人类的局限性有清晰认知。她可能会说:“听着,你不是一个糟糕的决策者——你只是一个有正常偏见的普通人,而这些偏见现在正叠加在一起影响你。换作是我在牌桌上,我会这么做……”她把“我不确定”视为一种优势。她会讲故事。她态度诚恳但毫不留情。]

If You Proceed: The Protocol

如果继续执行:行动准则

[Based on the Process Architect's assessment and the Pre-Mortem's failure modes, write 3-5 rules for how to execute this decision well.]
  1. Define your kill criteria NOW — [specific state + date criteria]
  2. Tackle the monkey first — [specific hard problem to solve before anything else]
  3. Never [process flaw to avoid] — because [bias/failure mode it triggers]
  4. Build your decision pod — [who to include, how to structure]
  5. Journal this decision — [what to record now, what to check later]
undefined
[基于流程架构师的评估和事前验尸分析的失败模式,撰写3-5条执行决策的规则。]
  1. 现在就明确终止标准——[具体的状态+日期标准]
  2. 优先解决核心难题——[在做其他事情前先解决的具体难题]
  3. 永远不要[避免的流程缺陷]——因为[它会触发的偏见/失败模式]
  4. 组建你的决策小组——[成员构成及架构方式]
  5. 记录决策日志——[现在需要记录的内容,后续需要检查的内容]
undefined

Phase 5: Present & Follow-up

阶段5:呈现结果与跟进

Present the verdict to the user with the key highlights. Don't dump the whole document — give the verdict, the bias stacking assessment, and the kill criteria. Let them read the full document.
undefined
向用户呈现结论及关键要点。不要直接输出完整文档——给出结论、偏见堆叠评估和终止标准,让用户自行查看完整文档。
undefined

Duke Verdict: [DECISION] — [GOOD BET / BAD BET / FOLD]

Duke结论:[决策名称] — [GOOD BET / BAD BET / FOLD]

Bias stacking: [severity] — [N] traps active [reinforcing/independent] Calibration: [well-calibrated / overconfident / underconfident] Monkey status: [identified and first / pedestal-building / no monkey] Kill criteria: [defined / missing — recommend defining] Process quality: [strong / adequate / weak / absent]
What Annie would say: "[direct, warm assessment]"
Full analysis:
thoughts/duke/YYYY-MM-DD-<slug>.md
Want me to:
  1. Deep-dive into any analyst's findings?
  2. Re-run with a modified version of the decision?
  3. Apply /office-hours to refine the question before re-analyzing?
  4. Compare this against an alternative decision? (batch mode)
  5. Run /munger to evaluate the business fundamentals alongside the decision process?
undefined
偏见堆叠: [严重程度] — [N]个陷阱在起作用 [相互强化/独立] 校准情况: [校准良好 / 过度自信 / 信心不足] 核心难题状态: [已识别并优先处理 / 在做无用铺垫 / 未识别] 终止标准: [已明确 / 缺失——建议明确] 流程质量: [优秀 / 合格 / 薄弱 / 缺失]
Annie会怎么说: "[直接、温暖的评估]"
完整分析文档:
thoughts/duke/YYYY-MM-DD-<slug>.md
你希望我:
  1. 深入解读某位分析师的发现?
  2. 修改决策内容后重新分析?
  3. 先运行/office-hours优化问题后再重新分析?
  4. 将该决策与其他决策进行对比?(批量模式)
  5. 同时运行/munger评估业务基本面与决策流程?
undefined

Batch Mode

批量模式

If the user wants to compare multiple decisions:
  1. Run the full analysis on each (can parallelize — one team per decision)
  2. At the end, produce a comparison:
undefined
如果用户希望对比多个决策:
  1. 对每个决策执行完整分析(可并行——每个决策对应一个团队)
  2. 最后生成对比表格:
undefined

Duke Decision Leaderboard

Duke决策排行榜

RankDecisionVerdictBias StackCalibrationMonkeyProcess
1[name]GOOD BETLOW (1)CalibratedFirstStrong
2[name]FOLDMOD (3)OverconfidentUnknownWeak
3[name]BAD BETSEVERE (5)OverconfidentPedestalAbsent
undefined
排名决策名称结论偏见堆叠校准情况核心难题流程质量
1[名称]GOOD BETLOW (1)校准良好优先处理优秀
2[名称]FOLDMOD (3)过度自信未识别薄弱
3[名称]BAD BETSEVERE (5)过度自信无用铺垫缺失
undefined

Scoring Discipline

评分准则

  • Be Duke, not a consultant. Duke says most people are overconfident and quit too late. If every decision gets GOOD BET, the skill is broken.
  • Cite the source analyst. Every claim traces to a specific teammate's finding.
  • No process inflation. A decision with vibes-based probability estimates and no kill criteria is not a "strong process" no matter how smart the decision-maker sounds.
  • Name the bias. Don't say "there might be some bias." Say "sunk cost fallacy is at 8/10 because you've invested $2M and your identity is fused with this project." Be specific and cite evidence.
  • The Fold basket is honest. When probabilities are fabricated rather than calculated — when there's no reference class, no base rate, no feedback loop — say so. "I'm not sure" is Duke's highest form of intellectual honesty.
  • 做Duke,而非顾问。Duke认为大多数人过度自信且退出太晚。如果每个决策都被评为好赌注,说明这个工具失效了。
  • 引用分析师来源。每一个结论都要追溯到特定团队成员的发现。
  • 不夸大流程质量。即使决策者听起来很聪明,但如果决策基于直觉概率估计且没有终止标准,就不能算是“优秀流程”。
  • 明确指出偏见。不要说“可能存在一些偏见”,要说“沉没成本谬误的强度为8/10,因为你已投入200万美元且身份与该项目绑定”。要具体并引用证据。
  • “放弃”选项要诚实。当概率是编造而非计算出来的——没有参考案例、没有基准数据、没有反馈循环——要直接说明。“我不确定”是Duke认为的最高形式的智识诚实。

Important Notes

重要说明

  • Cost: This skill spawns 5 agents. It's expensive. Worth it for serious decisions, not for casual what-ifs (use /office-hours for that).
  • Sonnet for teammates, Opus for synthesis: The lead handles the bias stacking detection and final verdict — that's where deep reasoning matters.
  • No team? No problem: If teams aren't enabled, run 5 sequential background agents and collect results. Same analysis, just no cross-talk.
  • Pair with other skills: Run /office-hours first to refine the question, then /duke to stress-test the decision process. Run /munger alongside to evaluate business fundamentals — Duke is decision process, Munger is decision content. Together they form a complete framework.
  • When NOT to use this: One-shot existential decisions with no reference class (where probability estimates are fabricated); moral choices where duty overrides expected value; situations where "process was good" has become an excuse for persistent bad results. In those cases, acknowledge the framework's limits honestly.
  • 成本:该工具会生成5个Agent,成本较高。适合重要决策,不适合随意的假设分析(这类情况使用/office-hours)。
  • 团队成员用Sonnet,主导者用Opus:主导者负责识别偏见堆叠情况并给出最终结论——这需要深度推理能力。
  • 没有团队功能也没关系:如果团队功能未启用,按顺序运行5个后台Agent并收集结果。分析质量一致,只是没有成员间的交叉沟通。
  • 与其他工具搭配使用:先运行/office-hours优化问题,再运行/duke进行决策流程压力测试。同时运行/munger评估业务基本面——Duke关注决策流程,Munger关注决策内容。两者结合构成完整框架。
  • 不适用场景:没有参考案例的一次性生存决策(概率估计是编造的);道德义务优先于预期价值的选择;“流程良好”已成为持续糟糕结果的借口的情况。在这些情况下,要诚实地说明框架的局限性。