Loading...
Loading...
Compare original and translation side by side
Phase 0: Setup
mkdir -p .specs/reports
│
Phase 1: Independent Analysis
┌─ Judge 1 → {name}.1.md ─┐
Solution ┼─ Judge 2 → {name}.2.md ─┼─┐
└─ Judge 3 → {name}.3.md ─┘ │
│
Phase 2: Debate Round (iterative) │
Each judge reads others' reports │
↓ │
Argue + Defend + Challenge │
↓ │
Revise if convinced ─────────────┤
↓ │
Check consensus │
├─ Yes → Final Report │
└─ No → Next Round ─────────┘Phase 0: Setup
mkdir -p .specs/reports
│
Phase 1: Independent Analysis
┌─ Judge 1 → {name}.1.md ─┐
Solution ┼─ Judge 2 → {name}.2.md ─┼─┐
└─ Judge 3 → {name}.3.md ─┘ │
│
Phase 2: Debate Round (iterative) │
Each judge reads others' reports │
↓ │
Argue + Defend + Challenge │
↓ │
Revise if convinced ─────────────┤
↓ │
Check consensus │
├─ Yes → Final Report │
└─ No → Next Round ─────────┘mkdir -p .specs/reports.specs/reports/{solution-name}-{YYYY-MM-DD}.[1|2|3].md{solution-name}users-apisrc/api/users.ts{YYYY-MM-DD}[1|2|3]mkdir -p .specs/reports.specs/reports/{solution-name}-{YYYY-MM-DD}.[1|2|3].md{solution-name}src/api/users.tsusers-api{YYYY-MM-DD}[1|2|3].specs/reports/{solution-name}-{date}.[1|2|3].mdYou are Judge {N} evaluating a solution independently.
<solution_path>
{path to solution file(s)}
</solution_path>
<task_description>
{what the solution was supposed to accomplish}
</task_description>
<evaluation_criteria>
{criteria with descriptions and weights}
</evaluation_criteria>
<output_file>
.specs/reports/{solution-name}-{date}.{N}.md
</output_file>
Read ${CLAUDE_PLUGIN_ROOT}/tasks/judge.md for evaluation methodology and execute using following criteria.
Instructions:
1. Read the solution thoroughly
2. For each criterion:
- Find specific evidence (quote exact text)
- Score on the defined scale
- Justify with concrete examples
3. Calculate weighted overall score
4. Write comprehensive report to {output_file}
5. Generate verification 5 questions about your evaluation.
6. Answer verification questions:
- Re-examine solutions for each question
- Find counter-evidence if it exists
- Check for systematic bias (length, confidence, etc.)
7. Revise your report file and update it accordingly.
Add to report begining `Done by Judge {N}`.specs/reports/{solution-name}-{date}.[1|2|3].mdYou are Judge {N} evaluating a solution independently.
<solution_path>
{path to solution file(s)}
</solution_path>
<task_description>
{what the solution was supposed to accomplish}
</task_description>
<evaluation_criteria>
{criteria with descriptions and weights}
</evaluation_criteria>
<output_file>
.specs/reports/{solution-name}-{date}.{N}.md
</output_file>
Read ${CLAUDE_PLUGIN_ROOT}/tasks/judge.md for evaluation methodology and execute using following criteria.
Instructions:
1. Read the solution thoroughly
2. For each criterion:
- Find specific evidence (quote exact text)
- Score on the defined scale
- Justify with concrete examples
3. Calculate weighted overall score
4. Write comprehensive report to {output_file}
5. Generate verification 5 questions about your evaluation.
6. Answer verification questions:
- Re-examine solutions for each question
- Find counter-evidence if it exists
- Check for systematic bias (length, confidence, etc.)
7. Revise your report file and update it accordingly.
Add to report begining `Done by Judge {N}`.specs/reports/{solution-name}-{date}.[1|2|3].md.specs/reports/{solution-name}-{date}.[1|2|3].md## Debate Round {R}You are Judge {N} in debate round {R}.
<your_previous_report>
{path to .specs/reports/{solution-name}-{date}.{N}.md}
</your_previous_report>
<other_judges_reports>
Judge 1: .specs/reports/{solution-name}-{date}.1.md
...
</other_judges_reports>
<task_description>
{what the solution was supposed to accomplish}
</task_description>
<solution_path>
{path to solution}
</solution_path>
<output_file>
.specs/reports/{solution-name}-{date}.{N}.md (append to existing file)
</output_file>
Read ${CLAUDE_PLUGIN_ROOT}/tasks/judge.md for evaluation methodology principles.
Instructions:
1. Read your previous assessment from {your_previous_report}
2. Read all other judges' reports
3. Identify disagreements (where your scores differ by >1 point)
4. For each major disagreement:
- State the disagreement clearly
- Defend your position with evidence
- Challenge the other judge's position with counter-evidence
- Consider whether their evidence changes your view
5. Update your report file by APPENDING:
6. Reply whether you are reached agreement, and with which judge. Include revisited scores and criteria scores.
---.specs/reports/{solution-name}-{date}.[1|2|3].md.specs/reports/{solution-name}-{date}.[1|2|3].md## Debate Round {R}You are Judge {N} in debate round {R}.
<your_previous_report>
{path to .specs/reports/{solution-name}-{date}.{N}.md}
</your_previous_report>
<other_judges_reports>
Judge 1: .specs/reports/{solution-name}-{date}.1.md
...
</other_judges_reports>
<task_description>
{what the solution was supposed to accomplish}
</task_description>
<solution_path>
{path to solution}
</solution_path>
<output_file>
.specs/reports/{solution-name}-{date}.{N}.md (append to existing file)
</output_file>
Read ${CLAUDE_PLUGIN_ROOT}/tasks/judge.md for evaluation methodology principles.
Instructions:
1. Read your previous assessment from {your_previous_report}
2. Read all other judges' reports
3. Identify disagreements (where your scores differ by >1 point)
4. For each major disagreement:
- State the disagreement clearly
- Defend your position with evidence
- Challenge the other judge's position with counter-evidence
- Consider whether their evidence changes your view
5. Update your report file by APPENDING:
6. Reply whether you are reached agreement, and with which judge. Include revisited scores and criteria scores.
---undefinedundefined.specs/reports/{solution-name}-{date}.[1|2|3].md.specs/reports/{solution-name}-{date}.[1|2|3].mdundefinedundefined| Criterion | Judge 1 | Judge 2 | Judge 3 | Final |
|---|---|---|---|---|
| {Name} | {X}/5 | {X}/5 | {X}/5 | {X}/5 |
| ... |
| Criterion | Judge 1 | Judge 2 | Judge 3 | Final |
|---|---|---|---|---|
| {Name} | {X}/5 | {X}/5 | {X}/5 | {X}/5 |
| ... |
<output>
The command produces:
1. **Reports directory**: `.specs/reports/` (created if not exists)
2. **Initial reports**: `.specs/reports/{solution-name}-{date}.1.md`, `.specs/reports/{solution-name}-{date}.2.md`, `.specs/reports/{solution-name}-{date}.3.md`
3. **Debate updates**: Appended sections in each report file per round
4. **Final synthesis**: Replied to user (consensus or disagreement summary)
</output>
<output>
本命令将生成:
1. **报告目录**:`.specs/reports/`(不存在则创建)
2. **初始报告**:`.specs/reports/{solution-name}-{date}.1.md`、`.specs/reports/{solution-name}-{date}.2.md`、`.specs/reports/{solution-name}-{date}.3.md`
3. **辩论更新**:每轮辩论后在各报告文件中追加的章节
4. **最终综合结果**:返回给用户的(共识或分歧总结)
</output>/judge-with-debate \
--solution "src/api/users.ts" \
--task "Implement REST API for user management" \
--criteria "correctness:30,design:25,security:20,performance:15,docs:10".specs/reports/users-api-2025-01-15.1.md.specs/reports/users-api-2025-01-15.2.md.specs/reports/users-api-2025-01-15.3.mdCorrectness: 4.3/5
Design: 4.5/5
Security: 4.0/5 (3 rounds to consensus)
Performance: 4.7/5
Documentation: 4.0/5
Overall: 4.3/5 - PASS/judge-with-debate \
--solution "src/api/users.ts" \
--task "Implement REST API for user management" \
--criteria "correctness:30,design:25,security:20,performance:15,docs:10".specs/reports/users-api-2025-01-15.1.md.specs/reports/users-api-2025-01-15.2.md.specs/reports/users-api-2025-01-15.3.mdCorrectness: 4.3/5
Design: 4.5/5
Security: 4.0/5 (3 rounds to consensus)
Performance: 4.7/5
Documentation: 4.0/5
Overall: 4.3/5 - PASS