Debate (Irving, Christiano, Amodei 2018) is a proposed scalable-oversight scheme. Two AI debaters argue opposing positions on a question; a human judge picks the winner. The argument: truth has an evidential advantage in adversarial debate, a deceiver must defend false claims against an honest opponent who can search the entire space of inconsistencies.
Theoretical claim: in the limit of unbounded debate length and a perfectly-rational judge, debate equilibrium produces honest answers. Even when the question is too complex for the judge to evaluate directly, dishonesty is exposed because every false sub-claim can be challenged with truth.
Practical concerns:
- Bounded debates: real debates are short. Whether the equilibrium is reached at length $\leq T$ depends on the structure of the question.
- Imperfect judges: humans have known biases (length, formality, confidence); a debater can exploit them. Empirical work tests whether AI debate helps humans converge on correct answers in spite of judge weaknesses.
- Collusion: if both debaters share a misalignment, they may coordinate to deceive the judge rather than expose each other's weaknesses.
Empirical research:
AI Safety via Debate: original 2018 Irving-Christiano paper, theoretical with toy examples.
Debate scales to evaluation tasks (Khan, Hughes et al. 2024): empirical evaluation of debate on reading-comprehension tasks where humans don't have access to the source text. Found that AI-debate-judged answers correlate better with ground truth than non-debate judging in some regimes.
Self-correcting language models via debate: more recent work tests whether weaker judges using stronger debaters reach correct answers more often than the same judges working alone.
Status: debate remains a research direction rather than a deployed alignment scheme. Combined with process supervision and scalable oversight, it forms part of the active research agenda on producing useful training signal for tasks beyond direct human verification.
Related terms: Scalable Oversight, paul-christiano, geoffrey-irving, RLHF
Discussed in:
- Chapter 16: Ethics & Safety, AI Safety