AI Safety Fundamentals: Alignment

Two-Turn Debate Doesn’t Help Humans Answer Hard Reading Comprehension Questions

June 17, 2024 BlueDot Impact Season 13
Two-Turn Debate Doesn’t Help Humans Answer Hard Reading Comprehension Questions
AI Safety Fundamentals: Alignment
More Info
AI Safety Fundamentals: Alignment
Two-Turn Debate Doesn’t Help Humans Answer Hard Reading Comprehension Questions
Jun 17, 2024 Season 13
BlueDot Impact

Using hard multiple-choice reading comprehension questions as a testbed, we assess whether presenting humans with arguments for two competing answer options, where one is correct and the other is incorrect, allows human judges to perform more accurately, even when one of the arguments is unreliable and deceptive. If this is helpful, we may be able to increase our justified trust in language-model-based systems by asking them to produce these arguments where needed. Previous research has shown that just a single turn of arguments in this format is not helpful to humans. However, as debate settings are characterized by a back-and-forth dialogue, we follow up on previous results to test whether adding a second round of counter-arguments is helpful to humans. We find that, regardless of whether they have access to arguments or not, humans perform similarly on our task. These findings suggest that, in the case of answering reading comprehension questions, debate is not a helpful format.


Source:

https://arxiv.org/abs/2210.10860


Narrated for AI Safety Fundamentals by Perrin Walker of TYPE III AUDIO.

---

A podcast by BlueDot Impact.

Learn more on the AI Safety Fundamentals website.

Show Notes Chapter Markers

Using hard multiple-choice reading comprehension questions as a testbed, we assess whether presenting humans with arguments for two competing answer options, where one is correct and the other is incorrect, allows human judges to perform more accurately, even when one of the arguments is unreliable and deceptive. If this is helpful, we may be able to increase our justified trust in language-model-based systems by asking them to produce these arguments where needed. Previous research has shown that just a single turn of arguments in this format is not helpful to humans. However, as debate settings are characterized by a back-and-forth dialogue, we follow up on previous results to test whether adding a second round of counter-arguments is helpful to humans. We find that, regardless of whether they have access to arguments or not, humans perform similarly on our task. These findings suggest that, in the case of answering reading comprehension questions, debate is not a helpful format.


Source:

https://arxiv.org/abs/2210.10860


Narrated for AI Safety Fundamentals by Perrin Walker of TYPE III AUDIO.

---

A podcast by BlueDot Impact.

Learn more on the AI Safety Fundamentals website.

Two turn debate
Abstract
1 Introduction
2 Counter-Argument Writing Protocol
2.1 Multi-Turn Writing Task
2.2 Multi-Turn Judging Protocols
3 Results
4 Conclusion