IPZ28ZqD4I@OpenReview

Total: 1

#1 Faithful and Unfaithful Error Recovery in Chain of Thought [PDF1] [Copy] [Kimi2] [REL]

Authors: Evelyn Yee ; Alice Li ; Chenyu Tang ; Yeon Ho Jung ; Ramamohan Paturi ; Leon Bergen

Large language models (LLMs) often improve their performance in downstream tasks when they generate Chain of Thought reasoning text before producing an answer. We investigate how LLMs recover from errors in Chain of Thought. Through analysis of error recovery behaviors, we find evidence for unfaithfulness in Chain of Thought, which occurs when models arrive at the correct answer despite invalid reasoning text. We identify factors that shift LLM recovery behavior: LLMs recover more frequently from obvious errors and in contexts that provide more evidence for the correct answer. Critically, these factors have divergent effects on faithful and unfaithful recoveries. Our results indicate that there are distinct mechanisms driving faithful and unfaithful error recoveries. Selective targeting of these mechanisms may be able to drive down the rate of unfaithful reasoning and improve model interpretability.