Skip to yearly menu bar Skip to main content


Poster

Benchmarking the Reasoning Robustness against Noisy Rationales in Chain-of-thought Prompting

Zhanke Zhou · Rong Tao · Jianing Zhu · Yiwen Luo · Zengmao Wang · Bo Han

East Exhibit Hall A-C #2207
[ ] [ Project Page ]
Wed 11 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

This paper investigates an under-explored challenge in large language models (LLMs): chain-of-thought prompting with noisy rationales—irrelevant or inaccurate reasoning steps—despite advancements in in-context learning. We construct the NoRa dataset, specifically designed to evaluate LLMs’ robustness to noisy rationales, based on which we reveal a widespread vulnerability among LLMs to such noise, with limited efficacy from existing reasoning methods. To combat this, we propose the contrastive denoising with noisy chain-of-thought (CD-CoT) method to enhance denoising-reasoning capabilities by contrasting noisy rationales with only one clean rationale, thereby advancing the robustness of LLMs in reasoning.

Live content is unavailable. Log in and register to view live content