Noisy Rationales (Noisy-R): Intermediate reasoning steps in few-shot examples that are either factually inaccurate or contextually irrelevant, despite the final answer being correct.
CD-CoT: Contrastive Denoising with noisy Chain-of-Thought—the proposed method that uses one clean example to filter and correct noisy rationales.
Irrelevant thoughts: Reasoning steps that are factually correct but unhelpful for the specific question (e.g., reciting biological facts during a logic puzzle).
Inaccurate thoughts: Reasoning steps containing factual errors (e.g., calculation errors) but leading to the correct final label in the example.
Self-consistency (SC): A technique where the model generates multiple reasoning paths and selects the most consistent answer via voting.
ICL: In-Context Learning—the ability of models to learn from a few examples provided in the prompt without parameter updates.
NoRa: The benchmark dataset constructed in this paper, containing 26,391 questions across Math, Symbolic, and Commonsense domains with controlled noise ratios.