Papers
arxiv:2311.09277

Contrastive Chain-of-Thought Prompting

Published on Nov 15, 2023
· Featured in Daily Papers on Nov 17, 2023
Authors:
,

Abstract

Despite the success of chain of thought in enhancing language model reasoning, the underlying process remains less well understood. Although logically sound reasoning appears inherently crucial for chain of thought, prior studies surprisingly reveal minimal impact when using invalid demonstrations instead. Furthermore, the conventional chain of thought does not inform language models on what mistakes to avoid, which potentially leads to more errors. Hence, inspired by how humans can learn from both positive and negative examples, we propose contrastive chain of thought to enhance language model reasoning. Compared to the conventional chain of thought, our approach provides both valid and invalid reasoning demonstrations, to guide the model to reason step-by-step while reducing reasoning mistakes. To improve generalization, we introduce an automatic method to construct contrastive demonstrations. Our experiments on reasoning benchmarks demonstrate that contrastive chain of thought can serve as a general enhancement of chain-of-thought prompting.

Community

interesting

While reading NLP papers, I always wondered what it would be like to train LM on incorrect data. However, after reading this paper, finally I was able to resolve my consideration! Thank you for conducting what a wonderful research!

Hi, this is a really nice paper.
I have a few questions:

  1. Have you tried mixing multiple "negative" methods? like irrelevant language with incoherent objects.
  2. Do the few-shot samples matter? If the LLM makes some mistake for some samples, applying them could be a stronger negative example.
  3. Why could irrelevant/incoherent objects help the model with the reasoning?

Sign up or log in to comment

Models citing this paper 0

No model linking this paper

Cite arxiv.org/abs/2311.09277 in a model README.md to link it from this page.

Datasets citing this paper 0

No dataset linking this paper

Cite arxiv.org/abs/2311.09277 in a dataset README.md to link it from this page.

Spaces citing this paper 0

No Space linking this paper

Cite arxiv.org/abs/2311.09277 in a Space README.md to link it from this page.

Collections including this paper 28