思维链(Chain-of-Thought Prompting Elicits Reasoning in Large Language Models)

generalize

The paper mainly describes a way to use thinking chain to improve the reasoning ability of LLM model, and through experiments, it proves the significant effect of thinking chain in tasks such as arithmetic, common sense and symbols. The sota effect on GSM8K can be achieved only through the 540B size PaLM model and 8 thought chain samples.
Insert image description here

Specific work

Insert image description here
Insert image description here
This paper mainly explores the reasoning ability of the LLM model. The reasoning ability of the LLM model can be enhanced through reinforcement learning and ICL, but the effect of reinforcement learning and ICL improvement is limited and the cost is relatively high. Based on the above status quo and inspiration, this paper proposes a method based on thinking chains to enhance the reasoning ability of LLM, and has achieved good results.
Insert image description here
And the thinking chain has the above attractive characteristics.

Supongo que te gusta

Origin blog.csdn.net/WitsMakeMen/article/details/133071438
Recomendado
Clasificación