View a PDF of the paper titled Do LLMs Really Think Step-by-step In Implicit Reasoning?, by Yijiong Yu
Abstract:It has been well-known that Chain-of-Thought can remarkably enhance LLMs’ performance on complex tasks. However, because it also introduces slower inference speeds and higher computational costs, many researches have attempted to use implicit CoT, which does not need LLMs to explicitly generate the intermediate steps. However, the invisible reasoning process leaves us a doubt that, can implicit CoT really be equal to explicit CoT? Therefore, in this study, we address this question through experiments. We probe the information of intermediate steps from the model’s hidden states when it is either trained or prompted to perform implicit CoT. The results surprisingly indicate that when prompted, LLMs hardly think about intermediate steps, suggesting they may just rely on experience rather than strict step-by-step reasoning. But when trained, they indeed calculate intermediate steps. Moreover, in both situations, we find the effect of using implicit CoT is susceptible to the format of the problem, reaffirming the current deficiency of implicit CoT.
Submission history
From: Yijiong Yu [view email]
[v1]
Sun, 24 Nov 2024 14:38:59 UTC (468 KB)
[v2]
Wed, 4 Dec 2024 05:52:03 UTC (468 KB)
[v3]
Fri, 27 Dec 2024 07:04:19 UTC (1,006 KB)
Source link
lol