Skip to yearly menu bar Skip to main content


Poster

Can Language Models Learn to Skip Steps?

Tengxiao Liu · Qipeng Guo · Xiangkun Hu · Cheng Jiayang · Yue Zhang · Xipeng Qiu · Zheng Zhang

East Exhibit Hall A-C #2900
[ ]
Thu 12 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

Trained on vast corpora of human language, language models demonstrate emergent human-like reasoning abilities. Yet they are still far from true intelligence, which opens up intriguing opportunities to explore the parallels of humans and model behaviors. In this work, we study the ability to skip steps in reasoning—a hallmark of human expertise developed through practice. Unlike humans, who may skip steps to enhance efficiency or to reduce cognitive load, models do not inherently possess such motivations to minimize reasoning steps. To address this, we introduce a controlled framework that stimulates step-skipping behavior by iteratively refining models to generate shorter and accurate reasoning paths. Empirical results indicate that models can develop the step skipping ability under our guidance. Moreover, after fine-tuning on expanded datasets that include both complete and skipped reasoning sequences, the models can not only resolve tasks with increased efficiency without sacrificing accuracy, but also exhibit comparable and even enhanced generalization capabilities in out-of-domain scenarios. Our work presents the first exploration into human-like step-skipping ability and provides fresh perspectives on how such cognitive abilities can benefit AI models.

Live content is unavailable. Log in and register to view live content