- Authors: Xilin Wei, Xiaoran Liu, Yuhang Zang, Xiaoyi Dong, Yuhang Cao, Jiaqi Wang, Xipeng Qiu, Dahua Lin
- Institutes: Fudan University; Shanghai AI Laboratory; The Chinese University of Hong Kong; Shanghai Innovation Institute;
- Resources: [📖Paper] [🏠Project Page] [🤗Huggingface]
-
🔥 Latent Instability in Implicit CoT: We systematically analyze the limitations of implicit Chain-of-Thought methods and reveal a latent instability issue—as the number of implicit tokens increases, models tend to collapse into homogeneous latent states that lose operator semantics.
-
🔥 Step-Level Supervision with SIM-CoT: We propose Supervised IMplicit-CoT (SIM-CoT), a plug-and-play module that introduces step-level supervision via an auxiliary decoder. This stabilizes optimization, prevents collapse, and ensures that latent tokens capture meaningful reasoning steps.
-
🔥 Strong and Consistent Performance: SIM-CoT consistently outperforms both explicit and implicit baselines. On GPT-2, it exceeds supervised CoT by +2.1%, Coconut by +8.2%, and CODI by +4.3%. Across larger LLaMA models (1B/3B/8B), it delivers +1.5% to +9.0% gains, and remains stable even with 8–16 implicit tokens, where prior methods collapse.
-
🔥 Efficiency and Interpretability: SIM-CoT adds no extra inference cost since the auxiliary decoder is discarded after training. It also provides interpretability, allowing each latent token to be decoded into a human-readable reasoning step.
[2025/9/24] Code and Paper are released!
- Code Release
- Checkpoint Release
- Usage Instructions Release
git clone https://github.com/InternLM/SIM-CoT.git
cd SIM-CoT
pip install -r requirements.txt
cd Coconut
torchrun --nnodes 1 --nproc_per_node 8 run.py args/gsm_coconut.yaml
Select a checkpoint that has been expanded to predefined implicit tokens, then continue training with SIM-CoT:
torchrun --nnodes 1 --nproc_per_node 8 run.py args/gsm_simcot.yaml
torchrun --nnodes 1 --nproc_per_node 8 run.py args/gsm_simcot_eval.yaml
cd CODI
bash scripts/train_llama3b_gsm8k-aug-decoder-2.sh
bash CODI/scripts/test_llama3b-copy.sh
If you find our work helpful for your research, please consider giving a star ⭐ and citation 📝
@article{wei2025simcot,
title={{SIM-COT}: Supervised Implicit Chain-of-Thought},
author={Wei, Xilin and Liu, Xiaoran and Zang, Yuhang and Dong, Xiaoyi and Cao, Yuhang and Wang, Jiaqi and Qiu, Xipeng and Lin, Dahua},
journal={arXiv preprint arXiv:2509.20317},
year={2025}
}
- Coconut: The codebase we built upon. Thanks for their wonderful work.
- CODI: Our work is based on this codebase; we are grateful for their valuable contribution.
- LLaMA series: The amazing open-sourced large language model!
- GPT2: An impressive open-source large language model!