Implicit Chain of Thought in Large Language Models
Published in Research Collaboration, 2024
This ongoing research project explores training language models that can reason internally using hidden states instead of articulating all reasoning steps explicitly, similar to human internal thought processes.
Project Overview
Objective: Develop language models capable of implicit reasoning without explicit step-by-step verbalization
Methodology:
- Training LMs to reason using internal hidden states
- Focus on simultaneous 2-digit multiplication tasks
- Knowledge distillation from explicit reasoning models
- Evaluation on mathematical reasoning benchmarks
Key Research Questions:
- Can language models learn to reason without explicitly verbalizing each step?
- How do implicit reasoning mechanisms compare to explicit Chain of Thought?
- What are the advantages of internal reasoning for complex tasks?
Current Work:
- Working on simultaneous 2-digit multiplication using Implicit Chain of Thought (ICoT)
- Exploring how models can develop internal reasoning mechanisms
- Analyzing the efficiency and accuracy trade-offs
Significance:
- More efficient reasoning without verbose explanations
- Better understanding of how LLMs process complex reasoning
- Potential for faster inference in production systems
- Insights into developing more human-like AI reasoning
Supervisor: Prof. Yuntian Deng
Institution: Deng’s Research Group, University of Waterloo
Duration: Fall 2024 - Present
Status: Ongoing Research
Recommended citation: Deng, Y., et al. (2024). "Implicit Chain of Thought Reasoning via Knowledge Distillation." arXiv preprint.
Download Paper
