TLDR: This research paper systematically revisits model interpolation, a simple method for merging AI models, and discovers a predictable three-stage evolutionary paradigm when blending ‘Instruct’ (short answers) and ‘Thinking’ (long reasoning) models. This framework provides a principled guide for balancing performance and computational cost. Empirical results show that a strategically interpolated model surprisingly outperforms more sophisticated model merging baselines on various reasoning benchmarks. Extensive ablation studies offer deep mechanistic insights, highlighting the roles of different model layers and modules in shaping reasoning capabilities. The work offers a practical framework for crafting models with precisely targeted reasoning behaviors.
Large Language Models (LLMs) have transformed natural language processing, showcasing impressive reasoning abilities, especially with techniques like Chain-of-Thought (CoT). However, these advanced reasoning methods often come with a trade-off: they can lead to ‘over-thinking’ and increased latency, making efficient reasoning a significant challenge.
To tackle this, a method called model merging has gained traction. This involves combining the strengths of two specialized models: an ‘Instruct’ model, which is optimized for quick, direct answers, and a ‘Thinking’ model, which excels at detailed, long-form reasoning. The goal is to create a hybrid model that balances both reasoning power and token efficiency.
This research paper, titled “Revisiting Model Interpolation for Efficient Reasoning” by Taiqiang Wu, Runming Yang, Tao Liu, Jiahao Wang, and Ngai Wong, delves into the simplest form of model merging: direct model interpolation. This method involves directly blending the weights of two models. The study uncovers a surprising finding: the performance of model interpolation doesn’t change linearly but follows a distinct three-stage evolutionary paradigm.
The Three-Stage Evolutionary Paradigm
The researchers observed that as the interpolation coefficient (λ) shifts from 0 (pure Instruct model) to 1 (pure Thinking model), the model’s behavior evolves through three predictable stages:
1. Stage #1 (Instruct Dominant): In this initial phase, the merged model is primarily influenced by the Instruct model. It starts generating longer outputs, and its basic performance (Pass@k) gradually improves. However, there’s almost no explicit reasoning (Think #R remains near zero), meaning it still favors direct answers without showing its thought process.
2. Stage #2 (Reasoning Emergence): This is a critical transition stage. The explicit reasoning pattern, characteristic of Thinking models, rapidly emerges. The ‘Think Ratio’ (Think #R) dramatically increases, indicating the model is now showing its step-by-step thought process. During this stage, the quality of reasoning (Mean@k) improves significantly, often faster than basic performance (Pass@k), which typically reaches its peak here. This stage often represents a ‘sweet spot’ for balancing effectiveness and efficiency.
3. Stage #3 (Thinking Dominant & Overthinking): In the final stage, the merged model closely resembles the pure Thinking model. The output responses become substantially longer, and the ‘Think Ratio’ saturates at 1.0. While reasoning quality (Mean@k) might see slight improvements, the gains in basic performance (Pass@k) diminish or even decline. This phenomenon is termed ‘overthinking,’ where longer reasoning doesn’t necessarily lead to better results. Interestingly, the interpolated model can sometimes outperform the pure Thinking model in this stage, suggesting that a slight blend with the Instruct model can regularize the reasoning process.
Also Read:
- Dynamic Temperature Control Enhances LLM Reasoning in Reinforcement Learning
- Streamlining LLM Reasoning: Introducing Upfront Chain-of-Thought Compression
Empirical Superiority and Mechanistic Insights
The study conducted extensive experiments using Qwen3 models, interpolating between their official Instruct and Thinking variants. A strategically interpolated model consistently surpassed more sophisticated model merging baselines across challenging benchmarks, including mathematical reasoning (AIME’25), instruction-following (IFEval), and scientific reasoning (GPQA-Diamond). This demonstrates that a simple interpolation can be more effective and efficient than complex merging techniques.
Further ablation studies provided deep insights into how interpolation works:
- Decoding Strategy: The performance of the interpolated models was found to be remarkably robust to changes in decoding hyperparameters like temperature and Top-p.
- Model Layers: The complex reasoning patterns of the Thinking model are predominantly stored in the middle and later layers of the neural network. Interpolating these specific layers was crucial for inducing thinking behavior.
- Transformer Modules: The Feed-Forward Network (FFN) modules from the Thinking model were identified as the primary drivers for generating long Chain-of-Thought reasoning patterns. Multi-Head Attention (MHA) modules, while not driving the pattern, were crucial for the quality and correctness of the reasoning itself.
- Alternative Backbones: The research also explored interpolating with other base models. It found that instruction-following alignment is vital for generating high-quality reasoning on complex problems.
In conclusion, this work demystifies model interpolation, revealing its predictable three-stage evolution. It offers a practical and interpretable framework for creating AI models with precisely targeted reasoning capabilities, providing a principled guide for navigating the performance-cost trade-off in efficient reasoning. For more details, you can read the full paper here.


