TLDR: Adaptive Query Reasoning (AdaQR) is a new framework that combines a fast Dense Reasoner with a powerful LLM Reasoner for complex search queries. A Reasoner Router dynamically directs queries to the most efficient and effective reasoning pathway, significantly reducing computational costs (by 28%) while maintaining or improving retrieval performance (by 7%) on reasoning-intensive benchmarks like BRIGHT. This hybrid approach offers a practical solution for balancing accuracy and efficiency in information retrieval.
In the rapidly evolving world of artificial intelligence, efficient and accurate information retrieval is paramount. Traditional search methods often fall short when faced with complex, reasoning-intensive queries. While advanced Large Language Models (LLMs) have shown remarkable capabilities in understanding and reformulating such queries, their computational cost can be a significant hurdle for widespread application.
A new research paper, titled “YOURDENSERETRIEVER ISSECRETLY AN EXPEDITIOUSREASONER,” introduces an innovative solution called Adaptive Query Reasoning (AdaQR). Authored by Yichi Zhang, Jun Bai, Zhixin Cai, Shuhan Qin, Zhuofan Chen, Jinghua Guan, Wenge Rong, and Jun Bai, this framework aims to strike a balance between the powerful reasoning of LLMs and the need for computational efficiency in retrieval systems.
At its core, AdaQR is a hybrid query rewriting framework designed to dynamically choose the most appropriate reasoning method for each query. It comprises three main components: an LLM Reasoner, a Dense Reasoner, and a Reasoner Router.
The LLM Reasoner leverages the deep understanding and reasoning capabilities of large language models to reformulate queries, making them more effective for complex searches. However, as mentioned, this comes at a high computational cost.
To address this, the researchers propose the Dense Reasoner. This ingenious component learns to mimic the reasoning style of an LLM directly within the embedding space – the mathematical representation of queries and documents. By doing so, it can perform LLM-style reasoning at a negligible cost, offering an extremely fast alternative for query rewriting. The Dense Reasoner is trained in two stages: first, on a large dataset to learn general reasoning patterns, and then fine-tuned on specific datasets to adapt to in-domain distributions.
The crucial element that ties these two reasoners together is the Reasoner Router. This intelligent mechanism acts as a gatekeeper, evaluating each incoming query. It determines whether a query can be reliably handled by the fast and efficient Dense Reasoner or if it requires the deeper, more resource-intensive reasoning of an LLM. This dynamic routing allows AdaQR to achieve a controllable trade-off between efficiency and accuracy, ensuring that complex queries receive the necessary attention without incurring unnecessary costs for simpler ones.
Experiments conducted on the BRIGHT benchmark, a challenging dataset for reasoning-intensive retrieval, demonstrated impressive results. AdaQR not only preserved, and in many cases improved, retrieval performance compared to using full LLM rewrites, but it also significantly reduced reasoning costs. On average, AdaQR achieved a 7% improvement in retrieval performance while cutting rewriting costs by an average of 28%.
The study also revealed interesting insights into the performance across different types of queries. While LLM Reasoners generally excelled in domains like StackExchange, which often involve distinctive linguistic styles, the Dense Reasoner showed superior performance in highly structured domains such as coding (e.g., LeetCode) and theorem-based tasks. This highlights the complementary strengths of both reasoning approaches within the AdaQR framework.
An ablation study confirmed the vital roles of both the Dense Reasoner and the Reasoner Router. Removing either component led to a noticeable drop in performance, underscoring their combined contribution to AdaQR’s effectiveness and efficiency. The decision threshold within the Reasoner Router also proved critical, allowing for fine-tuning the balance between cost and performance.
Also Read:
- FAIR-RAG: Enhancing LLM Accuracy with Evidence-Driven Iterative Refinement
- CustomIR: Enhancing Information Retrieval for Specialized Datasets with Unsupervised Fine-Tuning
This work represents a significant step forward in making advanced reasoning capabilities more accessible and practical for real-world information retrieval systems. By adaptively leveraging the strengths of both dense and large language models, AdaQR paves the way for smarter, faster, and more cost-effective search experiences. For more details, you can read the full research paper here.


