TLDR: A research paper titled “Small is Sufficient: Reducing the World AI Energy Consumption Through Model Selection” proposes that by strategically choosing smaller, more energy-efficient AI models during inference, global AI energy consumption could be reduced by 27.8%. This shift from a “bigger is better” to a “small is sufficient” paradigm, focusing on model selection, offers substantial energy savings (equivalent to the annual output of five nuclear power reactors in 2025) with minimal impact on model performance. The study analyzes various AI tasks, model sizes, and adoption patterns to demonstrate the feasibility and significant environmental benefits of this approach.
The rapid expansion of Artificial Intelligence (AI) has brought about remarkable advancements across various sectors, from medicine to robotics. However, this growth comes with a significant environmental cost: escalating energy consumption and a growing carbon footprint. A recent research paper titled Small is Sufficient: Reducing the World AI Energy Consumption Through Model Selection by Tiago da Silva Barros, Frédéric Giroire, Ramon Aparicio-Pardo, and Joanna Moulierac, explores a promising solution to this challenge: the “small is sufficient” paradigm, which advocates for energy sobriety through intelligent model selection during AI inference.
The Shifting Paradigm: From “Bigger is Better” to “Small is Sufficient”
For years, the AI community has largely operated under a “bigger is better” philosophy, where larger models with billions or even trillions of parameters were pursued for marginal gains in performance. While these massive models have driven impressive capabilities, their computational demands are immense. The researchers highlight a critical shift towards “small is sufficient,” emphasizing the use of smaller, more efficient models that can achieve comparable utility with significantly less energy.
This approach focuses specifically on model selection during the inference phase, which is when AI models are used to make predictions or decisions. Although a single inference task consumes less energy than training a model, the sheer volume of inference requests means it accounts for an estimated 60% of machine learning’s total energy usage. Unlike methods requiring new hardware or complex architectural changes, model selection is a straightforward and immediately applicable strategy for any AI user.
Analyzing AI Tasks and Model Efficiency
To understand the potential for energy savings, the study systematically analyzed popular AI tasks in data centers, examining their model sizes, utility (performance), and adoption patterns. Platforms like Hugging Face and Papers With Code were instrumental in gathering data on thousands of models for tasks such as text generation, image classification, speech recognition, and object detection.
A key observation was the “law of diminishing returns” in AI model utility. As models grow larger, the marginal gains in performance become increasingly smaller. This means that beyond a certain point, a significantly larger model might offer only a tiny improvement in accuracy or capability compared to a much smaller, more energy-efficient alternative. For instance, in speech recognition, the best-performing model might be 14 times larger than an energy-efficient one, yet provide only a 7.8% improvement in utility.
Quantifying Energy Savings Through Model Selection
The researchers developed a methodology to estimate the energy consumption of AI models during inference, using tools like CarbonTracker and validating measurements against hardware-based power meters. They found a consistent relationship: energy consumption scales linearly with model size on a log-log scale, allowing for reliable energy estimations even for models too large to measure directly.
The findings are compelling: applying model selection globally could reduce AI energy consumption by 27.8%. This translates to a saving of 31.9 TWh worldwide in 2025, an amount equivalent to the annual output of five nuclear power reactors. By 2028, these savings could reach 106 TWh, matching the annual production of 17 nuclear power reactors.
These significant energy reductions come with a surprisingly small impact on overall model utility, averaging a loss of only 3.9%. In some cases, switching to an energy-efficient model even led to an improvement in utility, as some widely used models are larger but less performant than more efficient alternatives due to factors like brand popularity or hardware limitations.
The Role of Task Maturity and Model Adoption
The study also highlighted how the maturity of an AI task and user adoption patterns influence potential energy savings. For mature tasks like image classification or speech recognition, extensive research has already led to the development of highly efficient, smaller models that perform nearly as well as their larger counterparts. However, widespread adoption of these efficient models is not always immediate, often due to user habits or the popularity of established, larger models.
The paper projects three future scenarios for AI inference energy consumption: a business-as-usual scenario, a pessimistic scenario where large, best-performing models are universally adopted (leading to a 111% increase in energy consumption), and a sobriety scenario where model selection is widely implemented. The sobriety scenario clearly demonstrates the immense potential for sustainable AI practices.
Also Read:
- Adaptive AI Agents: Bringing Advanced Intelligence to Mobile and Edge Devices
- Adaptive Reasoning Suppression: Making Large Language Models Think Smarter, Not Longer
A Path Towards Sustainable AI
In conclusion, the research underscores that model selection during AI inference is a powerful and practical strategy for achieving energy sobriety. By consciously choosing energy-efficient models that maintain high utility, the AI community can significantly mitigate its environmental impact without compromising performance. This work provides a clear roadmap for reducing AI’s carbon footprint, advocating for responsible and efficient AI use as a critical step in addressing climate change.


