TLDR: EvoTest is a new framework that allows AI agents to learn and improve continuously during test time without traditional fine-tuning. It uses an “Evolver Agent” to analyze game transcripts and evolve the entire agent’s configuration—including its strategy, memory, and tool use—after each attempt. This approach significantly outperforms existing methods on a new benchmark called J-TTL, enabling agents to adapt and win complex text-based games.
A significant challenge for artificial intelligence today is the inability of AI agents to learn and adapt quickly in new situations. Often, these agents are deployed with a fixed set of instructions, performing like highly capable but inexperienced assistants who can follow orders but struggle to improve their own methods from experience. This limitation severely impacts their usefulness in dynamic, real-world scenarios.
To tackle this, researchers have introduced a new evaluation framework called the Jericho Test-Time Learning (J-TTL) benchmark. This benchmark is designed to measure how well an agent can improve its performance over several consecutive attempts at the same complex, text-based adventure game. The goal is for the agent to learn and get better from one episode to the next, using only the experience gained within that single session.
Initial findings on the J-TTL benchmark revealed that existing adaptation methods, such as those relying on reflection, memory, or traditional reinforcement learning, often fall short. These methods either don’t fundamentally change the agent’s core decision-making logic or are too slow and data-intensive for rapid, in-session improvement.
Introducing EvoTest: A New Paradigm for Self-Improving Agents
To overcome these hurdles, a novel framework called EvoTest has been developed. EvoTest stands for Evolutionary Test-Time Learning, and it offers a unique approach to improving an agent without the need for traditional fine-tuning or complex gradient calculations. Instead, EvoTest evolves the entire agentic system after every episode of the game.
The EvoTest framework operates with two distinct roles:
- The Actor Agent: This agent is responsible for playing the game, interacting with the environment, and attempting to achieve the game’s objectives.
- The Evolver Agent: After each game episode, the Evolver Agent steps in. It meticulously analyzes the entire transcript of the episode, looking for successes, failures, and patterns. Based on this analysis, it proposes a revised configuration for the Actor Agent’s next run.
This revised configuration is comprehensive, encompassing several key aspects of the agent’s operation:
- Rewriting the Prompt: The Evolver can update the guiding instructions or ‘prompt’ that dictates the Actor Agent’s high-level strategy.
- Updating Memory: It logs effective state-action choices into a structured memory, allowing the agent to recall successful actions and avoid known pitfalls.
- Tuning Hyperparameters: The Evolver can adjust decision-making parameters, such as the ‘temperature’ (which influences how exploratory or conservative the agent is).
- Learning Tool-Use Routines: It refines how the agent uses its internal tools, including how it accesses memory or processes game information.
Also Read:
- Empowering AI Agents Through Dynamic Environment-Based Learning
- A New Standard for Assessing AI Agent Performance
Why EvoTest Excels
EvoTest’s strength lies in its holistic, whole-system evolution. Unlike methods that only tweak one part of the agent, EvoTest concurrently optimizes multiple components. This allows it to identify and resolve complex performance bottlenecks that single-channel adaptations cannot. For instance, it can learn to increase exploration in early episodes while simultaneously adding a new strategic rule to its prompt based on its discoveries.
On the J-TTL benchmark, EvoTest consistently showed significant performance increases. It outperformed not only reflection and memory-only baselines but also more complex online fine-tuning methods. Notably, EvoTest was the only method capable of winning two specific games (Detective and Library), while all other baselines failed to win any.
Furthermore, EvoTest addresses a fundamental challenge in test-time learning: data scarcity. Traditional reinforcement learning often struggles with sparse rewards in complex environments like Jericho. EvoTest bypasses this by leveraging the entire episode transcript as a rich, narrative feedback signal. The Evolver Agent performs ‘credit assignment’ through semantic analysis of the game’s story, identifying causal chains of failure and success, and making explicit, targeted edits to the agent’s configuration. This makes it far more data-efficient than methods relying on scalar rewards.
The practical implications are also significant. While traditional online reinforcement learning methods require substantial hardware and can take 5-10 minutes for a single learning update, EvoTest’s learning step is much faster, typically taking only 20-30 seconds via a single API call to a large language model. This makes it a more practical solution for real-time adaptation.
This research marks a concrete step towards building truly autonomous AI agents that can learn and self-improve continuously from their own experiences. For more details, you can refer to the full research paper here.


