TLDR: The Game-Time Benchmark is introduced to evaluate the temporal dynamics of Spoken Language Models (SLMs), focusing on their ability to manage timing, tempo, and simultaneous speaking in real-time conversations. The benchmark includes basic instruction-following tasks and advanced tasks with temporal constraints. Evaluations show that while state-of-the-art SLMs perform well on basic tasks, nearly all models struggle significantly when temporal constraints are introduced, highlighting a critical lack of “time-awareness” and full-duplex interaction capabilities in current systems. The benchmark aims to guide future research towards more temporally-aware conversational AI.
Conversational Spoken Language Models (SLMs) are rapidly advancing, promising more natural and real-time speech interactions. However, a significant challenge remains in their ability to handle the intricate temporal dynamics of human conversation, such as precise timing, consistent tempo, and simultaneous speaking. A new research paper introduces the Game-Time Benchmark, a novel framework designed to systematically evaluate these crucial temporal capabilities in SLMs.
The researchers behind Game-Time highlight that while existing benchmarks focus on content quality, style, and basic turn-taking, they often overlook the critical aspect of temporal fluency. This gap means that even advanced SLMs can struggle with the subtle cues and precise timing that are essential for truly human-like dialogue. The inspiration for Game-Time comes from how humans, particularly children, learn language through activities that require not just understanding words, but also developing a sense of timing and rhythm, like in games such as “rock-paper-scissors.”
The Game-Time Benchmark is structured into two main categories: Basic Tasks and Advanced Tasks. The Basic Tasks assess fundamental instruction-following abilities, such as generating sequences, repeating content, composing responses, recalling information, engaging in open-ended conversation, and role-playing. Surprisingly, even some contemporary SLMs show weaknesses in these foundational areas.
The Advanced Tasks build upon these basics by introducing explicit temporal constraints. These tasks are designed to test an SLM’s “time-awareness” and interactive fluency. They include:
Time Tasks
These evaluate a model’s ability to adjust the overall duration of its speech. This includes completing tasks quickly within a specified duration (Time-Fast), performing tasks slowly over a minimum duration (Time-Slow), and inserting precise silent intervals before responding (Time-Silence).
Tempo Tasks
These probe an SLM’s capacity to maintain rhythmic consistency. Examples include following a specified tempo with a fixed pause between words (Tempo-Interval) or adhering to a tempo demonstrated by the user’s spoken example (Tempo-Adhere).
Also Read:
- Enhancing In-Game Voice Communication Accuracy with a New AI Framework
- VitaBench: A New Standard for Evaluating LLM Agents in Real-World Scenarios
Simultaneous Speaking Tasks (SimulSpeak)
These challenge models to overlap with user speech, requiring real-time listening and synchronization. This involves repeating each word with immediate, word-by-word overlap (Simul-Shadow) or speaking at a designated timing or cue, like in a game of “rock-paper-scissors” (Simul-Cue).
The evaluation protocol for Game-Time leverages an LLM-as-a-judge framework. This involves transcribing dual-channel audio (user and model) to obtain time-aligned text, which is then fed to a powerful LLM (Gemini 2.5 Pro) to assess performance against instruction-following criteria. This method was validated against human evaluations, showing a strong correlation.
The study evaluated various SLMs, including time-multiplexing models like Freeze-Omni and Unmute, dual-channel models like Moshi, and commercial voice agents such as Gemini-Live and GPT-realtime. An oracle system, SSML-LLM, which operates with future knowledge and precise timing control, was used as a theoretical performance ceiling.
The results reveal a clear performance disparity. While state-of-the-art models like GPT-realtime perform strongly on most Basic Tasks, many contemporary academic SLMs still struggle with fundamental instruction-following. More critically, nearly all models show a substantial drop in performance when temporal constraints are introduced. Models particularly struggle with tasks requiring precise time awareness, such as inserting silent intervals, adhering to specific tempos, and synchronizing speech with users. This indicates a persistent lack of “time-awareness” and full-duplex interaction capabilities in current SLMs, even in the most advanced systems.
The Game-Time Benchmark provides a crucial foundation for guiding future research toward developing more temporally-aware conversational AI. It shifts the focus from merely “what to say” to the equally important “when to say it,” paving the way for SLMs that can engage in truly fluid and natural human-like conversations. You can find more details about this research in the paper: GAME-TIME: EVALUATING TEMPORAL DYNAMICS IN SPOKEN LANGUAGE MODELS.


