TLDR: A new study identifies “Format Inertia,” a failure mechanism in Large Language Models (LLMs) used for medical pre-consultation. This occurs when LLMs, trained on datasets with a skewed distribution of short dialogues, generate repetitive and uninformative questions in longer conversations. The researchers propose a data-centric solution: creating a “Uniform Turn-Count Dataset” that balances dialogue lengths during training, significantly mitigating Format Inertia and improving diagnostic utility.
Large Language Models (LLMs) have made incredible strides, transforming various service domains, including medical pre-consultation. These AI systems are increasingly being adapted to assist in healthcare, particularly in generating multi-turn dialogues between patients and virtual doctors. However, a recent study by Seungseop Lim, Gibaeg Kim, Wooseok Han, Jean Seo, Hyunkyung Lee, Jaehyo Yoo, and Eunho Yang, titled “Format Inertia: A Failure Mechanism of LLMs in Medical Pre-Consultation,” sheds light on a critical challenge faced by these models.
The common method for training LLMs for multi-turn dialogue in healthcare is Supervised Fine-Tuning (SFT). While effective, the datasets used for SFT often have a skewed distribution of conversation lengths, with short dialogues being far more common than long ones. This imbalance, the researchers found, leads to a novel failure mechanism they call “Format Inertia.”
What is Format Inertia?
Imagine an AI doctor asking a patient a series of questions. In a long consultation, if the AI exhibits Format Inertia, it will start generating repetitive, format-correct, but diagnostically uninformative questions. It’s like the AI gets stuck in a loop, asking variations of the same question even after receiving answers. This happens because the model, having been trained mostly on short dialogues, lacks sufficient exposure to the complex contextual dependencies required for longer, more in-depth interviews. When faced with the uncertainty of a long dialogue, it defaults to familiar, safe question patterns.
This isn’t just a minor glitch; it significantly impacts the user experience. Patients can become confused and frustrated when asked redundant questions, undermining the overall effectiveness of the pre-consultation. More importantly, it stalls clinical progress, as the AI fails to gather new, crucial diagnostic information.
The Solution: A Data-Centric Approach
To combat Format Inertia, the researchers adopted a straightforward, data-centric method: rebalancing the turn-count distribution of the training dataset. They created a “Uniform Turn-Count Dataset” by ensuring an equal number of dialogues across different maximum turn-count bins. This means the model is exposed to a balanced mix of short and long conversations during training.
The process involves grouping dialogues by their maximum turn-count, determining a sampling quota based on the smallest bin, and then uniformly sampling dialogues from each bin. This balanced exposure helps the model develop more robust strategies for handling a wide range of consultation lengths, from minor conditions to complex history-taking.
Experimental Validation
Experiments using real-world medical pre-consultation dialogues demonstrated the effectiveness of this approach. Models fine-tuned on skewed datasets showed high adherence to format but a significant drop in their ability to ask clinically meaningful questions (Task-Constraint Satisfaction Rate, TCSR). Interestingly, increasing the volume of skewed data actually worsened this degradation.
In contrast, models trained on the Uniform Turn-Count Dataset showed substantial alleviation of Format Inertia. They were able to generate responses that were both formally correct and clinically meaningful, even with a smaller dataset size compared to the larger, skewed dataset. This highlights that for medical pre-consultation, the quality and distributional balance of data are more critical than sheer quantity.
The study also quantified Format Inertia by measuring the lexical and semantic similarity of generated questions. Models trained on skewed data showed a progressive increase in question similarity across dialogue turns, confirming their tendency to produce redundant questions. This inverse relationship between a turn length’s rarity in training data and its failure rate in evaluation further solidified the hypothesis.
Also Read:
- Bridging Language and Sensors: LLMs for Proactive Industrial Health Monitoring
- Beyond Stereotypes: Verbalized Sampling Unlocks LLM Diversity by Tackling Hidden Data Bias
Conclusion
The research underscores the critical role of data distribution, particularly turn-count distribution, in the robustness of multi-turn conversational AI systems. By identifying and mitigating Format Inertia through a simple data rebalancing strategy, this study paves the way for more reliable and effective LLM-based medical pre-consultation services. For more details, you can read the full paper here.


