TLDR: TransLLM is a new AI framework that combines traditional spatiotemporal modeling with large language models to better manage urban transportation tasks like traffic forecasting, EV charging prediction, and taxi dispatch. It uses a unique “learnable prompting” system to adapt to different situations and performs well even with limited data, demonstrating strong generalization across various scenarios.
Urban transportation systems are the backbone of modern cities, but they face constant challenges like traffic congestion, predicting electric vehicle (EV) charging needs, and efficiently dispatching taxis. Traditionally, these problems have been tackled with separate, specialized deep learning models. While effective for specific tasks, these models often require a lot of data and struggle to adapt to new scenarios or integrate different types of information.
On the other hand, large language models (LLMs) like GPT-4 and Gemini offer great flexibility through their natural language understanding. However, they typically struggle with the structured, numerical, and time-sensitive data common in transportation. Imagine trying to get a language model to accurately predict traffic flow numbers – it’s not what they’re built for.
To bridge this gap, researchers have introduced TransLLM, a unified foundation framework designed to integrate spatiotemporal modeling with large language models. This innovative approach aims to overcome the limitations of both traditional small-scale models and generalist LLMs in the urban transportation domain. TransLLM is built to handle diverse challenges, from predicting traffic flow and electric vehicle charging demand to optimizing taxi dispatching.
How TransLLM Works
TransLLM operates through a clever three-stage process. First, it uses a lightweight spatiotemporal encoder. Think of this as a specialized “translator” that takes complex, structured transportation data – like traffic sensor readings or geographic proximity – and converts it into a format that the LLM can understand. This encoder is designed to capture intricate dependencies across space and time, using techniques like dilated temporal convolutions and dual-adjacency graph attention networks.
Next, a novel “learnable prompt routing” mechanism comes into play. Instead of using fixed, pre-written instructions for the LLM, TransLLM dynamically creates personalized prompts for each specific situation. This mechanism is trained using reinforcement learning, allowing it to adapt and select the most suitable prompt based on the unique characteristics of the input data. This means the system can tailor its approach whether it’s dealing with a rush hour traffic prediction or a quiet period for EV charging.
Finally, after the LLM processes the information guided by the personalized prompt, TransLLM doesn’t let the LLM generate the final output directly in natural language. Instead, it projects the LLM’s reasoning through specialized output layers. This is crucial for handling continuous numerical values accurately and ensuring the predictions align with the specific requirements and evaluation metrics of each transportation task, avoiding the precision loss that can occur when LLMs try to generate numbers as text.
Key Features and Benefits
One of TransLLM’s significant contributions is its ability to handle multiple, diverse transportation tasks within a single framework. It supports both forecasting problems, like predicting future traffic or charging demand, and optimization problems, such as efficient taxi dispatching. The framework’s design allows it to learn shared knowledge across different transportation domains, making it highly adaptable.
The spatiotemporal encoder is task-agnostic, meaning it can model complex dependencies regardless of the specific transportation problem. The learnable prompt routing mechanism is a game-changer, moving beyond rigid templates to provide dynamic, instance-specific guidance to the LLM. This adaptability is key to its strong performance across varied scenarios.
Also Read:
- Optimizing Urban Flood Response with an AI-Powered Multi-Agent System
- STPFormer: A New Approach to Smarter Traffic Prediction
Performance and Generalization
Experiments conducted on seven different datasets covering traffic forecasting, charging demand prediction, and taxi dispatch optimization have shown TransLLM’s exceptional effectiveness. It consistently outperforms ten baseline models, including traditional GNN-based deep learning models and even generalist LLMs like Deepseek-v3 and GPT-4o, in both supervised and zero-shot settings.
A particularly impressive aspect is its generalization ability. TransLLM demonstrates strong performance even when faced with unseen scenarios and datasets, such as predicting traffic in new cities without additional fine-tuning. This “zero-shot” capability is vital for practical deployment, as it significantly reduces the cost and effort of adapting the model to new environments.
For more technical details, you can refer to the full research paper available at arXiv.org.
In conclusion, TransLLM represents a significant step forward in urban transportation management. By intelligently combining the strengths of spatiotemporal modeling with the flexibility of large language models, it offers a unified, adaptable, and highly effective solution for the complex challenges facing our cities’ lifelines.


