TLDR: EntroPE is a new time series forecasting framework that uses conditional entropy to dynamically identify natural transition points in data, creating variable-length patches. This “temporally informed” patching preserves data coherence, and an adaptive encoder converts these patches into fixed representations for a global Transformer. This approach significantly improves forecasting accuracy and efficiency over traditional fixed-length patching methods.
The field of time series forecasting has seen significant advancements with the introduction of Transformer-based models. These models are particularly effective when using patch-based input strategies, which divide long sequences of data into smaller, fixed-length segments. This approach has improved efficiency and the ability to model long-term patterns.
However, a common limitation in existing patch-based methods is their “temporally-agnostic” nature. This means they construct patches without considering the natural flow or transitions within the time series data. Imagine splitting a gradual trend change or a seasonal shift right in the middle of a patch – this can break the natural connections in the data, making it harder for the model to learn accurate representations. This arbitrary segmentation can disrupt short-term dependencies and lead to inconsistent input representations between training and inference, ultimately weakening the model’s ability to generalize.
To address these challenges, researchers have introduced a novel framework called EntroPE, which stands for Entropy-Guided Dynamic Patch Encoder. EntroPE offers a “temporally informed” approach by dynamically detecting natural transition points in the data and placing patch boundaries accordingly. This method preserves the inherent temporal structure of the time series while still benefiting from the computational advantages of patching.
How EntroPE Works
EntroPE is built around two main components:
-
Entropy-based Dynamic Patcher (EDP): This module uses a concept from information theory called conditional entropy to identify where natural temporal shifts occur. Essentially, it looks for points in the time series where the predictability of the next data point is low (high entropy), indicating a significant change or transition. These high-uncertainty points are then used to define the boundaries of the patches.
-
Adaptive Patch Encoder (APE): Once the variable-length patches are created by the EDP, the APE processes them. It uses pooling and cross-attention mechanisms to capture the relationships within each patch and convert them into fixed-size representations. This is crucial because standard Transformer models typically require fixed-size inputs.
After these patch embeddings are generated, they are fed into a global Transformer, which models the longer-range dependencies between the different patches. Finally, a Fusion Decoder combines the global context from the Transformer with the detailed temporal information from the encoder to produce accurate forecasts.
Also Read:
- Adaptive AI: How Neuroplasticity is Reshaping Multi-Task Demand Forecasting
- Next Event Prediction: Enhancing AI’s Understanding of Patient Journeys in Electronic Health Records
Performance and Benefits
Experiments conducted on various long-term forecasting benchmarks have shown that EntroPE significantly improves both accuracy and efficiency compared to existing methods. For instance, it achieved notable accuracy gains of approximately 20% on the ETTh1 dataset and 15% on the Electricity dataset compared to PatchTST, while also reducing the number of tokens processed. The dynamic patching approach also proved to be robust to threshold selection, which controls the number of patches created.
This research highlights that by respecting the intrinsic temporal structure of time series data through entropy-guided dynamic patching, forecasting models can achieve better performance. The code for EntroPE is publicly available for further exploration. You can find the full research paper here: EntroPE Research Paper.


