TY - JOUR
T1 - PRformer
T2 - Pyramidal recurrent transformer for multivariate time series forecasting
AU - Yu, Yongbo
AU - Yu, Weizhong
AU - Nie, Feiping
AU - Miao, Zongcheng
AU - Liu, Ya
AU - Li, Xuelong
N1 - Publisher Copyright:
© 2025 Elsevier Ltd
PY - 2025/11
Y1 - 2025/11
N2 - The self-attention mechanism in Transformer architecture, invariant to sequence order, necessitates positional embeddings to encode temporal order in time series prediction. We argue that this reliance on positional embeddings restricts the Transformer's ability to effectively represent temporal sequences, particularly when employing longer lookback windows. To address this, we introduce an innovative approach that combines Pyramid RNN embeddings (PRE) for univariate time series with the Transformer's capability to model multivariate dependencies. PRE, utilizing pyramidal one-dimensional convolutional layers, constructs multiscale convolutional features that preserve temporal order. Additionally, RNNs, layered atop these features, learn multiscale time series representations sensitive to sequence order. This integration into Transformer models with attention mechanisms results in significant performance enhancements. We present the PRformer, a model integrating PRE with a standard Transformer encoder, demonstrating state-of-the-art performance on various real-world datasets. This performance highlights the effectiveness of our approach in leveraging longer lookback windows and underscores the critical role of robust temporal representations in maximizing Transformer's potential for prediction tasks.
AB - The self-attention mechanism in Transformer architecture, invariant to sequence order, necessitates positional embeddings to encode temporal order in time series prediction. We argue that this reliance on positional embeddings restricts the Transformer's ability to effectively represent temporal sequences, particularly when employing longer lookback windows. To address this, we introduce an innovative approach that combines Pyramid RNN embeddings (PRE) for univariate time series with the Transformer's capability to model multivariate dependencies. PRE, utilizing pyramidal one-dimensional convolutional layers, constructs multiscale convolutional features that preserve temporal order. Additionally, RNNs, layered atop these features, learn multiscale time series representations sensitive to sequence order. This integration into Transformer models with attention mechanisms results in significant performance enhancements. We present the PRformer, a model integrating PRE with a standard Transformer encoder, demonstrating state-of-the-art performance on various real-world datasets. This performance highlights the effectiveness of our approach in leveraging longer lookback windows and underscores the critical role of robust temporal representations in maximizing Transformer's potential for prediction tasks.
KW - Multiscale representation learning
KW - Pyramidal recurrent transformer
KW - Time series forecasting
UR - http://www.scopus.com/inward/record.url?scp=105009163037&partnerID=8YFLogxK
U2 - 10.1016/j.neunet.2025.107769
DO - 10.1016/j.neunet.2025.107769
M3 - 文章
AN - SCOPUS:105009163037
SN - 0893-6080
VL - 191
JO - Neural Networks
JF - Neural Networks
M1 - 107769
ER -