Your browser doesn't support javascript.
loading
A Joint Time-Frequency Domain Transformer for multivariate time series forecasting.
Chen, Yushu; Liu, Shengzhuo; Yang, Jinzhe; Jing, Hao; Zhao, Wenlai; Yang, Guangwen.
  • Chen Y; Department of Computer Science and Technology, Tsinghua University, RM.3-126, FIT Building, Haidian District, Beijing, 100084, China.
  • Liu S; College of Computer Science and Mathematics, Fujian University of Technology, RM.213, Building C4, Fuzhou, Fujian, 350118, China.
  • Yang J; Techorigin, No. 581, Jianzhu West Road, Binhu District, Wuxi, Jiangsu, 214000, China.
  • Jing H; Earth System Modeling and Prediction Center, No. 46, Zhongguancun South Street, Haidian District, Beijing, 100081, China.
  • Zhao W; Department of Computer Science and Technology, Tsinghua University, RM.3-126, FIT Building, Haidian District, Beijing, 100084, China.
  • Yang G; Department of Computer Science and Technology, Tsinghua University, RM.3-126, FIT Building, Haidian District, Beijing, 100084, China. Electronic address: ygw@tsinghua.edu.cn.
Neural Netw ; 176: 106334, 2024 Aug.
Article en En | MEDLINE | ID: mdl-38688070
ABSTRACT
In order to enhance the performance of Transformer models for long-term multivariate forecasting while minimizing computational demands, this paper introduces the Joint Time-Frequency Domain Transformer (JTFT). JTFT combines time and frequency domain representations to make predictions. The frequency domain representation efficiently extracts multi-scale dependencies while maintaining sparsity by utilizing a small number of learnable frequencies. Simultaneously, the time domain (TD) representation is derived from a fixed number of the most recent data points, strengthening the modeling of local relationships and mitigating the effects of non-stationarity. Importantly, the length of the representation remains independent of the input sequence length, enabling JTFT to achieve linear computational complexity. Furthermore, a low-rank attention layer is proposed to efficiently capture cross-dimensional dependencies, thus preventing performance degradation resulting from the entanglement of temporal and channel-wise modeling. Experimental results on eight real-world datasets demonstrate that JTFT outperforms state-of-the-art baselines in predictive performance.
Asunto(s)
Palabras clave

Texto completo: 1 Banco de datos: MEDLINE Asunto principal: Predicción Límite: Humans Idioma: En Año: 2024 Tipo del documento: Article

Texto completo: 1 Banco de datos: MEDLINE Asunto principal: Predicción Límite: Humans Idioma: En Año: 2024 Tipo del documento: Article