系列(地层学)
时间序列
计算机科学
建筑
人工智能
机器学习
地理
地质学
古生物学
考古
作者
Yuan Feng,Kai Xia,Xiaoyu Qu,Xuwei Hu,Long Sun,Zilong Zhang
标识
DOI:10.1109/ijcnn60899.2024.10651438
摘要
In real-life scenarios, long sequence time series forecasting (LSTF) has broad applications, such as electricity demand planning and abnormal weather prediction. In recent years, transformer-based models have achieved significant advancements in LSTF tasks, as they excel at capturing long-term dependencies in time series data. Long-term forecasting capability is particularly crucial in LSTF tasks. However, transformer-based models suffer from high computational costs, extensive memory usage, and limitations of the inherent encoder-decoder architecture, making them inefficient and challenging to directly apply in real-world scenarios, especially for long-period forecasting. Moreover, recent research has indicated that cross-attention mechanisms may cause temporal information loss, and sparse attention mechanisms can create information utilization bottlenecks. To address these issues, this paper proposes a novel architecture named "Halveformer," which combines linear models with the encoder to enhance both model performance and efficiency. Through experimental research on seven benchmark datasets, we demonstrate that Halveformer significantly outperforms existing advanced methods. We hope this innovative concept can pave the way for new research directions in LTSF tasks and prompt a reevaluation of the effectiveness of solutions based on the inherent encoder-decoder architecture.
科研通智能强力驱动
Strongly Powered by AbleSci AI