SVSeq2Seq: An Efficient Computational Method for State Vectors in Sequence-to-Sequence Architecture Forecasting
-
Published:2024-01-13
Issue:2
Volume:12
Page:265
-
ISSN:2227-7390
-
Container-title:Mathematics
-
language:en
-
Short-container-title:Mathematics
Author:
Sun Guoqiang1ORCID, Qi Xiaoyan2, Zhao Qiang1ORCID, Wang Wei1, Li Yujun1
Affiliation:
1. School of Information Science and Engineering, Shandong University, Qingdao 266237, China 2. State Key Laboratory of Microbial Technology, Shandong University, Qingdao 266237, China
Abstract
This study proposes an efficient method for computing State Vectors in Sequence-to-Sequence (SVSeq2Seq) architecture to improve the performance of sequence data forecasting, which associates each element with other elements instead of relying only on nearby elements. First, the dependency between two elements is adaptively captured by calculating the relative importance between hidden layers. Second, tensor train decomposition is used to address the issue of dimensionality catastrophe. Third, we further select seven instantiated baseline models for data prediction and compare them with our proposed model on six real-world datasets. The results show that the Mean Square Error (MSE) and Mean Absolute Error (MAE) of our SVSeq2Seq model exhibit significant advantages over the other seven baseline models in predicting the three datasets, i.e., weather, electricity, and PEMS, with MSE/MAE values as low as 0.259/0.260, 0.186/0.285 and 0.113/0.222, respectively. Furthermore, the ablation study demonstrates that the SVSeq2Seq model possesses distinct advantages in sequential forecasting tasks. It is observed that replacing SVSeq2Seq with LPRcode and NMTcode resulted in an increase under an MSE of 18.05 and 10.11 times, and an increase under an MAE of 16.54 and 9.8 times, respectively. In comparative experiments with support vector machines (SVM) and random forest (RF), the performance of the SVSeq2Seq model is improved by 56.88 times in the weather dataset and 73.78 times in the electricity dataset under the MSE metric, respectively. The above experimental results demonstrate both the exceptional rationality and versatility of the SVSeq2Seq model for data forecasting.
Funder
National Key Research and Development Program of China
Reference28 articles.
1. A context-independent ontological linked data alignment approach to instance matching;Barbosa;Int. J. Semant. Web. Inf.,2022 2. Long short-term memory;Hochreiter;Neural Comput.,2010 3. Cho, K., Merrienboer, B.V., Gulcehre, C., Bahdanau, D., Bougares, F., Schwenk, H., and Bengio, Y. (2014, January 25–29). Learning phrase representations using RNN encoder-decoder for statistical machine translation. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing, Doha, Qatar. 4. Wu, H., Xu, J., Wang, J., and Long, M. (2021, January 6–14). Autoformer: Decomposition transformers with auto-correlation for long-term series forecasting. Proceedings of the 35th Conference on Neural Information Processing Systems, Online. 5. Zhang, Y., and Yan, J. (2023, January 1–5). Crossformer: Transformer utilizing cross-dimension dependency for multivariate time series forecasting. Proceedings of the Eleventh International Conference on Learning Representations, Kigali, Rwanda.
|
|