Affiliation:
1. Bidhan Chandra College, Rishra, India
Abstract
Historical weather data shows time-sensitive patterns and exhibits seasonality and trends. Machine learning (ML) models can learn to identify these patterns and predict future events. Weather forecasting is a popular time series problem. Historically, RNNs like long short-term memory and gated recurrent unit addressed weather prediction. Yet, the vanishing gradient (VG) problem hindered long-term context connections in extended input sequences. The transformer addresses the VG problem and was initially designed for the natural language processing tasks, where it has proven highly successful. However, there have been few works to unfold the effectiveness of the transformer in regression-based problems. This study introduces a transformer encoder-based ML framework for weather prediction. Analysing historical data and augmenting the dataset with essential temporal information, the model enhances accuracy in forecasting air temperature one step ahead. The results of the experiments are compared to modern RNN-based models and with other similar works previously done in the same field.
Reference42 articles.
1. A transformer neural network for predicting
near‐surface
temperature
2. Autoregressive integrated moving average. (n.d.). In Wikipedia. https://en.wikipedia.org/wiki/Autoregressive_integrated_moving_average
3. Bahdanau, D., Cho, K., & Bengio, Y. (2014). Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473.
4. Gradient amplification: An efficient way to train deep neural networks
5. Bayesian Optimization Tuner. (n.d.). Keras 3 API documentation, Keras Tuner, Tuners. https://keras.io/api/keras_tuner/tuners/bayesian/