The National Oceanic and Atmospheric Administration reports a 95% decline in the oldest Arctic ice over the last 33 years [1], while the National Aeronautics and Space Administration states that summer Arctic Sea Ice Extent (SIE) is shrinking by 12.2% per decade since 1979 due to warmer temperatures [2]. Given the rapidly changing Arctic conditions, accurate prediction models are crucial. Deep learning models developed for Arctic forecasts primarily focus on exploring convolutional neural networks (CNN) and convolutional Long Short-Term Memory (LSTM) networks, while the exploration of the power of LSTM networks is limited. In this research, we focus on enhancing the performance of an LSTM network for predicting monthly Arctic SIE. We leverage five climate and atmospheric variables, validated for their correlation with SIE in prior studies [3]. We utilize the Spearman’s rank correlation and ExtraTrees regressor to enhance our understanding of the importance of the five variables in predicting SIE. We further enhance our predictor variables with seasonal information, lagged time steps, and a linear regression simulated SIE that accounts for the influence of past SIE on current SIE. Statistical methods guide our selection of data scalers and best evaluation metrics for our model. By experimenting with hyperparameter optimization and advanced deep learning training techniques, such as batch sizes, number of neurons, early stopping, and model checkpoint, our model achieved a Mean Absolute Error (MAE) of 0.191 and R2 of 0.996, underscoring its ability to account for nearly all the variance in our data and holds great promise for the prediction of SIE.
References
[1]
The National Oceanic and Atmospheric Administration (2018) Arctic Report Card: Update for 2018. https://arctic.noaa.gov/report-card/report-card-2018/executive-summary-5/
[2]
The National Aeronautics and Space Administration (2022) Arctic Sea Ice Minimum Extent Vital Signs—Climate Change: Vital Signs of the Planet. NASA Climate Change. https://climate.nasa.gov/vital-signs/arctic-sea-ice/
[3]
Chen, S., Li, K., Fu, H., Wu, Y.C. and Huang, Y. (2023) Sea Ice Extent Prediction with Machine Learning Methods and Subregional Analysis in the Arctic. Atmosphere, 14, Article 1023.
[4]
Serreze, M.C. and Meier, W.N. (2019) The Arctic’s Sea Ice Cover: Trends, Variability, Predictability, and Comparisons to the Antarctic. AnnalsoftheNewYorkAcademyofSciences, 1436, 36-53.
[5]
NASA Earth Observatory (2021) World of Change: Arctic Sea Ice. Earth Observatory. https://earthobservatory.nasa.gov/world-of-change/sea-ice-arctic
[6]
Arctic Monitoring and Assessment Programme (2017) Snow, Water, Ice and Permafrost in the Arctic (SWIPA). Assessment Summary for Policy-Makers. https://swipa.amap.no
[7]
Mu, B., Luo, X., Yuan, S. and Liang, X. (2023) IceTFT v1.0.0: Interpretable Long-Term Prediction of Arctic Sea Ice Extent with Deep Learning. GeoscientificModelDevelopment, 16, 4677-4697.
[8]
Chi, J. and Kim, H. (2017) Prediction of Arctic Sea Ice Concentration Using a Fully Data Driven Deep Neural Network. RemoteSensing, 9, Article 1305.
[9]
Pegkou Christofi, V., Li, A. and Lin, A. (2024) Wavelet Based Multiscale Deep Learning Algorithms for Arctic Sea Ice Melting Prediction. Proceedingsofthe 2024 6thInternationalSymposiumonSignalProcessingSystems, Xi’an, 22-24 March 2024, 29-39. https://doi.org/10.1145/3665053.3665054
[10]
Sherstinsky, A. (2020) Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) Network. PhysicaD: NonlinearPhenomena, 404, Article 132306. https://doi.org/10.1016/j.physd.2019.132306
[11]
TensorFlow. https://www.tensorflow.org/
[12]
Fetterer, F., Knowles, K., Meier, W.N., Savoie, M. and Windnagel, A.K. (2017) Sea Ice Index, Version 3 [Data Set]. Boulder, Colorado USA. National Snow and Ice Data Center.
[13]
Copernicus Climate Change Service (2023) Reanalysis ERA5 Single Levels Monthly Means. https://cds.climate.copernicus.eu/datasets/reanalysis-era5-single-levels-monthly-means?tab=download
[14]
Sluiter, R. (2009) Interpolation Methods for Climate Data: Literature Review. KNMI (Royal Netherlands Meteorological Institute), R&D Information and Observation Technology. https://uaf-snap.org/wp-content/uploads/2020/08/Interpolation_methods_for_climate_data.pdf
[15]
National Center for Atmospheric Research, Dennis, S. and National Center for Atmospheric Research Staff (2014) The Climate Data Guide: Statistical & Diagnostic Methods Overview. https://climatedataguide.ucar.edu/climate-tools/statistical-diagnostic-methods-overview
[16]
Schober, P., Boer, C. and Schwarte, L.A. (2018) Correlation Coefficients: Appropriate Use and Interpretation. Anesthesia&Analgesia, 126, 1763-1768. https://doi.org/10.1213/ane.0000000000002864
[17]
Keskar, N.S., Mudigere, D., Nocedal, J., Smelyanskiy, M. and Tang, P. (2017) On Large-Batch Training for Deep Learning: Generalization Gap and Sharp Minima.
[18]
Masters, D. and Luschi, C. (2018) Revisiting Small Batch Training for Deep Neural Networks.
[19]
Huilgol, P. (2024) Analytics Vidhya. Feature Transformation and Scaling Techniques to Boost Your Model Performance. https://www.analyticsvidhya.com/blog/2020/07/types-of-feature-transformation-and-scaling/
[20]
de Dieu Nyandwi, J. (2021) The Ultimate and Practical Guide on Feature Scaling. https://jeande.medium.com/the-ultimate-and-practical-guide-on-feature-scaling-d03fbe2cb25e
[21]
Scikit-Learn Developers (2023) Plot Different Scalers on the Same Data. https://scikit-learn.org/stable/auto_examples/preprocessing/plot_all_scaling.html#plot-all-scaling-quantile-transformer-section
[22]
Neptune Labs (2023) Fighting Overfitting with L1 or L2 Regularization: Which One Is Better? https://neptune.ai/blog/fighting-overfitting-with-l1-or-l2-regularization
[23]
Van Otten, N. (2023) L1 And L2 Regularization Explained, When to Use Them & Practical How to Examples. https://spotintelligence.com/2023/05/26/l1-l2-regularization/
[24]
Towards Data Science: Comparing Robustness of MAE, MSE, and RMSE. https://towardsdatascience.com/comparing-robustness-of-mae-mse-and-rmse-6d69da870828
[25]
Van Wyk, A. (2022) Encoding Cyclical Features for Deep Learning. https://www.kaggle.com/code/avanwyk/encoding-cyclical-features-for-deep-learning