Purpose: This study aimed to enhance the prediction of container dwell time, a crucial factor for optimizing port operations, resource allocation, and supply chain efficiency. Determining an optimal learning rate for training Artificial Neural Networks (ANNs) has remained a challenging task due to the diverse sizes, complexity, and types of data involved. Design/Method/Approach: This research used a RandomizedSearchCV algorithm, a random search approach, to bridge this knowledge gap. The algorithm was applied to container dwell time data from the TOS system of the Port of Tema, which included 307,594 container records from 2014 to 2022. Findings: The RandomizedSearchCV method outperformed standard training methods both in terms of reducing training time and improving prediction accuracy, highlighting the significant role of the constant learning rate as a hyperparameter. Research Limitations and Implications: Although the study provides promising outcomes, the results are limited to the data extracted from the Port of Tema and may differ in other contexts. Further research is needed to generalize these findings across various port systems. Originality/Value: This research underscores the potential of RandomizedSearchCV as a valuable tool for optimizing ANN training in container dwell time prediction. It also accentuates the significance of automated learning rate selection, offering novel insights into the optimization of container dwell time prediction, with implications for improving port efficiency and supply chain operations.
References
[1]
Obermeyer, Z. and Emanuel, E.J. (2016) Predicting the Future—Big Data, Machine Learning, and Clinical Medicine. New England Journal of Medicine, 375, 1216-1219. https://doi.org/10.1056/nejmp1606181
[2]
Rajkomar, A., Dean, J. and Kohane, I. (2019) Machine Learning in Medicine. New England Journal of Medicine, 380, 1347-1358. https://doi.org/10.1056/nejmra1814259
[3]
Mazzei, D. and Ramjattan, R. (2022) Machine Learning for Industry 4.0: A Systematic Review Using Deep Learning-Based Topic Modelling. Sensors, 22, Article 8641. https://doi.org/10.3390/s22228641
[4]
Liang, J., Xu, Y., Bao, C., Quan, Y. and Ji, H. (2019) Barzilai-Borwein-Based Adaptive Learning Rate for Deep Learning. Pattern Recognition Letters, 128, 197-203. https://doi.org/10.1016/j.patrec.2019.08.029
[5]
Takase, T., Oyama, S. and Kurihara, M. (2018) Effective Neural Network Training with Adaptive Learning Rate Based on Training Loss. Neural Networks, 101, 68-78. https://doi.org/10.1016/j.neunet.2018.01.016
[6]
Kourounioti, I., Polydoropoulou, A. and Tsiklidis, C. (2016) Development of Models Predicting Dwell Time of Import Containers in Port Container Terminals—An Artificial Neural Networks Application. Transportation Research Procedia, 14, 243-252. https://doi.org/10.1016/j.trpro.2016.05.061
[7]
Kourounioti, I. and Polydoropoulou, A. (2017) Identification of Container Dwell Time Determinants Using Aggregate Data. International Journal of Transport Economics, 44, 567-588.
[8]
Zhao, H., Liu, F., Zhang, H. and Liang, Z. (2019) Research on a Learning Rate with Energy Index in Deep Learning. Neural Networks, 110, 225-231. https://doi.org/10.1016/j.neunet.2018.12.009
[9]
López, O.A.M., López, A.M. and Crossa, J. (2022) Multivariate Statistical Machine Learning Methods for Genomic Prediction. Springer. https://doi.org/10.1007/978-3-030-89010-0
[10]
Montesinos López, O.A., Montesinos López, A. and Crossa, J. (2022) Fundamentals of Artificial Neural Networks and Deep Learning. In: Montesinos López, O.A., Montesinos López, A. and Crossa, J., Eds., Multivariate Statistical Machine Learning Methods for Genomic Prediction, Springer International Publishing, 379-425. https://doi.org/10.1007/978-3-030-89010-0_10
[11]
McCulloch, W.S. and Pitts, W. (1943) A Logical Calculus of the Ideas Immanent in Nervous Activity. The Bulletin of Mathematical Biophysics, 5, 115-133. https://doi.org/10.1007/bf02478259
[12]
Choe, Y. (2014) Hebbian Learning. In: Jaeger, D. and Jung, R., Eds., Encyclopedia of Computational Neuroscience, Springer, 1-5. https://doi.org/10.1007/978-1-4614-7320-6_672-1
[13]
Werbos, P. (1974) Beyond Regression: New Tools for Prediction and Analysis in the Behavioral Sciences. Harvard University Press. https://www.researchgate.net/publication/35657389
[14]
Samuel, A.L. (1959) Some Studies in Machine Learning Using the Game of Checkers. IBM Journal of Research and Development, 3, 210-229. https://doi.org/10.1147/rd.33.0210
[15]
Sutton, R.S. and Barto, A.G. (2018) Reinforcement Learning: An Introduction. 2nd Edition, MIT Press. https://web.stanford.edu/class/psych209/Readings/SuttonBartoIPRLBook2ndEd.pdf
[16]
Gerstner, W. and Kistler, W.M. (2002) Spiking Neuron Models: Single Neurons, Populations, Plasticity. Cambridge University Press. https://doi.org/10.1017/cbo9780511815706
[17]
Mcclelland, J.L. (2005) Hebbian Learning 1 Running Head: HEBBIAN LEARNING How Far Can You Go with Hebbian Learning, and When Does It Lead You Astray? https://www.semanticscholar.org/paper/Running-head-%3A-HEBBIAN-LEARNING-How-Far-Can-You-Go-McClelland/881ec0684974e5ee986d56bfd770f0ff7e638c50#extracted
[18]
Bergstra, J. and Bengio, Y. (2012) Random Search for Hyper-Parameter Optimization. http://scikit-learn.sourceforge.net
[19]
Liu, H., Fu, Q., Du, L., Zhang, T., Yu, G., Han, S., et al. (2022) Learning Rate Perturbation. Proceedings of the 31st ACM International Conference on Information & Knowledge Management, Atlanta, 17-21 October 2022, 4234-4238. https://doi.org/10.1145/3511808.3557626
[20]
Cankaya, B., Eren Tokgoz, B., Dag, A. and Santosh, K.C. (2021) Development of a Machine-Learning-Based Decision Support Mechanism for Predicting Chemical Tanker Cleaning Activity. Journal of Modelling in Management, 16, 1138-1165. https://doi.org/10.1108/jm2-12-2019-0284
[21]
Shankar, S., Punia, S. and Ilavarasan, P.V. (2021) Deep Learning-Based Container Throughput Forecasting: A Triple Bottom Line Approach. Industrial Management & Data Systems, 121, 2100-2117. https://doi.org/10.1108/imds-12-2020-0704
[22]
Fabianpedregosa, F.P., et al. (2011) Scikit-Learn: Machine Learning in Python. Journal of Machine Learning Research, 12, 2825-2830. http://scikit-learn.sourceforge.net
[23]
Fabianpedregosa, F.P., et al. (n.d.) Comparing Randomized Search and Grid Search for Hyperparameter Estimation. https://scikit-learn.org/stable/auto_examples/model_selection/plot_randomized_search.html
[24]
Yoo, S. and Oh, D. (2018) An Artificial Neural Network-Based Fall Detection. International Journal of Engineering Business Management, 10, 1-7. https://doi.org/10.1177/1847979018787905
[25]
Olawoyin, A. and Chen, Y. (2018) Predicting the Future with Artificial Neural Network. Procedia Computer Science, 140, 383-392. https://doi.org/10.1016/j.procs.2018.10.300
[26]
Basheer, I.A. and Hajmeer, M. (2000) Artificial Neural Networks: Fundamentals, Computing, Design, and Application. Journal of Microbiological Methods, 43, 3-31. https://www.sciencedirect.com/journal/journal-of-microbiological-methods https://doi.org/10.1016/s0167-7012(00)00201-3
[27]
Kamiri, J. and Mariga, G. (2021) Research Methods in Machine Learning: A Content Analysis. International Journal of Computer and Information Technology (2279-0764), 10, 78-91. https://doi.org/10.24203/ijcit.v10i2.79
[28]
Jolliffe, I.T. (2002) Principal Component Analysis. 2nd Edition, Springer-Verlag. https://doi.org/10.1007/b98835
[29]
Smith, L.E. (2018) Convergence and Divergence in Statistical and Programmatic Approaches to Address Child Stunting and Wasting. The Journal of Nutrition, 148, 823-824. https://doi.org/10.1093/jn/nxy098
[30]
Bergstra, J., Bardenet, R., Bengio, Y. and Kégl, B. (2011) Algorithms for Hyper-Parameter Optimization. Proceedings of the 24th International Conference on Neural Information Processing Systems, Granada, 12-15 December 2011, 2546-2554. https://proceedings.neurips.cc/paper_files/paper/2011/file/86e8f7ab32cfd12577bc2619bc635690-Paper.pdf