Financial Time Series Forecasting based on Novel CEEMDAN-LSTM-BN Network
DOI: https://doi.org/10.62381/ACS.SDIT2024.08
Author(s)
Shunwei Dou
Affiliation(s)
Xi’an Jiaotong-liverpool University, Suzhou, China
Abstract
Long-short term memory(LSTM) is a state-of-art and widely used model to forecast financial time series. However, primitive LSTM networks do not perform well due to over-fitting problems of the deep learning model and non-linear and non-stationary characteristics of financial time series data. Thus, this paper proposed a novel hybrid network based on LSTM to solve the two problems. To avoid over-fitting, the modified LSTM-BN network consists of two LSTM layers, two Batch Normalization(BN) layers following each LSTM layer and a dropout layer. Complete Ensemble Empirical Mode Decomposition with Adaptive Noise (CEEMDAN) is an outstanding data frequency decomposition technique which can decompose original time series into several intrinsic mode functions and a residue. Each of the intrinsic mode function and the residue would be processed by LSTM-BN and the final prediction results are obtained by reconstructing each predictive series. The advantages of the proposed networks are verified by comparing to primitive LSTM, other hybrid models and some famous machine learning model such as Support Vector Machine(SVM), Autoregressive(AR) and Random Forest. Moreover, the robustness of the networks are assessed by numerical experiments on different stock indices datasets including “Standard’s & Poor’s 500 Index”, “Nikkei 225”, “Heng Seng Index” and “Deutscher Aktienindex Index”.
Keywords
LSTM Networks; SVM; CEEMDAN
References
[1]Introduction to time series and forecasting[M]. New York, NY: Springer New York, 2002.
[2]Poon S H, Granger C W J. Forecasting volatility in financial markets: A review[J]. Journal of economic literature, 2003, 41(2): 478-539.
[3]Abu-Mostafa Y S, Atiya A F. Introduction to financial forecasting[J]. Applied intelligence, 1996, 6: 205-213.
[4]Wall L D. Some financial regulatory implications of artificial intelligence[J]. Journal of Economics and Business, 2018, 100: 55-63.
[5]Huang W, Nakamori Y, Wang S Y. Forecasting stock market movement direction with support vector machine[J]. Computers & operations research, 2005, 32(10): 2513-2522.
[6]Dash R K, Nguyen T N, Cengiz K, et al. Fine-tuned support vector regression model for stock predictions[J]. Neural Computing and Applications, 2023, 35(32): 23295-23309.
[7]Qin S J, Chiang L H. Advances and opportunities in machine learning for process data analytics[J]. Computers & Chemical Engineering, 2019, 126: 465-473.
[8]Chen Y, Hao Y. A feature weighted support vector machine and K-nearest neighbor algorithm for stock market indices prediction[J]. Expert Systems with Applications, 2017, 80: 340-355.
[9]Guresen E, Kayakutlu G, Daim T U. Using artificial neural network models in stock market index prediction[J]. Expert systems with Applications, 2011, 38(8): 10389-10397.
[10]Rezaei H, Faaljou H, Mansourfar G. Stock price prediction using deep learning and frequency decomposition[J]. Expert Systems with Applications, 2021, 169: 114332.
[11]Guresen E, Kayakutlu G, Daim T U. Using artificial neural network models in stock market index prediction[J]. Expert systems with Applications, 2011, 38(8): 10389-10397.
[12]Kim K, Han I. Genetic algorithms approach to feature discretization in artificial neural networks for the prediction of stock price index[J]. Expert systems with Applications, 2000, 19(2): 125-132.
[13]Devadoss A V, Ligori T A A. Forecasting of stock prices using multi layer perceptron[J]. International journal of computing algorithm, 2013, 2(1): 440-449.
[14]Kamijo K, Tanigawa T. Stock price pattern recognition-a recurrent neural network approach[C]//1990 IJCNN international joint conference on neural networks. IEEE, 1990: 215-221.
[15]Fischer T, Krauss C. Deep learning with long short-term memory networks for financial market predictions[J]. European journal of operational research, 2018, 270(2): 654-669.
[16]Chen S, He H. Stock prediction using convolutional neural network[C]//IOP Conference series: materials science and engineering. IOP Publishing, 2018, 435: 012026.
[17]Siami-Namini S, Tavakoli N, Namin A S. A comparison of ARIMA and LSTM in forecasting time series[C]//2018 17th IEEE international conference on machine learning and applications (ICMLA). IEEE, 2018: 1394-1401.
[18]Xue H, Huynh D Q, Reynolds M. SS-LSTM: A hierarchical LSTM model for pedestrian trajectory prediction[C]//2018 IEEE Winter Conference on Applications of Computer Vision (WACV). IEEE, 2018: 1186-1194.
[19]Han S, Qiao Y, Yan J, et al. Mid-to-long term wind and photovoltaic power generation prediction based on copula function and long short term memory network[J]. Applied energy, 2019, 239: 181-191.
[20]Wu Z, Huang N E. Ensemble empirical mode decomposition: a noise-assisted data analysis method[J]. Advances in adaptive data analysis, 2009, 1(01): 1-41.
[21]Yeh J R, Shieh J S, Huang N E. Complementary ensemble empirical mode decomposition: A novel noise enhanced data analysis method[J]. Advances in adaptive data analysis, 2010, 2(02): 135-156.
[22]Colominas M A, Schlotthauer G, Torres M E. Improved complete ensemble EMD: A suitable tool for biomedical signal processing[J]. Biomedical Signal Processing and Control, 2014, 14: 19-29.
[23]Nelson D M Q, Pereira A C M, De Oliveira R A. Stock market's price movement prediction with LSTM neural networks[C]//2017 International joint conference on neural networks (IJCNN). Ieee, 2017: 1419-1426.
[24]Hao W, Sun X, Wang C, et al. A hybrid EMD-LSTM model for non-stationary wave prediction in offshore China[J]. Ocean Engineering, 2022, 246: 110566.
[25]Huang N E, Shen Z, Long S R, et al. The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis[J]. Proceedings of the Royal Society of London. Series A: mathematical, physical and engineering sciences, 1998, 454(1971): 903-995.
[26]Hochreiter S, Schmidhuber J. Long short-term memory[J]. Neural computation, 1997, 9(8): 1735-1780.
[27]Hinton G E, Srivastava N, Krizhevsky A, et al. Improving neural networks by preventing co-adaptation of feature detectors[J]. arXiv preprint arXiv:1207.0580, 2012.
[28]Ioffe S, Szegedy C. Batch normalization: Accelerating deep network training by reducing internal covariate shift[C]//International conference on machine learning. pmlr, 2015: 448-456.
[29]Sapankevych N I, Sankar R. Time series prediction using support vector machines: a survey[J]. IEEE computational intelligence magazine, 2009, 4(2): 24-38.