A Financial Embedded Vector Model and Its Applications to Time Series Forecasting


Inspired by the embedding representation in Natural Language Processing (NLP), we develop a financial embedded vector representation model to abstract the temporal characteristics of financial time series. Original financial features are discretized firstly, and then each set of discretized features is considered as a “word” of NLP, while the whole financial time series corresponds to the “sentence” or “paragraph”. Therefore the embedded vector models in NLP could be applied to the financial time series. To test the proposed model, we use RBF neural networks as regression model to predict financial series by comparing the financial embedding vectors as input with the original features. Numerical results show that the prediction accuracy of the test data is improved for about 4-6 orders of magnitude, meaning that the financial embedded vector has a strong generalization ability.


[1] Scheffer M, Carpenter S R, Lenton T M, Bascompte J, Brock W, Dakos V, Van De Koppel J, Van De Leemput I A, Levin S A, Van Nes E H, Pascual M, Vandermeer J. Anticipating Critical Transitions [J]. Science, 2012, 338(6105): 344-348.

[2] Cavalcante R C, Brasileiro R C, Souza V L F, Nobrega J P, Oliveira A L I. Computational Intelligence and Financial Markets: A Survey and Future Directions [J]. Expert Systems with Applications, 2016, 55: 194-211.

[3] Tsai C F, Hsiao Y C. Combining multiple feature selection methods for stock prediction: Union, intersection, and multi-intersection approaches [J]. Decision Support Systems, 2010, 50(1): 258-269.

[4] Jasemi M, Kimiagari A M, Memariani A. A modern neural network model to do stock market timing on the basis of the ancient investment technique of Japanese Candlestick [J]. Expert Systems with Applications, 2011, 38(4): 3884-3890.

[5] Shen W, Guo X, Wu C, Wu D. Forecasting stock indices using radial basis function neural networks optimized by artificial fish swarm algorithm [J]. Knowledge-Based Systems, 2011, 24(3): 378-385.

[6] Xiong T, Bao Y, Hu Z, Chiong R. Forecasting interval time series using a fully complex-valued RBF neural network with DPSO and PSO algorithms [J]. Information Sciences, 2015, 305: 77-92.

[7] Akbilgic O, Bozdogan H, Balaban M E. A novel Hybrid RBF Neural Networks model as a forecaster [J]. Statistics and Computing, 2014, 24(3): 365-375.

[8] Cao L. Support vector machines experts for time series forecasting [J]. Neurocomputing, 2003, 51: 321-339.

[9] Wang J, Hou R, Wang C, Shen L. Improved v-Support vector regression model based on variable selection and brain storm optimization for stock price forecasting [J]. Applied Soft Computing, 2016, 49: 164-178.

[10] Shen F, Chao J, Zhao J. Forecasting exchange rate using deep belief networks and conjugate gradient method [J]. Neurocomputing, 2015, 167: 243-253.

[11] Kuremoto T, Kimura S, Kobayashi K, Obayashi M. Time series forecasting using a deep belief network with restricted Boltzmann machines [J]. Neurocomputing, 2014, 137: 47-56.

[12] Pulido M, Melin P, Castillo O. Particle swarm optimization of ensemble neural networks with fuzzy aggregation for time series prediction of the Mexican Stock Exchange [J]. Information Sciences, 2014, 280: 188-204.

[13] Lecun Y, Bengio Y, Hinton G. Deep learning [J]. Nature, 2015, 521(7553): 436-444.

[14] Al-Ayyoub M, Nuseir A, Alsmearat K, Jararweh Y, Gupta B. Deep learning for Arabic NLP: A survey [J]. Journal of Computational Science.

[15] Khan W, Daud A, Nasir A A, Amjad T. A survey on the state-of-the-art machine learning models in the context of NLP [J]. Kuwait Journal of Science, 2016, 43(4): 95-113.

[16] Sun S, Luo C, Chen J. A review of natural language processing techniques for opinion mining systems [J]. Information Fusion, 2017, 36(Supplement C): 10-25.

[17] Turian J, Ratinov L, Bengio Y. Word representations: a simple and general method for semi-supervised learning[A]. In proceedings of the 48th Annual Meeting of the Association for Computational Linguistics[C]. Uppsala, Sweden: Association for Computational Linguistics, 2010:384-394

[18] Deerwester S, Dumais S T, Furnas G W, Landauer T K, Harshman R. Indexing by Latent Semantic Analysis. Journal of the American Society for Information Science, 1990, 41: 391-407.

[19] Blei D M, Ng A Y, Jordan M I. Latent Dirichlet Allocation. Journal of Machine Learning Research [J]. 2003, 3 (4-5): 993-1022.

[20] Bengio Y, Ducharme J, Vincent P, Janvin C. A neural probabilistic language model [J]. The Journal of Machine Learning Research, 2003, 3(2): 1137-1155.

[21] Mikolov T, Chen K, Corrado G, Dean J. Efficient Estimation of Word Representations in Vector Space[J]. In proceedings of 1st International Conference on Learning Representations (ICLR2013). Scottsdale, AZ, USA, 2013.

[22] Mikolov T, Sutskever I, Chen K, Corrado G, Dean J. Distributed Representations of Words and Phrases and their Compositionality[A]. In proceedings of the 26th International Conference on Neural Information Processing Systems[C]. Lake Tahoe, Nevada: Curran Associates Inc., 2013:3111-3119.

[23] Le Q V, Mikolov T. Distributed Representations of Sentences and Documents[A]. In proceedings of the 31st International Conference on Machine Learning[C]. Beijing, China, 2014:1188-1196.

[24] Singh S P, Kumar A, Darbari H, Singh L, Rastogi A, Jain S. Machine translation using deep learning: An overview[A]. In proceedings of 2017 International Conference on Computer, Communications and Electronics (Comptelix)[C], 2017:162-167.

[25] Frome A, Corrado G, Shlens J, Bengio S, Dean J, Ranzato M A, Mikolov T. DeViSE: A Deep Visual-Semantic Embedding Model[A]. In proceedings of Advances in Neural Information Processing Systems 26 (NIPS 2013)[C]. Lake Tahoe, Nevada, USA: Curran Associates, Inc., 2013:2121-2129.

[26] Moyano L G. Learning network representations [J]. The European Physical Journal Special Topics, 2017, 226(3): 499-518.

[27] Perozzi B, Al-Rfou R, Skiena S. DeepWalk: online learning of social representations[A]. In proceedings of the 20th ACM SIGKDD international conference on Knowledge discovery and data mining[C]. New York, USA: ACM, 2014:701-710.

[28] Tang J, Qu M, Wang M, Zhang M, Yan J, Mei Q. LINE: Large-scale Information Network Embedding[A]. In proceedings of the 24th International Conference on World Wide Web[C]. Florence, Italy: International World Wide Web Conferences Steering Committee, 2015:1067-1077.

[29] Yahoo finance [DB/OL]. http://finance.yahoo.com, 2017, 1,1.
How to Cite
SUN, Yanfeng et al. A Financial Embedded Vector Model and Its Applications to Time Series Forecasting. INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, [S.l.], v. 13, n. 5, p. 881-894, sep. 2018. ISSN 1841-9844. Available at: <http://univagora.ro/jour/index.php/ijccc/article/view/3286>. Date accessed: 11 july 2020. doi: https://doi.org/10.15837/ijccc.2018.5.3286.


Embedded Vector; Financial Daily Vector; Financial Weekly Vector; RBF Neural Network