APPLICATION OF A CONVOLUTIONAL NEURAL NETWORK TO CREATE A DETECTOR OF TECHNICAL ANALYSIS FIGURES ON EXCHANGE QUOTES CHARTS
Today, the use of artificial intelligence based on neural networks is the most effective approach to solving image recognition problems. The possibility of using a convolutional neural network to create a pattern detector for technical analysis based on stock chart data has been investigated. The found figures of technical analysis can serve as the basis for making trading decisions in the financial markets. In the conditions of an ever-growing array of various information, the use of visual data reading tools is becoming more and more expedient, as it allows to speed up the process of searching and processing the necessary information for decision-makers. The modeling process, analysis, and results of applying the pattern detector of technical analysis are presented. The general approach to the construction and learning of a convolutional neural network is also described, and the process of preliminary processing of input data is described. Using the created detector allows to automate the search for patterns and improve the accuracy of making trading decisions. After finding the patterns, it becomes possible to obtain additional stock statistics for each type of figure: the context in front of the figures, the percentage of successfully completed figures, volume analysis, etc. These technical solutions can be used as expert and trading systems in the stock market, as well as integrated into existing ones.
Simard, P. Y., Steinkraus, D., Platt, J. C. (2003). Best practices for convolutional neural networks applied to visual document analysis. Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings. doi: https://doi.org/10.1109/icdar.2003.1227801
Velay, M., Daniel, F. (2018). Stock Chart Pattern recognition with Deep Learning. Available at: https://arxiv.org/pdf/1808.00418.pdf
Tsai, Y.-C., Chen, J.-H., Wang, J.-J. (2018). Predict Forex Trend via Convolutional Neural Networks. Journal of Intelligent Systems. doi: https://doi.org/10.1515/jisys-2018-0074
Dingy, X., Zhangz, Y., Liuy, T., Duan, J. (2015). Deep Learning for Event-Driven Stock Prediction. Proceedings of the Twenty-Fourth International Joint Conference on Artificial Intelligence (IJCAI 2015), 2327–2333. Available at: https://www.ijcai.org/Proceedings/15/Papers/329.pdf
Galeshchuk, S., Mukherjee, S. (2017). Deep networks for predicting direction of change in foreign exchange rates. Intelligent Systems in Accounting, Finance and Management, 24 (4), 100–110. doi: https://doi.org/10.1002/isaf.1404
Di Persio, L., Honchar, O. (2018). Multitask machine learning for financial forecasting. INTERNATIONAL JOURNAL OF CIRCUITS, SYSTEMS AND SIGNAL PROCESSING, 12, 444–451. Available at: https://www.researchgate.net/publication/324984045_Multitask_machine_learning_for_financial_forecasting
Di Persio, L., Honchar, O. (2016). Artificial Neural Networks Approach to the Forecast of Stock Market Price Movements. International Journal of Economics and Management Systems, 1, 158–162. Available at: https://www.iaras.org/iaras/filedownloads/ijems/2016/007-0019.pdf
LeCun, Y., Bengio, Y. (1998). Convolutional networks for images, speech, and time series. The handbook of brain theory and neural networks, 255–258.
LeCun, Y., Kavukcuoglu, K., Farabet, C. (2010). Convolutional networks and applications in vision. Proceedings of 2010 IEEE International Symposium on Circuits and Systems. doi: https://doi.org/10.1109/iscas.2010.5537907
Girshick, R., Donahue, J., Darrell, T., Malik, J. (2014). Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. 2014 IEEE Conference on Computer Vision and Pattern Recognition. doi: https://doi.org/10.1109/cvpr.2014.81
Girshick, R., Donahue, J., Darrell, T., Malik, J. (2016). Region-Based Convolutional Networks for Accurate Object Detection and Segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38 (1), 142–158. doi: https://doi.org/10.1109/tpami.2015.2437384
Girshick, R. (2015). Fast R-CNN. 2015 IEEE International Conference on Computer Vision (ICCV). doi: https://doi.org/10.1109/iccv.2015.169
Ren, S. et. al. (2015). Faster R-CNN: Towards real-time object detection with region proposal networks. Advances in neural information processing systems, 91–99.
He, K., Gkioxari, G., Dollar, P., Girshick, R. (2017). Mask R-CNN. 2017 IEEE International Conference on Computer Vision (ICCV). doi: https://doi.org/10.1109/iccv.2017.322
Redmon, J., Divvala, S., Girshick, R., Farhadi, A. (2016). You Only Look Once: Unified, Real-Time Object Detection. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). doi: https://doi.org/10.1109/cvpr.2016.91
Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C.-Y., Berg, A. C. (2016). SSD: Single Shot MultiBox Detector. Lecture Notes in Computer Science, 21–37. doi: https://doi.org/10.1007/978-3-319-46448-0_2
Skuratov, V., Kuzmin, K., Nelin, I., Sedankin, M. (2019). Application of kohonen neural networks to search for regions of interest in the detection and recognition of objects. Eastern-European Journal of Enterprise Technologies, 3 (9 (99)), 41–48. doi: https://doi.org/10.15587/1729-4061.2019.166887
Copyright (c) 2019 Victor Skuratov, Konstantin Kuzmin, Igor Nelin, Mikhail Sedankin
This work is licensed under a Creative Commons Attribution 4.0 International License.
Our journal abides by the Creative Commons CC BY copyright rights and permissions for open access journals.
Authors, who are published in this journal, agree to the following conditions:
1. The authors reserve the right to authorship of the work and pass the first publication right of this work to the journal under the terms of a Creative Commons CC BY, which allows others to freely distribute the published research with the obligatory reference to the authors of the original work and the first publication of the work in this journal.
2. The authors have the right to conclude separate supplement agreements that relate to non-exclusive work distribution in the form in which it has been published by the journal (for example, to upload the work to the online storage of the journal or publish it as part of a monograph), provided that the reference to the first publication of the work in this journal is included.