COMPARATIVE ANALYSIS OF STATE-OF-THE-ART NEURAL NETWORKS FOR ART OBJECT RECOGNITION
https://doi.org/10.55452/1998-6688-2025-22-2-67-75
Abstract
Currently, information technology is rapidly developing and one of its branches can be called machine translation. The use of machine translation in the process of understanding each other by people from different countries is increasing every year. At the moment, Google and Yandex machine translations are among the best machine translations. The quality of machine translation from Yandex and Google is improving every year. However, according to the results of the experiment, when translating from English or Russian into Kazakh and Turkic languages, the quality of the translation decreases. This was shown by the translation result obtained from these two machine translations in March 2024. After all, translation has also shown that it is directly related to the structure of language. Since 2000, scientists from the state of Kazakhstan have been actively studying translations into the Kazakh language. The goal of the work is to improve the quality of translation from English into Kazakh. For this purpose, a transforming model was created for the Kazakh and Turkic languages for learning translation in neural machine translation OpenNMT(). The created model studied and learned an English-Kazakh parallel corpus of 180,000 words. Later, the document with a structure of 20,000 different English sentences was translated into Kazakh. The result is measured using the Blue() metric. The translation result showed a high level. It is shown that in order to improve the results of the experiment carried out in the work during model training, it is necessary to increase the number of parallel corpora created from the English-Kazakh language pair.
About the Authors
Y. KozhagulovKazakhstan
PhD, Acting Associate Professor
A. Maksutova
Kazakhstan
PhD student
D. Zhexebay
Kazakhstan
PhD
A. Skabylov
Kazakhstan
PhD
T. Kozhagulov
Kazakhstan
Cand. Ped. Sci., Professor
References
1. Kozhagulov E.T., Zheksebaj D.M., Sarmanbetov S.A., Maksutova A.A., Bazhaev N.A. Klassifikator izobrazhenij mikroshem pri pomoshhi svertochnoj nejronnoj seti. News of the National Academy of Sciences of the Republic of Kazakhstan. Physico-Mathematical Series, 6, 40, 59–65 (2021) https://doi.org/10.32014/2020.2518-1726.102.
2. Kozhagulov Y.T., Zhexebay D.M., Sarmanbetov S.A., Sagatbayeva A.A., Zholdas D. Comparative analysis of object detection processing speed on the basis of neuroprocessors and neuroaccelerators. News of the National Academy of Sciences of the Republic of Kazakhstan, Series physico-mathematical, 4 (332), 61–67 (2020). https://doi.org/10.32014/2020.2518-1726.66
3. Kozhagulov Y.T., Ibraimov M.K., Zhexebay D.M., Sarmanbetov S.A. Face Detection of Integral Image by Viola-Jones method. Recent Contributions to Physics, 3 (66) (2018).
4. Simonyan K. and Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition (2014). https://arxiv.org/abs/1409.1556
5. Kaiming He, Xiangyu Zhang, Shaoqing Ren, Jian Sun. Deep Residual Learning for Image Recognition. Microsoft Research (2016). https://arxiv.org/abs/1512.03385
6. Jiazhi Liang. Image classification based on RESNET. Journal of Physics Conference Series, 1634(1), 012110 (September 2020). http://dx.doi.org/10.1088/1742-6596/1634/1/012110
7. Szegedy Ch., Wei Liu, Yangqing Jia, Sermanet P., Reed S., Arbor A., Anguelov D., Erhan D., Vanhoucke V., Rabinovich A. Going deeper with convolutions. 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). https://doi.ieeecomputersociety.org/10.1109/CVPR.2015.7298594
8. Tan M. and Le Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. Proceedings of the 36th International Conference on Machine Learning, ICML 2019 (Long Beach, 9–15 June 2019), pp. 6105–6114. http://proceedings.mlr.press/v97/tan19a.html
9. Dosovitskiy A., Beyer L., Kolesnikov A., Weissenborn D., Xiaohua Zhai, Unterthiner T., Dehghani M., Minderer M., Heigold G., Gelly S., Uszkoreit Ja., Houlsby N. An image is worth 16x16 words: transformers for image recognition at scale. Published as a conference paper at ICLR 2021 https://openreview.net/forum?id=YicbFdNTTy
10. YOLOv8 Architecture: A Deep Dive into its Architecture (2024). URL: https://yolov8.org/yolov8-architecture/
11. What is YOLO? The Ultimate Guide. Roboflow (2024). URL: https://blog.roboflow.com/guidetoyolo-models.
12. Redmon Jo., Divvala S., Girshick R., Farhadi A. You Only Look Once: Unified, Real-Time Object Detection. Conference: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (June 2016). http://dx.doi.org/10.1109/CVPR.2016.91
Review
For citations:
Kozhagulov Y., Maksutova A., Zhexebay D., Skabylov A., Kozhagulov T. COMPARATIVE ANALYSIS OF STATE-OF-THE-ART NEURAL NETWORKS FOR ART OBJECT RECOGNITION. Herald of the Kazakh-British Technical University. 2025;22(2):67-75. (In Russ.) https://doi.org/10.55452/1998-6688-2025-22-2-67-75