HYBRID SATELIT IMAGE RECOGNITION SYSTEM COMBINING NEURAL NETWORK FEATURE EXTRACTION AND AN INFORMATION-EXTREMAL CLASSIFIER
DOI:
https://doi.org/10.15588/1607-3274-2026-1-6Keywords:
information-extreme machine learning, convolutional neural network, information criterion, optimization, hybrid model, recognition feature extraction, land cover imagesAbstract
Context. The study solves the relevant task of developing an interpretable and adaptive recognition system for semantic segmentation of satellite imagery by integrating neural network feature extractors with an information-extreme classifier.
Objective. To improve the accuracy of satellite land cover classification by developing a hybrid machine learning model that combines a deep convolutional neural network for extracting informative features with an information-extreme classifier, enabling the construction of highly reliable decision rules even in the presence of overlapping recognition classes in the feature space.
Method. A hybrid model is proposed that combines efficient spatial feature extraction using a convolutional neural network (CNN) with an information-extreme intelligent (IEI) technology for data analysis, based on maximizing the information capacity of the recognition system during machine learning. For feature aggregation, GlobalAveragePooling is applied instead of the classical Flatten operation. Additionally, regularization techniques such as weight decay and cyclical learning rate scheduling are implemented. The optimization of IEI model parameters is carried out using a modified Kullback information criterion, interpreted as a measure of recognition class diversity.
Results. The developed model achieves high classification accuracy (95%) on the test set and demonstrates stable performance, with improved efficiency of the neural feature extractor due to a reduced number of training epochs enabled by regularization techniques. As a result of the information-extreme machine learning process, the optimal geometric parameters of the hyperspherical recognition class containers were determined, allowing the construction of highly reliable decision rules even under conditions of recognition class overlap in the feature space.
Conclusions. The proposed hybrid model enables the construction of highly reliable decision rules through information-extreme machine learning, even in cases of a priori fuzzy partitioning of recognition classes in the feature space, based on the input training matrix formed during feature extraction.
References
Ab Wahab M. N., Nazir A., Ren A. T. Z., Noor M. H. M., Akbar M. F., Mohamed A. S. A. Efficientnet-Lite and Hybrid CNN-KNN Implementation for Facial Expression Recognition on Raspberry Pi, IEEE Access, 2021, Vol. 9, pp. 134065–134080. DOI:10.1109/ACCESS.2021.3113337
Ghosh S., Singh A., Kavita, Jhanjhi N. Z., Masud M., Aljahdali S. SVM and KNN Based CNN Architectures for Plant Classification, Computers, Materials and Continua, 2022, Vol. 71, № 3, P. 4257. DOI:10.32604/CMC.2022.023414
Lanjewar M. G., Parab J. S., Shaikh A. Y. Development of framework by combining CNN with KNN to detect Alzheimer’s disease using MRI images, Multimedia Tools and Applications, 2023, Vol. 82, № 8, pp. 12699–12717. DOI:10.1007/S11042-022-13935-4/METRICS
Moskalenko V., Kharchenko V., Moskalenko A., Petrov S. Model and Training Method of the Resilient Image Classifier Considering Faults, Concept Drift, and Adversarial Attacks, Algorithms, 2022, Vol. 15, № 384, pp. 1–24. DOI:10.3390/a15100384
Moskalenko V. V., Moskalenko A. S., Korobov A. G., Zaretsky M. O. Image Classifier Resilient To Adversarial Attacks, Fault Injections And Concept Drift – Model Architecture And Training Algorithm, Radio Electronics, Computer Science, Control, 2022, Vol. 3, № 86, pp. 1–16.
DOI:10.15588/1607-3274-2022-3-9
Shelehov I., Prylepa D., Khibovska Yu. Informationextreme machine learning of an ophthalmic diagnostic system with a hierarchical class structure, Artificial Intelligence, 2024, Vol. 29, №3, pp. 114–125. DOI:10.15407/JAI2024.03.114
Dovbysh A. S. Shelekhov I. V., Prylepa D. V., Khibovska Yu. O., Nikitenko K. O. Information-Extreme Method For Ball Detection In Intelligent Video Analysis Systems Of Volleyball Matches, Visnyk Kremenchutskoho natsionalnoho universytetu imeni Mykhaila Ostrohradskoho, 2024, Vol. 5, pp. 41–51. DOI:10.32782/1995-0519.2024.5.6
Dovbysh A. S., Budnyk M. M., Piatachenko V. Y., Myronenko M. I. Information-extreme machine learning of on-board vehicle recognition system, Cybernetics and Systems Analysis, 2020, Vol. 56, pp. 534–543. DOI:10.1007/s10559-020-00269-y
Naumenko I., Piatachenko V., Myronenko M., Savchenko T. Information-Extreme Machine Learning of an On-board Ground Object Recognition System with a Choice of a Base Recognition Class, 6th International Conference on Computational Linguistics and Intelligent Systems, Gliwice, 12–13 May 2022: proceedings. Gliwice, CEUR, 2022, pp. 1139–1148.
Tan M., Le Q. V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks, 36th International Conference on Machine Learning, 9th June 2019: proceedings. Long Beach:arXiv, 2019, pp. 10691–10700. DOI: 10.48550/arXiv.1905.11946
Sandler M., Howard A., Zhu M., Zhmoginov A., Chen L. C. MobileNetV2: Inverted Residuals and Linear Bottlenecks, Conference on Computer Vision and Pattern Recognition, Salt Lake City, 2018: proceedings. Salt Lake City, IEEE, 2018, pp. 4510–4520. DOI:10.1109/CVPR.2018.00474
Dasgupta R., Chowdhury Y. S., Nanda S. Performance Comparison of Benchmark Activation Function ReLU, Swish and Mish for Facial Mask Detection Using Convolutional Neural Network, Algorithms for Intelligent Systems, Singapore, 2021: proceedings. Singapore,
Springer, 2021, pp. 355–367. DOI:10.1007/978-981-16-2248-9_34
Helber P., Bischke B., Dengel A., Borth D. Eurosat: A novel dataset and deep learning benchmark for land use and land cover classification, IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2019, pp. 1–9. DOI: 10.48550/arXiv.1709.00029
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 A. S. Dovbysh, V. Y. Piatachenko, V. M. Serhieiev, O. M. Hrytsenko

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Creative Commons Licensing Notifications in the Copyright Notices
The journal allows the authors to hold the copyright without restrictions and to retain publishing rights without restrictions.
The journal allows readers to read, download, copy, distribute, print, search, or link to the full texts of its articles.
The journal allows to reuse and remixing of its content, in accordance with a Creative Commons license СС BY -SA.
Authors who publish with this journal agree to the following terms:
-
Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution License CC BY-SA that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
-
Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
-
Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) as it can lead to productive exchanges, as well as earlier and greater citation of published work.