Neural Technologies for Objects Classification with Mobile Applications
Ievgen Sidenko1,*, Galyna Kondratenko1, Oleksandr Heras1 and Yuriy Kondratenko1, 2
1Petro Mohyla Black Sea National University, 68th Desantnykiv Str., 10, Mykolaiv, 54003, Ukraine
2Institute of Artificial Intelligence Problems, Mala Zhytomyrs’ka Str., 11/5, Kyiv, 01001, Ukraine
E-mail: ievgen.sidenko@chmnu.edu.ua; halyna.kondratenko@chmnu.edu.ua; acseler16@gmail.com; yuriy.kondratenko@chmnu.edu.ua
*Corresponding Author
Received 08 November 2023; Accepted 29 February 2024
This paper is related to the study of the features of the neural technologies’ application, in particular, ResNet neural networks for the classification of objects in photographs. The work aims to increase the accuracy of recognition and classification of objects in photographs by using various models of the ResNet neural network. The paper analyzes the features of the application of the corresponding models in comparison with other architectures of deep neural networks and evaluates their efficiency and accuracy in the classification of objects in photographs. The process of data formation for training neural networks, their processing and sorting is described. A web application and a mobile application for recognizing and classifying objects in a photo were also developed. A system for classifying objects, in particular airplanes in photographs, was developed using neural network technologies. It gives a recognition and classification accuracy of about 95%. Research results of ResNet models are of great practical importance, as they can improve the classification accuracy of various images. Features of ResNet, such as the use of skip connections or residual connections, make it effective in the relevant tasks. The results of the study will help to implement ResNet in various fields, including medicine, automatic pattern recognition and other areas where the classification of objects in photographs is an important task.
Keywords: Neural technologies, object classification, ResNet neural network, mobile application.
The research and application of neural technologies for object classification in photographs are relevant in the field of computer vision and image processing [1–5]. The use of neural networks (NNs) for classification has become appropriate and necessary for solving various problems [2, 4, 6–8], and ResNet is one of the effective architectures of deep neural networks, which solves the problem of gradient disappearance and allows training deep models with many layers. ResNet models can effectively detect complex dependencies in data and provide high classification accuracy [1, 4, 9–11].
Computer vision is an actual direction in this work. It involves the use of algorithms and machine learning methods to analyze and process digital images and videos to obtain important information [2–4, 12, 13]. Computer vision has a wide range of practical applications, including image and video recognition, object detection and tracking, face recognition, medical image analysis, and more. With the increasing availability of digital images and videos, computer vision technologies are becoming increasingly necessary in industry, medicine, transportation, military affairs, etc. [1–6, 14–18]. Computer vision requires a lot of data. It analyzes the data again and again until it recognizes the differences and finally recognizes the image. The amount of data required to train a neural network for image classification depends on the number of different parameters, such as the difficulty of classification, the number of categories, the number of elements in each category, etc. [15, 16, 19].
Currently, there are a large number of publications and systems for recognizing and classifying objects in photos using neural technologies. Such objects can be cars, people, cancerous tumors on medical images, other diseases that are classified in the photo, interior items, military equipment, etc [20–23]. The task of recognizing and classifying aircraft is relevant today, since there may be classes of aircraft in the sky and on the ground that belong only to enemy forces. In addition, there is often a need to classify civil aircraft to understand what class they are dealing with for further technical support. It will also be useful to recognize aircraft and evaluate their flight and technical characteristics when training pilots in educational institutions [24–28].
The authors of the paper [29] proposed an intelligent method of recognizing types of aircraft based on audio features, aimed at eliminating the shortcomings of existing recognition methods. An advanced self-attention algorithm is used to intelligently recognize and classify aircraft. A nonlinear self-attention algorithm is proposed, which allows taking complex audio features into account. Collected audio data from 9 types of aircraft are used to train the algorithm and its recognition ability is tested through experiments.
In this work [30], the YOLOv5 detection algorithm and the MMAL-Net classification algorithm are combined for the detection of aircraft in remote images and the determination of their types. First, the YOLOv5 algorithm is used to determine the area where the aircraft is located in complex scenes. The MMAL-Net algorithm is then used to classify the aircraft at a fine-grained level and the type is determined. According to the results of the experiment, the proposed method can accurately detect aircraft and determine their types with an accuracy of 73.2%.
This paper [31] proposes a method for military aircraft recognition with only one training sample. First, HOG characteristics and invariant Hu moments are used to obtain the features. Then, a classification system based on sparse vectors is applied as a recognition algorithm. Experimental results show that the proposed method achieves better recognition accuracy compared to existing methods in the case where there is only one training sample.
In this work [32], a new method for recognizing aircraft types on multi-projection optical images is proposed. The task is complicated by the difference in the postures of the planes, their flight altitude and other factors. Experimental results show that the proposed approach achieves a recognition accuracy of 71.3% for the best variant and 90.8% for the three best variants on a set of test images of aircraft types.
This paper [33] considers the problem of identifying commercial aircraft on satellite images near large airports. An R-CNN neural network is used, where the images are passed through convolutional layers to extract aircraft features, after which these features are fed to a binary classifier. The model was trained on 2100 satellite images with aircraft and tested on 900 images. The system can determine and localize the position of aircraft in the images with high accuracy, although there is a slight deterioration in the model performance when the number of aircraft in the frame increases.
One example of a similar system for aircraft classification is the JetPhotos website (according to Figure 1). On the site, you can download a photographed aircraft and classify it. Among the restrictions is that the photo must be of a specified extension and no more than 250 KB, so you will need to make some additional adjustments in an external photo editor to make it suitable.
In addition, among the advantages is the ability to recognize absolutely any aircraft, but the process of recognition and classification takes a lot of time, and the photo undergoes moderation before it is displayed to site users.
Another variant for similar systems is the site for object recognition from the Aspose company. This site deals exclusively with the recognition of objects in the photo. Thus, it is not possible to get the detailed name of the aircraft, only to check whether the photo shows the aircraft or something else.
Considering the research results in the above-mentioned publications and the functionality of similar systems, it should be noted that the accuracy of recognition and classification can be increased, as well as the time for network training and aircraft recognition can be reduced. In addition, it would be advisable to develop a mobile application for the recognition and classification of aircraft with the possibility of receiving a photo of the aircraft from the camera of a mobile device.
Therefore, the purpose of this work is the recognition and classification of aircraft using neural networks and mobile technologies.
Currently, the following types of neural network architectures are distinguished in various studies: direct neural networks, recurrent neural networks, autoencoding neural networks, neural networks with arbitrary architecture, deep neural networks, neural networks with reinforcement, generative convolutional networks, networks with short message responses, neural networks with memory support [5, 6, 8, 14]. All of them are widely used for various tasks of prediction, recognition, classification, clustering, object detection, control, segmentation, etc. The spheres that effectively found the application of neural networks are economy, medicine, communication, security, transport, industry, etc. [6–8, 13–16, 34–42].
ResNet NNs are one of the most effective and widely used networks for image classification, which is why the study of their models is very important and necessary for the development of the field of computer vision and machine learning. ResNet-50 is a variant of the ResNet model that has 48 convolution layers as well as 1 Max Pool and 1 Average Pool. It has 3.8 10^9 floating point operations. This is a widely used ResNet model [9–11, 37, 43]. The block “residual” with connections is the main block of ResNet-50, which allows the network to effectively solve the problem of vanishing gradients when training deep neural networks [9–11]. This architecture can be used for computer vision tasks such as image classification, object localization, object detection. ResNet-101 is a deep neural network that was proposed in 2015. ResNet101 has 101 layers and is a deeper model than ResNet-50. It is a rather complex model that requires a large amount of computing resources for training and application. It is usually used for image classification tasks on large datasets, such as ImageNet [43].
A more detailed description of the elements of various architectures of ResNet models is given in Figure 2.
AlexNet is a convolutional neural network developed in 2012, which for the first time gave a significant advantage in image processing using deep learning. The architecture (according to Figure 3) of the network contains 5 convolutional layers. The first two layers have a large number of filters helping to detect simple and complex features in images. After each convolutional layer, a pooling layer is applied to reduce the size of the original image and increase network efficiency. The 5th convolutional layer is followed by 3 fully connected layers, each of which has a dimension of 4096 neurons. These layers perform the function of classifying the image generated from the previous convolutional layer [2–4, 44–46].
The VGG architecture consists of several blocks, each of which consists of a sequence of convolutional layers and pooling layers. Each block has a fixed number of convolutional and pooling layers, allowing for easy tuning of the network depth. The VGG architecture has a large number of parameters, which has made it very powerful, but at the same time requires a large amount of computing resources for training and application. The VGG16 architecture (according to Figure 4) consists of 16 layers, including 13 convolutional layers and 3 fully connected layers. Convolutional layers consist of filters of different sizes that perform image convolutions to extract various image features. Fully connected layers use these features to classify images into different categories [2, 5, 47–50].
Completing the given task requires sufficiently powerful computing resources, so deeper and more complex architectures such as ResNet and VGG can be used for image classification, as they usually give better results compared to AlexNet [47]. At the same time, ResNet, thanks to the use of feedback blocks, allows for building an even deeper network with fewer parameters, which makes it a popular option used for many tasks in the field of computer vision. Thus, based on the basis of the conducted study of different architectures of neural networks and the obtained results analysis, it can be concluded that the ResNet architecture is a more effective tool in solving problems of image classification that require high accuracy.
In this work, ResNet was employed as the primary neural network architecture for the task at hand. Notably, Few Shot Learning Models were not utilized in the experimentation process.
To fulfill the task, a ready-made dataset from the Internet resource robots.ox.ac.uk was selected, because it contains photos of more than 50 different types of aircraft. However, in the basic dataset, all airplane photos are stored in one folder, which makes the data unusable (according to Figure 5). As you know, for the correct operation of the system, you need to sort all the photos into folders with the names of specific types of aircraft. The authors of the dataset left a text file in it, where you can see what type of aircraft is shown in the photo.
To sort the dataset, the authors of the work developed a console application that creates folders with the names of the desired types of aircraft and arranges photos in these folders. After dividing all the photos into classes, it was found that there were only 60 photos for each type of aircraft in the dataset. Therefore, for the effective training of the neural network and the operation of the classification system, it was decided to choose 4 types of aircraft and expand the existing dataset with photos from the Internet, to bring the collection of each aircraft to more than 200 photos for each type. The following 4 most popular types of aircraft were selected, such as An-12, ATR-42, Spitfire and Cessna 208. In turn, the folders containing photos of these aircraft were divided into 2 more, one for training and the other for testing neural networks. The photos between the training and test folders are split so that 73% of all aircraft photos are in the training folder and the other 27% in the test folder. It was this distribution that later turned out to be optimal from the viewpoint of training and classification accuracy.
To successfully train a neural network, photos must be of high quality and dimensionality to provide enough detail for the neural network to analyze. In addition, the photos must be representative of the classification task, that is, contain enough visual features to distinguish objects of different classes. To ensure successful learning, it is also important to have enough photos for each class. In our study, it was possible to collect a dataset of photos for training a neural network, the size of which is 802 photos for all 4 types of aircraft.
First, a model was created and trained in the ML.Net environment [3, 6, 51–53], and a web application was developed using Visual Studio 2022 to test the created neural network. A machine learning model was added to the created web application and the “image classification” scenario was selected in it. Next, the environment for training the image classification model was chosen, in our study it is local training of the model on an existing PC. The next step was to select data for training (according to Figure 6).
The next step was one of the most important, namely model training. The neural network coped very well with this task, training with an accuracy of 92.16% in 117.67 seconds, which is a very good result. Finally, the model ResNet-50 was checked for the correctness of image classification. As a result, the classification accuracy of this model was 89%.
To improve classification accuracy, the authors also created ResNet-50 and ResNet-101 models using the Python programming language [54–57].
After the program code was written, we ran it and received in the console (according to Figure 7) the number of photos that are used to train the ResNet-50 neural network model which is used to check the accuracy of aircraft classification. You can also see the progress of the training, which will last 40 iterations, each of which will take an average of 73 seconds and obtain a training accuracy of 92%.
Next, the ResNet-101 model was created. The results of training are shown in Figure 8.
The figures illustrate that ResNet-101 learns longer than ResNet-50, although the number of epochs and power are the same. This is because ResNet-101 has more layers (101) than ResNet-50 (50). Thus, ResNet-101 has more parameters to train, which usually results in longer training times. The web application was created using the ASP.Net MVC framework. The interface for users to upload their photos and get the results is shown in Figure 9.
Testing of the developed web application for aircraft classification showed that all developed models (ResNet-50 and ResNet-101) in ML.Net and Python environments demonstrated high accuracy, which is 95% on average. This proves the high efficiency of the corresponding models.
The results of testing, for example, in the Python environment are shown in Figure 10. It can be seen that the aircraft is classified with high accuracy (92.13% for the Spitfire class aircraft, 95.32% for the ATR-42, and 99.91% for the Cessna 208).
The results of aircraft classification, for example, Spitfire and ATR-42 using the developed web application are shown in Figures 11 and 12.
After creating various ResNet neural network models and testing the developed web application, we will analyze the results of model training to evaluate their effectiveness. For that purpose, a diagram of the accuracy of learning neural networks (according to Figure 13) was created. The diagram shows that the ML.Net neural network with the ResNet-50 architecture reached more than 80% accuracy the fastest, taking only 25 epochs, while the neural network with the same architecture in Python crossed the 80% training accuracy mark only on the 28th epoch, and the neural network with the ResNet-101 architecture in Python did it only on the 31st epoch.
Also, our neural networks underwent rigorous testing, including aggressive attacks, and consistently demonstrated stable performance. Geometric transformation models were also implemented, which not only accelerated the training process of the network, but also significantly increased accuracy and speed, exceeding performance benchmarks when testing a modified neural network outside of our study.
The obtained results of the accuracy of training and classification of aircraft by ResNet-50 and ResNet-101 models in comparison with other similar studies demonstrate an increase in accuracy and a decrease in training time, which are important indicators, especially when used on different mobile devices.
The authors also developed a mobile application for the recognition and classification of aircraft. The start window (according to Figure 14a) is the main page of the “Air recognizer” application, which describes the main capabilities of the system and what it can work with. The application can go to the “About us” window (according to Figure 14b) or press the “Recognize” button. The “About us” window (according to Figure 14b) is designed to briefly inform the user about the developed application and contact details of one of the authors. The “How it works” (according to Figure 14c) window is intended for a brief description of neural networks with the ResNet architecture used in this application.
The “Camera view” window (when clicking the “Recognize” button on the main window) allows users to take photos of aircraft in real-time through the camera of a mobile device for their further classification (according to Figure 15a). The “Picture view” window (according to Figure 15b) is created for viewing a photo taken in the “Camera view” window, with the subsequent possibility of recognizing this photo, sharing it in social networks, returning to the “Camera view” page and choosing another photo from the gallery.
The “Recognizing” window (according to Figure 15c) displays the progress of the process of recognizing the aircraft in the photo using 3 neural networks.
The “Result” window (according to Figure 16a) displays the result of the recognition and classification of the aircraft in the photo by each of the 3 neural networks. The “History” window (according to Figure 16b) is created for saving and viewing previous results of aircraft recognition and classification. The “History result” window (according to Figure 16c) allows you to view specific results of aircraft recognition and classification with detailed information.
As part of this work, research was conducted on ResNet neural network models for object classification in photographs. In the research process, three models were created, namely ResNet-50 in the Python programming language, ResNet-101 in the Python programming language, and ResNet-50 in the C# programming language (using ML.Net).
The obtained results showed that all three models demonstrate high accuracy when classifying objects in photographs. In particular, the best result was obtained using the ResNet-101 model, which showed a training accuracy of 94% and a classification accuracy of about 95%.
The results of the study showed that the best option for developing and training neural networks is to use the Python programming language, as it allows for high development speed and reduction of network training time.
[1] F. Manganiello, ‘Computer Vision with Maker Tech’, Apress, Berkeley, CA, 2021.
[2] M.S. Uddin, J.C. Bansal (Eds), ‘Computer Vision and Machine Learning in Agriculture’, Springer, Singapore, 2021.
[3] R. Szeliski, ‘Computer Vision. Algorithms and Applications’, Springer, Cham, 2022.
[4] B. Raman, et al. (Eds), ‘Computer Vision and Image Processing’, Springer, Cham, 2022.
[5] H. Sharma, M. Saraswat, S. Kumar, J.C. Bansal (Eds), ‘Intelligent Learning for Computer Vision’, Springer, Singapore, 2021.
[6] Y. Kondratenko, et al., ‘Machine Learning Techniques for Increasing Efficiency of the Robot’s Sensor and Control Information Processing’, in: Sensors, 22(3), 2022.
[7] A. Sheremet, et al., ‘Diagnosis of Lung Disease Based on Medical Images Using Artificial Neural Networks’, 3rd Ukraine Conference on Electrical and Computer Engineering, Lviv, Ukraine, 2021.
[8] I. Sidenko, et al., ‘Peculiarities of Human Machine Interaction for Synthesis of the Intelligent Dialogue Chatbot’, 10th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Metz, France, 2019.
[9] T.S. Pan, et al., ‘Multi-scale ResNet for real-time underwater object detection’, in: SIViP, 15, 2021.
[10] A. Sirco, A. Almisreb, N.M. Tahir, J. Bakri, ‘Liver Tumour Segmentation based on ResNet Technique’, 12th International Conference on Control System, Computing and Engineering, Penang, Malaysia, 2022.
[11] T. Liu, ‘Resnet-Based Intelligent Recognition Algorithm and Evaluation of Students’ Tennis Movement in Teaching Video’, 6th Eurasian Conference on Educational Innovation, Singapore, Singapore, 2023.
[12] J.V. Rissati, P.C. Molina, C.S. Anjos, ‘Hyperspectral Image Classification Using Random Forest and Deep Learning Algorithms’, Latin American GRSS & ISPRS Remote Sensing Conference, Santiago, Chile, 2020.
[13] V. Alekseeva, et al., ‘Intelligent Decision Support System for Differential Diagnosis of Chronic Odontogenic Rhinosinusitis Based on U-Net Segmentation’, in: Electronics, 12(5), 2023.
[14] M. Tetiana, Y. Kondratenko, I. Sidenko, G. Kondratenko, ‘Computer vision mobile system for education using augmented reality technology’, in: Journal of Mobile Multimedia, 17(4), 2021.
[15] Q. Li, ‘Application of Computer Vision Technology in Environmental Art Design’, 2nd International Conference on Artificial Intelligence and Computer Information Technology (AICIT), Yichang, China, 2023.
[16] T. Biloborodova, et al., ‘ECG Classification Using Combination of Linear and Non-Linear Features with Neural Network’, in: Studies in Health Technology and Informaticsthis link is disabled, 2022.
[17] X. Zhang, S. Xu, ‘Research on Image Processing Technology of Computer Vision Algorithm’, International Conference on Computer Vision, Image and Deep Learning (CVIDL), Chongqing, China, 2020.
[18] R. Chang, ‘Research on Computer Vision Image Multimedia Technology Based on Big Data’, IEEE 2nd International Conference on Civil Aviation Safety and Information Technology (ICCASIT, Weihai, China, 2020.
[19] A. Krueangsai, S. Supratid, ‘Effects of Shortcut-Level Amount in Lightweight ResNet of ResNet on Object Recognition with Distinct Number of Categories’, International Electrical Engineering Congress, Khon Kaen, Thailand, 2022.
[20] D. Pandey, et al., ‘Object Detection in Indian Food Platters using Transfer Learning with YOLOv4’, 38th International Conference on Data Engineering Workshops, Kuala Lumpur, Malaysia, 2022.
[21] Y. Railkar, et al., ‘Object Detection and Recognition System Using Deep Learning Method’, 8th International Conference for Convergence in Technology, Lonavla, India, 2023.
[22] T. He, et al., ‘Garbage Classification and Recognition System Based on Neural Network’, International Conference on Advanced Electrical Equipment and Reliable Operation, Beijing, China, 2021.
[23] E. Thenmozhi, et al., ‘An Efficient Object Detection and Classification from Restored Thermal Images based on Mask RCNN’, Sixth International Conference on I-SMAC (IoT in Social, Mobile, Analytics and Cloud), Dharan, Nepal, 2022.
[24] J. Fu, B. He, Z. Wang, ‘Aircraft recognition based on feature points and invariant moments’, 13th International Conference on Natural Computation, Fuzzy Systems and Knowledge Discovery, Guilin, China, 2017.
[25] K. Plupattanakit, W. Choensawat and K. Sookhanaphibarn, ‘Visual Aircraft Recognition via Mobile Virtual Reality’, 11th Global Conference on Consumer Electronics, Osaka, Japan, 2022.
[26] Y. Yi, Y. You, W. Zhou, G. Meng, ‘MHA-CNN: Aircraft Fine-Grained Recognition of Remote Sensing Image Based on Multiple Hierarchies Attention’, International Geoscience and Remote Sensing Symposium, Kuala Lumpur, Malaysia, 2022.
[27] T. Khajangthon, W. Choensawat and K. Sookhanaphibarn, “Visual Aircraft Recognition Training via Web Application,” 2022 IEEE 4th Global Conference on Life Sciences and Technologies (LifeTech), Osaka, Japan, 2022, pp. 540–541, doi: 10.1109/LifeTech53646.2022.9754802.
[28] J. Lin, X. Li and H. Pan, ‘Aircraft recognition in remote sensing images based on deep learning’, 33rd Youth Academic Annual Conference of Chinese Association of Automation, Nanjing, China, 2018.
[29] R. Ye, W. Wang, Z. Wang, J. Zhao, ‘Intelligent Recognition of Aircraft Type Using Audio Feature’, 3rd World Symposium on Artificial Intelligence, Guangzhou, China, 2021.
[30] J. Wang, L. Xia and S. Li, ‘Research on Aircraft Type Recognition from Remote Sensing Images in Complex Scenes’, International Conference on Computer Information Science and Artificial Intelligence, Kunming, China, 2021.
[31] Y. Ouyang, et al., ‘An Infrared image process approach for Military Aircraft Recognition with single training sample’, China Automation Congress, Xiamen, China, 2022.
[32] Z. Liang, R. Yu and Y. Li, ‘Aircraft Type Recognition in Multi-view Optical Images Using Coarse-to-Fine Framework’, International Conference on Control, Automation and Information Sciences, Xi’an, China, 2021.
[33] N. Wan Zulkipli, et al., ‘Aircraft recognition using convolutional neural network’, International Conference on Green Energy, Computing and Intelligent Technology, Hybrid Conference, Iskandar Puteri, Malaysia, 2023.
[34] O. Striuk, et al., ‘Implementation of Generative Adversarial Networks in Mobile Applications for Image Data Enhancement’, in: Journal of Mobile Multimedia, 19(3), 2023.
[35] V.M. Kuntsevich, et al. (Eds), ‘Control Systems: Theory and Applications’, River Publishers, Gistrup, Delft, 2018.
[36] O. Striuk, Y. Kondratenko, I. Sidenko, A. Vorobyova, ‘Generative Adversarial Neural Network for Creating Photorealistic Images’, IEEE 2nd International Conference on Advanced Trends in Information Theory, Kyiv, Ukraine, 2020.
[37] I. Sova, I. Sidenko, Y. Kondratenko, ‘Machine Learning Technology for Neoplasm Segmentation on Brain MRI Scans’, CEUR Workshop Proceedings, PhD Symposium at ICT in Education, Research, and Industrial Applications, 2791, Kharkiv, Ukraine, 2020.
[38] R. Duro, et al. (Eds), ‘Advances in intelligent robotics and collaborative automation’, River Publishers, Aalborg, 2015.
[39] M. Fan, Y. Liu, I., ‘Image processing technology based on computer vision algorithm’, 4th International Conference on Artificial Intelligence and Advanced Manufacturing (AIAM), Hamburg, Germany, 2022.
[40] V. Lytvyn, et al., ‘An intelligent system of the content relevance at the example of films according to user needs’, International Workshop on Information-Communication Technologies and Embedded Systems, ICT and ES, 2516, 2019.
[41] S. Kryvyi, O. Grinenko, V. Opanasenko, ‘Logical Approach to the Research of Properties of Software Engineering Ecosystem,’ 11th International Conference on Dependable Systems, Services and Technologies, Kyiv, Ukraine, 2020.
[42] R. Leizerovych, G. Kondratenko, I. Sidenko and Y. Kondratenko, ‘IoT-complex for Monitoring and Analysis of Motor Highway Condition Using Artificial Neural Networks’, IEEE 11th International Conference on Dependable Systems, Services and Technologies, Kyiv, Ukraine, 2020.
[43] X. Li, L. Rai, ‘Apple Leaf Disease Identification and Classification using ResNet Models’, IEEE 3rd International Conference on Electronic Information and Communication Technology, Shenzhen, China, 2020.
[44] A. Agarwal, K. Patni, R. D, ‘Lung Cancer Detection and Classification Based on Alexnet CNN’, 6th International Conference on Communication and Electronics Systems, Coimbatre, India, 2021.
[45] J.-F. Yeh, S.-Y. Wangm Y.-P. Chen, ‘Crop Disease Detection by Image Processing Using Modified Alexnet’, IEEE 3rd Eurasia Conference on Biomedical Engineering, Healthcare and Sustainability, Tainan, Taiwan, 2021.
[46] L.Y. Win Lwin, A.N. Htwe, ‘Image Classification for Rice Leaf Disease Using AlexNet Model’, IEEE Conference on Computer Applications, Yangon, Myanmar, 2023.
[47] J. Liu, ‘VGG, MobileNet and AlexNet on Recognizing Skin Cancer Symptoms’, 3rd International Conference on Electronic Communication and Artificial Intelligence, Zhuhai, China, 2022.
[48] A. Bagaskara, M. Suryanegara, ‘Evaluation of VGG-16 and VGG-19 Deep Learning Architecture for Classifying Dementia People’, 4th International Conference of Computer and Informatics Engineering, Depok, Indonesia, 2021.
[49] D. Kusumawati, A.A. Ilham, A. Achmad, I. Nurtanio, ‘Vgg-16 And Vgg-19 Architecture Models In Lie Detection Using Image Processing’, 6th International Conference on Information Technology, Information Systems and Electrical Engineering, Yogyakarta, Indonesia, 2022.
[50] A. Sokoliuk, et al., ‘Machine Learning Algorithms for Binary Classification of Liver Disease’, IEEE International Conference on Problems of Infocommunications. Science and Technology, Kharkiv, Ukraine, 2020.
[51] Y. Kondratenko, S. Sichevskyi, G. Kondratenko, I. Sidenko, ‘Manipulator’s Control System with Application of the Machine Learning’, 11th IEEE International Conference on Intelligent Data Acquisition and Advanced Computing Systems: Technology and Applications, Cracow, Poland, 2021.
[52] A. Alexan, A. Alexan, O. Ştefan, ‘Machine learning activity detection using ML.Net’, IEEE 26th International Symposium for Design and Technology in Electronic Packaging, Pitesti, Romania, 2020.
[53] S. Mukherjee, ‘ML.NET Revealed. Simple Tools for Applying Machine Learning to Your Applications’, Apress, Berkeley, CA, 2020.
[54] I. Sidenko, et al., ‘Machine Learning for Unmanned Aerial Vehicle Routing on Rough Terrain’, in: Z. Hu, I. Dychka, M. He (Eds) Advances in Computer Science for Engineering and Education VI. ICCSEEA 2023. Lecture Notes on Data Engineering and Communications Technologies, 181, Springer, Cham, 2023.
[55] S. Putatunda, ‘Practical Machine Learning for Streaming Data with Python’, Apress, Berkeley, CA, 2021.
[56] N. Sanghi, ‘Deep Reinforcement Learning with Python’, Apress, Berkeley, CA, 2021.
[57] J. Unpingco, ‘Python Programming for Data Analysis’, Springer, Cham, 2022.
Ievgen Sidenko is an Associate Professor, Ph.D., Associate Professor of the Intelligent Information Systems Department at Petro Mohyla Black Sea National University (PMBSNU), Ukraine. He has received master’s degree in specialty “Intelligent decision-making systems” (2010) at PMBSNU and a Ph.D. degree in “Information technologies” (2015) at PMBSNU. His research interests include fuzzy sets and fuzzy logic, decision-making, optimization methods, neural networks, data mining, clustering, and classification.
Galyna Kondratenko is an Associate Professor, Ph.D., Associate Professor of the Intelligent Information Systems Department, and Senior Researcher at Petro Mohyla Black Sea National University, Ukraine. She is a specialist in control systems, decision-making, and fuzzy logic. She worked in the framework of international scientific university cooperation during the implementation of international projects with the European Union: TEMPUS (Cabriolet), Erasmus (Aliot) and DAAD-Ostpartnerschaftsprogramm (project with the University of Saarland, Germany). Her research interests include computer control systems, fuzzy logic, decision-making, and intelligent robotic devices.
Oleksandr Heras is a master’s student of the Intelligent Information Systems Department at Petro Mohyla Black Sea National University (PMBSNU). He has received his bachelor degree with a major in 122 “Computer Science” at PMBSNU. Oleksandr participated in the implementation of various projects related to the creation of mobile applications and management systems using machine learning technologies. Oleksandr is interested in studying machine learning and computer vision technologies.
Yuriy Kondratenko is a Doctor of Science, Professor, Honour Inventor of Ukraine (2008), Corr. Academician of the Royal Academy of Doctors (Barcelona, Spain), Head of the Intelligent Information Systems Department at Petro Mohyla Black Sea National University (PMBSNU), and Leading Researcher at the Institute of Artificial Intelligence Problems under MES and NAS of Ukraine. He has received (a) a Ph.D. (1983) and a Dr.Sc. (1994) in Elements and Devices of Computer and Control Systems from Odessa National Polytechnic University, (b) several international grants and scholarships for conducting research at Institute of Automation of Chongqing University, P.R.China (1988–1989), Ruhr-University Bochum, Germany (2000, 2010), Nazareth College and Cleveland State University, USA (2003), (c) Fulbright Scholarship for researching in USA (2015/2016) at the Dept. of Electrical Engineering and Computer Science in Cleveland State University. Research interests include robotics, automation, sensors and control systems, intelligent decision support systems, and fuzzy logic.
Journal of Mobile Multimedia, Vol. 20_3, 727–748.
doi: 10.13052/jmm1550-4646.2039
© 2024 River Publishers