E3sconf Iconnect2023 04032

Download as pdf or txt
Download as pdf or txt
You are on page 1of 11

E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.

1051/e3sconf/202339904032
ICONNECT-2023

Deep Learning Techniques for Image Recognition


and Object Detection
Dr K Sharada1 Wajdi Alghamdi2 K Karthika3 Ahmed Hussein Alawadi4 Gulomova Nozima5Dr. V.
Vijayan6
1
Associate Professor Department of Computer Science and Engineering Gitam(Deemed to be
University)Visakhapatnam. Email : [email protected] Id : https://orcid.org/0000-0003-
4953-4938
2
Department of Information Technology, Faculty of Computing and Information Technology, King
Abdulaziz University, Jeddah, 21589, Saudi ArabiaE-mail [email protected]
3
Assistant Professor, Prince Shri Venkateshwara Padmavathy Engineering College, Chennai – 127
[email protected]
4
College of technical engineering, The Islamic university, Najaf, [email protected]
5
Tashkent State Pedagogical University, Tashkent, Uzbekistan. E-mail: [email protected]
6
Department of mechanical Engineering, K. Ramakrishnan college of technology, Tiruchirappalli,
[email protected]

Abstract-Particularly in the fields of object identification and picture


recognition, deep learning approaches have transformed the science of
computer vision. This abstract provides a summary of recent
developments and cutting-edge methods in deep learning for
applications like object identification and picture recognition. The
automated identification and classification of objects or patterns inside
digital photographs is known as image recognition. Convolutional
neural networks (CNNs), for example, have displayed outstanding
performance in image identification tests. By directly learning
hierarchical representations of visual characteristics from raw pixel
data, these algorithms are able to recognize complex patterns and
provide precise predictions. The ability for models to learn
sophisticated visual representations straight from raw pixel data has
transformed applications like object identification and picture
recognition. The development of extremely accurate and effective
systems has been accelerated by advances in deep learning
architectures and large-scale annotated datasets. Further advances in
object identification and picture recognition are anticipated as deep
learning develops, with applications in a variety of fields including
autonomous driving, surveillance, and medical imaging.
Keywords: machine learning, convolutional neural network,
object detection.

INTRODUCTION

© The Authors, published by EDP Sciences. This is an open access article distributed under the terms of the Creative Commons
Attribution License 4.0 (https://creativecommons.org/licenses/by/4.0/).
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

Deep learning has developed into a powerful approach in the field of artificial
intelligence that is revolutionizing a number of industries, including object
identification and picture recognition. These methods have considerably improved
computer vision systems' capacities, allowing them to comprehend and interpret
visual input with astounding precision and effectiveness. We will delve into the
fundamental theories, methodology, and applications of deep learning approaches
for object and image identification in this article as we explore the intriguing world
of these techniques. Image recognition is the method of automatically locating and
classifying items or patterns in digital stills or moving pictures. It is essential to a
variety of real-world applications, including augmented reality, autonomous driving,
and surveillance systems. Convolutional neural networks (CNNs), in particular,
have demonstrated outstanding effectiveness in image identification tasks using
deep learning approaches by learning hierarchical representations of visual data
directly from raw pixel values. This enables them to automatically extract
meaningful features and classify images with unprecedented accuracy.
As opposed to image recognition, object detection locates items inside an image as
well as identifies them. It involves drawing bounding boxes around detected
objects, providing precise spatial information. Deep learning-based object detection
algorithms combine the power of image recognition with additional techniques such
as region proposal methods and spatial transformations to achieve accurate and
efficient object localization. These techniques have numerous practical applications,
including autonomous robotics, video surveillance, and image search.
The majority of deep learning-based image identification and object detection
systems are built on convolutional neural networks (CNNs). In order to comprehend
more complicated data, CNNs use numerous layers of linked neurons, simulating
the visual processing method of the human brain. The first layers collect low-level
properties like edges and textures, while the following layers pick up more high-
level, abstract information. CNNs can better perform in image interpretation tasks
thanks to this hierarchical feature extraction, which enables them to represent
complicated connections.
The availability of expansive annotated datasets, like as ImageNet and COCO, is
one of the major developments in deep learning for object detection and picture
recognition. Millions of photos in these datasets have been annotated, allowing
CNNs to learn a variety of visual representations for various item categories.
Additionally, deep learning models' training and inference speeds have increased
because to the development of potent graphics processing units (GPUs) and
distributed computing frameworks, making them useful for real-time applications.
Many deep learning architectures have been introduced in recent years to improve
the efficiency of object and picture detection. As an illustration, consider well-
known designs like AlexNet, VGGNet, GooLeNet, and ResNet, which have
produced cutting-edge outcomes on benchmark datasets. For example, deeper
networks, skip connections, and residual learning are frequently used in these
designs to boost model capacity and handle problems like disappearing gradients.
Researchers have also looked at a number of methods to enhance the effectiveness
and stability of deep learning-based image recognition and object detection systems.

2
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

Transfer learning makes use of information from massive datasets to enable pre-
trained models to be improved on domain-specific datasets. Data augmentation
techniques, which includes image rotation, scaling, and cropping, increase the
diversity of training data, enhancing the model's generalization ability. Moreover,
attention mechanisms, which selectively focus on salient features, have been
employed to improve the interpretability and performance of deep learning models.
[31] [32]
The continuous advancements in deep learning techniques for image recognition
and object detection have paved the way for groundbreaking applications across
various domains. These include autonomous vehicles that can perceive their
surroundings, medical systems that can accurately diagnose diseases from medical
images, and smart surveillance systems that can detect and track objects in real-
time. As deep learning techniques continue to evolve, we can expect even more
remarkable breakthroughs in the field of computer vision, enabling machines to
comprehend visual data with human-level accuracy and beyond. [33]
One prominent deep learning approach for image recognition and object detection is
the region-based convolutional neural network (R-CNN) family. R-CNN models
employ a two-stage pipeline that first generates region proposals and then classifies
these proposals using CNNs. This approach has achieved remarkable results in
various benchmark datasets and has been widely adopted in many practical
applications. [34]
To overcome the computational inefficiency of the two-stage R-CNN approach,
subsequent works have introduced single-stage models, such as You Only Look
Once (YOLO) and Single Shot MultiBox Detector (SSD). These models directly
predict object classes and bounding box coordinates in a single pass through the
network, making them faster and more efficient. The trade-off is that they may
sacrifice some accuracy compared to two-stage models, but recent iterations have
shown significant improvements in both speed and accuracy. [35]
Another important aspect of deep learning for image recognition and object
detection is the availability of large-scale annotated datasets. Datasets such as
ImageNet and COCO have played a crucial role in training and evaluating deep
learning models. These datasets consist of millions of labeled images and provide a
diverse range of object categories, enabling models to learn robust and
generalizable representations. Furthermore, recent advancements in network
architectures, such as residual connections, attention mechanisms, and feature
pyramid networks, have further enhanced the capabilities of deep learning models.
These techniques enable models to capture fine-grained details, handle scale
variations, and focus on relevant regions, resulting in improved performance on
challenging tasks.

LITERATURE REVIEW
Title Description Reference

3
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

1. Abhinav, A., & A comprehensive Abhinav, A., & Agrawal, A. (2019). A


Agrawal, A. survey of deep learning comprehensive survey of deep learning
(2019). techniques for image techniques for image recognition. Journal
recognition. of Pattern Recognition and Artificial
Intelligence, 32(1), 47-63.
2. Chen, Z., & Deep learning for Chen, Z., & Gupta, S. (2020). Deep
Gupta, S. (2020). object detection: A learning for object detection: A
comprehensive review. comprehensive review. Journal of Visual
Communication and Image
Representation, 68, 102768.
3. Simonyan, K., Deep convolutional Simonyan, K., & Zisserman, A. (2014).
& Zisserman, A. networks for Very deep convolutional networks for
(2014). recognizing images on large-scale image recognition. arXiv
a huge scale. preprint arXiv:1409.1556.
4. Girshick, R., For precise object Girshick, R., Donahue, J., Darrell, T., &
Donahue, J., identification and Malik, J. (2014). Rich feature hierarchies
Darrell, T., & semantic segmentation, for accurate object detection and
Malik, J. (2014). rich feature hierarchies. semantic segmentation. Proceedings of
the IEEE conference on computer vision
and pattern recognition, 580-587.
5. Ren, S., He, K., Faster R-CNN: Ren, S., He, K., Girshick, R., & Sun, J.
Girshick, R., & Towards real-time (2015). Faster R-CNN: Towards real-
Sun, J. (2015). object detection with time object detection with region
region proposal proposal networks. Advances in Neural
networks. Information Processing Systems, 91-99.
6. Redmon, J., You only look once: Redmon, J., Divvala, S., Girshick, R., &
Divvala, S., Unified, real-time Farhadi, A. (2016). You only look once:
Girshick, R., & object detection. Unified, real-time object detection.
Farhadi, A. Proceedings of the IEEE conference on
(2016). computer vision and pattern recognition,
779-788.
7. He, K., Mask R-CNN. He, K., Gkioxari, G., Dollár, P., &
Gkioxari, G., Girshick, R. (2017). Mask R-CNN.
Dollár, P., & Proceedings of the IEEE international
Girshick, R. conference on computer vision, 2961-
(2017). 2969.
8. Redmon, J., & YOLOv3: An Redmon, J., & Farhadi, A. (2018).
Farhadi, A. incremental YOLOv3: An incremental improvement.
(2018). improvement. arXiv preprint arXiv:1804.02767.
9. Szegedy, C., Rethinking the Szegedy, C., Vanhoucke, V., Ioffe, S.,
Vanhoucke, V., inception architecture Shlens, J., & Wojna, Z. (2016).
Ioffe, S., Shlens, for computer vision. Rethinking the inception architecture for
J., & Wojna, Z. computer vision. Proceedings of the
(2016). IEEE conference on computer vision and
pattern recognition, 2818-2826.

4
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

10. He, K., Deep residual learning He, K., Zhang, X., Ren, S., & Sun, J.
Zhang, X., Ren, for image recognition. (2016). Deep residual learning for image
S., & Sun, J. recognition. Proceedings of the IEEE
(2016). conference on computer vision and
pattern recognition, 770-778.
11. Lin, T. Y., Feature pyramid Lin, T. Y., Dollár, P., Girshick, R., He,
Dollár, P., networks for object K., Hariharan, B., & Belongie, S. (2017).
Girshick, R., He, detection. Feature pyramid networks for object
K., Hariharan, B., detection. Proceedings of the IEEE
& Belongie, S. conference on computer vision and
(2017). pattern recognition, 2117-2125.
12. Redmon, J., & YOLO9000: Better, Redmon, J., & Farhadi, A. (2017).
Farhadi, A. faster, stronger. YOLO9000: Better, faster, stronger.
(2017). Proceedings of the IEEE conference on
computer vision and pattern recognition,
7263-7271.
13. Liu, W., SSD: Single shot Liu, W., Anguelov, D., Erhan, D.,
Anguelov, D., multibox detector. Szegedy, C., Reed, S., Fu, C. Y., & Berg,
Erhan, D., A. C. (2016). SSD: Single shot multibox
Szegedy, C., detector. European conference on
Reed, S., Fu, C. computer vision, 21-37.
Y., & Berg, A. C.
(2016).
14. Redmon, J., & YOLOv3: An Redmon, J., & Farhadi, A. (2018).
Farhadi, A. incremental YOLOv3: An incremental improvement.
(2018). improvement. arXiv preprint arXiv:1804.02767.

PROPOSED SYSTEM
The proposed methodology will outline the system's architecture, highlighting the
deep learning techniques to be utilized for image recognition and object detection. It
will describe the selection of suitable deep learning models, such as convolutional
neural networks (CNNs) or recurrent neural networks (RNNs), and state-of-the-art
algorithms like Faster R-CNN or YOLO. The methodology will also detail the
training data acquisition, preprocessing, and model evaluation strategies, ensuring
an original approach, as detailed in figure 1.
System Implementation:
This section will describe the step-by-step implementation process of the proposed
system. It will cover the software and hardware requirements, along with the
programming languages and libraries utilized. Detailed explanations of data
preprocessing, model training, and fine-tuning procedures will be provided,
highlighting any modifications or customizations made to the existing
methodologies.

5
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

Figure 1: Key factors for choosing between deep learning and machine learning.
Performance Evaluation:
To assess the effectiveness of the proposed system, comprehensive performance
evaluation will be conducted. Measuring measures like accuracy, precision, recall,
and F1-score against current models or datasets will be a part of this evaluation. The
efficacy and novelty of the suggested approach will be highlighted as the findings
are contrasted with those produced using cutting-edge methods.
Table 1: CNN network dataset description.

Class Number of images


Bird 811811
Cat 11281128
Dog 13411341
Horse 526526
Sheep 357357
Total 4163

Ethical Considerations:
The ethical ramifications of deep learning algorithms for object identification and
picture recognition will be covered in this section. It will talk about potential biases,
privacy issues, and how important it is to create AI responsibly. The suggested
system would follow moral principles and make sure that any data utilized is
legitimately obtained and anonymous.
Convolutional Neural Networks (CNNs)
For image identification and object detection applications, convolutional neural
networks (CNNs) have become a potent tool. In-depth information about CNNs,
their design, and the techniques used to train and test them for object detection and
image recognition is provided in this study. The usefulness of CNNs in many real-

6
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

world circumstances is further demonstrated by experimental details that are


presented.

Figure 2: Faster RCNN acts as a single, unified network for object detection.
Among the most important computer vision tasks are object and image
recognition. By reaching state-of-the-art performance on several benchmark
datasets, CNNs have transformed these disciplines. As illustrated in figure 2, we
give a quick overview of CNNs and their uses in this section.
CNN Architecture:
Convolutional, pooling, and fully linked layers are just a few of the layers found in
CNNs. Each of these layers and their function within the CNN architecture are
discussed in this section. Additionally, we go through several activation
mechanisms frequently utilized in CNNs.
Training CNNs:
Two essential procedures are involved in training a CNN: backpropagation and
forward propagation. We explain the forward propagation method, which generates
feature maps by feeding inputs into the network. The network's parameters are then
modified depending on the obtained gradients via backpropagation. We also go
through well-known optimization techniques including stochastic gradient descent
(SGD), Adam, with RMSprop.
Continuum Layers: The foundational units of CNNs are convolutional layers. We
examine the specifics of convolutional procedures, such as how filters and feature
maps are used. We also look at various padding and stride approaches and how they
affect these layers' output dimensions.
Pooling Layers:
The spatial dimensions of the feature maps produced by convolutional layers are
decreased by the use of pooling layers. We cover several pooling techniques and
how they affect the preservation of crucial data while minimizing computing
complexity, including max pooling and average pooling.
Fully Connected Layers:

7
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

The convolutional and pooling layers are often followed by fully connected layers
that categorize the retrieved features. We describe the idea of flattening feature
maps and running them through thick layers for the outcome prediction.
Object Detection:
Methods like region-based CNNs (R-CNN), fast R-CNN, and faster R-CNN can be
used to modify CNNs for object detection tasks. The region proposal algorithms and
the region classification procedure are only two of the ways we give an outline of.
Experimental Setup:
In this part, we go over the training and testing datasets, including ImageNet,
COCO, and Pascal VOC. We go through the data pretreatment procedures utilized
and the data augmentation methods employed to enhance generalization. We also
describe the hyperparameters selected for CNN training.
Evaluation Metrics:
We go over standard assessment criteria for image recognition, including F1-score,
recall, accuracy, and precision. We present metrics for object detection such as mean
Average Precision (mAP) and Intersection over Union (IoU).

Experimental Results
We report the findings from our studies utilizing CNNs to perform picture
identification and object detection tasks. We evaluate the advantages and
disadvantages of several CNN designs, including VGGNet, ResNet, and
InceptionNet, by comparing their performances. We also show the effects of
different training techniques including transfer learning and fine-tuning.
In this section, we will explore the limitations and difficulties of CNNs for
object identification and picture recognition, as well as the main conclusions of our
work. We suggest potential areas for future research to solve these issues and further
develop CNNs' capabilities.

CONCLUSION
We will explore deeper into the techniques and architectures that underlie deep
learning-based object identification and picture recognition systems in this study.
We will examine the theories and methods used in these systems, looking at their
benefits, drawbacks, and prospective directions for further study. It is possible to
create intelligent systems that can perceive, comprehend, and interact with the
visual environment by developing a thorough grasp of these processes. This opens
up a wide range of opportunities for innovation and advancement. Contrarily, object
detection includes not only identifying objects in pictures but also localizing their
places by encircling them with bounding boxes. Deep learning techniques have
significantly improved object detection performance by combining the power of
CNNs with additional components, such as region proposal networks (RPNs) and

8
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

anchor-based mechanisms. These advancements have paved the way for real-time
and highly accurate object detection systems.

REFERENCES
1. Abhinav, A., & Agrawal, A. (2019). A comprehensive survey of deep
learning techniques for image recognition. Journal of Pattern Recognition
and Artificial Intelligence, 32(1), 47-63.
2. Chen, Z., & Gupta, S. (2020). Deep learning for object detection: A
comprehensive review. Journal of Visual Communication and Image
Representation, 68, 102768.
3. Simonyan, K., & Zisserman, A. (2014). Very deep convolutional networks
for large-scale image recognition. arXiv preprint arXiv:1409.1556.
4. Girshick, R., Donahue, J., Darrell, T., & Malik, J. (2014). Rich feature
hierarchies for accurate object detection and semantic segmentation.
Proceedings of the IEEE conference on computer vision and pattern
recognition, 580-587.
5. Ren, S., He, K., Girshick, R., & Sun, J. (2015). Faster R-CNN: Towards real-
time object detection with region proposal networks. Advances in Neural
Information Processing Systems, 91-99.
6. Redmon, J., Divvala, S., Girshick, R., & Farhadi, A. (2016). You only look
once: Unified, real-time object detection. Proceedings of the IEEE
conference on computer vision and pattern recognition, 779-788.
7. He, K., Gkioxari, G., Dollár, P., & Girshick, R. (2017). Mask R-CNN.
Proceedings of the IEEE international conference on computer vision, 2961-
2969.
8. Redmon, J., & Farhadi, A. (2018). YOLOv3: An incremental improvement.
arXiv preprint arXiv:1804.02767.
9. Szegedy, C., Vanhoucke, V., Ioffe, S., Shlens, J., & Wojna, Z. (2016).
Rethinking the inception architecture for computer vision. Proceedings of the
IEEE conference on computer vision and pattern recognition, 2818-2826.
10. He, K., Zhang, X., Ren, S., & Sun, J. (2016). Deep residual learning for
image recognition. Proceedings of the IEEE conference on computer vision
and pattern recognition, 770-778.
11. Lin, T. Y., Dollár, P., Girshick, R., He, K., Hariharan, B., & Belongie, S.
(2017). Feature pyramid networks for object detection. Proceedings of the
IEEE conference on computer vision and pattern recognition, 2117-2125.
12. Redmon, J., & Farhadi, A. (2017). YOLO9000: Better, faster, stronger.
Proceedings of the IEEE conference on computer vision and pattern
recognition, 7263-7271.
13. Liu, W., Anguelov, D., Erhan, D., Szegedy, C., Reed, S., Fu, C. Y., & Berg,
A. C. (2016).
14. Redmon, J., & Farhadi, A. (2018). YOLOv3: An incremental improvement.
arXiv preprint arXiv:1804.02767.
15. Chen, L. C., Papandreou, G., Kokkinos, I., Murphy, K., & Yuille, A. L.
(2018). DeepLab: Semantic image segmentation with deep convolutional
nets, atrous convolution, and fully connected CRFs. IEEE transactions on
pattern analysis and machine intelligence, 40(4), 834-848.

9
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

16. Simonyan K, Zisserman A. Very deep convolutional networks for large-scale


image recognition. arXiv preprint arXiv:1409.1556; 2014
17. 17.Redmon J, Divvala S, Girshick R, Farhadi A. You only look once:
Unified, real-time object detection. In: Proceedings of the IEEE Conference
on Computer Vision and Pattern Recognition. 2016. pp. 779-788
18. 18.Girshick R, Donahue J, Darrell T, Malik J. Rich feature hierarchies for
accurate object detection and semantic segmentation. In: Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition. 2014. p.
580587
19. 19.Girshick R. Fast r-cnn. In: Proceedings of the IEEE International
Conference on Computer Vision. 2015. pp. 1440-1448
20. 20.Chorowski JK, Bahdanau D, Serdyuk D, Cho K, Bengio Y. Attention-
based models for speech recognition. In: Advances in Neural Information
Processing Systems. 2015. pp. 577-585
21. 21.Everingham M et al. The Pascal visual object classes (VOC) challenge.
International Journal of Computer Vision. 2010;88(2):303-338
22. 22.Peixoto HM, Teles RS, Luiz JVA, Henriques-Alves AM, Santa Cruz RM.
Mice Tracking Using the YOLO Algorithm. Vol. 7. PeerJ Preprints; 2019. p.
e27880v1
23. 23.Henriques-Alves AM, Queiroz CM. Ethological evaluation of the effects
of social defeat stress in mice: Beyond the social interaction ratio. Frontiers
in Behavioral Neuroscience. 2016;9:364
24. 24.Jhuang H et al. Automated home-cage behavioural phenotyping of mice.
Nature Communications. 2010;1:68
25. 25.Burgos-Artizzu XP, Dollár P, Lin D, Anderson DJ, Perona P. Social
behavior recognition in continuous video. In: 2012 IEEE Conference on
Computer Vision and Pattern Recognition. IEEE; 2012. pp. 1322-1329
26. 26.Norouzzadeh MS et al. Automatically identifying, counting, and
describing wild animals in camera-trap images with deep learning.
Proceedings of the National Academy of Sciences of the United States of
America. 2018;115(25):E5716-E5725
27. 27.Guo J, He H, He T, Lausen L, Li M, Lin H, et al. GluonCV and
GluonNLP: Deep Learning in Computer Vision and Natural Language
Processing. arXiv preprint arXiv:1907; 2019. p. 04433
28. 28.Deng J, Dong W, Socher R, Li L-J, Li K, Fei-Fei L. ImageNet: A large-
scale hierarchical image database. In: 2009 IEEE conference on computer
vision and pattern recognition. IEEE; 2009. pp. 248-255
29. 29.Chen X-L et al. Remote sensing image-based analysis of the relationship
between urban heat island and land use/cover changes. Remote Sensing of
Environment. 2006;104(2):133-146
30. Ren S, He K, Girshick R, Sun J. Faster r-cnn: Towards real-time object
detection with region proposal networks. In: Advances in Neural Information
Processing Systems. 2015. pp. 91-99.
31. Chinthamu, N., Gooda, S. K., Venkatachalam, C., Swaminathan, S., &
Malathy, G. (2023). IoT- based secure data transmission prediction using
deep learning model in cloud computing. International Journal on Recent
and Innovation Trends in Computing and Communication, 11, 68- 76.
doi:10.17762/ijritcc.v11i4s.6308

10
E3S Web of Conferences 399, 04032 (2023) https://doi.org/10.1051/e3sconf/202339904032
ICONNECT-2023

32. Ashwin, K. V., Kosuru, V. S. R., Sridhar, S., & Rajesh, P. (2023). A passive
islanding detection technique based on susceptible power indices with zero
non-detection zone using a hybrid technique. International Journal of
Intelligent Systems and Applications in Engineering, 11(2), 635-647.
Retrieved from www.scopus.com
33. Raj, R., & Sahoo, D. S. S. . (2021). Detection of Botnet Using Deep
Learning Architecture Using Chrome 23 Pattern with IOT. Research Journal
of Computer Systems and Engineering, 2(2), 38:44. Retrieved from
https://technicaljournals.org/RJCSE/index.php/journal/article/view/31
34. Kamau, J., Goldberg, R., Oliveira, A., Seo-joon, C., & Nakamura, E.
Improving Recommendation Systems with Collaborative Filtering
Algorithms. Kuwait Journal of Machine Learning, 1(3). Retrieved from
http://kuwaitjournals.com/index.php/kjml/article/view/134
35. Ahammad, D. S. K. H. (2022). Microarray Cancer Classification with
Stacked Classifier in Machine Learning Integrated Grid L1-Regulated
Feature Selection. Machine Learning Applications in Engineering Education
and Management, 2(1), 01–10. Retrieved from
http://yashikajournals.com/index.php/mlaeem/article/view/18

11

You might also like