18Vol102No16
18Vol102No16
ABSTRACT
Finding and classifying brain tumors are important parts of medical image analysis that need
advanced deep-learning methods and optimization algorithms. Recognizing the urgent need for accurate
methods in brain tumor diagnosis, we present a comprehensive approach integrating various stages, including
data preprocessing. In this preprocessing phase, we employ techniques like aspect ratio normalization and
resizing to form a standardized dataset. By standardizing image dimensions, we aim to improve subsequent
processes like feature extraction and segmentation, reducing potential distortions. The suggested model is
made by using Convolutional Neural Networks (CNN) to find patterns and traits that make tumor and non-
tumor areas different from each other. To overcome the intricate sections and fine textures during down-
sampling, the proposed model is hybridized with U-Net architecture which gives accurate and robust results
of 98%. Furthermore, the Dice coefficient is measured using Intersection Over Union (IOU) to ensure
whether it is robust to class imbalance. This shows an intuitive interpretation, with higher values of 0.83 and
0.9 indicating strong and better segmentation performance. The model is further developed with VGG-16 to
classify the tumor grades. In terms of accurately segmenting the tumor grades, the learnt relevant
characteristics that are derived from the segmented tumor photos provide a 73% level of satisfaction. In
order to overcome the complexity and over-fitting problems, the Butterfly Optimization algorithm is
hybridized with VGG-16 which gives an enhanced output in classifying the grades. The proposed model
outperforms other Machine Learning (ML) and Deep Learning (DL) methods in tumor and non-tumor
identification and categorization with 99.99% accuracy. To further evaluate the suggested model's
performance, mobility, and energy economy, it is also implemented in JETSON Orin hardware.
Keywords-Deep Learning, Convolutional Neural Networks (CNNs), U-Net, VGG-16, Butterfly Optimization
Algorithm.
6258
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
6259
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
collaboration and creativity in the use of deep instructions. In 2020, Bhanothu et al. [9] introduced
learning to address difficulties in the healthcare a deep convolutional network to classify brain
industry. tumors in MRI scans. MRI scans are essential for
tumor diagnosis, therefore the scientists created
2. RELATED WORKS Faster R-CNN deep learning to automate tumor
identification and localization. It designates tumor
Many deep learning experiments have occurrence areas using Region Proposal Network
targeted brain tumor detection and classification. (RPN), accelerating review and reducing errors.
Dipu et al. [6] used YOLOv5 for object Regional proposal and categorization of glioma,
identification and FastAi for classification to achieve meningioma, and pituitary tumors utilizing VGG-16
good MRI scan accuracy (2021). To increase architecture. Evaluation results showed promising
detection accuracy and resilience, better CNN accuracy ratings of 75.18% for glioma, 89.45% for
architectures for tumor segmentation (2021), hybrid meningioma, and 68.18% for pituitary tumors. The
CNN-RNN models (2021), and ensemble learning mean average precision of automated brain tumor
(2021) have been investigated. Deep learning may detection and classification was 77.60% across all
change brain tumor diagnosis and therapy by classes.
enabling early detection and planning. In 2022, Abirami et al. [11] used
Sankaranarayaanan et al. [7] presented a VGG-16 Generative Adversarial Networks (GANs) to fuse
solution for brain tumor identification and PET and MRI images. They stressed the necessity of
classification at the 2023 International Conference multimodal picture fusion in medical applications to
on AI and Knowledge Discovery in Concurrent improve visual content and medical analysis. They
Engineering. As big data grows in medicine, the noted that multimodal fusion helps doctors view
authors stressed the importance of data analysis and hard and soft tissue, especially in brain imaging
mining on tumor prediction, monitoring, diagnosis, where tumor segmentation is critical. The suggested
and therapy. Brain tumors are aggressive and have method enhances tumor localization accuracy and
low survival rates, thus better diagnostic and diagnostic speed by fusing PET and MRI data. PET
treatment methods are needed. Deep learning, gives functional data while MRI provides
particularly CNNs, was used to solve brain cancer anatomical data to characterize brain cancers. The
diagnosis and treatment issues. Federated Learning proposed GAN-based fusion model blends different
(FL) was used to improve privacy and scalability in modalities to provide fused pictures for medical
centralized data collecting. The study optimized analysis and surgery planning. GAN-based model
model parameters for MRI tumor identification fared well with 0.8551 structural similarity and
using the VGG-16 architecture for brain cancer 2.8059 mutual information. Multimodal medical
detection and a CNN model framework. image fusion enhances brain tumor identification
Experimental findings showed that the suggested and treatment planning in this research. Haq et al.
approach outperformed traditional techniques with [12] suggested DACBT, a deep learning brain tumor
92% accuracy. This research advances brain tumor classification approach using MRI data in IoT
identification and shows how deep learning may healthcare, in 2022. Deep learning-based brain
improve healthcare. tumor classification was proposed to address the
Younis et al. [8] combined deep learning limitations of artificial diagnostic systems for brain
and ensembling learning to show that AI and neural cancer diagnosis in IoT-healthcare systems. DACBT
network algorithms may identify brain tumors early. classified brain tumors using brain magnetic
Segmentation improved brain tumor detection and resonance (MR) imaging data and an upgraded
categorization. The study trained models to reliably CNN. Adding data and transfer learning enhanced
identify brain cancers using MRI utilizing the model's categorization. The suggested model
convolutional neural networks (CNNs) and VGG-16 outperformed baseline brain cancer diagnostic
architecture. Popular for its simplicity and speed, methods in IoT-healthcare systems in experiments.
VGG-16 developed convolutional feature maps that This research improves brain tumor classification
were categorized to find tumor locations. The accuracy and efficiency using deep learning.
proposed method was tested on 253 MRI brain Dhakshnamurthy et al. (2024) classified
scans, 155 of which showed cancers. CNN's 96% brain cancers using transfer learning [13].
accuracy and 91.78% F1-score, VGG-16's 98.5% Understanding the time-consuming nature of brain
and 92.6%, and the ensemble model's 98.14% tumor identification and the limitations of
outperformed classical methods. The project closes conventional techniques in managing expanding
with deep learning brain tumor detection data volume, the scientists researched deep learning
6260
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
approaches to build automated systems for precise of tumors, categorization of tumors, segmentation of
and efficient diagnosis. The study found AlexNet, tumors, and prediction of survival rates. It is
VGG16, and ResNet-50 effective. The authors then necessary to collect data, perform preprocessing, and
created a hybrid VGG16–ResNet-50 model with make improvements to the Convolutional
99.98% accuracy, sensitivity, specificity, and F1 Normalized Mean Filter in order to implement the
score. The framework can accurately identify approach for the BraTS dataset. Multi-class
cerebral neoplasms when compared to other models, classification is accomplished by the utilization of
showing the promise of transfer learning in medical the cutting-edge DBT-CNN classifier model. A
imaging for brain tumor identification and survival rate prediction is made using a logistic
classification. Dhiman et al. (2022) proposed a regression model after the tumor has been delineated
machine learning-based hybrid CNN model for and features have been extracted using RU-Net2+.
medical image processing tumor detection [14]. The When it came to classification accuracy, tumor
study quickly extracted valuable malignant tumor segmentation precision, and survival rate prediction,
oncology medical data from electronic clinical the experimental findings were superior to the
medical records. The study provided a standard standards that are now in place. LGG tumors had a
extraction method for initial tumor site, size, and classification accuracy of 99.28%, whereas HGG
metastatic sites to address tumor-related medical malignancies had a classification accuracy of
events. Also, key-based and pseudo-data-generation 99.51%. The HGG had a tumor segmentation
algorithms improved transfer learning across tumor- accuracy of 98.39%, whereas the LGG had a 99.1%
related medical event extractions. Multiple accuracy. Long-term patient survival rates were
CCKS2020 dataset experiments demonstrated that predicted by the RU-Net2+ algorithm to be 85.71%,
the recommended approach placed third in clinical medium-term survival rates to be 72.72%, and short-
medical event extraction and assessment of the term survival rates to be 61.54%, with Mean Squared
electronic medical record. Errors of 0.13, 0.21, and 0.31 respectively. These
Deep learning and machine learning were findings provide medical practitioners with
utilized by Senan et al. in 2022 in order to investigate information that assists them in making decisions
the early identification of brain tumors using MRI regarding the treatment of brain tumors and
images [15]. As a result of the fact that cancer is one demonstrate that automated brain tumor
of the most aggressive and life-threatening diseases, identification may enhance patient care.
the purpose of the study was to improve patient Srinivasan et al. (2023) used deep learning
survival by properly detecting brain tumors. With to grade cancers from brain magnetic resonance
advances in deep and machine learning, computer- imaging [17]. In order to increase tumor
aided diagnostic systems can help doctors make identification accuracy, the study developed a
accurate diagnoses. The study integrated deep unique automated detection and classification
learning methods like AlexNet and ResNet-18 with approach. MRI scans were pre-processed,
classic machine learning methods like SVM for segmented, feature-extracted, and classified. During
brain tumor classification and diagnosis. Brain the pre-processing stage of magnetic resonance
tumor MRI pictures were improved using the imaging (MRI), an adaptive filter was utilized to
average filter. Deep convolutional layers extracted eliminate background noise. The local-binary grey
robust and crucial deep features, followed by level co-occurrence matrix (LBGLCM) was utilized
SoftMax and SVM classification. The 3,060 pictures for feature extraction, and enhanced fuzzy c-means
were sorted into three tumor kinds and one normal clustering was utilized for image segmentation. The
class. All systems performed well, but the images obtained from an MRI were categorized as
AlexNet+SVM hybrid approach had the highest either normal or glioma using a convolutional
accuracy (95.10%), sensitivity (95.25%), and recurrent neural network (CRNN). In terms of
specificity (98.50%). This study shows that hybrid accuracy, specificity, and sensitivity, the
methods can detect brain tumors early, advancing recommended method for classifying brain tumors
medical imaging and diagnostics. performed better than BP, U-Net, and ResNet. It
RU-Net2+ is a deep learning system that achieved 98.17 percent, 91.34%, and 98.79 percent
was created by Zaitoon and Syed (2023) for the respectively. In order to demonstrate that the CRNN
purpose of brain tumor segmentation and survival technique enhanced the accuracy of brain image
rate prediction [16]. The process of diagnosing and classification, the research utilized 620 testing and
treating brain tumors is challenging, but deep 2480 training MRI images taken from the
learning may make it possible to automate the REMBRANDT dataset that were employed.
process. The framework incorporates the detection
6261
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
ZainEldin et al. (2023) classified brain revolutionize medical imaging analysis [6–17].
cancers using deep learning and sine-cosine fitness However, despite these advancements, several gaps
gray wolf optimization [26]. Due to increased patient in the field remain to be addressed. One such
data volume, the study intended to develop reliable technical gap is the need for more robust and
and efficient automated brain tumor diagnosis generalizable deep learning models that can
methods, which are time-consuming and rely on effectively handle diverse patient populations and
radiologists. Pre-trained CNN models are used for imaging modalities. Additionally, there is a pressing
quicker brain tumor identification and segmentation need to develop more effective and secure federated
using deep learning. BCM-CNN used ADSCFGWO learning frameworks to address privacy concerns
to tune CNN hyperparameters. Hyperparameter associated with medical data sharing [7].
adjustment and model training with Inception- Furthermore, the establishment of standard
ResnetV2, a popular pre-trained model, provided evaluation protocols and benchmark datasets is
binary outcomes (0: Normal, 1: Tumor) to improve essential to enable fair comparisons between
brain tumor diagnosis. ADSCFGWO improved different automated brain tumor analysis systems
hyperparameters utilizing sine cosine and grey wolf [7]. Lastly, extensive validation studies are required
strengths. BCM-CNN obtained 99.98% accuracy on to evaluate the practicality and clinical utility of
the BRaTS 2021 Task 1 dataset, indicating these automated systems in real-world settings [8].
hyperparameter change increases CNN To overcome these technological shortcomings, our
performance. Butterfly optimization algorithm research provides an improved deep learning
(BOA) is a nature-inspired metaheuristic for global framework for brain tumor detection and
optimization presented by Arora and Singh (2019) classification using advanced methodologies. The
[28]. Computer scientists discovered more efficient approach uses Convolutional Neural Networks
ways to solve multidimensional and multimodal (CNNs) for precise tumor detection and hybrid
real-world problems. Metaheuristic algorithms architectures like U-Net to handle complicated
inspired by nature outperform traditional approaches sections and fine textures during down-sampling for
for comparable problems. BOA uses fragrance like robust segmentation [13]. Additionally, VGG-16 is
butterflies to discover nectar and mates. It was tested utilized for tumor grade classification, with relevant
on 30 benchmark functions and compared to other features extracted from segmented tumor images to
metaheuristic algorithms. BOA also addressed three achieve high accuracy [8]. To overcome complexity
old technical problems: spring, welded beam, and and overfitting issues, the Butterfly Optimization
gear train design. BOA was more efficient than algorithm is hybridized with VGG-16, enhancing the
previous metaheuristic algorithms, implying a global model's performance in classifying tumor grades
optimization revolution. Makhadmeh et al. (2023) [26]. By addressing these technical gaps, our
conducted a study of the butterfly optimization research contributes to the advancement of
algorithm (BOA), which is a metaheuristic swarm- automated brain tumor analysis systems, offering
based optimization approach that has proved to be improved accuracy and efficiency in tumor detection
successful in recent times [29]. Scholars have and classification. Moreover, the deployment of our
studied the BOA owing of its few adaptive model on JETSON Orin hardware ensures its
characteristics and good exploration-exploitation performance, portability, and energy efficiency,
balance. In a short time, BOA has been widely used further enhancing its practical applicability in
for optimization issues in several fields. The study clinical settings. Through multidisciplinary
begins by explaining BOA's fundamentals and collaboration and innovative research, we aim to
optimization ideas. It then examines BOA's translate these advancements into therapeutically
mathematical model and motivation, using an useful solutions while upholding patient privacy,
example to demonstrate its usefulness. Adaptation data security, and regulatory compliance.
forms divide the examined research into original,
modified, and hybridized. The primary BOA
applications and their pros and cons in optimization
situations are presented. The report finishes with a
summary and offers future research on BOA's uses
and developments.
6262
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
3. METHODOLOGY
3.1 Proposed Methodology
Figure 1 shows the proposed Work Flow Br35H[28], figshare[30], and the SARTAJ[31]
of Brain Tumor Detection Model. The proposed dataset for brain tumor identification and
methodology integrates Convolutional Neural classification. This dataset is well prepared and
Networks (CNNs) with U-Net for segmentation annotated to allow thorough study and model
and VGG-16 architecture with Butterfly training; it contains 7023 MRI pictures of the
Optimization for feature extraction. Data human brain. This dataset offers a broad picture of
preprocessing ensures standardization and noise brain tumor forms and pathophysiology. It
removal in brain MRI images. CNN+U-Net comprises four separate classes: glioma,
facilitates precise tumor segmentation by meningioma, no tumor, and pituitary. In particular,
integrating low and high-level features. VGG-16 the no-tumor class guarantees a complete depiction
extracts intricate image features, optimized by of brain areas free of tumors by using pictures
Butterfly Optimization for reduced complexity. drawn entirely from the Br35H dataset. Figure 2
Classification distinguishes tumor and non-tumor shows the overview of Brain Tumor Data.
regions using extracted features. Evaluation Upon careful examination, it was
metrics and clinical validation ensure the efficacy observed that the glioma class images from the
and applicability of the proposed framework in SARTAJ dataset exhibited categorization
accurate brain tumor diagnosis and treatment inconsistencies, prompting their exclusion from
planning. the dataset. To address this issue, the glioma class
images were replaced with those sourced from the
3.1.1 Dataset Description figshare repository, ensuring the integrity and
Our research relies on a dataset[29] that is accuracy of class labels for model training and
produced from a mixture of three main sources: evaluation.
6263
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
6264
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Table 1 presents the attributes encompassed where Width and Height denote the dimensions of
within the brain tumor. MRI dataset utilized for the image. This normalization technique ensures that
detection and classification tasks. The attributes are images are proportionally scaled, enhancing the
categorized into two main sections: Basic Image efficacy of subsequent feature extraction processes.
Features and Content Related Features. 3.1.4 Feature Extraction:
The distribution of images across classes is as Feature extraction is a crucial step in our
follows: methodology, aimed at capturing discriminative
➢ Glioma: 300 files spatial and spectral features from brain tumor
➢ Meningioma: 306 files images. Two main algorithms are employed for
➢ No tumor: 405 files feature extraction: the Multi-task Network and the
Global Net architecture.
➢ Pituitary: 300 files
3.2 Multi-task Network:
Additionally, a subset of 1,311 images from the
The Multi-task Network utilizes three
dataset is reserved for use in testing models. By
concurrent convolutional neural networks (CNNs) to
doing so, the model's performance may be validated
extract and process features with resilience. Table 2
using data that has not been seen before. The
presents a comprehensive summary of the design,
collection includes meticulously annotated images
specifying the quantity of filters, kernel sizes,
that indicate the presence or categorization of tumors
strides, and activation functions utilized in each
based on ground truth. This facilitates the
layer.
development of robust and precise deep-learning
𝑍 =𝑊∗𝑋+𝑏
models for the processing of medical images.
where Z represents the output feature map, W
Further, we manually categorized the data into the
denotes the convolutional filter weights, X denotes
following groups: tumor, tumor with grade II, tumor
the input image, and b represents the bias term. By
with grade III, tumor with grade IV, and no-tumor.
leveraging multiple CNNs in parallel, our model
Improved accuracy in tumor classification and
effectively captures diverse spatial and spectral
grading is a direct result of this manual separation,
features inherent in brain tumor images, enabling
which in turn allows for the development of more
comprehensive analysis and classification.
trustworthy deep-learning models for the detection
3.3 Global Net (U-Net) Architecture:
and classification of brain tumors.
The Global Net architecture, based on the
Our study results will be reliable and
renowned U-Net architecture, is specifically
informative because we used this vast dataset that
designed for semantic segmentation tasks. Table 3
was carefully picked. More than that, it enables us to
outlines the architectural details, including the
build therapeutically useful, high-powered deep-
number of filters, kernel sizes, strides, and activation
learning models for detecting and categorizing brain
functions employed in each layer. This architecture
cancers.
comprises both downsampling and upsampling
3.1.2 Feature Extraction & Preprocessing
layers, facilitating feature extraction and high-
In the early stages of our method, it is very
resolution feature map reconstruction.
important to carefully prepare the input data so that
By incorporating U-Net architecture into
our deep learning models can find and label brain
our model, we enhance its ability to capture intricate
tumors as well as possible. This section outlines the
structural details and spatial relationships within the
preprocessing techniques employed, followed by an
brain tumor images, thereby improving
in-depth discussion of the feature extraction process
segmentation accuracy and classification
using the Multi-task Network and the Global Net
performance.
architectures.
Together, the preprocessing techniques and
3.1.3 Preprocessing Techniques:
feature extraction algorithms outlined in this section
Before feature extraction, input MRI
form the foundation of our robust and efficient
images undergo preprocessing steps to standardize
approach to brain tumor detection and classification.
their aspect ratios and dimensions. Aspect ratio
By optimizing both data representation and feature
normalization and resizing techniques are applied to
extraction processes, we aim to develop a highly
ensure uniformity across the dataset, mitigating
accurate and clinically relevant deep learning model
potential distortions and facilitating consistent
capable of accurately identifying and classifying
model input. Mathematically, aspect ratio
brain tumors with unprecedented precision and
normalization is represented as:
𝑊𝑖𝑑𝑡ℎ reliability.
𝐴𝑠𝑝𝑒𝑐𝑡 𝑅𝑎𝑡𝑖𝑜 =
𝐻𝑒𝑖𝑔ℎ𝑡
6265
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Table 2: Multi-Task Network and U-Net Architecture for Brain Tumor Segmentation
6266
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Table 2 presents the Multi-Task Network In order to find the best answers, it mimics the way
and U-Net Architecture for Brain Tumor butterflies naturally move through their
Segmentation. This combined table represents the surroundings. By mimicking the dynamics of
multi-task network architecture and the U-Net butterfly movements, this algorithm offers a
architecture used in the proposed system for brain powerful approach to solving complex optimization
tumor segmentation. problems, including brain tumor detection and
classification in medical imaging analysis. At the
3.4 Butterfly Optimization Algorithm start of the optimization process, the search area is
The Butterfly Optimization algorithm is a filled with a random group of butterflies. Each
new type of metaheuristic optimization method that butterfly is a possible answer to the optimization
was inspired by the way butterflies flap their wings. problem. The positions of these butterflies
6267
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
𝑥 (𝑡 + 1) = 𝑥 (𝑡) + 𝑣 (𝑡)
𝑣 (𝑡 + 1) = 𝑣 (𝑡) + 𝑐 . 𝑟 . (𝑥 ∗ − 𝑥 (𝑡))
+ 𝑐 . 𝑟 . (𝑥 ∗∗ − 𝑥 (𝑡))
Where 𝑥 (𝑡) and 𝑣 (𝑡) denote the position and
Figure 3: Workflow Design of Butterfly Optimization velocity of the 𝑖 − 𝑡ℎ butterfly along the 𝑑-th
Algorithm dimension at time 𝑡,respectively.𝑥 ∗ and 𝑥 ∗∗
represent the local and global best positions
A distinguishing feature of the Butterfly encountered by the butterfly,𝑐 and 𝑐 are
Optimization algorithm is its ability to perform both acceleration coefficients, and 𝑟 and 𝑟 are random
local and global search. Local search involves numbers sampled from uniform distributions. The
refining solutions in the vicinity of promising positions of butterflies are updated iteratively based
regions, leveraging the gradient information on their velocities, ensuring convergence towards
obtained from the objective function evaluations. regions of higher fitness. The position update
Meanwhile, global search enables butterflies to equation is given by:
explore distant regions of the search space, thereby
𝑥 (𝑡 + 1) = 𝑥 (𝑡) + 𝑣 (𝑡)
6268
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Algorithm 1
Step 1: Generate random positions for each butterfly within the search space.
Step 2: Evaluate the objective function for each butterfly's position.
Step 3: Update the position and velocity of each butterfly using:
Butterfly movement equation:𝑥 (𝑡 + 1) = 𝑥 (𝑡) + 𝑣 (𝑡)
Velocity update equation:
𝑣 (𝑡 + 1) = 𝑣 (𝑡) + 𝑐 . 𝑟 . (𝑥 ∗ − 𝑥 (𝑡)) + 𝑐 . 𝑟 . (𝑥 ∗∗ − 𝑥 (𝑡))
𝑐 and 𝑐 are acceleration coefficients, 𝑟 and 𝑟 are random numbers.
Step 4: Perform local search around the best positions.
Step 5: Explore new regions in the search space.
Step 6: Update the positions of butterflies based on their velocities.
Termination:
Step 7: Repeat steps 2-5 until the termination criterion is met.
The above algorithm 1 is the Butterfly Our initial step involves adapting the
Optimization Algorithm (BOA). It initializes a conventional VGG-16 architecture to accommodate
swarm of butterflies in a search space and evaluates reduced weights obtained through Butterfly
their positions against an objective function. Optimization. Mathematically, this modification can
Through iterative updates guided by acceleration be expressed as:
coefficients and random numbers, butterflies 3.7 Weight Adjustment
navigate toward promising solutions while exploring 𝑊 =𝑊 ×𝛼
new regions. This process involves both local Where 𝑊 represents teh original weights of
refinement and global exploration to optimize the VGG-16, and 𝛼 denotes the weight reduction factor
objective function. Positions are continuously obtained through Butterfly Optimization.
updated based on butterfly velocities until a Using the Butterfly Optimization
termination criterion is met, leading to efficient technique, the VGG-16 model's weights are
optimization of complex problems. optimized iteratively. To fine-tune the model's
parameters for the goal of brain tumor detection and
3.5 Implementing Hybrid Visual Geometry classification, this optimization procedure involves
Group-16 with Reduced Weights via Butterfly modifying the weights depending on their
Optimization contribution to the overall classification accuracy.
In our pursuit of refining the efficacy of Our goal is to improve the model's diagnosis
brain tumor detection and classification, we accuracy by repeatedly tweaking the weights until it
implement a novel approach leveraging the synergy can better distinguish between tumor and non-tumor
between the renowned Visual Geometry Group-16 areas.
(VGG-16) architecture and the Butterfly Mathematically, the weight adjustment can be
Optimization algorithm. This fusion aims to enhance formulated as:
the discriminative power of VGG-16 while 𝑊 = 𝑊 + 𝛥𝑊
mitigating the computational burden associated with Where 𝑊 represents the current weights and 𝛥𝑊
its standard implementation. denotes the change in weights computed using
3.6 Architecture Modification of VGG-16 Butterfly Optimization.
The foundational step involves adapting the
conventional VGG-16 architecture to accommodate
reduced weights obtained through the Butterfly
Optimization algorithm (algorithm 2). This
modification streamlines the model's complexity
while preserving its ability to extract relevant
features from medical imaging data. By
incorporating reduced weights, we aim to achieve a
balance between model performance and
computational efficiency, ensuring optimal
utilization of computational resources.
6269
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Algorithm 2
Figure 4: Sequence Flow For The Proposed Hybrid Visual Geometry Group-16 With Reduced Weights Via Butterfly
Optimization
The figure 4 illustrates the sequence flow and classification, this figure offers a structured
for the proposed Hybrid Visual Geometry Group-16 visual representation of the process.
(VGG-16) with Reduced Weights via Butterfly 4. EXPERIMENTATION RESULTS AND
Optimization. It outlines the sequential interactions DISCUSSION
within the algorithm, commencing with the Experiments were conducted on a dataset
initialization phase, where the VGG-16 architecture consisting of 7023 brain MRI images to see how
is initialized, and Butterfly Optimization is initiated. well the suggested hybrid method worked.
The reduction factor obtained from Butterfly
Optimization is then applied to update the weights of 4.1 Experimental Setup
VGG-16. The design then goes on to classify and A computer with an Intel Core i7 processor,
extract features, and finally, during training, it 16 GB of RAM, and an NVIDIA GeForce GTX 1080
updates the network's biases and weights using Ti graphics card was used for the experiments, as
labeled datasets of brain tumors. Lastly, validation was an NVIDIA Jetson AGX Orin module, a low-
datasets are used to evaluate the model's power AI computing platform designed for edge AI
performance in terms of classification accuracy. In applications. The proposed approach was
order to better understand how to apply the implemented using the Keras deep learning
suggested hybrid method for brain tumor detection framework with a TensorFlow backend. The
approach's adaptability and potential for deployment
6270
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
in multiple situations were demonstrated by our platform. Figure 5 shows the Hardware
evaluation of its performance on both a high- Specification Used for Experimentation.
performance computer and a low-power edge AI
6271
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
The outcomes of tuning the shows the Optimized Parameters for the Proposed
hyperparameters of a deep learning model through Model.
the utilization of various combinations of dropout Table 3: Optimized Parameters for the Proposed Model
rates, learning rates, loss functions, and optimizers Optimized
Hyperparameter
are illustrated in Figure 7. The plot displays the Value
relationships between these hyperparameters and the Learning Rate 0.00123
resulting model performance, measured by the loss Batch Size 32
function. The color bar indicates the loss value, with Number of Epochs 150
lower values indicating better model performance. Dropout Rate 0.25
The scatter points represent individual Weight Decay 0.0001
hyperparameter combinations, with the size of each Momentum 0.9
point corresponding to the dropout rate. The y-axis Activation Function (Hidden
ReLU
denotes the optimizer executed, while the x-axis Layers)
represents the learning rate. The plot can be used to Activation Function (Output
Sigmoid
Layer)
identify the most promising hyperparameter
Number of Hidden Layers 3
combinations and to visualize the relationships
between the different hyperparameters. Table 3 Number of Neurons (Hidden
128, 64, 32
Layers)
Figure 8 Accuracy Comparison of Deep Learning Models with and without Butterfly Optimization on Local PC and
Nvidia Jetson AGX ORIN
6272
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Figure 8 shows the accuracy Comparison of and Proposed VGG-16. In this graph, we can
Deep Learning Models with and without Butterfly observe how well each model performed with and
Optimization on Local PC and Nvidia Jetson AGX without Butterfly Optimization, a unique
ORIN. On two separate hardware platforms, a local optimization method developed to boost the
PC (NVIDIA GTX 1080) and an Nvidia Jetson AGX efficiency of DL models. The outcomes show that
ORIN, this graphic compares the accuracy of Butterfly Optimization is a powerful tool for
multiple deep learning models, such as Simple CNN, enhancing the precision of deep learning models on
Mix-pooling CNN with FCRF Segmentation, Fine- both hardware types.
tuned YOLOv7, BCM-CNN, Inception+resnetV2,
Figure 9 shows the Performance Metrics of whose performance metrics are compared. Among
Deep Learning Models. Simple CNN, Mix-pooling the measures are F1 Score, Accuracy (ACC),
CNN with FCRF Segmentation, Fine-tuned Precision, and Recall. The outcomes prove that
YOLOv7, BCM-CNN, Inception+resnetV2, and Butterfly Optimization is a powerful tool for
Proposed VGG-16 (with and without Butterfly enhancing the efficiency of deep learning models.
Optimization) are some of the deep learning models
6273
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
This figure 10 compares the performance of (IOU), Dice Coefficient, and Jaccard Distance. The
various segmentation algorithms, including U-Net, findings show that the CNN+UNet model works
FCN, SegNet, CNN+UNet (Proposed), well, with competitive performance on Dice
CNN+ResNet50, DeepLab, and PSPNet. The Coefficient and Jaccard Distance and excellent
metrics include Accuracy, Intersection over Union accuracy and IOU.
6274
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Figure 11 illustrates the output predicted by tumor regions with high fidelity. Furthermore, the
our proposed system for brain tumor detection and classification of tumor grades using VGG-16,
classification, run on a local PC. The image enhanced by the Butterfly Optimization algorithm,
showcases the segmented tumor regions accurately ensures accurate characterization of tumor severity.
identified by the optimized deep learning The integration of these methodologies results in a
framework. Utilizing advanced techniques such as comprehensive and efficient system for brain tumor
Convolutional Neural Networks (CNNs) and hybrid analysis, demonstrating the feasibility of deploying
architectures like U-Net, the system achieves precise deep learning models on local hardware for medical
segmentation, distinguishing tumor areas from non- imaging diagnostics.
6275
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
Figure 12 presents the predicted outputs of tumor segmentation and grade classification. The
our brain tumor detection and classification system predicted outputs include critical, mild, and
deployed on the NVIDIA Jetson AGX Orin module. moderate-grade tumors, with corresponding sizes of
The image illustrates accurately segmented tumor 18360, 2688, and 4256, respectively. Additionally,
regions, showcasing the system's robust the absence of tumors is accurately identified,
performance in real-world hardware settings. ensuring reliable diagnosis and efficient deployment
Utilizing advanced deep learning techniques and in clinical environments. Table 4 shows the
hybrid architectures like U-Net and VGG-16 with Segmentation Models with Data Augmentation and
Butterfly Optimization, the system achieves precise Computational Efficiency.
6276
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
6277
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
6278
Journal of Theoretical and Applied Information Technology
31st August 2024. Vol.102. No. 16
© Little Lion Scientific
6279