Deep Learning Based Classification On Fruit Diseases An Application For Precision Agriculture
Deep Learning Based Classification On Fruit Diseases An Application For Precision Agriculture
Deep Learning Based Classification On Fruit Diseases An Application For Precision Agriculture
DOI:10.32604/cmc.2020.012945
Article
1
Department of Computer Science, HITEC University, Taxila, 47040, Pakistan
2
Department of Computer Science, COMSATS University Islamabad, Wah Campus, Wah Cantt, 47080, Pakistan
3
Department of Computer Science, COMSATS University Islamabad, Attock Campus, Attock, Pakistan
4
Department of Computer Science and Engineering, Soonchunhyang University, Asan, South Korea
5
Department of Mathematics and Computer Science, Faculty of Science, Beirut Arab University, Beirut, Lebanon
Corresponding Author: Seifedine Kadry. Email: [email protected]; [email protected]
Received: 18 July 2020; Accepted: 05 October 2020
Abstract: Agriculture is essential for the economy and plant disease must be
minimized. Early recognition of problems is important, but the manual inspection
is slow, error-prone, and has high manpower and time requirements. Artificial
intelligence can be used to extract fruit color, shape, or texture data, thus aiding
the detection of infections. Recently, the convolutional neural network (CNN)
techniques show a massive success for image classification tasks. CNN extracts
more detailed features and can work efficiently with large datasets. In this work,
we used a combined deep neural network and contour feature-based approach to
classify fruits and their diseases. A fine-tuned, pretrained deep learning model
(VGG19) was retrained using a plant dataset, from which useful features were
extracted. Next, contour features were extracted using pyramid histogram of
oriented gradient (PHOG) and combined with the deep features using serial based
approach. During the fusion process, a few pieces of redundant information were
added in the form of features. Then, a “relevance-based” optimization technique
was used to select the best features from the fused vector for the final classifica-
tions. With the use of multiple classifiers, an accuracy of up to 99.6% was
achieved on the proposed method, which is superior to previous techniques.
Moreover, our approach is useful for 5G technology, cloud computing, and the
Internet of Things (IoT).
1 Introduction
Early plant disease detection is economically vital. State-of-the-art computer vision and machine-
learning can recognize and identify infections at a very early stage, thus reducing disease spread and
increasing cure rates. Some disease symptoms appear too late for effective action to be taken, but in other
This work is licensed under a Creative Commons Attribution 4.0 International License, which
permits unrestricted use, distribution, and reproduction in any medium, provided the original
work is properly cited.
1950 CMC, 2021, vol.66, no.2
cases diseases are apparent at a very early stage [1]. Apples and cherries, among other fruits, can develop
powdery mildew, rust, and black rot that are often identified during manual inspection, or in the
laboratory. Manual inspection is slow, error-prone, and has high manpower and time requirements.
Artificial intelligence can be used to extract fruit color, shape, or texture data, thus aiding the detection of
infections. Over 374 million fruits are harvested annually; in 2011, 50–55% of these were grown in Asia
[2]. Fruit appearance affects customer decisions, and varies by the point of sale (hawkers, shop, malls,
etc.). Most fruits are packed by hand, which can affect the quality and imposes a manpower cost.
Automated techniques must be fast, work effectively in poor light, meet various client requirements, and
be able to handle new fruit varieties. Low-quality fruit can be sent to clients who prefer it for juicing and
pickling; the rotten parts are removed and the remainder is sliced or pressed. High-quality fruits
command a good price. A robot that can distinguish among fruit types and quality is required.
Many automated strategies based on digital image processing are used to check fruit type and quality [3].
Computer vision techniques can be used to detect shapes, and have applications including breast cancer
classification [4], object detection [5,6], and medical image evaluation [7,8]. Few machine-learning
techniques have been used to analyze fruit type, ripeness, or disease, but could aid automatic harvesting,
fruit-picking, and fruit identification.
Mobile devices place unprecedented demand on computing, edge computing, and the Internet of Things
(IoT). In 2015, monthly data consumption was 3.7 exabytes; content providers (CPs) such as Netflix, Vimeo,
and YouTube predict that the figure will reach 30.6 exabytes by the end of 2020 [9]. Capacity shortage and
data consumption will become significant issues for mobile network operators (MNOs). The small cells
(SCs) of 5G systems have many academic and industrial applications. The SCs are backhaul-connected to
the core networks of mobile operators. Mobile edge caching avoids possible bottlenecks. The primary
purpose of caching is to ensure fast and adaptable linkages between limited backhaul and radio
connections [10]. 5G-enabled mobile edge computing (MEC) efficiently handles many SCs and their
associated cache memories. In an MEC, services required by increasing numbers of users run inside a
radio access network (RAN). MECs not only improve user experience, but also reduce the time taken to
transfer data from a server to an end-user. Increases in cache memory, the number of CPs, and SC spatial
density enhance the performance of mobile edge caching systems [11].
In a store, a clerk identifies and prices fruit. Standardized barcode tags are used for packed items;
typically, fruit is not packed and thus not barcoded. Fruit classification is important for automated juicing
and pricing. Fruits must first be identified, and then graded in terms of quality. In Chandwani et al. [12],
robotic harvesting reduced labor costs and improved fruit quality. Robots are increasingly used for
harvesting. Automated fruit classification is less well-established, and its efficacy is hotly debated; there
is a need to rapidly identify and price fruit, which can be achieved by cashiers [13]. Automated fruit-
packing and grading, via machine-learning algorithms, has attracted much interest [14]. The typical
features extracted are color, texture, and shape. Trained classifiers may be restricted to unacceptably
simple classifications.
A plant disease classification technique was proposed using pretrained GoogleNet and AlexNet models
to extract deep features. The models were tested using the publicly available Plant Village dataset; the
classification accuracy was 99.3% [15]. Another model was used to assess rice fields, baaed on high-
definition images captured by unmanned aerial vehicles (UAVs). A deep convolutional neural network
(DCNN) was used to extract features and the data were labeled by multiple classifiers. The overall
classification accuracy was 72% [16]. In another study, a hybrid classification/detection model evaluated
apple diseases. This method extracted and combined color, shape, and texture data and fed them to a
Support Vector Machine (SVM) classifier; the maximum accuracy was 95.94% [17]. Tomato diseases
were detected and classified using a robust DCNN model with a metaheuristic architecture including a
CMC, 2021, vol.66, no.2 1951
single-shot multi-box detector, a region-based fast CNN, a regional CNN (which was even faster), and a data
augmentation module; the classification accuracy was 85.9% [18].
Given the enormous increase in fruit production, an accurate classification method is vital; manpower
demands must be reduced and efficiency increased [19]. Machine-learning approaches, which employ
computational and mathematical models trained on previous data, are essential and are now replacing
manual methods in many industries. Previously, gas sensors, infrared imaging, and liquid chromatography
devices were used to scan fruit. However, these methods require skilled operators and are expensive.
Image-based fruit classification avoids these drawbacks, and requires only a digital camera and efficient
model. The neural network (NN)-based artificial bee colony (ABC) algorithm was used to classify four
types of fruits, with an accuracy of 76.39% [20]. A hybrid technique using an enhanced fitness-scaled
chaotic ABC, feedforward NNs, and an optimized genetic algorithm had an accuracy of 87.9% [21].
Two-step classification extracted shape, color, and texture data that were fed to a fast NN trained by a
biogeography-based optimization (BBO) algorithm. Principal component analysis (PCA) had a
classification accuracy of 89.11% when applied to 18 different fruits [22]. Fruit texture, shape, and color
data in Hue, Saturation, and Value (HSV) color space were used to evaluate a dataset including
20 different fruit types. The overall classification accuracy was 91:52% [23]. A custom dataset of
1,653 images of 18 fruit types (classes) was used to validate a method that extracted and classified FRFE
features using a single-layer feedforward NN. The accuracy was 88.99% after five-fold cross-validation [24].
Classification of the 971 fruit images of the FIDS30 dataset was 94% accurate using a CNN with
featured parameter optimization. The model had eight layers and classified images with dimensions of
256 256 3 pixels [25]. Fruit images were classified based on edge distances and zoning features,
combined via discrete Fourier transform. The highest accuracy (97:5%Þ was achieved by the neighbor-
joining technique [26]. A shape-based method of fruit classification independent of fruit appearance was
proposed. The technique was validated for seven fruits (210 images) and the overall accuracy was 88–
95% [27]. Another technique used an SVM classifier to classify fruit based on color and textural features;
a classification accuracy of 95.3% was achieved [28]. One real-time classification system trained an
artificial NN using over 1,900 images; the accuracy was 97.4% [29]. The main issues for classifiers are as
follows: i) Extracting the features that are most useful for classifying fruit types and diseases;
ii) Eliminating irrelevant features that compromise accuracy; and iii) reducing the training time.
This article is structured as follows. The objectives and contributions are detailed in Section 2 and the
classification technique is described in Section 3. Experimental results validating the technique are presented
in Section 4, and Section 5 concludes provides the discussion and future research directions.
to ensure the inclusion of relevant features only. No previous system has utilized 5G technology and cloud
computing to classify real-time data with such high accuracy.
The device layer includes mobile phones with cameras that capture images or video frames that are
saved locally and then transferred via 5G to the local edge cache unit. The 5G network contains multiple
SCs connected to separate cache memory units, which are reserved by CPs for MNOs. The MNOs
operate in a multi-tenant environment and store a variety of content at different prices. All available
cache memory slots are allocated to different over-the-top (OTT) CPs. A cache memory slot is reserved
by a single optimized CP; different types of content can be stored based on the particular SC spatial
distribution. This helps to identify the minimum missed cache rate as a function of the purchased slot.
The order in which different types of content are sent over a 5G network is determined by demand and
CP reputation (including the number of CPs with similar reputations).
The resource engine layer optimizes resources by learning about network parameters such as
communication quality, data flow, and network type. This is critical for green communication when big
data must be transferred but bandwidth is limited. The resource engine layer also influences 5G
reliability, scalability, and flexibility. Network operations in data centers that control the cloud
infrastructure are linked to the hardware via the NFV. Intelligent programmable networks, such as SDNs,
CMC, 2021, vol.66, no.2 1953
provide consistent decoupling of network intellect, ultimately enabling the system to request and modulate
network services. The data engine layer uses 5G technology to transfer images to data centers. Human-like
intelligent systems, ML, and DL algorithms learn useful features and make detection/classification decisions.
Large images can be resized to dimensions of 224 224 3 pixels, which enables MNOs to rapidly transfer
them to cache memory. The image with the highest demand is prioritized in the cache memory of the adjacent
SC so that it can be effectively and quickly moved to the data engine for further processing.
fruit classification in this study, the results were outstanding. The normalized colors and gamma values of
input images increase search efficiency. For fruit detection, a small sliding window probes the input
image, region-by-region. The resultant histograms at all grid levels are concatenated to yield spatial
information. We divided the PHOG descriptor into 40 orientation bins between 0 and 360 , and set l ¼ 2
and Q ¼ 40 to give a N 840 feature vector for the N images.
layer as the output layer. Supervised CNN training uses comprehensive datasets such as ImageNet. The
VGG19 model has 16 convolutional layers, 5 pooling layers, 3 FCLs [33], and 2 dropout layers with
dropout probabilities of 0.5. VGG-19 also features an extractor and a classifier. The extractor changes an
image to a vector that can be connected to various classifiers. VGG-19 is very accurate and efficient.
Here, we used the penultimate fully connected “fc7” layer to extract 4; 096 features.
The repetition of every feature within an S gives the total mean value of f i and f j , calculated as:
1 X
RðSÞ ¼ 2 Iðfi ; fj Þ (2)
jS j fi ;fj 2 S
The mRMR technique maps the two different feature values described above as follows:
mRMR ¼ maxs ½ZðS; cÞ RðSÞ (3)
Suppose there are n features and xi denotes the applicability of a fitness function to feature f i . As mi ¼ 1
denotes the features that are present in the globally optimized feature set, and mj ¼ 0 denotes bes those that
are absent, the above becomes:
2 3
P
n P
n
a x x
6 ci xi i¼1;j¼1 ij i j 7
6 7
mRMR ¼ maxs 6i¼1 2 7 (4)
4P n
Pn 5
xi xi
i¼1 i¼1
Theoretically optimal features (in terms of maximum dependency) can be approximated using the
mRMR technique. The information shared by selected features with neighboring features and
1956 CMC, 2021, vol.66, no.2
classification variables are maximized and optimized. The mRMR technique creates a feature vector denoted
by FVS 3,000, which has 3,000 unique features that are fed to the classifiers to ensure accuracy.
4 Experimental Results
4.1 Datasets and Experimental Setup
The Fruit-360 dataset [36] contains 65,429 images of 95 different classes. The dataset is separated into
two parts: a training set of 48,905 images and a testing set of 16,421 images. For the DCNN models, the
minimum number of (random) images per class was set to 300. Ten-fold cross-validation and seven
classification methods were used to test the performance of our method, measured in terms of sensitivity,
precision, area under the curve, the format-and-restore (FNR) rate, and accuracy.
The publicly available Plant Village dataset [37] was also used to test the model in terms of disease
classification performance. The dataset contains several classes of leaf disease; we were interested in the
apple black rot (ABR), apple scab (AS), apple rust (AR), apple healthy (AH), cherry powdery mildew
(CPM), cherry healthy (CH), peach bacterial spot (PBS), and peach healthy (PH) classes. A total of
7,733 images were divided into training and testing datasets using the standard 70/30% split. The
minimum number of (random) images per class was set to 360. To evaluate model performance, the
accuracy, FNR rate, and training time was recorded for all seven classifiers. We used 10-fold cross-
validation to measure performance. All simulations were performed in the MATLAB environment
(2018a; MathWorks, Inc., Natick, MA, USA) using a PC with an i7 core, 16 GB of RAM and a 256 GB SSD.
4.2.2 Discussion
As the updated Fruit-360 dataset was only published in 2018, very few studies have used it. The dataset
proposed in Mureşan et al. [36] was used to train a CNN that employed the TensorFlow library [38]
for deep learning. A six-layered DCNN model was also used; this analyzed input images
100 100 3 pixels in size, and featured four convolutional layers connected by max-pooling layers,
two FCLs, and a final softmax output layer. Various combinations of color transformations were tested;
the [Grayscale, HSI, HSV, and Flips] combination provided accuracy rates of 97.04% and 99.86% for the
test and training datasets, respectively. Another technique used only 25 classes of fruit in the Fruit-
360 dataset; of the five cases that were tested, case 1 had the highest accuracy, of 99.94% [39].
CMC, 2021, vol.66, no.2 1957
A classification technique based on shape, color, and textural features was used to classify 24 fruit classes
(2,400 images) in the Fruit-360 dataset. The kNN classifier achieved an accuracy of 97.5% [40].
However, the accuracy rates in our study were higher, at 99.6% and 99.97% for the test and training data
sets, respectively. The classifier training time was 23.6 s, and the average prediction time was 0.09 s (Tab. 2).
4.3.2 Discussion
To determine the validity of the model, it was compared with existing techniques (Tab. 4). In a previous
study, color, shape, and textural features were extracted to classify Google images of apples as healthy or
diseased (blotches, rot, and scabs); the accuracy was 95.47% [41]. Another DCNN-based technique
distinguished among the apple diseases Alternaria, brown rust, and mosaicism. An AlexNet model was
CMC, 2021, vol.66, no.2 1959
used to evaluate 13,689 images and the accuracy was 97.62% [42]. A DCNN-based approach was used to
classify plant leaf disease in another dataset, with 96.3% accuracy [43]. Finally, DCNN models (AlexNet and
VGG) were used to extract features from the Plant Village dataset. Multi-level fusion followed by entropy
selection provided an overall classification accuracy of 97.80% [44].
Figure 4: Accuracy rates and training loss after fine-tuning of both datasets
1960 CMC, 2021, vol.66, no.2
The results show that a fine-tuned DCNN model converges faster and is more accurate. Fruit
classification does not require preprocessing when a DCNN model is combined with structural features.
Our method is very accurate even when evaluating a very large and challenging dataset.
5 Conclusion
Our hybrid classification method combines a DCNN model with PHOG features to classify fruits using
5G and cloud technology. Images are captured on a smart device and transferred to the cloud via the 5G
network. Large images are resized before transfer, and the fine-tuned hybrid model classifies them. The
combined DCNN and PHOG features are optimized using the mRMR technique. We employed the public
Fruit-360 dataset to determine the validity of our method, which was considerably better than currently
available classifiers. To the best of our knowledge, this is the first attempt to combine DCNN and PHOG
features for fruit classification based on 5G technology and the IoT. Fusion was associated with good
classification accuracy. In the future, we will add features for multi-fruit classification, which should aid
fully automated robotic harvesting.
Funding Statement: This research was supported by the MSIT(Ministry of Science and ICT), Korea, under
the ICAN (ICT Challenge and Advanced Network of HRD) program(IITP-2020-2020-0-01832) supervised
by the IITP (Institute of Information & Communications Technology Planning & Evaluation) and the
Soonchunhyang University Research Fund.
Conflicts of Interest: On the behalf of corresponding author, all authors declare that they have no conflict of
interest.
References
[1] S. S. Abu-Naser, K. A. Kashkash and M. Fayyad, “Developing an expert system for plant disease diagnosis,”
Journal of Artificial Intelligence, 2010.
[2] E. Tsormpatsidis, M. Ordidge, R. G. C. Henbest, A. Wagstaffe, N. H. Battey et al., “Harvesting fruit of equivalent
chronological age and fruit position shows individual effects of UV radiation on aspects of the strawberry ripening
process,” Environmental and Experimental Botany, vol. 74, pp. 178–185, 2011.
[3] P. C. Marchal, D. M. Gila, J. G. García and J. G. Ortega, “Expert system based on computer vision to estimate the
content of impurities in olive oil samples,” Journal of Food Engineering, vol. 119, no. 2, pp. 220–228, 2013.
[4] I. M. Nasir, M. Rashid, J. H. Shah, M. Sharif, M. Y. H. Awan et al., “An optimized approach for breast cancer
classification for histopathological images based on hybrid feature set,” Current Medical Imaging, 2020.
[5] M. Rashid, M. A. Khan, M. Alhaisoni, S. H. Wang, S. R. Naqvi et al., “A sustainable deep learning framework for
object recognition using multi-layers deep features fusion and selection,” Sustainability, vol. 12, no. 12, pp. 5037, 2020.
[6] A. Daniel, K. Subburathinam, B. Muthu, N. Rajkumar, S. Kadry et al., “Procuring cooperative intelligence in
autonomous vehicles for object detection through data fusion approach,” IET Intelligent Transport Systems,
vol. 14, no. 11, 2020.
[7] B. Muthu, C. Sivaparthipan, G. Manogaran, R. Sundarasekar, S. Kadry et al., “IOT based wearable sensor for
diseases prediction and symptom analysis in healthcare sector,” Peer-to-Peer Networking and Applications,
pp. 1–12, 2020.
[8] I. M. Nasir, M. A. Khan, M. Alhaisoni, T. Saba, A. Rehman et al., “A hybrid deep learning architecture for the
classification of superhero fashion products: An application for medical-tech classification,” Computer Modeling
in Engineering & Sciences, vol. 124, no. 3, pp. 1017–1033, 2020.
[9] CISCO, “Index, Global mobile data traffic forecast update, 2015-2020,” White Paper, 2016.
[10] X. Wang, M. Chen, T. Taleb, A. Ksentini and V. C. Leung, “Cache in the air: Exploiting content caching and
delivery techniques for 5G systems,” IEEE Communications Magazine, vol. 52, no. 2, pp. 131–139, 2014.
CMC, 2021, vol.66, no.2 1961
[11] B. Bharath, K. G. Nagananda and H. V. Poor, “A learning-based approach to caching in heterogenous small cell
networks,” IEEE Transactions on Communications, vol. 64, no. 4, pp. 1674–1686, 2016.
[12] V. Chandwani, V. Agrawal and R. Nagar, “Modeling slump of ready mix concrete using genetic algorithms assisted
training of Artificial Neural Networks,” Expert Systems with Applications, vol. 42, no. 2, pp. 885–893, 2015.
[13] B. Zhang, W. Huang, J. Li, C. Zhao, S. Fan et al., “Principles, developments and applications of computer
vision for external quality inspection of fruits and vegetables: A review,” Food Research International, vol.
62, pp. 326–343, 2014.
[14] M. A. Khan, T. Akram, M. Sharif, M. Awais, K. Javed et al., “CCDF: Automatic system for segmentation and
recognition of fruit crops diseases based on correlation coefficient and deep CNN features,” Computers and
Electronics in Agriculture, vol. 155, pp. 220–236, 2018.
[15] S. P. Mohanty, D. P. Hughes and M. Salathé, “Using deep learning for image-based plant disease detection,”
Frontiers in Plant Science, vol. 7, pp. 1419, 2016.
[16] N. C. Tri, T. Van Hoai, H. N. Duong, N. T. Trong, V. Van Vinh et al., “A novel framework based on deep learning
and unmanned aerial vehicles to assess the quality of rice fields,” in Int. Conf. on Advances in Information and
Communication Technology, Cham: Springer, pp. 84–93, 2016.
[17] S. R. Dubey and A. S. Jalal, “Apple disease classification using color, texture and shape features from images,”
Signal, Image and Video Processing, vol. 10, no. 5, pp. 819–826, 2016.
[18] A. Fuentes, S. Yoon, S. C. Kim and D. S. Park, “A robust deep-learning-based detector for real-time tomato plant
diseases and pests recognition,” Sensors, vol. 17, no. 9, pp. 2022, 2017.
[19] A. Khan, J. P. Li, R. A. Shaikh and I. Khan, “Vision based classification of fresh fruits using fuzzy logic,” in
2016 3rd Int. Conf. on Computing for Sustainable Global Development (INDIACom), New Delhi, India,
pp. 3932–3936, 2016.
[20] M. Adak and N. Yumusak, “Classification of e-nose aroma data of four fruit types by ABC-based neural network,”
Sensors, vol. 16, no. 3, pp. 304, 2016.
[21] Y. Zhang, S. Wang, G. Ji and P. Phillips, “Fruit classification using computer vision and feedforward neural
network,” Journal of Food Engineering, vol. 143, pp. 167–177, 2014.
[22] Y. Zhang, P. Phillips, S. Wang, G. Ji, J. Yang et al., “Fruit classification by biogeography-based optimization and
feedforward neural network,” Expert Systems, vol. 33, no. 3, pp. 239–253, 2016.
[23] F. Garcia, J. Cervantes, A. Lopez and M. Alvarado, “Fruit classification by extracting color chromaticity, shape
and texture features: Towards an application for supermarkets,” IEEE Latin America Transactions, vol. 14, no. 7,
pp. 3434–3443, 2016.
[24] S. Wang, Z. Lu, J. Yang, Y. Zhang, J. Liu et al., “Fractional fourier entropy increases the recognition rate of fruit
type detection,” BMC Plant Biology, vol. 16, 2016.
[25] Z. M. Khaing, Y. Naung and P. H. Htut, “Development of control system for fruit classification based on
convolutional neural network,” in 2018 IEEE Conf. of Russian Young Researchers in Electrical and Electronic
Engineering, Moscow, Russia, pp. 1805–1807, 2018.
[26] P. A. Macanhã, D. M. Eler, R. E. Garcia and W. E. M. Junior, “Handwritten feature descriptor methods applied to
fruit classification, ” in Latifi S. (eds.) Information Technology-New Generations. Advances in Intelligent Systems
and Computing, Cham: Springer, vol. 558, pp. 699–705, 2018.
[27] S. Jana and R. Parekh, “Shape-based fruit recognition and classification,” in Int. Conf. on Computational
Intelligence, Communications, and Business Analytics, Singapore: Springer, pp. 184–196, 2017.
[28] R. S. S. Kumari and V. Gomathy, “Fruit Classification using Statistical Features in SVM Classifier,” in 2018 4th
Int. Conf. on Electrical Energy Systems, Chennai, India, pp. 526–529, 2018.
[29] H. S. Choi, J. B. Cho, S. G. Kim and H. S. Choi, “A real-time smart fruit quality grading system classifying by
external appearance and internal flavor factors,” in 2018 IEEE Int. Conf. on Industrial Technology, Lyon, France,
pp. 2081–2086, 2018.
[30] K. Alex, I. Sutskever and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,”
Advances in Neural Information Processing Systems, pp. 1097–1105, 2012.
1962 CMC, 2021, vol.66, no.2
[31] A. Bosch, A. Zisserman and X. Munoz, “Representing shape with a spatial pyramid kernel,” in Proc. of the 6th
ACM Int. Conf. on Image and Video Retrieval, pp. 401–408, 2007.
[32] N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,” in 2005 IEEE Computer Society
Conf. on Computer Vision and Pattern Recognition, San Diego, CA, USA, vol. 1, pp. 886–893, 2005.
[33] K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition. arXiv
preprint arXiv:1409.1556, 2014.
[34] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens and Z. Wojna, “Rethinking the inception architecture for computer
vision,” in Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition, pp. 2818–2826, 2016.
[35] H. Peng, F. Long and C. Ding, “Feature selection based on mutual information: Criteria of max-dependency, max-
relevance, and min-redundancy,” IEEE Transactions on Pattern Analysis & Machine Intelligence, no. 8,
pp. 1226–1238, 2005.
[36] H. Mureşan and M. Oltean, “Fruit recognition from images using deep learning,” Acta Universitatis Sapientiae
Informatica, vol. 10, no. 1, pp. 26–42, 2018.
[37] K. P. Ferentinos, “Deep learning models for plant disease detection and diagnosis,” Computers and Electronics in
Agriculture, vol. 145, pp. 311–318, 2018.
[38] M. Abadi, P. Barham, J. Chen, Z. Chen, A. Davis et al., “Tensorflow: A system for large-scale machine learning,”
in 12th {USENIX} Sym. on Operating Systems Design and Implementation, pp. 265–283, 2016.
[39] S. Sakib, Z. Ashrafi, M. Siddique and A. Bakr, “Implementation of fruits recognition classifier using
convolutional neural network algorithm for observation of accuracies for various hidden layers. arXiv preprint
arXiv:1904.00783, 2019.
[40] L. Rajasekar and D. Sharmila, “Performance analysis of soft computing techniques for the automatic classification
of fruits dataset,” Soft Computing, vol. 23, no. 8, pp. 2773–2788, 2019.
[41] Y. Duan, F. Liu, L. Jiao, P. Zhao and L. Zhang, “SAR Image segmentation based on convolutional-wavelet neural
network and markov random field,” Pattern Recognition, vol. 64, pp. 255–267, 2017.
[42] B. Liu, Y. Zhang, D. He and Y. Li, “Identification of apple leaf diseases based on deep convolutional neural
networks,” Symmetry, vol. 10, no. 1, pp. 11, 2018.
[43] S. Sladojevic, M. Arsenovic, A. Anderla, D. Culibrk and D. Stefanovic, “Deep neural networks based recognition
of plant diseases by leaf image classification,” Computational Intelligence and Neuroscience, vol. 2016, 2016.
[44] T. Akram, M. Sharif and T. Saba, “Fruits diseases classification: Exploiting a hierarchical framework for deep
features fusion and selection,” Multimedia Tools and Applications, pp. 1–21, 2020.