Handwritten Hindi Character Recognition Using Deep Learning Techniques
Handwritten Hindi Character Recognition Using Deep Learning Techniques
net/publication/332114646
CITATIONS READS
5 1,193
2 authors, including:
SEE PROFILE
All content following this page was uploaded by Vijaya Kumar Reddy .R on 12 June 2020.
Keywords: DFFNN, CNN, Softmax classifier, RMSprop and Adam Estimation, Deep Learning.
which was not easily available at that time. Since then, the
I. INTRODUCTION researchers have drawn their attention to finding the
technique which needs less power by converting the images
Deep learning Techniques have been effectively applied to into feature vectors. In the last few decades, a lot of feature
various areas like image classification, speech recognition, extraction technique have been proposed such as HOG
Medical Images detection, face detection, satellite images, (histogram of oriented gradients) [4] and many others
recognizing traffic signs and pedestrian detection and so on. techniques are used as prominent feature extraction methods,
The outcome of deep learning techniques is also prominent, which have been experimented for many problems like
and in some cases the results are superior to human experts image recognition, character recognition, face detection, etc.
[1,2] in the past years. From last few years most of the Feature extraction [5] is one type of dimensionality
problems are also being re-experimented with deep learning reduction technique that represents the important parts of a
techniques with the view to achieving improvements in the large image into a feature vector. These features are
existing findings. Different architectures of deep learning handcrafted and clearly designed by the research
have been introduced in recent years, such as convolutional community. The robustness and performance of these
neural networks, deep networks, and recurrent neural features depend on the skill and the knowledge of each
networks. The entire architecture has shown the expertise in researcher. There are the cases where some vital features
different areas. Character recognition is one of the areas may be unseen by the researchers while extracting the
where machine learning techniques have been extensively features from the image and this may result in a high
experimented. The first deep learning technique, which is classification error. Deep learning inverts the process of
one of the leading machine learning techniques, was handcrafting and designing features for a particular problem
proposed for character recognition in 1998 on MNIST into an automatic process to compute the best features for
database [3]. that problem. A convolutional neural network has multiple
convolutional layers to extract the features automatically.
The deep learning techniques are basically composed of The features are extracted only once in most of the shallow
multiple hidden layers, and each hidden layer consists of learning models, but in the case of deep learning models,
multiple neurons, which compute the suitable weights for the multiple convolutional layers have been adopted to extract
deep network. A lot of computing power is needed to discriminating features multiple times. This is one of the
compute these weights, and a powerful system was needed, reasons that deep learning models are generally successful.
And also in Deep feed forward neural networks the features
are compute automatically by using different number of and used Euclidean similarity measure to classify the
hidden layer in it. numerals. KNN classifier obtained 90 % as recognition rate
whereas PCA obtained recognition rate of 84%. After the
The LeNet [3] is an example of deep convolutional neural comparison it is observed that KNN classifier has shown
network for character recognition. Recently, many other better results as compared to PCA classifier.
examples of deep learning models can be listed such as
AlexNet [2], ZFNet [6], VGGNet [7] and spatial transformer Sonu Varghese K et al. [14] proposed a Novel Tri-Stage
networks [8]. These models have been successfully applied Recognition Scheme for Handwritten Malayalam Character
for image classification and character recognition. Owing to Recognition. In the first stage we are grouping characters
their great success, many leading companies have also into different classes based on the number of corners,
introduced deep models. Google Corporation has made a bifurcations, loops and endings. In the second phase we are
GoogLeNet having 22 layers of convolutional and pooling identifying exact character in the class based on the different
layers alternatively. Apart from this model, Google has also feature extraction technique specially defined for each class.
developed an open source software library named In the third stage we are checking the probability of
Tensorflow to conduct deep learning research. Microsoft occurrence of the current character in the given position
also introduced its own deep convolutional neural network based on defined rules for the formation of words. we are
architecture named ResNet in 2015. ResNet has 152-layer implementing a three stage feature extraction technique
network architectures which made a new record in detection, which uses structural, statistical and moment variant features
localization, and classification. This model introduced a new of the character. Recognition conducted in different stages
idea of residual learning that makes the optimization of improves the efficiency, recognition rate and accuracy of the
different data. given system.
Character recognition is a field of image processing where Parshuram M. Kamble [15] proposed a method for
the image is recognized and converted into a machine- handwritten Marathi character recognition using R-HOG
readable format. As discussed above, the deep learning Feature. The system has been evaluated on a large amount of
techniques and especially convolutional neural networks handwritten Marathi characters. From the results as shown it
have been used for image detection and recognition. It has can be concluded that the use of R-HOG based feature
also been successfully applied on Roman (MNIST) [3], extraction method and FFANN based classification will be
Chinese [9], Bangla [10] and Arabic [11] languages. In this more effective with increased processing speed and
work, a convolutional neural network and Deep Feed accuracy.
Forward neural network is applied for handwritten Hindi
characters recognition. Verma [3] used radial basis function and multilayer
perceptron neural networks for recognizing the handwritten
Rest of the paper is organized as follows; Section 2 contains characters of Devanagari script. Back propagation error
related work. The proposed algorithm is presented in Section algorithm is also used to improve the recognition rate. In this
3. The experimental details and results obtained are proposed system, they compare the results obtained from
presented in Section 4. Section 5 contains the conclusion radial basis Function (RBF) networks and multi-layer
part. perceptron. Dataset consists of 245 samples written by five
different users. The results so obtained show that multilayer
II. RELATED WORK perceptron (MLP) networks performs better than that of
radial basis functions. But MLP networks training time is
Amit Choudhary et al. [12] proposed an Off-Line more as compared to radial basis function networks. Highest
Handwritten Character Recognition using Features Extracted recognition rate so obtained using radial basis function and
from Binarization Technique. This work is to extract multilayer perceptron (MLP) networks are 85.01% and
features obtained by Binarization technique for recognition 70.8% respectively.
of handwritten characters of English language. The
recognition of handwritten character images have been done III. PROPOSED APPROACH
by using multi-layered feed forward artificial neural network
as a classifier. This algorithm delivers outstanding The proposed method is mainly consists of 4 phases steps. In
classification accuracy of 85.62 %. the primary phase, collecting the characters data from kaggle
dataset and gathering images from different users. After
Baheti M. J et al. [13] proposed a comparison of the offline collecting the character data of gray scale images will be
handwritten character recognition system for the isolated pre-processing by checking null and missing values. Using
Gujarati numerals. They used affine invariant moments the normalization techniques to convert the gray level values
based model for the feature extraction. They used KNN to range of 0 to 1 values, and then labelling Hindi characters
classifier and PCA to reduce dimensions of feature space from 0 to 35 with one hot coding which will generate vector
form of data. In the 3rd phase, we extracted the features Many attempts, such as decreasing an adaptive learning rate,
atomically from Different Deep learning algorithm like adopting a big batch size, incorporating a temporal
Convolutional Neural Network (CNN) and Deep Feed decorrelation technique, seeking an analogous surrogate,
Forward Neural Networks (DFFNN) for recognition of etc., In addition, we illustrate that Adam is essentially a
handwritten character system. In finally phase we applied an specifically weighted AdaGrad with exponential moving
optimization technique like Optimizer RMSprop (Root average momentum, which provides a novel perspective for
Mean Square Propagation), Adaptive Moment (Adam) understanding Adam and RMSProp. At last, we validate the
Estimation to get very promising results. The proposed sufficient condition by applying Adam and RMSProp to
method block diagram is shown below in Figure 1. tackle the counterexamples and train deep neural networks.
B. RMSprop Optimizer
The RMSprop (Root Mean Square Propagation) optimizer is
similar to the gradient descent algorithm with momentum.
The RMSprop optimizer restricts the oscillations in the
vertical direction. Therefore, we can increase our learning
rate and our algorithm could take larger steps in the
horizontal direction converging faster. It has shown success
for training Recurrent Models.
(1)
(3)
(4)
mt and vt are estimates of the first moment (the mean) and
the second moment (the uncentered variance) of the
gradients respectively, hence the name of the method.
As mt and vt are initialized as vectors of 0's, the authors of
Figure 2: The overall structural design of the CNN Model of our Adam observe that they are biased towards zero, especially
proposed system with different layer. during the initial time steps, and especially when the decay
Adam and RMSProp, as two of the most influential adaptive rates are small (i.e. β1 and β2 are close to 1). They counteract
stochastic algorithms for training deep neural networks.
In this testing phase we used five layered Convolutional 2D array into 1D array and finally two fully connected
Neural Networks (CNN) model with Adam Optimization. layers for classification. All the parameters with respect to
On them one layers for convolutional, one layers for max the corresponding layers are stated in Table 2.
pooling or sub sampling, one Flatten layer which converts
Table 2. Parameters setup for CNN with Adam Optimization.
Layer Type layer Operation No of feature Maps feature maps Size window Size Total parameters
In this testing phase we used five layered Convolutional Optimization. On them one layers for convolutional, one
Neural Networks (CNN) model with RMSProp layers for max pooling or sub sampling, one Flatten layer
which converts 2D array into 1D array and finally two fully 28 waw 378 352 26 93.12
connected layers for classification. All the parameters with 418 404 14 96.65
respect to the corresponding layers are stated in Table 3. 29 mot
30 pet 422 406 16 96.21
Table 3. Parameters setup for CNN with Adam 31 pat 391 368 23 94.12
Optimization. 383 354 29 92.43
32 ha
bha ra dhaa dha ta
33 chhya 430 418 12 97.21
34 tra 378 360 18 95.24
35 gya 382 360 22 94.24
Average recognition percentage 95.57
Table 5: Individual results of the test database using CNN-Adam Estimation
Table 4: Individual results of the test database DFFNN Not
Labele No. Correctl Correctl
Label No. Not
d Hindi of y y %
ed Hindi of Correctly Correctly %
numbe Charact imag Classifi Classifi Accura
numb Charact image Classifie Classifie Accu
r er es ed ed cy
er er s d d racy
396 383 13 96.72 0 Ka 396 383 13 96.72
0 Ka
413 401 12 97.09 1 kha 413 401 12 97.09
1 kha
2 ga 418 403 15 96.41 2 ga 408 403 5 98.77
3 gha 402 387 15 96.27 3 gha 402 387 15 96.27
4 kna 368 352 16 95.65 4 kna 368 352 16 95.65
5 cha 396 370 26 93.43 5 cha 396 370 26 93.43
6 chha 394 374 20 94.92 6 chha 394 374 20 94.92
7 ja 427 413 14 96.72 7 ja 417 413 4 99.04
8 jha 389 370 19 95.12 8 jha 389 370 19 95.12
9 yna 430 415 15 96.51 9 yna 420 415 5 98.81
10 taa 438 421 17 96.12 10 taa 428 421 7 98.36
11 thaa 405 389 16 96.05 11 thaa 405 389 16 96.05
12 daa 413 400 13 96.85 12 daa 403 400 3 99.26
13 dhaa 402 390 12 97.01 13 dhaa 402 390 12 97.01
14 adna 430 414 16 96.28 14 adna 420 414 6 98.57
15 ta 395 384 11 97.22 15 ta 395 384 11 97.22
16 tha 380 358 22 94.21 16 tha 380 358 22 94.21
17 da 409 394 15 96.33 17 da 409 394 15 96.33
18 dha 378 353 25 93.39
18 dha 378 353 25 93.39
19 na 405 383 22 94.57
19 na 405 383 22 94.57
20 pa 393 370 23 94.15
20 pa 393 370 23 94.15
21 pha 407 393 14 96.56
21 pha 407 393 14 96.56
22 ba 395 379 16 95.95
22 ba 395 379 16 95.95
23 bha 401 372 29 92.77
23 bha 401 372 29 92.77
24 ma 398 387 11 97.24
24 ma 398 387 11 97.24
25 yaw 410 387 23 94.39
25 yaw 410 387 23 94.39
26 ra 421 409 12 97.15
26 ra 421 409 12 97.15
27 la 405 390 15 96.30