Arora 2020
Arora 2020
Arora 2020
Abstract—Artificial Neural networks have been proved most Reverse transmission means back propagation and is used to
efficient in Deep Learning mainly because of large number of calculate error between the results of the generated output
datasets it can handle. The most widely used is the Convolutional transmission and the actual sample [4].
Neural Network (CNN). It has been proved useful for computer
vision, pattern recognition and Natural Language Processing Because of the performance of CNN in image processing, the
(NLP). CNN is so vastly used, as, unlike traditional Neural Nets, network usage has increased exponentially. Earlier image
it reduces number of parameters and focus more on domain retrieval systems were based on text; the images were
specific features. There are various CNN architectures proposed, manually analyzed and then indexed accordingly. As the
such as LeNet, AlexNet, GoogleNet. In this paper, we talk about image database increased per user the task became
structure of CNN and all the models of CNN which are proposed burdensome and hard.
till date.
CNN is a feed forward neural network, that’s why it can
Keywords—Artificial Neural Networks, Convolutional Neural
Network, CNN models, deep learning
extract topological properties from image and due to this it can
recognize extreme variable patterns. That’s the reason why
I. INTRODUCTION many researchers have used Deep CNN for image processing
purpose [5].
Deep Convolution Neural Network has now achieved a
state-of-the-art work in the area of computer vision and image CNN is a collection of neurons arranged in an acyclic graph,
recognition. CNN is so successful because of the hidden layers with multiple hidden layer neurons which is connected to a
which are not fully connected to the previous layers [1], and subset of neurons (only to a few not to all) from previous
do layer. This arrangement is done to promote the network to
multiple successive computations between convolution and learn implicitly.
pooling (subsampling layer). Apart from other deep learning
neural networks, CNN is easy to train due to backpropagation Convolutional Neural Net or ConvNet is a class of feed
because they have very sparse connectivity at each layer. A forward, deep artificial neural networks, applied to analyze
linear filter is used for convolution purpose [2]. The visual imagery. CNN was inspired by biological neurons
Convolution Neural Network got its name due to a connectivity found in animal’s visual cortex.
mathematical operation Convolution which means roll
together (use together) two or more mathematical operations Just like other Artificial Neural Networks it has neurons with
such as sigmoid, leaky ReLU or Tanh. learnable weights and bias. Although not every neuron in the
layer is connected with the neuron in previous layer, the
The first ever architecture was proposed by LeCun et al. in neurons receives many inputs, take the weights of sum and
1990. It was designed to predict handwritten digits, but the passes it to an activation function, and produces the desired
architecture was not very successful due to less amount of output.
training data and computing power. In 2012, AlexNet was
proposed by Krizhevsky et al. which successfully reduced the
CNN has shown most efficient responses in image and video
error rate and won ILSVRC 2012. After that CNN became the
processing and recognition, image classifier, medical image
most used Neural Network, it is not only used in object
analysis and natural language processing.
recognition, but also in object tracking [3], pose estimation,
text recognition and many more.
Convolutional Neural Net has four basic layers naming
Convolution Layer, Pooling Layer, Fully Connected Layer and
II. STRUCTURE OF CONVOLUTIONAL
Loss Layer [6]. The detailed description of each layer is as
NEURAL NETWORK
follows:
Convolutional Neural Network, have a multiple hierarchical
network structure. It can have feed forward as well as back A. Convolution Layer
propagation structure. Feed forward propagation implies that It is the basic layer of ConvNet which involves all the
input goes forward in the structure through multiple layers and computational part. The parameters are learnable filters or
is operated with respective activation function and finally the kernels. It gets its name from mathematical term convolve
output is obtained in the output layer. which means combining more than one functions, here also
Authorized licensed use limited to: Raytheon Technologies. Downloaded on May 20,2021 at 08:52:25 UTC from IEEE Xplore. Restrictions apply.
2020 2nd International Conference on Advances in Computing, Communication Control and Networking (ICACCCN)
more the activation functions are combined. The need for E. Activation Functions
convolution is that, suppose an image is of 32X32 pixels and Sigmoid: It is a logistic regression function which squashes
we need pass it to the next layer with 3 neurons, so we have to the number between -1 and +1 and returns the probability of
make 32 X 32 X 3 connections, if we add 2 more neurons the the output being positive.
connections become more complex. And the input can even
be larger than this. So rather than making so many 1
f(x)=
connections if we focus on connecting only local points say 1+𝑒 −𝑥
image than the connections become less [7].
Tanh: tanh squashes the output between -1 and +1.
32X32
tanh(x) = 2f(2x) - 1
Hidden layer
ReLU: ReLU function is mostly used as the basic function for
Neural Nets. It returns 0 if output is negative and if its greater
than zero than returns the number itself.
rect(x) = max(0, x)
C. Fully Connected Layer The architecture got many new versions in which original
LeNet is always included; the most used is LeNet 5. Total of 7
In this layer the neurons are connected fully with the neurons layers are included which includes an input layer, convolution
of previous layer. Since the neurons are not arranged spatially layer, pooling layer, fully connected layer and an output layer
and all the neurons are connected with previous neurons there [8].
cannot be a convolution layer after a FC Layer.
D. Loss Layer
It calculates the overall loss and error between the actual and
desired output.
750
Authorized licensed use limited to: Raytheon Technologies. Downloaded on May 20,2021 at 08:52:25 UTC from IEEE Xplore. Restrictions apply.
2020 2nd International Conference on Advances in Computing, Communication Control and Networking (ICACCCN)
V. ACKNOWLEDGEMENT
We would like to thank the staff of UIRC, GGSIPU, Delhi for
providing the resources and support pivotal for the research
work of this document. We would also like to extend our
thanks to the varied authors of the research papers used for the
reference for the work on this document and providing valuable
insights into the fields of Neural Nets and its implementations.
VI. REFERENCES
751
Authorized licensed use limited to: Raytheon Technologies. Downloaded on May 20,2021 at 08:52:25 UTC from IEEE Xplore. Restrictions apply.