advancements in image classification using convolutional

9
This paper has been accepted and presented on 2018 Fourth International Conference on Research in Computational Intelligence and Communication Networks. This is preprint version and original proceeding will be published in IEEE Xplore. Advancements in Image Classification using Convolutional Neural Network Farhana Sultana Department of Computer Science University of Gour Banga West Bengal, India Email: [email protected] Abu Sufian Department of Computer Science University of Gour Banga West Bengal, India Email: sufi[email protected] Paramartha Dutta Department of CSS Visva-Bharati University West Bengal, India Email: [email protected] Abstract—Convolutional Neural Network (CNN) is the state- of-the-art for image classification task. Here we have briefly discussed different components of CNN. In this paper, We have explained different CNN architectures for image classification. Through this paper, we have shown advancements in CNN from LeNet-5 to latest SENet model. We have discussed the model description and training details of each model. We have also drawn a comparison among those models. Keywords—AlexNet, Capsnet, Convolutional Neural Network, Deep learning, DenseNet, Image classification, ResNet, SENet. I. I NTRODUCTION Computer vision consists of different problems such as image classification, localization, segmentation and object detection. Among those, image classification can be consid- ered as the fundamental problem and forms the basis for other computer vision problems. Until ’90s only traditional machine learning approaches were used to classify image. But the accuracy and scope of the classification task were bounded by several challenges such as hand-crafted feature extraction process etc. In recent years, the deep neural network (DNN), also entitled as deep learning [1][2], finds complex formation in large data sets using the backpropagation [3] algorithm. Among DNNs, convolutional neural network has demonstrated excellent achievement in problems of computer vision, especially in image classification. Convolutional Neural Network (CNN or ConvNet) is a especial type of multi-layer neural network inspired by the mechanism of the optical system of living creatures. Hubel and Wiesel [4] discovered that animal visual cortex cells detect light in the small receptive field. Motivated by this work, in 1980, Kunihiko Fukushima introduced neocognitron [5] which is a multi-layered neural network capable of recognizing visual pattern hierarchically through learning. This network is considered as the theoretical inspiration for CNN. In 1990 LeCun et al. introduced the practical model of CNN [6] [7] and developed LeNet-5 [8]. Training by backpropagation [9] algorithm helped LeNet-5 recognizing visual patterns from raw pixels directly without using any separate feature engi- neering mechanism. Also fewer connections and parameters of CNN than conventional feedforward neural networks with similar network size, made model training easier. But at that time in spite of several advantages, the performance of CNN in intricate problems such as classification of high-resolution image, was limited by the lack of large training data, lack of better regularization method and inadequate computing power. Nowadays we have larger datasets with millions of high resolution labelled data of thousands category like ImageNet [10], LabelMe [11] etc. With the advent of powerful GPU machine and better regularization method, CNN delivers out- standing performance on image classification tasks. In 2012 a large deep convolution neural network, called AlexNet [12], designed by Krizhevsky et al. showed excellent performance on the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) [13]. The success of AlexNet has become the inspiration of different CNN model such as ZFNet [14], VGGNet [15], GoogleNet [16], ResNet [17], DenseNet [18], CapsNet [19], SENet [20] etc in the following years. In this study, we have tried to give a review of the advance- ments of the CNN in the area of image classification. We have given a general view of CNN architectures in section II. Section III describes architecture and training details of different models of CNN. In Section IV we have drawn a comparison between various CNN models. Finally, we have concluded our paper in Section V. II. CONVOLUTIONAL NEURAL NETWORK A typical CNN is composed of single or multiple blocks of convolution and sub-sampling layers, after that one or more 978-1-5386-7638-7/18/$31.00 © 2018 IEEE arXiv:1905.03288v1 [cs.CV] 8 May 2019

Upload: others

Post on 16-Oct-2021

20 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Advancements in Image Classification using Convolutional

This paper has been accepted and presented on 2018 Fourth International Conference on Research inComputational Intelligence and Communication Networks. This is preprint version and original

proceeding will be published in IEEE Xplore.

Advancements in Image Classification usingConvolutional Neural Network

Farhana SultanaDepartment of Computer Science

University of Gour BangaWest Bengal, India

Email: [email protected]

Abu SufianDepartment of Computer Science

University of Gour BangaWest Bengal, India

Email: [email protected]

Paramartha DuttaDepartment of CSS

Visva-Bharati UniversityWest Bengal, India

Email: [email protected]

Abstract—Convolutional Neural Network (CNN) is the state-of-the-art for image classification task. Here we have brieflydiscussed different components of CNN. In this paper, We haveexplained different CNN architectures for image classification.Through this paper, we have shown advancements in CNN fromLeNet-5 to latest SENet model. We have discussed the modeldescription and training details of each model. We have alsodrawn a comparison among those models.

Keywords—AlexNet, Capsnet, Convolutional Neural Network,Deep learning, DenseNet, Image classification, ResNet, SENet.

I. INTRODUCTION

Computer vision consists of different problems such asimage classification, localization, segmentation and objectdetection. Among those, image classification can be consid-ered as the fundamental problem and forms the basis forother computer vision problems. Until ’90s only traditionalmachine learning approaches were used to classify image.But the accuracy and scope of the classification task werebounded by several challenges such as hand-crafted featureextraction process etc. In recent years, the deep neural network(DNN), also entitled as deep learning [1][2], finds complexformation in large data sets using the backpropagation [3]algorithm. Among DNNs, convolutional neural network hasdemonstrated excellent achievement in problems of computervision, especially in image classification.

Convolutional Neural Network (CNN or ConvNet) is aespecial type of multi-layer neural network inspired by themechanism of the optical system of living creatures. Hubeland Wiesel [4] discovered that animal visual cortex cells detectlight in the small receptive field. Motivated by this work,in 1980, Kunihiko Fukushima introduced neocognitron [5]which is a multi-layered neural network capable of recognizingvisual pattern hierarchically through learning. This network isconsidered as the theoretical inspiration for CNN. In 1990

LeCun et al. introduced the practical model of CNN [6] [7]and developed LeNet-5 [8]. Training by backpropagation [9]algorithm helped LeNet-5 recognizing visual patterns fromraw pixels directly without using any separate feature engi-neering mechanism. Also fewer connections and parametersof CNN than conventional feedforward neural networks withsimilar network size, made model training easier. But at thattime in spite of several advantages, the performance of CNNin intricate problems such as classification of high-resolutionimage, was limited by the lack of large training data, lack ofbetter regularization method and inadequate computing power.

Nowadays we have larger datasets with millions of highresolution labelled data of thousands category like ImageNet[10], LabelMe [11] etc. With the advent of powerful GPUmachine and better regularization method, CNN delivers out-standing performance on image classification tasks. In 2012 alarge deep convolution neural network, called AlexNet [12],designed by Krizhevsky et al. showed excellent performanceon the ImageNet Large Scale Visual Recognition Challenge(ILSVRC) [13]. The success of AlexNet has become theinspiration of different CNN model such as ZFNet [14],VGGNet [15], GoogleNet [16], ResNet [17], DenseNet [18],CapsNet [19], SENet [20] etc in the following years.

In this study, we have tried to give a review of the advance-ments of the CNN in the area of image classification. Wehave given a general view of CNN architectures in sectionII. Section III describes architecture and training details ofdifferent models of CNN. In Section IV we have drawn acomparison between various CNN models. Finally, we haveconcluded our paper in Section V.

II. CONVOLUTIONAL NEURAL NETWORK

A typical CNN is composed of single or multiple blocks ofconvolution and sub-sampling layers, after that one or more978-1-5386-7638-7/18/$31.00 © 2018 IEEE

arX

iv:1

905.

0328

8v1

[cs

.CV

] 8

May

201

9

Page 2: Advancements in Image Classification using Convolutional

fully connected layers and an output layer as shown in figure1.

Fig. 1: Building block of a typical CNN

A. Convolutional Layer

The convolutional layer (conv layer) is the central part of aCNN. Images are generally stationary in nature. That meansthe formation of one part of the image is same as any otherpart. So, a feature learnt in one region can match similarpattern in another region. In a large image, we take a smallsection and pass it through all the points in the large image(Input). While passing at any point we convolve them into asingle position (Output). Each small section of the image thatpasses over the large image is called filter (Kernel). The filtersare later configured based on the back propagation technique.Figure 2 shows typical convolutional operation.

Fig. 2: Convolutional Layer

B. Sub-sampling or Pooling Layer

Pooling simply means down sampling of an image. Ittakes small region of the convolutional output as input andsub-samples it to produce a single output. Different poolingtechniques are there such as max pooling, mean pooling,average pooling etc. Max pooling takes largest of the pixelvalues of a region as shown in figure 3. Pooling reduces thenumber of parameter to be computed but makes the networkinvariant to translations in shape, size and scale.

Fig. 3: Max Pooling operation

C. Fully-connected Layer (FC Layer)

Last section of CNN are basically fully connected layers asdepicted in figure 4. This layer takes input from all neuronsin the previous layer and performs operation with individualneuron in the current layer to generate output.

Fig. 4: Fully-connected layer

III. DIFFERENT MODELS OF CNN FOR IMAGECLASSIFICATION

A. LeNet-5(1998):

In 1998 LeCun et al. introduced the CNN to classifyhandwritten digit. Their CNN model, called LeNet-5 [8] asshown in figure 5, has 7 weighted (trainable) layers. Amongthem, three (C1, C3, C5) convolutional layers, two (S2, S4)average pooling layers, one (F6) fully connected layer andone output layer. Sigmoid function was used to include non-linearity before a pooling operation. The output layer usedEuclidean Radial Basis Function units (RBF) [21] to classify10 digits.

Fig. 5: Architecture of LeNet-5 [8]

In table I we have shown different layers, size of the filterused in each convolution layer, output feature map size andthe total number of parameters required per layer of LeNet-5.

TABLE I: Architecture of LeNet-5

Layer filtersize/stride

# filter output size #Para-meters

Convolution(C1) 5× 5/1 6 28× 28× 6 156Sub-sampling(S2) 2× 2/2 14× 14× 6 12Convolution(C3) 5× 5/1 16 10×10×16 1516Sub-sampling(S4) 2× 2/2 5× 5× 16 32Convolution(C5) 5× 5 120 1× 1× 120 48120Fully Connected(F6) 2× 2 14× 14× 6 10164OUTPUT 84

1) Dataset used: To train and test LeNet-5, LeCun et al.used the MNIST [22] database of handwritten digits. Thedatabase contains 60k training and 10k test data. The inputimage size of this model is basically 32× 32 pixels which islarger than the largest character (20×20 pixels) in the databaseas center part of the receptive field is rich in features. Inputimages are size normalized and centred in a 28 × 28 field.They have used data augmentation like horizontal translation,vertical translation, scaling, squeezing and horizontal shearing.

2) Training Details: The authors trained several versionsof LeNet-5 using stochastic gradient descent (SGD) [23]approach with 20 iterations for entire training data per sessionwith a decreased rate of global learning rate and a momentum

Page 3: Advancements in Image Classification using Convolutional

of 0.02. In 1990’s LeNet-5 was sufficiently good. LeNet-5 andLeNet-5 (with distortion) achieved test error rate of 0.95% and0.8% respectively on MNIST data set.

But as the amount of data, resolution of an image and thenumber of classes of a classification problem got increasedwith time, we needed deeper convolutional network and pow-erful GPU machine to train the model.

B. AlexNet-2012:

In 2012 Krizhevky et al. designed a large deep CNN,called AlexNet [12], to classify ImageNet [10] data. Thearchitecture of AlexNet is same as LeNet-5 but much bigger. Itis made up of 8 trainable layers. Among them, 5 convolutionallayers (conv layer) and 3 fully connected layers are there.Using rectified linear unit (ReLU) [24] non-linearity afterconvolutional and FC layers helped their model to be trainedfaster than similar networks with tanh units. They haveused local response normalization (LRN), called ”brightnessnormalization”, after the first and second convolutional layerwhich aids generalization. They have used max-pooling layerafter each LRN layer and fifth convolutional layer. In figure 6architectural details of AlexNet is shown. In table II we haveshown different elements of AlexNet.

Fig. 6: Architecture of AlexNet [12]

TABLE II: Details of different layers of AlexNet

Layer filtersize/stride

padding # filter output size #Parameters

Conv-1 11×11/4 0 96 55× 55× 96 34848pool-1 3× 3/2 27× 27× 96Conv-2 5× 5/1 2 256 27×27×256 614400pool-2 3× 3/2 13×13×256Conv-3 3× 3/1 1 384 13×13×384 981504Conv-4 3× 3/1 1 384 13×13×384 1327104Conv-5 3× 3/1 1 256 13×13×256 884736pool3 3× 3/2 6× 6× 256FC6 1× 1× 4096 37748736FC7 1× 1× 4096 16777216FC8 1× 1× 1000 4096000

1) Dataset used: Krizhevsky et al. designed AlexNet forclassification of 1.2 million high-resolution images of 1000classes for ILSVRC - 2010 and ILSVRC - 2012 [25] . Thereare around 1.2 million/50K/150K training/validation/testingimages. On ILSVRC, competitors submit two kinds of errorrates: top-1 and top-5.

2) Training Details: From the variable resolution image ofImageNet, AlexNet used down-sampled and centred 256×256pixels image. To reduce overfitting they have used runtime dataaugmentation as well as a regularization method called dropout[26]. In data augmentation, they have extracted translatedand horizontally reflected 10 random patches of 224 × 224images and also used principal component analysis (PCA)[27] for RGB channel shifting of training images. The authorstrained AlexNet using stochastic gradient descent (SGD) withbatch size of 128, weight decay of 0.0005 and momentum of0.9. The weight decay works as a regularizer and it reducestraining error also. Their initial learning rate was 0.01 reducedmanually three times by 1/10 when value accuracy plateaus.AlexNet was trained on two NVIDIA GTX-580 3 GB GPUsusing cross-GPU parallelization for five to six days.

The authors have noticed that removing any middle layerdegrades network’s performance. So, the result depends on thedepth of the network. Also, they have used purely supervisedlearning approach to simplify their experiment, but they haveexpected that unsupervised pre-training would help if we canhave adequate computational power to remarkably increase thenetwork size without increasing the amount of the correspond-ing labelled dataset.

C. ZFNetIn 2014 Zeiler and Fergus presented a CNN called ZFNet

[14]. The Architecture of AlexNet and ZFNet is almost similarexcept that the authors have reduced 1st layer filter size to 7×7instead of 11×11 and used stride 2 convolutional layer in bothfirst and second layers to retain more information in thoselayers’ features. In their paper, the authors tried to explainthe reason behind the outstanding performance of large deepCNN. They have used a novel visualization technique whichis a deconvolutional network with multiple networks, calleddeconvnet [28], to map activation at higher layers back to thespace of input pixel to recognize which pixels of the inputlayer is accountable for a given activation in the feature map.Basically, deconvnet is a reversely ordered convnet. It acceptsfeature map as input and applies unpooling using a switch. Aswitch is basically the position of maxima within a poolingregion recorded during convolution. Then they rectify it usingReLU non-linearity and uses the transpose version of filtersto rebuild the activity in the layer below which activates thechosen activation.

Fig. 7: Architecture of ZFNet [14]

1) Training Details: ZFNet used the ImageNet dataset of1.3 million/50k/100k training/validation/testing images. Theauthors trained their model following [12]. The slight differ-ence is that they have substituted the sparse connection of

Page 4: Advancements in Image Classification using Convolutional

layers 3, 4 and 5 of AlexNet with a dense connection in theirmodel and trained it on single GTX-580 GPU for 12 dayswith 70 epochs. They have also experimented their model withdifferent depths and different filter sizes on Caltech 101 [29],Caltech-256 [30] and PASCAL-2012 [31] data set and shownthat their model also generalizes these datasets well.

During training their visualization technique discovers dif-ferent properties of CNN such as the projections from eachlayer in ascending order shows that the nature of the featuresare hierarchical in the network. For this reason, firstly, theupper layers need a higher number of epochs than lower layersto converge and secondly, the network output is stable totranslation and scaling. They have used a bunch of occlusionexperiments to check whether the model is sensitive to localor global information.

D. VGGNet

Simonyan and Zisserman used deeper configuration ofAlexNet [12], and they proposed it as VGGNet [15]. Theyhave used small filters of size 3 × 3 for all layers and madethe network deeper keeping other parameters fixed. They haveused total 6 different CNN configurations: A, A-LRN, B, C, D(VGG16) and E (VGG19) with 11, 11, 13, 16, 16, 19 weightedlayers respectively. Figure 8 shows configuration of model D.

Fig. 8: Architecture of VGGNet (configuration D, VGG16)

The authors have used three 1 × 1 filters in the sixth,ninth and twelfth convolution layer in model C to increasenon-linearity. Also, a pack of three 3 × 3 convolution layers(with stride 1) has same effective receptive field as one 7× 7convolution layer. So, They have substituted a single 7 × 7layer with a pack of three 3 × 3 convolution layers and thischange increases non-linearity and decreases the number ofparameters of the network.

1) Training Details: The training procedure of VGGNetfollows AlexNet except the cropping and scaling sizes of inputimage for training and testing. Pre-initialization of certainlayers and uses of small filters helps their model to convergeafter 74 epoch in spite of having a large number of parametersand greater depth. They have trained configuration VGG Awith random initialisation. Then using its first 4 convolutionlayers and last 3 FC layers as pre-initialised layers, theygradually increased the number of weighted layers up to 19and trained VGG A-LRN to E. They have randomly croppedimage to 224×224 from isotropically rescaled training images.They perform horizontal flipping, random RGB colour shiftingand scale jittering as data augmentation technique. The scalejittering in train/test phase, the blending of cropped (multi-crop) and uncropped (dense) test images result in betteraccuracy.

The authors experienced that a deep network with small fil-ters performs better than a shallower one with larger filters. Sothe depth of the network is important in visual representation.

E. GoogLeNet

The architecture of GoogLeNet [16], proposed by Szegedyet al., is different from conventional CNN. They have increasedthe number of units in each layer using parallel filters calledinception module [32] of size 1 × 1, 3 × 3 and 5 × 5 ineach convolution layer (conv layer). They have also increasedthe layers to 22. Figure 10 shows the 22 layers GoogLeNet.While designing this model, they have considered the com-putational budget fixed. So that the model can be used inmobile and embedded systems. They have used a series ofweighted Gabor filters [33] of various size in the inceptionarchitecture to handle multiple scales. To make the architecturecomputationally efficient they have used inception modulewith dimensionality reduction instead of the naive versionof inception module. Figure 9a and figure 9b are showingboth inception modules. Despite 22 layers, the number ofparameters used in GoogLeNet is 12 times lesser than AlexNetbut its accuracy is significantly better. All the convolution,reduction and projection layers use ReLU non-linearity. Theyhave used average pooling layer instead of the fully connectedlayers. On top of some inception modules, they have usedauxiliary classifiers which are basically smaller CNNs, tocombat vanishing gradient problem and overfitting.

(a)

(b)

Fig. 9: (a) Naive version Inception Module (b) Dimensionalityreduction version [16]

1) Training Details: GoogLeNet, a CPU based implemen-tation, was trained using DistBelief [34] distributed machinelearning system by using moderate amount of model and dataparallelization. They used asynchronous SGD with momentum0.9 and a constant learning rate schedule. Using differentsampling and random ordering of input images, they have

Page 5: Advancements in Image Classification using Convolutional

Fig. 10: The architecture of GoogLeNet [16]

trained 7 ensemble GoogLeNet with same initialization. Un-like AlexNet they have used resized image of 4 scales withshorter dimension of 256, 288, 320 and 352 respectively. Thetotal number of crops per image is 4 (scales) ×3 (left, rightand centre square/scale) ×6 (4 corner and centre 224 × 224crop and the square resized to 224 × 224) ×2 (mirror imageof all six crops)=144.

The result of inception architecture has proved that movingtowards sparser architecture is realistic and competent idea.

F. ResNet

He et al. experienced that a deeper CNN stacked up withmore layers suffers from vanishing gradient problem. Thoughthis problem is handled by normalized and intermediate initial-ization, the deeper model shows worse performance on bothtrain and test errors and it is not caused by overfitting. Thisindicates that optimization of deeper network is hard. To solvethis problem the authors used pre-trained shallower modelwith additional layers to perform identity mapping. So thatthe performance of deeper network and the shallower networkshould be similar. They have proposed deep residual learningframework [17] as a solution to the degradation problem. Theyhave included residual mapping (H(x) = F (x) + x) insteadof desired underlying mapping (H(x)) into their network andnamed their model as ResNet [17].

(a) (b)

Fig. 11: (a) Plain layer (b) Residual block [17]

ResNet architecture consists of stacked residual blocks of3 × 3 convolutional layers. They have periodically doubledthe number of filters and used a stride of 2. Figure 11aand 11b shows a plain layer and residual block. As a first

layer, they have used a 7 × 7 conv layer. They have notused any fully connected layers at the end. They have useddifferent depth (34, 50, 101 and 152) ResNet in ILSVRC-2014 competition. For the CNN with depth more than 50 theyhave used ’bottleneck’ layer for dimensionality reduction andto improve efficiency as GoogLeNet. Their bottleneck designconsists of 1× 1, 3× 3 and 1× 1 convolution layer. Althoughthe 152 Layer ResNet is 8 times deeper than VGG nets, it haslower complexity than VGG nets (16/19).

1) Training Details: To train ResNet, He et al. used SGDwith batch size of 128, weight decay of 0.0001 and momentumof 0.9. They have used a learning rate of 0.1 reduced manuallytwo times at 32k and 48k iterations by 1/10 when valueaccuracy plateaus and stopped at 64k iterations. They usedweight initialization and Batch Normalization after every convlayer. The did not use dropout regularization method.

The experiment of ResNet shows the ability to train deepernetwork without degrading the performance. The authors havealso shown that with increased depth the ResNet, it is easierto optimize and it gains accuracy.

G. DenseNet

Huang et al. introduced Dense Convolutional Networks(DenseNet) [18], which includes dense block in conventionalCNN. The input of a certain layer in a dense block is theconcatenation of the output of all the previous layers as shownin figure 12. Here, each layer is reusing the features of all pre-vious layers, strengthening feature propagation and reducingvanishing gradient problem. Also uses of small number offilters reduced the number of parameters as well.

Figure 13 shows a DenseNet with three dense blocks. Ina dense block, the non-linear transformation functions are acomposite function of batch normalization, ReLU and 3 × 3convolution operation. They have also used the 1×1 bottlenecklayer to reduce dimensionality.

1) Training Details: Huang et al. trained DenseNet onCIFAR [35], SVHN [36] and ImageNet dataset using SGDwith batch size 64 on both CIFAR and SVHN dataset, andwith batch size 256 on ImageNet dataset. Initial learning ratewas 0.1 and is decreased two times by 1/10. They have used

Page 6: Advancements in Image Classification using Convolutional

Fig. 12: A typical dense block with 5 layers [18]

Fig. 13: A DenseNet with 3 Dense Block

weight decay of 0.0001, Nesterov momentum [37] of 0.9 anddropout of 0.2.

On C10 [38], C100 [39], SVHN dataset DenseNet,DenseNet-BC outperforms the error rates of previous CNNarchitectures. A DenseNet, doubly deeper than ResNet, givessimilar accuracy on ImageNet datasets with very less (factorof 2) number of parameters. The authors experienced thatDenseNet can be scaled to hundreds layers without optimiza-tion difficulty. It also gives consistent improvement if numberof parameters increases without degrading performance andoverfitting. Also, it requires comparatively fewer parametersand less computational power for better performance.

H. CapsNet

Conventional CNNs, described above, suffer from twoproblems. Firstly, Sub-sampling loses the spatial informationbetween higher-level features. Secondly, it faces difficulty ingeneralizing to novel view points. It can deal with translationbut can not detect different dimension of affine transformation.In 2017, Geoffrey E. Hinton proposed CapsNet [19] to handlethese problems. CapsNet has components called capsule. Acapsule is a group of neurons. So a layer of CapsNet isbasically composed with nested neurons. Unlike a typicalneural network, a capsule is squashed as a whole vector ratherthan individual output unit squashing. So scalar output featuredetector of CNN is replaced by vector output capsules. Alsomax-pooling is replaced by ”dynamic routing by agreement”which makes each capsule in each layer to go to the next mostrelevant capsules at the time of forward propagation.

Architecture of a simple CapsNet is shown in figure 14.The CapsNet, proposed by Sabour et al, is composed with

three layers - two conv layers and one FC layer. First convlayer consist of 256 convolutional unit (CU) with 9×9 kernels

Fig. 14: A 3 layer CapsNet, used for handwritten digit recog-nition [19]

of stride 1 and uses ReLU as activation function. This layerdetects local features and then sends it to the primary capsulesof second layer as input. Each primary capsule contains 8CU with 9× 9 kernel of stride of 2. In total primary capsulelayer has 32× 6× 6 8D capsules. The final layer (DigitCaps)has one 16D capsule per digit class. The authors have usedrouting between primary layer and DigitCaps layer. As the firstconvolutional layer is a 1D layer, no routing is used betweenthis layer and primary capsule layer.

1) Training details: Training of CapsNet is performed onMNIST images. To compare the test accuracy, they have usedone standard CNN (baseline) and two CapsNets with 1 and 3routing iterations respectively. They have used reconstructionloss as regularization method. Using a 3 layer CapsNet with3 routing iterations and with added reconstruction the authorsget a test error of 0.25%.

Though CapsNet has shown outstanding performance onMNIST, it may not perform well with large scale image datasetlike ImageNet. It may also suffer from vanishing gradientproblem.

I. SENet

In 2017, Hu et al. have designed ”Squeeze-and-Excitationnetwork” (SENet) [20] and have become the winner ofILSVRC-2017. They have reduced the top-5 error rate to2.25%. Their main contribution is ”Squeeze-and-Excitation”(SE) block as shown in figure 15. Here, Ftr: X→U is aconvolutional operation. A squeeze function (Fsq) performsaverage pooling on individual channel of feature map Uand produce 1 × 1 × C dimensional channel descriptor. Anexcitation function (Fex) is a self-gating mechanism made upof three layers - two fully connected layers and a ReLU non-linearity layer in between. It takes squeezed output as inputand produce a per channel modulation weights. By applyingthe excited output on the feature map U, U is scaled (Fscale)to generate final output (X) of SE block.

Fig. 15: A Sqeeze and excitation block [20]

This SE block can be stacked together to make SENet whichgeneralise different data set very well. The authors developed

Page 7: Advancements in Image Classification using Convolutional

TABLE III: Comparative performance of different CNN configurations. The + indicates- DenseNet with Bottleneck layer andcompression (10 crop testing result).

Nameof TheCNN

Dataset Year Type of CNN #trained layer Top-1(val)

Top-5(val)

Top-5(test)

AlexNet ImageNet 2012

1 CNN 8 40.7% 18.2%5 CNN - 38.1% 16.4% 16.4%1 CNN - 39.0% 16.6% -7 CNN - 36.7% 15.4% 15.3%

ZFNet ImageNet 2013

1 CNN 8 38.4 % 16.5%5 CNN - (a) - 36.7 % 15.3% 15.3%1 CNN with layers 3, 4, 5: 512, 1024, 512 maps-(b) - 37.5 % 16.0% 16.1%6 CNN, combination of (a) & (b) - 36.0 % 14.7% 14.8%

VGGNet ImageNet 2014

ensemble of 7 ConvNets (3-D,2-C & 2-E) - 24.7% 7.5% 7.3%ConvNet- D( multi-crop & dense) 16 24.4 % 7.2% -ConvNet-E (Multi-crop & dense ) 19 24.4 % 7.1% -ConvNet-E (Multi-crop & dense ) 19 24.4 % 7.1% 7.0%Ensemble of multi-scale ConvNets D & E (multi-crop &dense)

- 23.7% 6.8% 6.8%

GoogLeNet ImageNet 2014

1 CNN with 1 crop 22 - - 10.07%1 CNN with 10 crops - - - 9.15%1 CNN with 144 crops - - - 7.89%7 CNN with 1 crop - - - 8.09%1 CNN with 10 crops - - - 7.62%1 CNN with 144 crops - - - 6.67%

ResNet ImageNet 2015

plain layer 18 27.94% -ResNet-18 18 27.88% -Plain layer 34 28.54% 10.02ResNet-34 (zero-padding shortcuts), 10 crop testing -(a) 34 25.03% 7.76ResNet-34 (projection shortcuts to increase dimension, oth-ers are identity shortcuts ), 10 crop testing-(b)

34 24.52% 7.46%

ResNet-34 (all shortcuts are projection), 10 crop testing-(c) 34 24.52% 7.46%ResNet-50 (with bottleneck layer), 10 crop testing 50 22.85% 6.71%ResNet-101 (with bottleneck layer), 10 crop testing 101 21.75% 6.05%ResNet-152 (with bottleneck layer), 10 crop testing 152 21.43% 5.71%1 ResNet-34 (b) 34 21.84% 5.71%1 ResNet-34 (c) 34 21.53% 5.60%1 ResNet-50 50 20.74% 5.25%1 ResNet-101 101 19.87% 4.60%1 ResNet-152 152 19.38% 4.49%Ensemble of 6 models - 3.57%

DenseNet ImageNet 2016

DensNet-121 + 121 23.61% 6.66%DenseNet-169 + 169 22.80% 5.92%DenseNet-201 + 201 22.58% 5.54%DenseNet-264 + 264 20.80% 5.29%

SENet ImageNet 2017

SE-ResNet-50 50 23.29% 6.62%SE-ResNext-50 50 21.10% 5.49%SENet-154 (crop size 320× 320/299× 229) - 17.28% 3.79%SENet-154(crop size 320× 320) - 16.88% 3.58%

different SENets including these blocks into several complexCNN models such as VGGNet [15], GoogLeNet [16], ResNext(Variant of ResNet) [40], Inception-ResNet [41], MobileNet[42], ShuffleNet [43].

1) Training Details: The authors have trained and testtheir model variants on ImageNet, CIFAR-10 and CIFAR-100.They have trained original CNN models and those modelswith SE blocks, and compare speed accuracy trade-off. Theyhave shown that their models outperform original models byincreasing a little bit training/testing time.

IV. COMPARATIVE RESULT

In table III, we have shown comparative performance ofdifferent CNN (AlexNet to DenseNet) on ImageNet dataset.Top-1 and top-5 error rate on validation dataset and top-5 errorrates on test dataset are also shown.

V. CONCLUSION

In this study, we have discussed the advancements ofCNN in image classification tasks. We have shown herethat although AlexNet, ZFNet and VGGNet followed thearchitecture of conventional CNN model such as LeNet-5their networks are larger and deeper. We have experiencedthat combining inception module and residual blocks withconventional CNN model, GoogLeNet and ResNet gainedbetter accuracy than stacking the same building blocks againand again. DenseNet focused on feature reusing to strengthenthe feature propagation. Though CapsNet reached state-of-the-art achievement on MNIST but it is yet to perform as well asprevious CNNs performance on high resolution image datasetsuch as ImageNet. The result of SENet on ImageNet datasetgives us the hope that it may turn out useful for other taskwhich requires strong discriminative features.

Page 8: Advancements in Image Classification using Convolutional

REFERENCES

[1] Y. Lecun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521,no. 7553, pp. 436–444, 5 2015.

[2] I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning. MIT Press,2016, http://www.deeplearningbook.org.

[3] R. Hecht-Nielsen, “Theory of the backpropagation neural network,” inInternational 1989 Joint Conference on Neural Networks, 1989, pp. 593–605 vol.1.

[4] D. H. Hubel and T. N. Wiesel, “Receptive fields and functional archi-tecture of monkey striate cortex,” Journal of Physiology (London), vol.195, pp. 215–243, 1968.

[5] K. Fukushima, “Neocognitron: A self-organizing neural network modelfor a mechanism of pattern recognition unaffected by shift in position,”Biological Cybernetics, vol. 36, no. 4, pp. 193–202, Apr 1980. [Online].Available: https://doi.org/10.1007/BF00344251

[6] Y. LeCun, B. Boser, J. S. Denker, D. Henderson, R. E. Howard,W. Hubbard, and L. D. Jackel, “Backpropagation applied to handwrittenzip code recognition,” Neural Computation, vol. 1, no. 4, pp. 541–551,Dec 1989.

[7] Y. LeCun, B. E. Boser, J. S. Denker, D. Henderson, R. E. Howard,W. E. Hubbard, and L. D. Jackel, “Handwritten digit recognitionwith a back-propagation network,” in Advances in Neural InformationProcessing Systems 2, D. S. Touretzky, Ed. Morgan-Kaufmann,1990, pp. 396–404. [Online]. Available: http://papers.nips.cc/paper/293-handwritten-digit-recognition-with-a-back-propagation-network.pdf

[8] Y. Lecun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learningapplied to document recognition,” Proceedings of the IEEE, vol. 86,no. 11, pp. 2278–2324, Nov 1998.

[9] Y. L. Cun, “A theoretical framework for back-propagation,” 1988.[10] J. Deng, W. Dong, R. Socher, L.-J. Li, K. Li, and L. Fei-Fei, “ImageNet:

A Large-Scale Hierarchical Image Database,” in CVPR09, 2009.[11] B. C. Russell, A. Torralba, K. P. Murphy, and W. T. Freeman, “Labelme:

A database and web-based tool for image annotation,” InternationalJournal of Computer Vision, vol. 77, no. 1, pp. 157–173, May 2008.[Online]. Available: https://doi.org/10.1007/s11263-007-0090-8

[12] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classificationwith deep convolutional neural networks,” in Advances in NeuralInformation Processing Systems 25, F. Pereira, C. J. C. Burges,L. Bottou, and K. Q. Weinberger, Eds. Curran Associates, Inc.,2012, pp. 1097–1105. [Online]. Available: http://papers.nips.cc/paper/4824-imagenet-classification-with-deep-convolutional-neural-networks.pdf

[13] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma,Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, andL. Fei-Fei, “Imagenet large scale visual recognition challenge,” Int. J.Comput. Vision, vol. 115, no. 3, pp. 211–252, Dec. 2015. [Online].Available: http://dx.doi.org/10.1007/s11263-015-0816-y

[14] M. D. Zeiler and R. Fergus, “Visualizing and understanding convolu-tional networks,” in Computer Vision – ECCV 2014, D. Fleet, T. Pajdla,B. Schiele, and T. Tuytelaars, Eds. Cham: Springer InternationalPublishing, 2014, pp. 818–833.

[15] K. Simonyan and A. Zisserman, “Very deep convolutional networksfor large-scale image recognition,” CoRR, vol. abs/1409.1556, 2014.[Online]. Available: http://arxiv.org/abs/1409.1556

[16] C. Szegedy, W. Liu, Y. Jia, P. Sermanet, S. Reed, D. Anguelov, D. Erhan,V. Vanhoucke, and A. Rabinovich, “Going deeper with convolutions,”in The IEEE Conference on Computer Vision and Pattern Recognition(CVPR), June 2015.

[17] K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for imagerecognition,” in The IEEE Conference on Computer Vision and PatternRecognition (CVPR), June 2016.

[18] G. Huang, Z. Liu, and K. Q. Weinberger, “Densely connectedconvolutional networks,” CoRR, vol. abs/1608.06993, 2016. [Online].Available: http://arxiv.org/abs/1608.06993

[19] S. Sabour, N. Frosst, and G. E. Hinton, “Dynamic routing betweencapsules,” CoRR, vol. abs/1710.09829, 2017. [Online]. Available:http://arxiv.org/abs/1710.09829

[20] J. Hu, L. Shen, and G. Sun, “Squeeze-and-excitation networks,” CoRR,vol. abs/1709.01507, 2017. [Online]. Available: http://arxiv.org/abs/1709.01507

[21] M. D. Buhmann, “Radial basis functions,” Acta Numerica, vol. 9, p.138, 2000.

[22] Y. LeCun and C. Cortes, “MNIST handwritten digit database,” 2010.[Online]. Available: http://yann.lecun.com/exdb/mnist/

[23] L. Bottou, “Large-scale machine learning with stochastic gradientdescent,” in Proceedings of COMPSTAT’2010, Y. Lechevallier andG. Saporta, Eds. Heidelberg: Physica-Verlag HD, 2010, pp. 177–186.

[24] V. Nair and G. E. Hinton, “Rectified linear units improverestricted boltzmann machines,” in Proceedings of the 27th InternationalConference on International Conference on Machine Learning, ser.ICML’10. USA: Omnipress, 2010, pp. 807–814. [Online]. Available:http://dl.acm.org/citation.cfm?id=3104322.3104425

[25] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma,Z. Huang, A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, andL. Fei-Fei, “ImageNet Large Scale Visual Recognition Challenge,”International Journal of Computer Vision (IJCV), vol. 115, no. 3, pp.211–252, 2015.

[26] G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, andR. Salakhutdinov, “Improving neural networks by preventing co-adaptation of feature detectors,” CoRR, vol. abs/1207.0580, 2012.[Online]. Available: http://arxiv.org/abs/1207.0580

[27] I. Jolliffe, Principal Component Analysis. Berlin, Heidelberg: SpringerBerlin Heidelberg, 2011, pp. 1094–1096. [Online]. Available: https://doi.org/10.1007/978-3-642-04898-2 455

[28] M. D. Zeiler, G. W. Taylor, and R. Fergus, “Adaptive deconvolutionalnetworks for mid and high level feature learning,” in 2011 InternationalConference on Computer Vision, Nov 2011, pp. 2018–2025.

[29] L. Fei-Fei, R. Fergus, and P. Perona, “Learning generative visual modelsfrom few training examples: An incremental bayesian approach testedon 101 object categories,” in 2004 Conference on Computer Vision andPattern Recognition Workshop, June 2004, pp. 178–178.

[30] G. Griffin, A. Holub, and P. Perona, “Caltech256 imagedataset,” 2006. [Online]. Available: http://www.vision.caltech.edu/Image Datasets/Caltech256/

[31] M. Everingham, L. Van Gool, C. K. I. Williams, J. Winn, and A. Zisser-man, “Pascal visual object classes challenge 2012 (voc2012) completedataset.”

[32] M. Lin, Q. Chen, and S. Yan, “Network in network,” CoRR, vol.abs/1312.4400, 2013. [Online]. Available: http://arxiv.org/abs/1312.4400

[33] T. Serre, L. Wolf, S. Bileschi, M. Riesenhuber, and T. Poggio, “Robustobject recognition with cortex-like mechanisms,” IEEE Transactions onPattern Analysis and Machine Intelligence, vol. 29, no. 3, pp. 411–426,March 2007.

[34] J. Dean, G. Corrado, R. Monga, K. Chen, M. Devin, M. Mao,M. aurelio Ranzato, A. Senior, P. Tucker, K. Yang, Q. V.Le, and A. Y. Ng, “Large scale distributed deep networks,” inAdvances in Neural Information Processing Systems 25, F. Pereira,C. J. C. Burges, L. Bottou, and K. Q. Weinberger, Eds. CurranAssociates, Inc., 2012, pp. 1223–1231. [Online]. Available: http://papers.nips.cc/paper/4687-large-scale-distributed-deep-networks.pdf

[35] A. Krizhevsky and G. Hinton, “Learning multiple layers of features fromtiny images,” Tech. Rep., 2009.

[36] Y. Netzer, T. Wang, A. Coates, A. Bissacco, B. Wu, and A. Y. Ng,“Reading digits in natural images with unsupervised feature learning,” inNIPS Workshop on Deep Learning and Unsupervised Feature Learning2011, 2011. [Online]. Available: http://ufldl.stanford.edu/housenumbers/nips2011 housenumbers.pdf

[37] I. Sutskever, J. Martens, G. Dahl, and G. Hinton, “On theimportance of initialization and momentum in deep learning,” inProceedings of the 30th International Conference on Machine Learning,ser. Proceedings of Machine Learning Research, S. Dasgupta andD. McAllester, Eds., vol. 28, no. 3. Atlanta, Georgia, USA:PMLR, 17–19 Jun 2013, pp. 1139–1147. [Online]. Available:http://proceedings.mlr.press/v28/sutskever13.html

[38] A. Krizhevsky, V. Nair, and G. Hinton, “Cifar-10 (canadian institutefor advanced research).” [Online]. Available: http://www.cs.toronto.edu/∼kriz/cifar.html

[39] A. Krizhevsky, V. Nair, and G. E. Hinton, “Cifar-100 (canadian institutefor advanced research).” [Online]. Available: http://www.cs.toronto.edu/∼kriz/cifar.html

[40] S. Xie, R. B. Girshick, P. Dollar, Z. Tu, and K. He, “Aggregated residualtransformations for deep neural networks,” CoRR, vol. abs/1611.05431,2016. [Online]. Available: http://arxiv.org/abs/1611.05431

[41] C. Szegedy, S. Ioffe, and V. Vanhoucke, “Inception-v4, inception-resnetand the impact of residual connections on learning,” CoRR, vol.

Page 9: Advancements in Image Classification using Convolutional

abs/1602.07261, 2016. [Online]. Available: http://arxiv.org/abs/1602.07261

[42] A. G. Howard, M. Zhu, B. Chen, D. Kalenichenko, W. Wang, T. Weyand,M. Andreetto, and H. Adam, “Mobilenets: Efficient convolutional neuralnetworks for mobile vision applications,” CoRR, vol. abs/1704.04861,2017. [Online]. Available: http://arxiv.org/abs/1704.04861

[43] X. Zhang, X. Zhou, M. Lin, and J. Sun, “Shufflenet: An extremelyefficient convolutional neural network for mobile devices,” CoRR, vol.abs/1707.01083, 2017. [Online]. Available: http://arxiv.org/abs/1707.01083