Skip to main content
. 2022 Sep 24;82(9):14193–14217. doi: 10.1007/s11042-022-13911-y

Table 2.

The table depicts the number of convolution, pooling layers, FC layers, and basic features of AlexNet, VggNet-16, ResNet-18, and VggNet-19 architectures

Model Year No. of Covoluti on Layers No. of Pooling Layers Fully Connect ed Layers Main Features
AlexNet 2012 5 3 3 ▪ First CNN architecture to win ImageNet challenge with top-5 error rate of 15.3%.
▪ Used ReLU as activation function instead of tanh or sigmoid.
▪ AlexNet has 60 million parameters.
▪ It had used the Stochastic Gradient descent as the learning algorithm.
VggNet-16 2014 13 5 3 ▪ The model achieved 92.7% top-5 test accuracy in ImageNet challenge.
▪ The model replaces the the large sized kernals used in AlexNet with 3✕3 sized multiple kernals enabling better learning.
▪ Main con of this network is that it is slow to train.
▪ And network architecture weights are quite large.
ResNet 2015 17 with 8 residual units 2 1 ▪ Main building blocks are residual blocks that increase the performance of the network.
▪ The identity connection helps the network to handle vanishing gradient problem.
▪ The batch normalisation used by network mitigates the problem of covariant shift.
▪ ResNet 18 has residual blocks of two layers deep.
VggNet-19 2014 16 5 3 ▪ Has 3 additional convolutional layers than that of Vgg-16.
▪ Deep network is believed to train more efficiently.
▪ Requires more memory than Vgg-16.