#76 Machine Learning & Data Science Challenge 76


What is VGG16 and explain the architecture of VGG16?
VGG-16 is a simpler architecture model since it’s not using many hyperparameters.
- It always uses 3 x 3 filters with the stride of 1 in the convolution layer and uses SAME padding in pooling layers 2 x 2 with a stride of 2.
This architecture is from the VGG group, Oxford. It improves AlexNet by replacing the large kernel-sized filter with multiple 3X3 kernel-sized filters one after another. With a given receptive field(the effective area size of the input image on which output depends), multiple stacked smaller size kernel is better than the one with a larger size kernel because multiple non-linear layers increase the depth of the network which enables it to learn more complex features, and that too at a lower cost.
Three fully connected layers follow the VGG convolutional layers.
The width of the networks starts at the small value of 64 and increases by a factor of 2 after every sub-sampling/pooling layer. It achieves the top-5 accuracy of 92.3 % on ImageNet.
Subscribe to my newsletter
Read articles from Bhagirath Deshani directly inside your inbox. Subscribe to the newsletter, and don't miss out.
Written by

Bhagirath Deshani
Bhagirath Deshani
Hello everyone! I am Machine Learning Engineer. I am from India. I have been interested in machine learning since my engineering days. I have completed Andrew NG’s original Machine Learning course from Stanford University at Coursera and also completed the IBM course on Machine Learning and Deep Learning. Currently, I am working on Machine Learning and Data Science project. My goal is to use the skills I have acquired to solve real-world problems and make a positive impact on the world.