#76 Machine Learning & Data Science Challenge 76

What is VGG16 and explain the architecture of VGG16?

VGG-16 is a simpler architecture model since it’s not using many hyperparameters.

  • It always uses 3 x 3 filters with the stride of 1 in the convolution layer and uses SAME padding in pooling layers 2 x 2 with a stride of 2.

This architecture is from the VGG group, Oxford. It improves AlexNet by replacing the large kernel-sized filter with multiple 3X3 kernel-sized filters one after another. With a given receptive field(the effective area size of the input image on which output depends), multiple stacked smaller size kernel is better than the one with a larger size kernel because multiple non-linear layers increase the depth of the network which enables it to learn more complex features, and that too at a lower cost.

  • Three fully connected layers follow the VGG convolutional layers.

  • The width of the networks starts at the small value of 64 and increases by a factor of 2 after every sub-sampling/pooling layer. It achieves the top-5 accuracy of 92.3 % on ImageNet.

0
Subscribe to my newsletter

Read articles from Bhagirath Deshani directly inside your inbox. Subscribe to the newsletter, and don't miss out.

Written by

Bhagirath Deshani
Bhagirath Deshani

Greetings. I am a machine learning engineer based in India, possessing a sustained interest in machine learning since my undergraduate studies. I have completed Stanford University's machine learning course (Andrew Ng) via Coursera, and IBM's machine learning and deep learning curriculum. My current focus is on machine learning and data science projects, aiming to leverage my expertise for impactful, real-world problem-solving.