Welcome to my notebook! Iβm Daniel, a spanish πͺπΈ aerospace engineer living in France π«π· with a passion for learning. This blog documents my journey exploring machine learning and other topics that interest me. I believe that writing about what you learn can be not only useful for yourself but also for others following a similar path. Moreover, learning in public can spark discussions and positive exchanges.
Let's code and train VGGNet from scratch! In this post, I will explain the process of implementing this iconic CNN from designing a general architecture and using dense evaluation to optimizing training speed and actually training the network to obtain a validation top-1 and top-5 error rates of 28.33% and 9.66% respectively. I will also compare the error rates and training performance against the original paper and AlexNet.
Continuing our exploration of foundational deep learning models in computer vision, we will dive into the 2014 paper Very Deep Convolutional Networks for Large-Scale Image Recognition by Karen Simonyan and Andrew Zisserman, which introduced VGGNet, a set of simple yet highly performant networks. We will examine its architecture, data processing, training, testing, and analysis of the results as a preliminary step toward implementing it.
-
In this post, we will go through all the elements required to create and train AlexNet following the original paper. We will cover data processing, architecture definition, coding of training and validation loops, optimizations to speed up and training. Achieving comparable results with a top-1 error rate of 39.9% and top-5 error rate of 17.7%.
Thank you for visiting my site! Donβt hesitate to reach out to discuss more, share your ideas or just introduce yourself. Feel free to reach me either via the contact form below or connecting with me using the links at the bottom of the page.