Implement Adam and compare it with Vanilla GD
- Implement Adam and compare its convergence with that of Vanilla Gradient Descent
- Details of the problem statement , data set , summary of the code/solution , sample output/Prediction from the program and final result of the project are listed in the sections to follow.
This Note book build upon our previous notebook "Vanilla Gradient Decsent From Scratch" which is present in my repository .In this notebook ,we build on top of that notebook , implement Adam and observe its effects on convergence.We jump staright into Adam , so if the reader needs more of a background on gradient descent then I would recommend going through the previous notebook(which has detailed comments on Gradient Descent)
The dataset is manually created for the purpose of this exercise
Deep Learning :Proof of concept
- We implement Vanilla Gradient Descent on a dataset and log observations
- We implement Adam on the same data set and log observations
- We compare results and demostrate which algorithm is superior
- Refer python worksheet Adam_Vs_VanillaGD.ipynb for the solution
The above image depicts the following
- We start at a random line (marked in RED)
- The BLUE line indicates the best fit line which is our target
- We need to arrive at this line through the Gradient Descent Algorithm
- Let us see how changing learning rate affects this convergence
The following references were used while creating this notebook:
- https://machinelearningmastery.com/adam-optimization-from-scratch/ by Jason Brownlee
- Post Graduation AI/ML Study Material by GL/UAT