Skip to content

jnkien/optimization

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Optimization algorithms

equ

A description of algorithms is available here.

This repository contains my naive implementations in python in order to understand the different methods.

Stochastic Gradient Descent

Talk here

SGD is like GD but with a "partial" gradient computation. This "partial" part can be done by randomness + bagging and this allows parallelization for GD.

SGD is more sensitive to the step-size than GD.

SGD convergence is sensitive to the step-size. It is stable at start and the closer it goes to the solution the more fluctuating it is. It is good in ML to go to a solution quickly.

SGD is unbiased with respect to the randomness used.

The speed of the SGD convergence depends of the level of noise it has => the variance of the SGD.

Back-propagation in neural networks IS SGD.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages