paint-brush
Demystifying Different Variants of Gradient Descent Optimization Algorithmby@NKumar
5,375 reads
5,375 reads

Demystifying Different Variants of Gradient Descent Optimization Algorithm

by NiranjanKumar12mApril 7th, 2019
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

The choice of optimization algorithms in deep learning can influence the network training speed and its performance. In this article, we will discuss the need for improving the gradient descent optimization technique. Gradient descent algorithm updates the parameters by moving in the direction opposite to the gradient of the objective function with respect to the network parameters. The gradient descent algorithm for single sigmoid neuron works like this,Initialize the parameters randomly w and b and iterate over all the observations in the data. Then update the value of each parameter based on its gradient value. Then continue doing step 2 and 3 till loss function gets minimized.

Company Mentioned

Mention Thumbnail

Coin Mentioned

Mention Thumbnail
featured image - Demystifying Different Variants of Gradient Descent Optimization Algorithm
NiranjanKumar HackerNoon profile picture
NiranjanKumar

NiranjanKumar

@NKumar

DeepLearning Enthusiast. Data Science Writer @marktechpost.com

About @NKumar
LEARN MORE ABOUT @NKUMAR'S
EXPERTISE AND PLACE ON THE INTERNET.
L O A D I N G
. . . comments & more!

About Author

NiranjanKumar HackerNoon profile picture
NiranjanKumar@NKumar
DeepLearning Enthusiast. Data Science Writer @marktechpost.com

TOPICS

THIS ARTICLE WAS FEATURED IN...

Permanent on Arweave
Read on Terminal Reader
Read this story in a terminal
 Terminal
Read this story w/o Javascript
Read this story w/o Javascript
 Lite