paint-brush
ResNet: Block Level Design with Deep Learning Studio |PART 1|by@maniksoni653
1,280 reads
1,280 reads

ResNet: Block Level Design with Deep Learning Studio |PART 1|

by Manik Soni4mJanuary 8th, 2019
Read on Terminal Reader
Read this story w/o Javascript
tldt arrow

Too Long; Didn't Read

The main benefit of a very deep network is that it can represent very complex functions. It can also learn features at many different levels of abstraction, from edges (at the lower layers) to very complex features (at the deeper layers). However, using a deeper network doesn’t always help. A huge barrier to training them is vanishing gradients: very deep networks often have a gradient signal that goes to zero quickly, thus making gradient descent unbearably slow. More specifically, during gradient descent, as you backprop from the final layer back to the first layer, you are multiplying by the weight matrix on each step, and thus the gradient can decrease exponentially quickly to zero (or, in rare cases, grow exponentially quickly and “explode” to take very large values).

Company Mentioned

Mention Thumbnail
featured image - ResNet: Block Level Design with Deep Learning Studio |PART 1|
Manik Soni HackerNoon profile picture
Manik Soni

Manik Soni

@maniksoni653

SDE @Amazon

Learn More
LEARN MORE ABOUT @MANIKSONI653'S
EXPERTISE AND PLACE ON THE INTERNET.
L O A D I N G
. . . comments & more!

About Author

Manik Soni HackerNoon profile picture
Manik Soni@maniksoni653
SDE @Amazon

TOPICS

THIS ARTICLE WAS FEATURED IN...

Permanent on Arweave
Read on Terminal Reader
Read this story in a terminal
 Terminal
Read this story w/o Javascript
Read this story w/o Javascript
 Lite
Bitcoininsider