Optimizers - EXPLAINED!
From Gradient Descent to Adam. Here are some optimizers you should know. And an easy way to remember them.
SUBSCRIBE to my channel for more good stuff!
REFERENCES
[1] Have fun plotting equations : https://academo.org/demos/3d-surface-plotter
[2] Original paper on the Adam optimizer: https://arxiv.org/pdf/1412.6980.pdf
[3] Blog on types of optimizers: https://towardsdatascience.com/types-of-optimization-algorithms-used-in-neural-networks-and-ways-to-optimize-gradient-95ae5d39529f
[4] Blog on optimizing gradient descent: https://ruder.io/optimizing-gradient-descent/index.html#adagrad
[5] Github gist of code for rending animation of a math function: https://gist.github.com/ajhalthor/33533b4673ad6955e08a4005850b512f
[6] Another Blog to quench your thirst for knowledge on optimizers cuz the other links weren't good enough: https://machinelearningmastery.com/adam-optimization-algorithm-for-deep-learning/
Видео Optimizers - EXPLAINED! канала CodeEmporium
SUBSCRIBE to my channel for more good stuff!
REFERENCES
[1] Have fun plotting equations : https://academo.org/demos/3d-surface-plotter
[2] Original paper on the Adam optimizer: https://arxiv.org/pdf/1412.6980.pdf
[3] Blog on types of optimizers: https://towardsdatascience.com/types-of-optimization-algorithms-used-in-neural-networks-and-ways-to-optimize-gradient-95ae5d39529f
[4] Blog on optimizing gradient descent: https://ruder.io/optimizing-gradient-descent/index.html#adagrad
[5] Github gist of code for rending animation of a math function: https://gist.github.com/ajhalthor/33533b4673ad6955e08a4005850b512f
[6] Another Blog to quench your thirst for knowledge on optimizers cuz the other links weren't good enough: https://machinelearningmastery.com/adam-optimization-algorithm-for-deep-learning/
Видео Optimizers - EXPLAINED! канала CodeEmporium
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
Loss Functions - EXPLAINED!Batch Normalization - EXPLAINED!Transformer Neural Networks - EXPLAINED! (Attention is all you need)LSTM Networks - EXPLAINED!RMSProp (C2W2L07)The Evolution of Gradient DescentAdam Optimization Algorithm (C2W2L08)Variational Autoencoders - EXPLAINED!Lecture 3 | Loss Functions and OptimizationOptimization Tricks: momentum, batch-norm, and more | Lecture 10Activation Functions - EXPLAINED!Deep Learning-All Optimizers In One Video-SGD with Momentum,Adagrad,Adadelta,RMSprop,Adam OptimizersHyperparameter Optimization - The Math of Intelligence #7Variational Autoencoders134 - What are Optimizers in deep learning? (Keras & TensorFlow)Batch Normalization (“batch norm”) explainedGradient Descent With Momentum (C2W2L06)Gradient Descent, Step-by-StepWhy Does Batch Norm Work? (C2W3L06)Tutorial 16- AdaDelta and RMSprop optimizer