Optimization Tricks: momentum, batch-norm, and more
Deep Learning Crash Course playlist: https://www.youtube.com/playlist?list=PLWKotBjTDoLj3rXBL-nEIPRN9V3a9Cx07
How to Design a Convolutional Neural Network
https://www.youtube.com/watch?v=fTw3K8D5xDs&t=596s
Highlights:
Stochastic Gradient Descent
Momentum Algorithm
Learning Rate Schedules
Adaptive Methods: AdaGrad, RMSProp, and Adam
Internal Covariate Shift
Batch Normalization
Weight Initialization
Local Minima
Saddle Points
References and further reading:
Deep Learning by Ian Goodfellow:
http://www.deeplearningbook.org/
Stochastic gradient descent
https://en.wikipedia.org/wiki/Stochastic_gradient_descent
Adaptive subgradient methods for online learning and stochastic optimization
http://jmlr.org/papers/volume12/duchi11a/duchi11a.pdf
RMSProp Lecture Notes by Geoffrey Hinton
https://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf
Adam: A method for stochastic optimization
https://arxiv.org/pdf/1412.6980
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
https://arxiv.org/pdf/1502.03167.pdf
Saddle point
https://en.wikipedia.org/wiki/Saddle_point
Understanding the difficulty of training deep feedforward neural networks
http://proceedings.mlr.press/v9/glorot10a/glorot10a.pdf
#deeplearning #machinelearning
Видео Optimization Tricks: momentum, batch-norm, and more канала Leo Isikdogan
How to Design a Convolutional Neural Network
https://www.youtube.com/watch?v=fTw3K8D5xDs&t=596s
Highlights:
Stochastic Gradient Descent
Momentum Algorithm
Learning Rate Schedules
Adaptive Methods: AdaGrad, RMSProp, and Adam
Internal Covariate Shift
Batch Normalization
Weight Initialization
Local Minima
Saddle Points
References and further reading:
Deep Learning by Ian Goodfellow:
http://www.deeplearningbook.org/
Stochastic gradient descent
https://en.wikipedia.org/wiki/Stochastic_gradient_descent
Adaptive subgradient methods for online learning and stochastic optimization
http://jmlr.org/papers/volume12/duchi11a/duchi11a.pdf
RMSProp Lecture Notes by Geoffrey Hinton
https://www.cs.toronto.edu/~tijmen/csc321/slides/lecture_slides_lec6.pdf
Adam: A method for stochastic optimization
https://arxiv.org/pdf/1412.6980
Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift
https://arxiv.org/pdf/1502.03167.pdf
Saddle point
https://en.wikipedia.org/wiki/Saddle_point
Understanding the difficulty of training deep feedforward neural networks
http://proceedings.mlr.press/v9/glorot10a/glorot10a.pdf
#deeplearning #machinelearning
Видео Optimization Tricks: momentum, batch-norm, and more канала Leo Isikdogan
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
Tutorial 14- Stochastic Gradient Descent with MomentumBatch Normalization (“batch norm”) explainedRegularization and Data AugmentationDeep Learning-All Optimizers In One Video-SGD with Momentum,Adagrad,Adadelta,RMSprop,Adam OptimizersCan AI Create Original Art?Recurrent Neural NetworksHow to Predict Stock Market Crashes using Mathematical ModelsGradient Descent With Momentum (C2W2L06)Can deep learning predict the stock market?Batch Normalization - EXPLAINED!How to Design a Neural Network | 2020 EditionGradient Descent, Step-by-StepOptimization in Deep Learning | All Major Optimizers Explained in DetailFORMATION DEEP LEARNING COMPLETE (2021)10000 Hz | INSTANT THIRD EYE STIMULATION (WARNING!!!) 100% MOST POWERFUL THIRD EYE BINAURAL BEATSAdam Optimization Algorithm (C2W2L08)Unbiased Learning to Rank: Counterfactual and Online Approaches - The Web Conference 2020 Tutorial6. Design Definition and Multidisciplinary Optimization