Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis
In this video from 2018 Swiss HPC Conference, Torsten Hoefler from (ETH) Zürich presents: Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis.
"Deep Neural Networks (DNNs) are becoming an important tool in modern computing applications. Accelerating their training is a major challenge and techniques range from distributed algorithms to low-level circuit design. In this talk, we describe the problem from a theoretical perspective, followed by approaches for its parallelization.
Specifically, we present trends in DNN architectures and the resulting implications on parallelization strategies. We discuss the different types of concurrency in DNNs; synchronous and asynchronous stochastic gradient descent; distributed system architectures; communication schemes; and performance modeling. Based on these approaches, we extrapolate potential directions for parallelism in deep learning."
Learn more: http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Видео Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis канала InsideHPC Report
"Deep Neural Networks (DNNs) are becoming an important tool in modern computing applications. Accelerating their training is a major challenge and techniques range from distributed algorithms to low-level circuit design. In this talk, we describe the problem from a theoretical perspective, followed by approaches for its parallelization.
Specifically, we present trends in DNN architectures and the resulting implications on parallelization strategies. We discuss the different types of concurrency in DNNs; synchronous and asynchronous stochastic gradient descent; distributed system architectures; communication schemes; and performance modeling. Based on these approaches, we extrapolate potential directions for parallelism in deep learning."
Learn more: http://hpcadvisorycouncil.com
Sign up for our insideHPC Newsletter: http://insidehpc.com/newsletter
Видео Demystifying Parallel and Distributed Deep Learning: An In-Depth Concurrency Analysis канала InsideHPC Report
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![Distributed TensorFlow training (Google I/O '18)](https://i.ytimg.com/vi/bRMGoPqsn20/default.jpg)
![Hogwild for Machine Learning on Multicore](https://i.ytimg.com/vi/l5JqUvTdZts/default.jpg)
![Geoffrey Hinton talk "What is wrong with convolutional neural nets ?"](https://i.ytimg.com/vi/rTawFwUvnLE/default.jpg)
![Study Techniques - The Good, Bad, & Useless](https://i.ytimg.com/vi/w1cX89XgqUM/default.jpg)
![Parallel Streams, CompletableFuture, and All That: Concurrency in Java 8](https://i.ytimg.com/vi/x5akmCWgGY0/default.jpg)
![Distributed Systems in One Lesson by Tim Berglund](https://i.ytimg.com/vi/Y6Ev8GIlbxc/default.jpg)
![PyTorch Lightning - William Falcon](https://i.ytimg.com/vi/TM_jRrXYXxc/default.jpg)
![An Introduction to GPU Programming with CUDA](https://i.ytimg.com/vi/1cHx1baKqq0/default.jpg)
![The Issue with Machine Learning in Finance](https://i.ytimg.com/vi/vKgnxf1NFB8/default.jpg)
![Ian Goodfellow: Adversarial Machine Learning (ICLR 2019 invited talk)](https://i.ytimg.com/vi/sucqskXRkss/default.jpg)
![Distributed Training in the Cloud](https://i.ytimg.com/vi/J_d4bEKUG2Q/default.jpg)
![SysML 19: Jia Zhihao, Beyond Data and Model Parallelism for Deep Neural Networks](https://i.ytimg.com/vi/81l6kkV-OkE/default.jpg)
![Lecture 17.6 — Large Scale Machine Learning | Map Reduce And Data Parallelism — [ Andrew Ng ]](https://i.ytimg.com/vi/TCA2VuHTHcM/default.jpg)
![Training on multiple GPUs and multi-node training with PyTorch DistributedDataParallel](https://i.ytimg.com/vi/a6_pY9WwqdQ/default.jpg)
![Machine Learning in R: Speed up Model Building with Parallel Computing](https://i.ytimg.com/vi/CMTlVuq6Wgw/default.jpg)
![Lecture 17.2 — Large Scale Machine Learning | Stochastic Gradient Descent — [ Andrew Ng ]](https://i.ytimg.com/vi/W9iWNJNFzQI/default.jpg)
![Understanding Mini-Batch Gradient Dexcent (C2W2L02)](https://i.ytimg.com/vi/-_4Zi8fCZO4/default.jpg)
![Distributed Machine Learning with Python](https://i.ytimg.com/vi/eVvjbTZc1CM/default.jpg)
![[deep learning] Federated Learning - training on decentralized data](https://i.ytimg.com/vi/KxZXhzfDgik/default.jpg)
![How to Manage GPU Resource Utilization in Tensorflow and Keras](https://i.ytimg.com/vi/1bEqRvJrZgE/default.jpg)