Загрузка страницы

41. Pyspark: How to run Spark Application on Amazon EMR ElasticMapReduce cluster

Apache Spark is a data processing framework that can quickly perform processing tasks on very large data sets, and can also distribute data processing tasks across multiple computers, either on its own or in tandem with other distributed computing tools. These two qualities are key to the worlds of big data and machine learning, which require the marshalling of massive computing power to crunch through large data stores. Spark also takes some of the programming burdens of these tasks off the shoulders of developers with an easy-to-use API that abstracts away much of the grunt work of distributed computing and big data processing.

Видео 41. Pyspark: How to run Spark Application on Amazon EMR ElasticMapReduce cluster канала Analytics Excellence
Показать
Комментарии отсутствуют
Введите заголовок:

Введите адрес ссылки:

Введите адрес видео с YouTube:

Зарегистрируйтесь или войдите с
Информация о видео
9 сентября 2020 г. 11:00:09
00:15:11
Яндекс.Метрика