Anna Veronika Dorogush - CatBoost - the new generation of Gradient Boosting
CatBoost - the new generation of Gradient Boosting
[EuroPython 2018 - Talk - 2018-07-26 - PyCharm [PyData]]
[Edinburgh, UK]
By Anna Veronika Dorogush
Gradient boosting is a powerful machine-learning technique that achieves state-of-the-art results
in a variety of practical tasks. For a number of years, it has remained the primary method for
learning problems with heterogeneous features, noisy data, and complex dependencies: web search,
recommendation systems, weather forecasting, and many others.
CatBoost (http://catboost.yandex) is a new open-source gradient boosting library, that outperforms existing publicly available implementations of gradient boosting in terms of quality. It has a set of addional advantages.
CatBoost is able to incorporate categorical features in your data (like music genre, URL, search query, etc.) in predictive models with no additional preprocessing. For more details on our approach please refer to our NIPS 2017 ML Systems Workshop paper (http://learningsys.org/nips17/assets/papers/paper_11.pdf).
CatBoost inference is 20-60 times faster then in other open-source gradient boosting libraries, which makes it possible to use CatBoost for latency-critical tasks.
CatBoost has the fastest GPU and multi GPU training implementations of all the openly available gradient boosting libraries.
CatBoost requires no hyperparameter tunning in order to get a model with good quality.
CatBoost is highly scalable and can be efficiently trained using hundreds of machines
The talk will cover a broad description of gradient boosting and its areas of usage and the differences between CatBoost and other gradient boosting libraries. We will also briefly explain the details of the proprietary algorithm that leads to a boost in quality.
License: This video is licensed under the CC BY-NC-SA 3.0 license: https://creativecommons.org/licenses/by-nc-sa/3.0/
Please see our speaker release agreement for details: https://ep2018.europython.eu/en/speaker-release-agreement/
Видео Anna Veronika Dorogush - CatBoost - the new generation of Gradient Boosting канала EuroPython Conference
[EuroPython 2018 - Talk - 2018-07-26 - PyCharm [PyData]]
[Edinburgh, UK]
By Anna Veronika Dorogush
Gradient boosting is a powerful machine-learning technique that achieves state-of-the-art results
in a variety of practical tasks. For a number of years, it has remained the primary method for
learning problems with heterogeneous features, noisy data, and complex dependencies: web search,
recommendation systems, weather forecasting, and many others.
CatBoost (http://catboost.yandex) is a new open-source gradient boosting library, that outperforms existing publicly available implementations of gradient boosting in terms of quality. It has a set of addional advantages.
CatBoost is able to incorporate categorical features in your data (like music genre, URL, search query, etc.) in predictive models with no additional preprocessing. For more details on our approach please refer to our NIPS 2017 ML Systems Workshop paper (http://learningsys.org/nips17/assets/papers/paper_11.pdf).
CatBoost inference is 20-60 times faster then in other open-source gradient boosting libraries, which makes it possible to use CatBoost for latency-critical tasks.
CatBoost has the fastest GPU and multi GPU training implementations of all the openly available gradient boosting libraries.
CatBoost requires no hyperparameter tunning in order to get a model with good quality.
CatBoost is highly scalable and can be efficiently trained using hundreds of machines
The talk will cover a broad description of gradient boosting and its areas of usage and the differences between CatBoost and other gradient boosting libraries. We will also briefly explain the details of the proprietary algorithm that leads to a boost in quality.
License: This video is licensed under the CC BY-NC-SA 3.0 license: https://creativecommons.org/licenses/by-nc-sa/3.0/
Please see our speaker release agreement for details: https://ep2018.europython.eu/en/speaker-release-agreement/
Видео Anna Veronika Dorogush - CatBoost - the new generation of Gradient Boosting канала EuroPython Conference
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![Gradient Boosting : Data Science's Silver Bullet](https://i.ytimg.com/vi/en2bmeB4QUo/default.jpg)
![Anna Veronika Dorogush: Mastering gradient boosting with CatBoost | PyData London 2019](https://i.ytimg.com/vi/usdEWSDisS0/default.jpg)
![](https://i.ytimg.com/vi/DiMozPkN8Ho/default.jpg)
![Y-DATA Tel Aviv #11 - Anna Veronica Dorogush: New capabilities of CatBoost](https://i.ytimg.com/vi/BgDmuvPaUBo/default.jpg)
![Can one do better than XGBoost? - Mateusz Susik](https://i.ytimg.com/vi/5CWwwtEM2TA/default.jpg)
![SHAP - What Is Your Model Telling You? Interpret CatBoost Regression and Classification Outputs](https://i.ytimg.com/vi/ZkIxZ5xlMuI/default.jpg)
![CatBoost VS XGboost - It's Modeling Cat Fight Time! Welcome to 5 Minutes for Data Science](https://i.ytimg.com/vi/dvZLk7LxGzc/default.jpg)
![CatBoost: Fast Open-Source Gradient Boosting Library For GPU - Vasily Ershov](https://i.ytimg.com/vi/aHSw4BQkK3s/default.jpg)
![LA Data Science Meetup, March 30, 2021 - Stanislav Kirillov: CatBoost](https://i.ytimg.com/vi/aTtD2QlQ9n0/default.jpg)
![Let's take a look at CatBoost - gradient boosted regression & classification (webinar recorded 1/5)](https://i.ytimg.com/vi/h8pJqpfo8uk/default.jpg)
![AdaBoost, Clearly Explained](https://i.ytimg.com/vi/LsK-xG1cLYA/default.jpg)
![How Distributed LightGBM on Dask Works - James Lamb| Dask Summit 2021](https://i.ytimg.com/vi/XFVcoBu1rNw/default.jpg)
![17. Learning: Boosting](https://i.ytimg.com/vi/UHBmv7qCey4/default.jpg)
![Jaroslaw Szymczak - Gradient Boosting in Practice: a deep dive into xgboost](https://i.ytimg.com/vi/s3VmuVPfu0s/default.jpg)
![CatBoost - The New Generation of Gradient Boosting - Vasily Ershov](https://i.ytimg.com/vi/ZAGXnXmDCT8/default.jpg)
![I Quit React Development - JavaScript Fatigue and Burnout](https://i.ytimg.com/vi/GENjgsim2SE/default.jpg)
![Y-DATA Tel Aviv #11 - Aviv Nahon: CatBoost A-Z: from experiments to production](https://i.ytimg.com/vi/PZxN-27NuDs/default.jpg)
![CatBoost - the new generation of gradient boosting - Anna Veronika Dorogush](https://i.ytimg.com/vi/8o0e-r0B5xQ/default.jpg)
![How to deal with Imbalanced Datasets in PyTorch - Weighted Random Sampler Tutorial](https://i.ytimg.com/vi/4JFVhJyTZ44/default.jpg)
![Better than Deep Learning: Gradient Boosting Machines (GBM)](https://i.ytimg.com/vi/9GCEVv94udY/default.jpg)