Загрузка страницы

Multilingual BERT - Part 3 - Multi vs. Monolingual on XNLI

In Part 3, I'll walk you through some code for applying BERT-like models to Arabic.

We'll start by applying XLM-R (a multilingual model), and using the technique of Cross-Lingual Transfer.

Then, for comparison, we'll look through a second Notebook where I've applied a monolingual Arabic BERT model, and augmented its training data using machine-translated text.

If you'd like a copy of these two Notebooks, they're available to purchase from my site here:
https://www.chrismccormick.ai/multilingual-models?utm_source=youtube&utm_medium=vid_desc&utm_campaign=bert_mulitlingual&utm_content=vid6

==== Receive Updates ====
Sign up to hear about new content across my blog and channel: https://www.chrismccormick.ai/subscribe

Видео Multilingual BERT - Part 3 - Multi vs. Monolingual on XNLI канала ChrisMcCormickAI
Показать
Комментарии отсутствуют
Введите заголовок:

Введите адрес ссылки:

Введите адрес видео с YouTube:

Зарегистрируйтесь или войдите с
Информация о видео
6 октября 2020 г. 20:58:28
00:20:10
Яндекс.Метрика