Yoshua Bengio: Deep Learning Priors Associated With Conscious Processing | ICLR 2020
Join the channel membership:
https://www.youtube.com/c/AIPursuit/join
Subscribe to the channel:
https://www.youtube.com/c/AIPursuit?sub_confirmation=1
Support and Donation:
Paypal ⇢ https://paypal.me/tayhengee
Patreon ⇢ https://www.patreon.com/hengee
BTC ⇢ bc1q2r7eymlf20576alvcmryn28tgrvxqw5r30cmpu
ETH ⇢ 0x58c4bD4244686F3b4e636EfeBD159258A5513744
Doge ⇢ DSGNbzuS1s6x81ZSbSHHV5uGDxJXePeyKy
Wanted to own BTC, ETH, or even Dogecoin? Kickstart your crypto portfolio with the largest crypto market Binance with my affiliate link:
https://accounts.binance.com/en/register?ref=27700065
Video is reposted for educational purposes
Abstract:
Recent efforts in Dialogue State Tracking (DST) for task-oriented dialogues have progressed toward open-vocabulary or generation-based approaches where the models can generate slot value candidates from the dialogue history itself. These approaches have shown good performance gain, especially in complicated dialogue domains with dynamic slot values. However, they fall short in two aspects: (1) they do not allow models to explicitly learn signals across domains and slots to detect potential dependencies among \textit{(domain, slot)} pairs; and (2) existing models follow auto-regressive approaches which incur high time cost when the dialogue evolves over multiple domains and multiple turns. In this paper, we propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST) which can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots. In particular, the non-autoregressive nature of our method not only enables decoding in parallel to significantly reduce the latency of DST for real-time dialogue response generation, but also detect dependencies among slots at token level in addition to slot and domain level. Our empirical results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus, and the latency of our model is an order of magnitude lower than the previous state of the art as the dialogue history extends over time.
Видео Yoshua Bengio: Deep Learning Priors Associated With Conscious Processing | ICLR 2020 канала AIPursuit
https://www.youtube.com/c/AIPursuit/join
Subscribe to the channel:
https://www.youtube.com/c/AIPursuit?sub_confirmation=1
Support and Donation:
Paypal ⇢ https://paypal.me/tayhengee
Patreon ⇢ https://www.patreon.com/hengee
BTC ⇢ bc1q2r7eymlf20576alvcmryn28tgrvxqw5r30cmpu
ETH ⇢ 0x58c4bD4244686F3b4e636EfeBD159258A5513744
Doge ⇢ DSGNbzuS1s6x81ZSbSHHV5uGDxJXePeyKy
Wanted to own BTC, ETH, or even Dogecoin? Kickstart your crypto portfolio with the largest crypto market Binance with my affiliate link:
https://accounts.binance.com/en/register?ref=27700065
Video is reposted for educational purposes
Abstract:
Recent efforts in Dialogue State Tracking (DST) for task-oriented dialogues have progressed toward open-vocabulary or generation-based approaches where the models can generate slot value candidates from the dialogue history itself. These approaches have shown good performance gain, especially in complicated dialogue domains with dynamic slot values. However, they fall short in two aspects: (1) they do not allow models to explicitly learn signals across domains and slots to detect potential dependencies among \textit{(domain, slot)} pairs; and (2) existing models follow auto-regressive approaches which incur high time cost when the dialogue evolves over multiple domains and multiple turns. In this paper, we propose a novel framework of Non-Autoregressive Dialog State Tracking (NADST) which can factor in potential dependencies among domains and slots to optimize the models towards better prediction of dialogue states as a complete set rather than separate slots. In particular, the non-autoregressive nature of our method not only enables decoding in parallel to significantly reduce the latency of DST for real-time dialogue response generation, but also detect dependencies among slots at token level in addition to slot and domain level. Our empirical results show that our model achieves the state-of-the-art joint accuracy across all domains on the MultiWOZ 2.1 corpus, and the latency of our model is an order of magnitude lower than the previous state of the art as the dialogue history extends over time.
Видео Yoshua Bengio: Deep Learning Priors Associated With Conscious Processing | ICLR 2020 канала AIPursuit
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
The Rise of Artificial Intelligence through Deep Learning | Yoshua Bengio | TEDxMontrealYoshua Bengio: From Deep Learning to Consciousness - NeurIPS 2019 Robin.ly InterviewYoshua Bengio | A Meta-Transfer Objective for Learning to Disentangle Causal Mechanisms | ICLR 2020The Heidelberg Laureate Forum Foundation presents the HLF Portraits: Yoshua BengioYoshua Bengio: "Representation Learning and Deep Learning, Pt. 1"Deep Learning State of the Art (2020)Building AGI: Promising Approaches, Remaining Milestones, and Likely Obstacles | Yoshua BengioLecture 1 | Machine Learning (Stanford)AI DEBATE : Yoshua Bengio | Gary MarcusCS50P - Lectures 0, 1, 2 - Live ShootYoshua Bengio: Deep Learning | Lex Fridman Podcast #4Yoshua Bengio - Deep learning and Backprop in the Brain (CCN 2017)Yoshua Bengio: From Deep Learning to Consciousness - Robin.ly Interview @ NeurIPS 2019Consciousness and COVID-19 with Yoshua Bengio - #361Yoshua Bengio Extra Footage 1: Brainstorm with students 🔴Uniform convergence may be unable to explain generalization in deep learning | NeurIPSYoshua Bengio: From System 1 Deep Learning to System 2 Deep Learning (NeurIPS 2019)Yoshua Bengio: Deep Learning Cognition | Full Keynote - AI in 2020 & BeyondYoshua Bengio - More Hardware-Friendly Deep LearningFireside Chat with Daphne Koller (ICLR 2018)