Mirror-Generative Neural Machine Translation | NLP Journal Club
Paper link: https://openreview.net/forum?id=HkxQRTNYPH
Abstract: Training neural machine translation models (NMT) requires a large amount of parallel corpus, which is scarce for many language pairs. However, raw non-parallel corpora are often easy to obtain. Existing approaches have not exploited the full potential of non-parallel bilingual data either in training or decoding. In this paper, we propose the mirror-generative NMT (MGNMT), a single unified architecture that simultaneously integrates the source to target translation model, the target to source translation model, and two language models. Both translation models and language models share the same latent semantic space, therefore both translation directions can learn from non-parallel data more effectively. Besides, the translation models and language models can collaborate together during decoding. Our experiments show that the proposed MGNMT consistently outperforms existing approaches in all a variety of scenarios and language pairs, including resource-rich and low-resource languages.
Keywords: neural machine translation, generative model, mirror
Subscribe for some more upcoming NLP videos!
Видео Mirror-Generative Neural Machine Translation | NLP Journal Club канала The NLP Lab
Abstract: Training neural machine translation models (NMT) requires a large amount of parallel corpus, which is scarce for many language pairs. However, raw non-parallel corpora are often easy to obtain. Existing approaches have not exploited the full potential of non-parallel bilingual data either in training or decoding. In this paper, we propose the mirror-generative NMT (MGNMT), a single unified architecture that simultaneously integrates the source to target translation model, the target to source translation model, and two language models. Both translation models and language models share the same latent semantic space, therefore both translation directions can learn from non-parallel data more effectively. Besides, the translation models and language models can collaborate together during decoding. Our experiments show that the proposed MGNMT consistently outperforms existing approaches in all a variety of scenarios and language pairs, including resource-rich and low-resource languages.
Keywords: neural machine translation, generative model, mirror
Subscribe for some more upcoming NLP videos!
Видео Mirror-Generative Neural Machine Translation | NLP Journal Club канала The NLP Lab
Показать
Комментарии отсутствуют
Информация о видео
Другие видео канала
![CTRLsum: Towards Generic Controllable Text Summarization](https://i.ytimg.com/vi/6DsvJiFverQ/default.jpg)
![Jukebox: A Generative Model for Music | NLP Journal Club](https://i.ytimg.com/vi/OlpEdLavHlc/default.jpg)
![Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks | NLP journal club](https://i.ytimg.com/vi/E2AcqHuqFuk/default.jpg)
![Memorizing Transformers](https://i.ytimg.com/vi/aWOkC9US7EY/default.jpg)
![Answering Complex Open-Domain Questions with Multi-Hop Dense Retrieval | NLP Journal Club](https://i.ytimg.com/vi/ltFe40taFqM/default.jpg)
![Pitfalls of Static Language Modelling | NLP Journal Club](https://i.ytimg.com/vi/ybq7Sb9A76U/default.jpg)
![How Close is ChatGPT to Human Experts?](https://i.ytimg.com/vi/XkT9PN6zqpE/default.jpg)
![Best General-purpose NLP Libraries to Use in 2021](https://i.ytimg.com/vi/sFw1mJ4b5R0/default.jpg)
![Who will Win the Large Language Model App Race? with @Slator](https://i.ytimg.com/vi/7o1NWN2pmsQ/default.jpg)
![Learning to Reason and Memorize with Self-Notes | Paper summary](https://i.ytimg.com/vi/qT5kUc3vwzg/default.jpg)
![Falcon LLM: the Best Open-source LLM Available at the Moment](https://i.ytimg.com/vi/daU8q9eyzfY/default.jpg)
![Tree of Thoughts: Deliberate Problem Solving with Large Language Models | Paper summary](https://i.ytimg.com/vi/RndhsZvr-cI/default.jpg)
![When to use a large language model? 4 points to consider in 2023.](https://i.ytimg.com/vi/SR4LzhS52v0/default.jpg)
![QURIOUS: Question Generation Pretraining for Text Generation | NLP Journal Club](https://i.ytimg.com/vi/K36dW4uJ4ho/default.jpg)
![GreaseLM: Graph REASoning Enhanced Language Models for Question Answering](https://i.ytimg.com/vi/5woxVibfR4U/default.jpg)
![A Distributional Approach to Controlled Text Generation | NLP Journal Club](https://i.ytimg.com/vi/RJ9TT81i338/default.jpg)
![REALM: Retrieval-Augmented Language Model Pre-Training | NLP Journal Club](https://i.ytimg.com/vi/mFoEig-Xi_0/default.jpg)
![Multi-scale Transformer Language Models](https://i.ytimg.com/vi/SO37e1Ho-Y0/default.jpg)
![*Paper summary* ByT5: Towards a Token-Free Future with Pre-trained Byte-to-Byte Models](https://i.ytimg.com/vi/JHOvr_SF5PI/default.jpg)
![Deep Subjecthood: Higher-Order Grammatical Features in Multilingual BERT | NLP Journal Club](https://i.ytimg.com/vi/EwKreP-pXAQ/default.jpg)