Consistency by Agreement in Zero-shot Neural Machine Translation | AISC

Published on ● Video Link: https://www.youtube.com/watch?v=2vR06ih4010



Duration: 1:02:43
436 views
6


For slides and more information on the paper, visit https://aisc.ai.science/events/3919-08-28

Discussion lead: Maruan Al-Shedivat


Motivation:
Generalization and reliability of multilingual translation often highly depend on the amount of available parallel data for each language pair of interest. In this paper, we focus on zero-shot generalization---a challenging setup that tests models on translation directions they have not been optimized for at training time. To solve the problem, we (i) reformulate multilingual translation as probabilistic inference, (ii) define the notion of zero-shot consistency and show why standard training often results in models unsuitable for zero-shot tasks, and (iii) introduce a consistent agreement-based training method that encourages the model to produce equivalent translations of parallel sentences in auxiliary languages. We test our multilingual NMT models on multiple public zero-shot translation benchmarks (IWSLT17, UN corpus, Europarl) and show that agreement-based learning often results in 2-3 BLEU zero-shot improvement over strong baselines without any loss in performance on supervised translation directions.




Other Videos By LLMs Explained - Aggregate Intellect - AI.SCIENCE


2019-09-24Lookahead Optimizer: k steps forward, 1 step back
2019-09-24Similarity of neural network representations revisited
2019-09-23Detecting Customer Complaint Escalation w/ Recurrent Neural Networks & Manually-Engineered Features
2019-09-23Graph Normalizing Flows
2019-09-23CNN Architectures for Large-Scale Audio Classification | AISC
2019-09-222019 AI Squared Forum Paper Track | AISC
2019-09-16Making of a conversational agent platform | AISC
2019-09-09A Survey of Singular Learning | AISC
2019-09-04Overview of Reinforcement Learning | AISC
2019-09-03Ernie 2.0: A Continual Pre-Training Framework for Language Understanding | AISC
2019-08-28Consistency by Agreement in Zero-shot Neural Machine Translation | AISC
2019-08-26TensorFuzz: Debugging Neural Networks with Coverage-Guided Fuzzing | AISC
2019-08-21Science of science: Identifying Fundamental Drivers of Science | AISC
2019-08-19AI Product Stream Meet and Greet | AISC
2019-08-12[Original ResNet paper] Deep Residual Learning for Image Recognition | AISC
2019-08-11[GAT] Graph Attention Networks | AISC Foundational
2019-08-06XLNet: Generalized Autoregressive Pretraining for Language Understanding | AISC
2019-07-31Overview of Generative Adversarial Networks | AISC
2019-07-29Multi-Armed Bandit Strategies for Non-Stationary Reward Distributions and Delayed Feedback Processes
2019-07-22AISC Abstract Night
2019-07-15The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words & Sentences From Natural Supervision