[BERT] Pretranied Deep Bidirectional Transformers for Language Understanding (discussions) | TDLS

Published on ● Video Link: https://www.youtube.com/watch?v=rMQMHA-uv_E



Category:
Discussion
Duration: 37:20
3,251 views
27


Toronto Deep Learning Series, 6 November 2018

Paper: https://arxiv.org/abs/1810.04805

Speaker: Danny Luo (Dessa)
https://dluo.me/

Host: Ada
Date: Nov 6th, 2018

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers. Unlike recent language representation models, BERT is designed to pre-train deep bidirectional representations by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT representations can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications.
BERT is conceptually simple and empirically powerful. It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE benchmark to 80.4% (7.6% absolute improvement), MultiNLI accuracy to 86.7 (5.6% absolute improvement) and the SQuAD v1.1 question answering Test F1 to 93.2 (1.5% absolute improvement), outperforming human performance by 2.0%.




Other Videos By LLMs Explained - Aggregate Intellect - AI.SCIENCE


2018-12-16Automated Deep Learning: Joint Neural Architecture and Hyperparameter Search (discussions) | AISC
2018-12-16Automated Deep Learning: Joint Neural Architecture and Hyperparameter Search (algorithm) | AISC
2018-12-09Automated Vulnerability Detection in Source Code Using Deep Learning (discussions) | AISC
2018-12-09Automated Vulnerability Detection in Source Code Using Deep Learning (algorithm) | AISC
2018-12-05[DQN] Human-level control through deep reinforcement learning (discussions) | AISC Foundational
2018-12-05Deep Q-Learning paper explained: Human-level control through deep reinforcement learning (algorithm)
2018-12-03SMOTE, Synthetic Minority Over-sampling Technique (discussions) | AISC Foundational
2018-12-02TDLS - Classics: SMOTE, Synthetic Minority Over-sampling Technique (algorithm)
2018-11-30Visualizing Data using t-SNE (algorithm) | AISC Foundational
2018-11-30Visualizing Data using t-SNE (discussions) | AISC Foundational
2018-11-27[BERT] Pretranied Deep Bidirectional Transformers for Language Understanding (discussions) | TDLS
2018-11-27[BERT] Pretranied Deep Bidirectional Transformers for Language Understanding (algorithm) | TDLS
2018-11-27Neural Image Caption Generation with Visual Attention (algorithm) | AISC
2018-11-27Neural Image Caption Generation with Visual Attention (discussion) | AISC
2018-11-17PGGAN | Progressive Growing of GANs for Improved Quality, Stability, and Variation (part 2) | AISC
2018-11-16PGGAN | Progressive Growing of GANs for Improved Quality, Stability, and Variation (part 1) | AISC
2018-11-16(Original Paper) Latent Dirichlet Allocation (discussions) | AISC Foundational
2018-11-15(Original Paper) Latent Dirichlet Allocation (algorithm) | AISC Foundational
2018-10-31[Transformer] Attention Is All You Need | AISC Foundational
2018-10-25[Original attention] Neural Machine Translation by Jointly Learning to Align and Translate | AISC
2018-10-16[StackGAN++] Realistic Image Synthesis with Stacked Generative Adversarial Networks | AISC



Tags:
deep learning
nlp
machine learning
ai
artificial intelligence
representation learning
bert
bert nlp
google bert
bert google
bert model
bert deep learning
bert language model
bert explained
bert transformer
transformer