[BERT] Pretranied Deep Bidirectional Transformers for Language Understanding (algorithm) | TDLS

Published on ● Video Link: https://www.youtube.com/watch?v=BhlOGGzC0Q0



Duration: 53:07
83,060 views
1,000


Toronto Deep Learning Series
Host: Ada + @ML Explained - Aggregate Intellect - AI.SCIENCE
Date: Nov 6th, 2018

Aggregate Intellect is a Global Marketplace where ML Developers Connect, Collaborate, and Build.
-Connect with peers & experts at https://ai.science
-Join our Slack Community: https://join.slack.com/t/aisc-to/shared_invite/zt-f5zq5l35-PSIJTFk4v60FML177PgsPg
-Check out the user generated Recipes that provide step by step, and bite sized guides on how to do various tasks: https://ai.science/recipes

For details including slides, visit https://aisc.ai.science/events/2018-11-06

Paper: https://arxiv.org/abs/1810.04805

Speaker: Danny Luo (Dessa)
https://dluo.me/

BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding

We introduce a new language representation model called BERT, which stands for Bidirectional Encoder Representations from Transformers. Unlike recent language representation models, BERT is designed to pre-train deep bidirectional representations by jointly conditioning on both left and right context in all layers. As a result, the pre-trained BERT representations can be fine-tuned with just one additional output layer to create state-of-the-art models for a wide range of tasks, such as question answering and language inference, without substantial task-specific architecture modifications.
BERT is conceptually simple and empirically powerful. It obtains new state-of-the-art results on eleven natural language processing tasks, including pushing the GLUE benchmark to 80.4% (7.6% absolute improvement), MultiNLI accuracy to 86.7 (5.6% absolute improvement) and the SQuAD v1.1 question answering Test F1 to 93.2 (1.5% absolute improvement), outperforming human performance by 2.0%.




Other Videos By LLMs Explained - Aggregate Intellect - AI.SCIENCE


2018-12-16Automated Deep Learning: Joint Neural Architecture and Hyperparameter Search (algorithm) | AISC
2018-12-09Automated Vulnerability Detection in Source Code Using Deep Learning (discussions) | AISC
2018-12-09Automated Vulnerability Detection in Source Code Using Deep Learning (algorithm) | AISC
2018-12-05[DQN] Human-level control through deep reinforcement learning (discussions) | AISC Foundational
2018-12-05Deep Q-Learning paper explained: Human-level control through deep reinforcement learning (algorithm)
2018-12-03SMOTE, Synthetic Minority Over-sampling Technique (discussions) | AISC Foundational
2018-12-02TDLS - Classics: SMOTE, Synthetic Minority Over-sampling Technique (algorithm)
2018-11-30Visualizing Data using t-SNE (algorithm) | AISC Foundational
2018-11-30Visualizing Data using t-SNE (discussions) | AISC Foundational
2018-11-27[BERT] Pretranied Deep Bidirectional Transformers for Language Understanding (discussions) | TDLS
2018-11-27[BERT] Pretranied Deep Bidirectional Transformers for Language Understanding (algorithm) | TDLS
2018-11-27Neural Image Caption Generation with Visual Attention (algorithm) | AISC
2018-11-27Neural Image Caption Generation with Visual Attention (discussion) | AISC
2018-11-17PGGAN | Progressive Growing of GANs for Improved Quality, Stability, and Variation (part 2) | AISC
2018-11-16PGGAN | Progressive Growing of GANs for Improved Quality, Stability, and Variation (part 1) | AISC
2018-11-16(Original Paper) Latent Dirichlet Allocation (discussions) | AISC Foundational
2018-11-15(Original Paper) Latent Dirichlet Allocation (algorithm) | AISC Foundational
2018-10-31[Transformer] Attention Is All You Need | AISC Foundational
2018-10-25[Original attention] Neural Machine Translation by Jointly Learning to Align and Translate | AISC
2018-10-16[StackGAN++] Realistic Image Synthesis with Stacked Generative Adversarial Networks | AISC
2018-10-11Bayesian Deep Learning on a Quantum Computer | TDLS Author Speaking



Tags:
deep learning
meachnie learning
ai
artificial intelligence
natural language processing
nlp
bert
bert nlp
google bert
bert google
bert model
bert deep learning
bert language model
bert explained
bert transformer