BottleSum: Unsupervised & Self-supervised Sentence Summarization w/ Information Bottleneck Principle

Published on ● Video Link: https://www.youtube.com/watch?v=4D_0SZHrkK0



Category:
Guide
Duration: 55:28
833 views
19


For slides and more information on the paper, visit https://aisc.ai.science/events/2020-01-22

Discussion lead: Peter West

Motivation:
The principle of the Information Bottleneck (Tishby et al., 1999) is to produce a summary of information X optimized to predict some other relevant information Y . In this paper, we propose a novel approach to unsupervised sentence ummarization by mapping the Information Bottleneck principle to a conditional language modelling objective: given a sentence, our approach seeks a compressed sentence that can best predict the next sentence. Our iterative algorithm under the Information Bottleneck objective searches gradually shorter subsequences of the given sentence while maximizing the probability of the next sentence conditioned on the summary. Using only pretrained language models with no direct supervision, our approach can efficiently perform extractive sentence summarization over a large corpus. Building on our unsupervised extractive summarization (BottleSumEx), we then present a new approach to self-supervised abstractive summarization (BottleSumSelf ), where a transformer-based language model is trained on the output summaries of our unsupervised method. Empirical results demonstrate that our extractive method outperforms other unsupervised models on multiple automatic metrics. In addition, we find that our selfsupervised abstractive model outperforms unsupervised baselines (including our own) by human evaluation along multiple attributes.




Other Videos By LLMs Explained - Aggregate Intellect - AI.SCIENCE


2020-02-25Protein secondary structure detection in intermediate-resolution cryo-EM maps using deep learning |
2020-02-24[AI Product] How to build products people actually want to use | AISC
2020-02-20Build and Deploy Machine Learning Models | MLOps Overview
2020-02-19Quantifying the dynamics of failure across science, startups and security | AISC
2020-02-18Deep Learning for Symbolic Mathematics | AISC
2020-02-12Visualizing and measuring the geometry of BERT | AISC
2020-02-11Attention is not not explanation + Character Eyes: Seeing Language through Character-Level Taggers |
2020-02-10Single Headed Attention RNN: Stop Thinking With Your Head | AISC
2020-01-27Identifying Big ML product opportunities inside Big organizations | AISC
2020-01-23Machine Learning in Cyber Security, Overview | AISC
2020-01-22BottleSum: Unsupervised & Self-supervised Sentence Summarization w/ Information Bottleneck Principle
2020-01-20A Hybrid GA-PSO Method for Evolving Architecture and Short Connections of Deep Convolutional Neural
2020-01-13Plug and Play Language Models: A Simple Approach to Controlled Text Generation | AISC
2020-01-08Overview of Modern Anomaly and Novelty Detection | AISC
2020-01-06Annotating Object Instances With a Polygon RNN | AISC
2019-12-11Predicting translational progress in biomedical research | AISC
2019-12-09AlphaStar explained: Grandmaster level in StarCraft II with multi-agent RL
2019-12-04How Can We Be So Dense? The Benefits of Using Highly Sparse Representations | AISC
2019-12-02[RoBERT & ToBERT] Hierarchical Transformers for Long Document Classification | AISC
2019-11-25[OpenAI] Solving Rubik's Cube with a Robot Hand | AISC
2019-11-18Top-K Off-Policy Correction for a REINFORCE Recommender System | AISC