The Linear Algebraic Structure of Word Meanings

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=gaVR3WnczOQ



Duration: 1:31:56
396 views
1


Word embeddings are often constructed with discriminative models such as deep nets and word2vec. Mikolov et al (2013) showed that these embeddings exhibit linear structure that is useful in solving "word analogy tasks". Subsequently, Levy and Goldberg (2014) and Pennington et al (2014) tried to explain why such linear structure should arise in embeddings derived from nonlinear methods. We provide a new generative model "explanation" for various word embedding methods as well as the above-mentioned linear structure. It also gives a generative explanation of older vector space methods such as the PMI method of Church and Hanks (1990). The model has surprising predictions (e.g., the spatial isotropy of word vectors), which are empirically verified. It also directly leads to a linear algebraic understanding of how a word embedding behaves when the word is polysemous (has multiple meanings), and to recover the different meanings from the embedding. This methodology and generative model may be useful for other NLP tasks and neural models. Joint work with Sanjeev Arora, Yuanzhi Li, Yingyu Liang, and Andrej Risteski (listed in alphabetical order).




Other Videos By Microsoft Research


2016-06-13How Much Information Does a Human Translator Add to the Original and Multi-Source Neural Translation
2016-06-13Opportunities and Challenges in Global Network Cameras
2016-06-13Nature in the City: Changes in Bangalore over Time and Space
2016-06-13Making Small Spaces Feel Large: Practical Illusions in Virtual Reality
2016-06-13Machine Learning as Creative Tool for Designing Real-Time Expressive Interactions
2016-06-13Recent Developments in Combinatorial Optimization
2016-06-13Computational Limits in Statistical Inference: Hidden Cliques and Sum of Squares
2016-06-13Coloring the Universe: An Insider's Look at Making Spectacular Images of Space
2016-06-13Towards Understandable Neural Networks for High Level AI Tasks - Part 6
2016-06-13The 37th UW/MS Symposium in Computational Linguistics
2016-06-13The Linear Algebraic Structure of Word Meanings
2016-06-13Machine Learning Algorithms Workshop
2016-06-13Interactive and Interpretable Machine Learning Models for Human Machine Collaboration
2016-06-13Improving Access to Clinical Data Locked in Narrative Reports: An Informatics Approach
2016-06-13Representation Power of Neural Networks
2016-06-13Green Security Games
2016-06-13e-NABLE: A Global Network of Digital Humanitarians on an Infrastructure of Electronic Communications
2016-06-10Microsoft Research New England: An introduction
2016-06-06Python+Machine Learning tutorial - Data munging for predictive modeling with pandas and scikit-learn
2016-06-06Symposium: Deep Learning - Xiaogang Wang
2016-06-06Symposium: Deep Learning - Leon Gatys



Tags:
microsoft research
word2vec
algorithms
nlp