Research talk: Focal Attention: Towards local-global interactions in vision transformers

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=dhAz_hgoG1M



Duration: 7:40
432 views
0


Speaker: Jianwei Yang, Senior Researcher, Microsoft Research Redmond

At present, deep neural networks have become prevalent for building AI systems for vision, language and multimodality. However, how to build efficient and task-oriented models are still challenging problems for researchers. In these lightning talks, Senior RSDE Baolin Peng and Senior Researcher Jianwei Yang from the Deep Learning Group at Microsoft Research Redmond, will discuss end-to-end dialog systems and new architecture for vision systems, respectively. For dialog systems, an end-to-end learning system is achieved by using self-learning from the conversations with a human in the loop. For vision systems, a sparse attention mechanism has been developed for the Vision Transformer to cope with high-resolution image inputs for image classification, object detection and semantic segmentation.

Learn more about the 2021 Microsoft Research Summit: https://Aka.ms/researchsummit




Other Videos By Microsoft Research


2022-02-08Tutorial: Best practices for prioritizing fairness in AI systems
2022-02-08Demo: RAI Toolbox: An open-source framework for building responsible AI
2022-02-08Opening remarks: Responsible AI
2022-02-08Closing remarks: Deep Learning and Large Scale AI
2022-02-08Roundtable discussion: Beyond language models: Knowledge, multiple modalities, and more
2022-02-08Research talk: Closing the loop in natural language interfaces to relational databases
2022-02-08Just Tech: Bringing CS, the social sciences, and communities together for societal resilience
2022-02-08Research talk: WebQA: Multihop and multimodal
2022-02-08Opening remarks: Tech for resilient communities
2022-02-08Research talk: Towards Self-Learning End-to-end Dialog Systems
2022-02-08Research talk: Focal Attention: Towards local-global interactions in vision transformers
2022-02-08Research talk: Knowledgeable pre-trained language models
2022-02-08Opening remarks: Deep Learning and Large-Scale AI
2022-02-08Closing remarks: Cloud Intelligence/AIOps
2022-02-08Research talk: Optimizing the cloud supply chain
2022-02-08Research talk: Automating and Optimizing IT Operations Management with AI
2022-02-08Research talk: An intelligent data-driven paradigm towards cloud reliability
2022-02-08Talk: Multidimensional analysis of cloud-native software based on large-scale operation data
2022-02-08Keynote: Cloud Intelligence: Infusing AI into cloud computing systems
2022-02-08Opening remarks: Cloud Intelligence/AIOps
2022-02-08Research talk: NUWA: Neural visual world creation with multimodal pretraining



Tags:
deep learning
large-scale models
large-scale AI models
AI
artificial intelligence
microsoft research summit