Research talk: Post-contextual-bandit inference

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=UnJdPrjSr7M



Duration: 0:00
513 views
15


Speaker: Nathan Kallus, Associate Professor, Cornell University

Contextual bandit algorithms are increasingly replacing non-adaptive A/B tests in e-commerce, healthcare, and policymaking because they can both improve outcomes for study participants and increase the chance of identifying good or even best policies. To support credible inference on novel interventions at the end of the study, nonetheless, we still want to construct valid confidence intervals on average treatment effects, subgroup effects, or value of new policies. The adaptive nature of the data collected by contextual bandit algorithms, however, makes this difficult: standard estimators are no longer asymptotically normally distributed and classic confidence intervals fail to provide correct coverage. While this has been addressed in non-contextual settings by using stabilized estimators, the contextual setting poses unique challenges that we tackle for the first time in this paper. We propose the Contextual Adaptive Doubly Robust (CADR) estimator, the first estimator for policy value that is asymptotically normal under contextual adaptive data collection. The main technical challenge in constructing CADR is designing adaptive and consistent conditional standard deviation estimators for stabilization. Extensive numerical experiments using 57 OpenML datasets demonstrate that confidence intervals based on CADR uniquely provide correct coverage.

Learn more about the 2021 Microsoft Research Summit: https://Aka.ms/researchsummit




Other Videos By Microsoft Research


2022-02-08Closing Remarks: Health and Life Sciences - Delivery
2022-02-08Tutorial: Translating real-world data into evidence
2022-02-08Future of technology in combatting disease and disparities in treatment: A cardiovascular case study
2022-02-08Research talks: Showcasing health equity, access, and resilience collaborations
2022-02-08Opening remarks: Health & Life Sciences - Discovery
2022-02-08The role of tech in decreasing health inequities, improving access, and strengthening resilience
2022-02-08Opening remarks: Health and Life Sciences - Delivery
2022-02-08Closing remarks: Causal Machine Learning
2022-02-08Demo: Enabling end-to-end causal inference at scale
2022-02-08Research Talk: Enhancing the robustness of massive language models via invariant risk minimization
2022-02-08Research talk: Post-contextual-bandit inference
2022-02-08Research talk: Causal ML and fairness
2022-02-08Research talk: Causal learning: Discovering causal relations for out-of-distribution generalization
2022-02-08Research talk: Can causal learning improve the privacy of ML models?
2022-02-08Research talk: Causal ML and business
2022-02-08Research talk: Challenges and opportunities in causal machine learning
2022-02-08Opening remarks: Causal Machine Learning
2022-02-08Closing remarks: The Future of Privacy and Security
2022-02-08Demo: Generating formally proven low-level parsers with EverParse
2022-02-08Demo: EverParse: Automatic generation of formally verified secure parsers for cloud integrity
2022-02-08Research talk: DARPA SafeDocs: an approach to secure parsing and information interchange formats