Bringing Order to Chaos: Navigating the Disagreement Problem in Explainable ML

Published on ● Video Link: https://www.youtube.com/watch?v=AUTxyTbij-Q



Duration: 47:25
601 views
3


Hima Lakkaraju (Harvard University)
https://simons.berkeley.edu/node/22930
Societal Considerations and Applications

As various post hoc explanation methods are increasingly being leveraged to explain complex models in high-stakes settings, it becomes critical to develop a deeper understanding of if and when the explanations output by these methods disagree with each other, why these disagreements occur, and how to address these disagreements in a rigorous fashion. However, there is little to no research that provides answers to these critical questions. In this talk, I will present some of our recent research which addresses the aforementioned questions. More specifically, I will discuss i) a novel quantitative framework to formalize the disagreement between state-of-the-art feature attribution based explanation methods (e.g., LIME, SHAP, Gradient based methods). I will also touch upon on how this framework was constructed by leveraging inputs from interviews and user studies with data scientists who utilize explanation methods in their day-to-day work; ii) an online user study to understand how data scientists resolve disagreements in explanations output by the aforementioned methods; iii) a novel function approximation framework to explain why explanation methods often disagree with each other. I will demonstrate that all the key feature attribution based explanation methods are essentially performing local function approximations albeit, with different loss functions and notions of neighborhood. (iv) a set of guiding principles on how to choose explanation methods and resulting explanations when they disagree in real-world settings. I will conclude this talk by presenting a brief overview of an open source framework that we recently developed called Open-XAI which enables researchers and practitioners to seamlessly evaluate and benchmark both existing and new explanation methods based on various characteristics such as faithfulness, stability, and fairness.




Other Videos By Simons Institute for the Theory of Computing


2022-11-22Introduction and Research Fellows Lightning Talks
2022-11-16Using Theories of Decision-Making Under Uncertainty to Improve Data Visualization
2022-11-15Graphons and Graphexes as Models of Large Sparse Networks
2022-11-11Re-designing Recommendation on VolunteerMatch: Theory and Practice
2022-11-11Efficient and Targeted COVID-19 Border Testing via Reinforcement Learning
2022-11-10Unpacking the Black Box: Regulating Algorithmic Decisions
2022-11-10Decision-Aware Learning for Global Health Supply Chains
2022-11-10Supply-Side Equilibria in Recommender Systems
2022-11-10What Really Matters for Fairness in Machine Learning: Delayed Impact and Other Desiderata
2022-11-10Predictive Modeling in Healthcare – Special Considerations
2022-11-10Bringing Order to Chaos: Navigating the Disagreement Problem in Explainable ML
2022-11-09Pipeline Interventions
2022-11-09Algorithmic Challenges in Ensuring Fairness at the Time of Decision
2022-11-09Improving Refugee Resettlement
2022-11-09Learning to Predict Arbitrary Quantum Processes
2022-11-09A Kerfuffle: Differential Privacy and the 2020 Census
2022-11-08Chasing the Long Tail: What Neural Networks Memorize and Why
2022-11-08Concurrent Composition Theorems for all Standard Variants of Differential Privacy
2022-11-08Privacy Management: Achieving the Possimpible
2022-11-07Privacy-safe Measurement on the Web: Open Questions From the Privacy Sandbox
2022-10-29Transmission Neural Networks: From Virus Spread Models to Neural Networks



Tags:
Simons Institute
theoretical computer science
UC Berkeley
Computer Science
Theory of Computation
Theory of Computing
Epidemics and Information Diffusion
Hima Lakkaraju