Trustworthy AI: Towards Robust and Reliable Model Explanations | AI FOR GOOD DISCOVERY
As machine learning black boxes are increasingly being deployed in domains such as healthcare and criminal justice, there is growing emphasis on building tools and techniques for explaining these black boxes in an interpretable manner. Such explanations are being leveraged by domain experts to diagnose systematic errors and underlying biases of black boxes. In this talk, I will present some of our recent research that sheds light on the vulnerabilities of popular post hoc explanation techniques such as LIME and SHAP, and also introduce novel methods to address some of these vulnerabilities. More specifically, I will first demonstrate that these methods are brittle, unstable, and are vulnerable to a variety of adversarial attacks. Then, I will discuss two solutions to address some of the vulnerabilities of these methods – (i) a framework based on adversarial training that is designed to make post hoc explanations more stable and robust to shifts in the underlying data; (ii) a Bayesian framework that captures the uncertainty associated with post hoc explanations and in turn allows us to generate explanations with user specified levels of confidences. I will conclude the talk by discussing results from real world datasets to both demonstrate the vulnerabilities in post hoc explanation techniques as well as the efficacy of our aforementioned solutions.
🔴 Watch the latest #AIforGood videos:
Explore more #AIforGood content:
1️ ⃣ • Top Hits
2️ ⃣ • AI for Good Webinars
3️ ⃣ • AI for Good Keynotes
📅 Discover what's next on our programmhttps://aiforgood.itu.int/programme/me/
Social Media:
Websithttps://aiforgood.itu.int/nt/
Twittehttps://twitter.com/ITU_AIForGoodd
LinkedIn Paghttps://www.linkedin.com/company/265119077
LinkedIn Grouhttps://www.linkedin.com/groups/85677488
Instagrahttps://www.instagram.com/aiforgoodd
Faceboohttps://www.facebook.com/AIforGoodd
WHAT IS TRUSTWORTHY AI SERIES?
Artificial Intelligence (AI) systems have steadily grown in complexity, gaining predictivity often at the expense of interpretability, robustness and trustworthiness. Deep neural networks are a prime example of this development. While reaching “superhuman” performances in various complex tasks, these models are susceptible to errors when confronted with tiny (adversarial) variations of the input – variations which are either not noticeable or can be handled reliably by humans. This expert talk series will discuss these challenges of current AI technology and will present new research aiming at overcoming these limitations and developing AI systems which can be certified to be trustworthy and robust.
What is AI for Good?
The AI for Good series is the leading action-oriented, global & inclusive United Nations platform on AI. The Summit is organized all year, always online, in Geneva by the ITU with XPRIZE Foundation in partnership with over 35 sister United Nations agencies, Switzerland and ACM. The goal is to identify practical applications of AI and scale those solutions for global impact.
Disclaimer:
The views and opinions expressed are those of the panelists and do not reflect the official policy of the ITU.
#trustworthyAI #reliableAI