32. Bahador Khaleghi - Explainable AI and AI interpretability

Published: May 6, 2020, 3:14 p.m.

If I were to ask you to explain why you\u2019re reading this blog post, you could answer in many different ways.

\n

For example, you could tell me \u201cit\u2019s because I felt like it\u201d, or \u201cbecause my neurons fired in a specific way that led me to click on the link that was advertised to me\u201d. Or you might go even deeper and relate your answer to the fundamental laws of quantum physics.

\n

The point is, explanations need to be targeted to a certain level of abstraction in order to be effective.

\n

That\u2019s true in life, but it\u2019s also true in machine learning, where explainable AI is getting more and more attention as a way to ensure that models are working properly, in a way that makes sense to us. Understanding explainability and how to leverage it is becoming increasingly important, and that\u2019s why I wanted to speak with Bahador Khaleghi, a data scientist at H20.ai whose technical focus is on explainability and interpretability in machine learning.