One of the biggest criticisms of the field of post hoc #XAI is that each method "does its own thing", it is unclear how these methods relate to each other & which methods are effective under what conditions. Our #NeurIPS2022 paper provides (some) answers to these questions. [1/N]
In our #NeurIPS2022 paper, we unify eight different state-of-the-art local post hoc explanation methods, and show that they are all performing local linear approximations of the underlying models, albeit with different loss functions and notions of local neighborhoods. [2/N]
By doing so, we are able to explain the similarities & differences between these methods. These methods are similar in the sense that they all perform local linear approximations of models, but they differ considerably in "how" they perform these approximations [3/N]
Literature on local post hoc explanation methods is also commonly criticized for the fact that the definition of "locality" is often unclear. Our work provides one of the first formal characterizations of the notion of local neighborhoods employed by various SOTA methods. [4/N]
We also introduce a **no free lunch theorem for explanation methods** to demonstrate that no single explanation method can perform optimally across all (notions of) local neighborhoods. This result further highlights the need for choosing among explanation methods [5/N]
We also provide a guiding principle to choose among explanation methods -- a method is considered effective if it perfectly recovers the underlying model when it can i.e., when the (local) underlying model is a member of the explanation's function class. [6/N]
By leveraging the above principle, we theoretically analyze the conditions under which popular explanation methods are effective, and provide recommendations for choosing among them based on the characteristics/distributions of the input variables. [7/N]
Our framework also enables researchers and practitioners to readily construct **new post hoc explanation methods** based on the requirements of their applications and the characteristics of their input variables. More details about this in the paper. [8/N]
We are very grateful to reviewers at #NeurIPS2022 & #ICML Workshop on Interpretable Machine Learning in Healthcare (IMLH) for their invaluable feedback/suggestions which helped us improve the work. This paper also won the best paper award at IMLH.
Last but not the least, this work is a result of the hard work and efforts of some amazing students in our group @ai4life_harvard -- #TessaHan and @Suuraj. Many congratulations #TessaHan and @Suuraj !! :) [N/N]
PS If you find the above work exciting & are interested in joining our efforts on #XAI and @trustworthy_ml, do check out our open postdoc positions --
Explainable ML is a rather nascent field with lots of exciting work and discourse happening around it. But, it is very important to separate actual findings and results from hype. Below is a thread with some tips for navigating discourse and scholarship in explainable ML [1/N]
Overarching claims: We all have seen talks/tweets/discourse with snippets such as "explanations dont work" or "explanations are the answer to all these critical problems". When we hear such claims, they are often extrapolating results or findings from rather narrow studies. [2/N]
When we hear overarching claims, it is helpful to step back/ask what is the evidence to back such claims? which studies are such claims being based on? what is the context/application? how were the studies carried out? how reasonable is it to extrapolate these claims? [3/N]
Excited to share our @AIESConf paper "Does Fair Ranking Improve Outcomes?: Understanding the Interplay of Human and Algorithmic Biases in Online Hiring". We investigate if fair ranking algorithms can mitigate gender biases in online hiring settings arxiv.org/pdf/2012.00423… [1/n]
More specifically, we were trying to examine the interplay between humans and fair ranking algorithms in online hiring settings, and assess if fair ranking algorithms can negate the effect of (any) gender biases prevalent in humans & ensure that the hiring outcomes are fair [2/n]
We found that fair ranking algorithms certainly help across all job contexts, but their effectiveness in mitigating gender biases (prevalent in online recruiters) heavily depends on the nature of the job. [3/n]
If you have less than 3 hours to spare & want to learn (almost) everything about state-of-the-art explainable ML, this thread is for you! Below, I am sharing info about 4 of our recent tutorials on explainability presented at NeurIPS, AAAI, FAccT, and CHIL conferences. [1/n]
NeurIPS 2020: Our longest tutorial (2 hours 46 mins) discusses various types of explanation methods, their limitations, evaluation frameworks, applications to domains such as decision making/nlp/vision, and open problems explainml-tutorial.github.io/neurips20@sameer_@julius_adebayo [2/n]
AAAI 2021: Can't spend 2 hours 46 mins on this topic? No problem! Our tutorial at AAAI 2021 is right here (1 hour 32 mins): explainml-tutorial.github.io/aaai21. This one discusses different explanation methods, their limitations, evaluation, and open problems. @sameer_@julius_adebayo [3/n]
As I struggled to deal with the impact of COVID on my family members in India, I got delayed by a day for submitting my reviews for a conference & I got a message from a senior reviewer with the blurb below. My humble request to everyone - pls don't say this to anyone ever! [1/n]
I dont typically share any of my personal experiences on social media. But, I strongly felt that I need to make an exception this time. I am so incredibly hurt, appalled, flabbergasted, and dumbfounded by that blurb. It shows how academia can lack basic empathy! [2/n]
What bothers me is that I am an assistant professor at Harvard & I am decently known in my area of work. If someone can say this to me, I can't even imagine what they can say to a grad student. I am so sad this is the state of the research community that I am a part of! [3/n]
Twitter might seem like a not-so-kind place especially if you are a young student who just had your paper rejected by #NeurIPS2020. You might be seeing all your peers/professors talking about their paper acceptances. Let me shed some light on the reality of the situation [1/N]
Twitter (and generally social media) paints a biased view of a lot of situations including this one thechicagoschool.edu/insight/from-t…. Looking at your twitter feed, you might be feeling that everyone else seems to have gotten their papers accepted except for you. That is so not true! [2/N]
#NeurIPS2020 has an acceptance rate of around 20% which means an overwhelming majority of the papers (80%) have been rejected. Also, a lot of the accepted papers might have already faced rejection(s) at other venues before being accepted at #NeurIPS2020. [3/N]
Excited to join the team of and contribute to @trustworthy_ml handle. We will be covering the latest developments and research in "Trustworthy ML" regularly. Follow us and don't forget to tag @trustworthy_ml if you want us to tweet about your work.
One of the goals of our @trustworthy_ml handle is to provide visibility to the work of researchers who are new to the field. Please RT widely & follow @trustworthy_ml. Don't forget to tag us if you want us to tweet about your work! @black_in_ai@_LXAI@QueerinAI@icmlconf
We are currently covering trustworthy ML papers being presented at @icmlconf. Excited to collaborate with @JaydeepBorkar and @sbmisi to curate content and ensure that we keep our followers up-to-date with the latest on fairness/explainability/causality/privacy/ethics.