One of the biggest criticisms of the field of post hoc #XAI is that each method "does its own thing", it is unclear how these methods relate to each other & which methods are effective under what conditions. Our #NeurIPS2022 paper provides (some) answers to these questions. [1/N]
In our #NeurIPS2022 paper, we unify eight different state-of-the-art local post hoc explanation methods, and show that they are all performing local linear approximations of the underlying models, albeit with different loss functions and notions of local neighborhoods. [2/N]
By doing so, we are able to explain the similarities & differences between these methods. These methods are similar in the sense that they all perform local linear approximations of models, but they differ considerably in "how" they perform these approximations [3/N]
Literature on local post hoc explanation methods is also commonly criticized for the fact that the definition of "locality" is often unclear. Our work provides one of the first formal characterizations of the notion of local neighborhoods employed by various SOTA methods. [4/N]
We also introduce a **no free lunch theorem for explanation methods** to demonstrate that no single explanation method can perform optimally across all (notions of) local neighborhoods. This result further highlights the need for choosing among explanation methods [5/N]
We also provide a guiding principle to choose among explanation methods -- a method is considered effective if it perfectly recovers the underlying model when it can i.e., when the (local) underlying model is a member of the explanation's function class. [6/N]
By leveraging the above principle, we theoretically analyze the conditions under which popular explanation methods are effective, and provide recommendations for choosing among them based on the characteristics/distributions of the input variables. [7/N]
Our framework also enables researchers and practitioners to readily construct **new post hoc explanation methods** based on the requirements of their applications and the characteristics of their input variables. More details about this in the paper. [8/N]
We are very grateful to reviewers at #NeurIPS2022 & #ICML Workshop on Interpretable Machine Learning in Healthcare (IMLH) for their invaluable feedback/suggestions which helped us improve the work. This paper also won the best paper award at IMLH. [9/N]
The link to our paper on arxiv is at arxiv.org/abs/2206.01254. Another detailed thread about this work is at [10/N]
Last but not the least, this work is a result of the hard work and efforts of some amazing students in our group @ai4life_harvard -- #TessaHan and @Suuraj. Many congratulations #TessaHan and @Suuraj !! :) [N/N]
PS If you find the above work exciting & are interested in joining our efforts on #XAI and @trustworthy_ml, do check out our open postdoc positions -- and other openings for new PhD students and interns. Details at himalakkaraju.github.io/openpositions.…

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with 𝙷𝚒𝚖𝚊 𝙻𝚊𝚔𝚔𝚊𝚛𝚊𝚓𝚞

𝙷𝚒𝚖𝚊 𝙻𝚊𝚔𝚔𝚊𝚛𝚊𝚓𝚞 Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @hima_lakkaraju

Jun 2
Explainable ML is a rather nascent field with lots of exciting work and discourse happening around it. But, it is very important to separate actual findings and results from hype. Below is a thread with some tips for navigating discourse and scholarship in explainable ML [1/N]
Overarching claims: We all have seen talks/tweets/discourse with snippets such as "explanations dont work" or "explanations are the answer to all these critical problems". When we hear such claims, they are often extrapolating results or findings from rather narrow studies. [2/N]
When we hear overarching claims, it is helpful to step back/ask what is the evidence to back such claims? which studies are such claims being based on? what is the context/application? how were the studies carried out? how reasonable is it to extrapolate these claims? [3/N]
Read 13 tweets
May 19, 2021
Excited to share our @AIESConf paper "Does Fair Ranking Improve Outcomes?: Understanding the Interplay of Human and Algorithmic Biases in Online Hiring". We investigate if fair ranking algorithms can mitigate gender biases in online hiring settings arxiv.org/pdf/2012.00423… [1/n]
More specifically, we were trying to examine the interplay between humans and fair ranking algorithms in online hiring settings, and assess if fair ranking algorithms can negate the effect of (any) gender biases prevalent in humans & ensure that the hiring outcomes are fair [2/n]
We found that fair ranking algorithms certainly help across all job contexts, but their effectiveness in mitigating gender biases (prevalent in online recruiters) heavily depends on the nature of the job. [3/n]
Read 8 tweets
May 7, 2021
If you have less than 3 hours to spare & want to learn (almost) everything about state-of-the-art explainable ML, this thread is for you! Below, I am sharing info about 4 of our recent tutorials on explainability presented at NeurIPS, AAAI, FAccT, and CHIL conferences. [1/n]
NeurIPS 2020: Our longest tutorial (2 hours 46 mins) discusses various types of explanation methods, their limitations, evaluation frameworks, applications to domains such as decision making/nlp/vision, and open problems explainml-tutorial.github.io/neurips20 @sameer_ @julius_adebayo [2/n]
AAAI 2021: Can't spend 2 hours 46 mins on this topic? No problem! Our tutorial at AAAI 2021 is right here (1 hour 32 mins): explainml-tutorial.github.io/aaai21. This one discusses different explanation methods, their limitations, evaluation, and open problems. @sameer_ @julius_adebayo [3/n]
Read 8 tweets
Apr 25, 2021
As I struggled to deal with the impact of COVID on my family members in India, I got delayed by a day for submitting my reviews for a conference & I got a message from a senior reviewer with the blurb below. My humble request to everyone - pls don't say this to anyone ever! [1/n] Image
I dont typically share any of my personal experiences on social media. But, I strongly felt that I need to make an exception this time. I am so incredibly hurt, appalled, flabbergasted, and dumbfounded by that blurb. It shows how academia can lack basic empathy! [2/n]
What bothers me is that I am an assistant professor at Harvard & I am decently known in my area of work. If someone can say this to me, I can't even imagine what they can say to a grad student. I am so sad this is the state of the research community that I am a part of! [3/n]
Read 12 tweets
Sep 26, 2020
Twitter might seem like a not-so-kind place especially if you are a young student who just had your paper rejected by #NeurIPS2020. You might be seeing all your peers/professors talking about their paper acceptances. Let me shed some light on the reality of the situation [1/N]
Twitter (and generally social media) paints a biased view of a lot of situations including this one thechicagoschool.edu/insight/from-t…. Looking at your twitter feed, you might be feeling that everyone else seems to have gotten their papers accepted except for you. That is so not true! [2/N]
#NeurIPS2020 has an acceptance rate of around 20% which means an overwhelming majority of the papers (80%) have been rejected. Also, a lot of the accepted papers might have already faced rejection(s) at other venues before being accepted at #NeurIPS2020. [3/N]
Read 12 tweets
Jul 15, 2020
Excited to join the team of and contribute to @trustworthy_ml handle. We will be covering the latest developments and research in "Trustworthy ML" regularly. Follow us and don't forget to tag @trustworthy_ml if you want us to tweet about your work.
One of the goals of our @trustworthy_ml handle is to provide visibility to the work of researchers who are new to the field. Please RT widely & follow @trustworthy_ml. Don't forget to tag us if you want us to tweet about your work! @black_in_ai @_LXAI @QueerinAI @icmlconf
We are currently covering trustworthy ML papers being presented at @icmlconf. Excited to collaborate with
@JaydeepBorkar and @sbmisi to curate content and ensure that we keep our followers up-to-date with the latest on fairness/explainability/causality/privacy/ethics.
Read 5 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(