Interested in interpretable and explainable machine learning? Check out our new blog post with opinions on the field and 70 summaries of recent papers, by @__Owen___ and me!

Link: alignmentforum.org/posts/GEPX7jgL…
Topics include Theory, Evaluation, Feature Importance, Interpreting Representations, Generating Counterfactuals, Finding Influential Data, Natural Language Explanations, Adversarial/Robust Explanations, Unit Testing, Explaining RL Agents, and others (note: not a formal taxonomy)
We're excited to highlight the wide array of research in interpretability/transparency/explainability. We hope this work can help others identify common threads across research areas and get up to speed on the latest work in different subareas.
Please feel free to leave any comments here :) and thanks to @_robertkirk and @mohitban47 for helpful feedback on the post!

#XAI #NLProc @trustworthy_ml

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Peter Hase

Peter Hase Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @peterbhase

4 Feb
This project has been a nice and long effort, but I’m excited to share a new paper: **When Can Models Learn From Explanations? A Formal Framework for Understanding the Roles of Explanation Data**

Work done with @mohitban47

Arxiv: arxiv.org/abs/2102.02201
Thread below 1/n
There are datasets where people explain why data point x gets label y, and the explanations look very helpful for solving the task. But what if models already know the relevant facts or can infer what they need to from a task input alone?
To test this question, we first design a synthetic task where we vary the num. of distinct hidden tasks in the data (we also test with existing datasets later). Our “explanations” of each point reveal what hidden task it belongs to & provide helpful info for predicting its label
Read 9 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!