, 8 tweets, 2 min read
My Authors
Read all threads
What does a pruned deep neural network "forget"?

Very excited to share our recent work w Aaron Courville, Yann Dauphin and @DreFrome

weightpruningdamage.github.io
At face value, deep neural network pruning appears to promise you can (almost) have it all — remove the majority of weights with minimal degradation to top-1 accuracy. In this work, we explore this trade-off by asking whether certain classes are disproportionately impacted.
We find that pruning is better described as "selective brain damage" -- performance on a tiny subset of classes and images is cannibalized in order to preserve overall performance. The interesting part is what makes certain images more likely to be forgotten...
The images impacted most by pruning are far more challenging for the original model. As model capacity is varied, the model appears to "forget" the examples it was already uncertain about. For real world datasets, this suggests caution should be used before pruning.
At the very least, we should consider additional measures to articulate this trade-off beyond just top-1 accuracy. :) We propose some tools to start thinking about this problem, for example PIE surfaces a subset of impacted images for a human expert to inspect...
Inspecting PIE images reveals a lot about the mapping the model learns. This may also explain why I have shared so many poorly structured ImageNet test set images over the last few months. :) Many PIE ImageNet images are multi-object or challenging for a human...
After this project, I am more convinced that a good data cleaning of ImageNet (and less rigid treatment of this a single object classification task) may compensate for much of the capacity we have thrown at the problem. But... that is a topic for another paper. :)
On real world data, the stakes are often higher than correctly classifying “guacamole”. For tasks like patient risk stratification, understanding the features pruning is cannibalizing is crucial. I hope our work highlight the need for better tooling to understand these tradeoffs.
Missing some Tweet in this thread? You can try to force a refresh.

Enjoying this thread?

Keep Current with Sara Hooker

Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

Twitter may remove this content at anytime, convert it as a PDF, save and print for later use!

Try unrolling a thread yourself!

how to unroll video

1) Follow Thread Reader App on Twitter so you can easily mention us!

2) Go to a Twitter thread (series of Tweets by the same owner) and mention us with a keyword "unroll" @threadreaderapp unroll

You can practice here first or read more on our help page!

Follow Us on Twitter!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just three indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3.00/month or $30.00/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!