Thread Reader
Share this page!
×
Post
Share
Email
Enter URL or ID to Unroll
×
Unroll Thread
You can paste full URL like: https://x.com/threadreaderapp/status/1644127596119195649
or just the ID like: 1644127596119195649
How to get URL link on X (Twitter) App
On the Twitter thread, click on
or
icon on the bottom
Click again on
or
Share Via icon
Click on
Copy Link to Tweet
Paste it above and click "Unroll Thread"!
More info at
Twitter Help
Jean-Baptiste Cordonnier
@jb_cordonnier
PhD student @epfl_en 🇨🇭supervised by Martin Jaggi. Interested in attention models + optimization. Mountain lover 🏔
Subscribe
Save as PDF
Jan 10, 2020
•
5 tweets
•
4 min read
Very happy to share our latest work accepted at
#ICRL2020
: we prove that a Self-Attention layer can express any CNN layer.
1/5
📄Paper:
openreview.net/pdf?id=HJlnC1r…
🍿Interactive website :
epfml.github.io/attention-cnn/
🖥Code:
github.com/epfml/attentio…
📝Blog:
jbcordonnier.com/posts/attentio…
Two *necessary* conditions (often met in practice):
(a) Multiple heads, ex: 3x3 kernel requires 9 heads,
(b) Relative positional encoding to allow translation invariance.
Each head can attend on pixels at a fixed shift from the query pixel forming the pixel receptive field.
2/5