My Authors
Read all threads
Computational neuroscience has lately had great success at modeling perception with ANNs - but it has been unclear if this approach translates to higher cognitive systems. We made some exciting progress in modeling human language processing biorxiv.org/content/10.110… #tweeprint 1/
This work is the result of a terrific collaboration with @ibandlank @GretaTuckute @KaufCarina @eghbalhosseini @Nancy_Kanwisher Josh Tenenbaum and @ev_fedorenko; @mitbrainandcog @MIT_CBMM @mcgovernmit 2/
Work by @ev_fedorenko and others has localized the language network as a set of regions that support high-level language processing (e.g. sciencedirect.com/science/articl…) BUT the actual mechanisms underlying human language processing have remained unknown. 3/
To evaluate model candidates of mechanisms, we use previously published human recordings: fMRI activations to short passages (Pereira et al., 2018), ECoG recordings to single words in diverse sentences (Fedorenko et al., 2016), fMRI to story fragments (Blank et al. 2014) 4/
More specifically, we present the same stimuli to models that were presented to humans and "record" model activations. We then compute a correlation score of how well the model recordings can predict human recordings with a regression fit on a subset of the stimuli. 5/
Since we also want to figure out how close model predictions are to the internal reliability of the data, we extrapolate a ceiling of how well an "infinite number of subjects" could predict individual subjects in the data. Scores are normalized by this estimated ceiling. 6/
So how well do models actually predict our recordings? We tested 43 diverse language models, incl. embedding, recurrent, and transformer models. Specific models (GPT2-xl) predict some of the data near perfectly, and consistently across datasets. Embeddings like GloVe do not. 7/
The scores of models are further predicted by the task performance of models to predict the next word on the WikiText-2 language modeling dataset (evaluated as perplexity, lower is better) - but NOT by task performance on any of the GLUE benchmarks. 8/
Since we only care about neurons because they support interesting behaviors, we tested how well models predict human reading times: specific models again do well and their success correlates with 1) their neural scores, and 2) their performance on the next-word prediction task 9/
We also explored the relative contributions to brain predictivity of two different aspects of model design: network architecture and training experience, ~akin to evolutionary and learning-based optimization. (see also ) 10/
Intrinsic architectural properties (like size and directionality) in some models already yield representational spaces that - without any training - reliably predict brain activity. These untrained scores predict scores after training. 11/
While deep learning is mostly focused on the learning part, architecture alone works surprisingly well even on the next-word prediction task. Critically for the brain datasets, a random embedding with the same number of features as GPT2-xl does not yield reliable predictions. 12/
Summary: 1) specific models accurately predict human language data; 2) their neural predictivity is correlated with task performance to predict the next word, 3) and with their ability to predict human reading times; 4) architecture alone already yields reasonable scores. 13/
These results suggest that predicting future inputs may shape human language processing, and they enable using ANNs as embodied hypotheses of brain mechanisms. To fuel future generations of neurally plausible models, we will soon release all our code and data. /fin
Special shout-out to the wonderful @huggingface transformers library which really sped up the evaluation of models. These resources are such a catalyst for community progress. @Thom_Wolf I wonder if you anticipated it ever being used to predict human brain recordings ;-)
Missing some Tweet in this thread? You can try to force a refresh.

Keep Current with Martin Schrimpf

Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

Twitter may remove this content at anytime, convert it as a PDF, save and print for later use!

Try unrolling a thread yourself!

how to unroll video

1) Follow Thread Reader App on Twitter so you can easily mention us!

2) Go to a Twitter thread (series of Tweets by the same owner) and mention us with a keyword "unroll" @threadreaderapp unroll

You can practice here first or read more on our help page!

Follow Us on Twitter!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3.00/month or $30.00/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!