Timo Flesch Profile picture
Mar 23, 2022 17 tweets 7 min read Read on X
1/ Preprint alert: Humans can learn continuously, while standard deep neural networks need interleaved data. Here, @dvgnagy, @SaxeLab @summerfieldlab and I propose a neural network model of human continual learning. #continuallearnig, #cata strophicinterference,#DeepLearning Image
2/ In 2018, we showed that humans perform better after blocked compared to interleaved training on multiple categorisation tasks pnas.org/doi/epdf/10.10….
3/ Back then, participants classified trees from a bivariate stimulus space according to a single dimension per task. Image
4/ They performed worse after interleaved compared to blocked training. We argued that interleaved training biases participants towards a single category boundary (“linear” solution), while blocked training promotes a “factorised” solution Image
5/ Neural networks can’t do this: They suffer from catastrophic forgetting under blocked curricula, but not after interleaved training. Bad news if you want to use them as models of human learning! Image
6/ Inspired by work on #cognitivecontrol, we introduce two algorithmic motifs to model human continual learning.
7/ We rarely switch contexts in everyday life (for example when leaving the office to go home). The brain should capitalise on this. We introduce “sluggish” task signals during training, where information is maintained over several trials. Image
8/ Under interleaved training, sluggishness controls task accuracy, whether a factorised or linear solution is learned, and how many hidden layer units are allocated to each task. Image
9/ With interleaved data, networks learn to allocate different tasks to different task-specific units, forming orthogonal representations. In cell.com/neuron/fulltex… we showed similar results in the brain after blocked training.
10/ Here, we replicate this, and demonstrate that a standard neural network ignores the task signal and only represents the most recently learned task under blocked training (MDS on hidden layer): Image
11/ However, if we hand-craft this gating, the network no longer suffers from catastrophic forgetting under blocked training: Image
12/ ... but we don’t have to do this by hand: Our 2nd proposal is a Hebbian learning step, applied to the task-signalling weights, which is alternated with standard SGD and strengthens connections between task-units and relevant hidden units Image
13/ This Hebbian step is sufficient to protect against forgetting, and enables the network to learn a “factorised” representation under blocked training: Image
14/ How does it compare to human performance: We reanalysed our data from Flesch et al 2018) and compared it to a baseline model and our sluggish Hebbian model (+decision noise in the output). Our network recreated the performance differences: Image
15/ Decomposing this into different sources of errors revealed that our model, like human participants, tended towards a linear solution under interleaved training: Image
16/ We first presented the results at #Cosyne2021, time flies! We also acknowledge other awesome work that has used gating for continual learning, and variants of sluggishness to model switch costs (Masse et al, 2018; Russin et al, 2022).
17/ Thanks for reading the tweeprint! Here’s a link to the paper, feedback most welcome!
arxiv.org/abs/2203.11560

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Timo Flesch

Timo Flesch Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(