"Training Machines to Learn the Way Humans Do: an Alternative to #Backpropagation"

Today's SFI Seminar by Sanjukta Krishnagopal
(@UCBerkeley & @UCLA)

Starting now — follow this 🧵 for highlights:
Image
"When we learn something new, we look for relationships with things we know already."

"I don't just forget Calculus because I learned something else."

"We automatically know what a 'cat-dog' would look like, if it were to exist."

"We learn by training on very few examples." Image
1, 2) "[#MachineLearning] is fundamentally different from the way humans learn things."

3) Re: #FeedForward #NeuralNetworks

"You choose some loss function...maybe I'm learning the wrong weights. So I define some goal and then I want to learn these weights, these thetas." ImageImageImage
"The reason that one-layer #networks don't really work is that they can only learn linear functions. With multilayer neural networks, you can learn decision boundaries through #backpropagation...so it's a fundamental part of how we train machines, these days." ImageImageImageImage
"The #brain learns [instead] by local #learning — instead of the error getting fed back through backpropagation, each #neuron does some kind of linear regression. It [consequently] works very fast. We have experimental evidence that the brain does something like this." Image
For dendritic gated networks in animal #brains:

"For each branch I pick a random hyperplane and draw [it] somewhere in this square, and say, 'If this input falls on one side, the gate will be open, and if it falls on the other side, the gate will be closed." ImageImage
"Each weight learns a different piecewise linear function, and then I aggregate as I go through the layers. This neuron is learning this section, this neuron is learning this section, and then the next layer is learning both sections." Image
2) "What do we want? Some desirable features of this model include that it is modeled on the #cerebellum. There isn't any ridiculous time delay due to forward and backward passes."

3) "Parallel fiber inputs go in through 'dendrites' and each branch has a gating key..." ImageImageImage
"The fact that the gates are significantly more correlated through learning than the error signal validates our decision to use [this approach]." Image
1) On the desirable features for computational experiments exploring a cerebellar model for #MachineLearning

2) "If you keep your finger in front of you and you move your head, you'll notice your eyes fixate very well on your fingernail...unlike if you move your finger around." ImageImage
"Both kinds of #NeuralNetworks learn this chaotic time series, but in different ways. DGNs learn this in very intuitive ways." ImageImage
On Mitigating "catastrophic #forgetting": ImageImage
"Instead of hyperplanes, can we have hypermanifolds?"

Re: Continuous learning with biologically plausible networks that don't suffer catastrophic forgetting...

Link to the pre-print c/o @DeepMind:
deepmind.com/publications/a… ImageImageImage

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Santa Fe Institute is on Mastodon

Santa Fe Institute is on Mastodon Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @sfiscience

Dec 14
"Compositionality in Vector Space Models of Meaning"

Today's SFI Seminar by @marthaflinders, streaming:


Follow this 🧵 for highlights! Image
"Scientists gather here
Santa Fe Institute, oh so near
Inquiring minds seek truth"

#haiku about SFI c/o @marthaflinders & #ChatGPT

...but still, #AI fails at simple tasks: ImageImage
"One way to represent the kind of #compositionality we want to do is with this kind of breakdown...eventually a kind of representation of a sentence. On the other hand, vector space models of #meaning or set-theoretical models put into a space have been very successful..." Image
Read 10 tweets
Dec 13
"Humans are prone to giving machines ambiguous or mistaken instructions, and we want them to do what we mean, not what we say. To solve this problem we must find ways to align AI with human preferences, goals & values."
- @MelMitchell1 at @QuantaMagazine:
quantamagazine.org/what-does-it-m…
“All that is needed to assure catastrophe is a highly competent machine combined with humans who have an imperfect ability to specify human preferences completely and correctly.”

- Stuart Russell (@UCBerkeley) as quoted by @MelMitchell1 in her latest @QuantaMagazine article
"It’s a familiar trope in #ScienceFiction — humanity threatened by out-of-control machines who have misinterpreted human desires. Now a not-insubstantial segment of the #AI research community is concerned about this kind of scenario playing out in real life."
- @MelMitchell1
Read 6 tweets
Nov 30
🧵 "#Criticality: A Balance Between #Robustness and #Adaptability"

Today's SFI Seminar by @cgershen @cgg_mx:
Image
Regarding systems at the #EdgeOfChaos
Identifying #PhaseTransition dynamics
Citing @WagnerEvolution on #Robustness
Contrasting it with #Chaos (fragile, hard to predict, etc.)
And then #Criticality, somewhere in the middle... ImageImageImageImage
With #criticality, due to #PowerLaw(-like) distributions:

"Most of the changes will be small, but for some, you can have large changes."

See also #fractals

Key history point: Stuart Kauffman's #random #Boolean #networks (1969!)

Updated by @cgershen in 2004 ImageImageImageImage
Read 12 tweets
Nov 29
"Liquid and Solid #Brains: Mapping the #Cognition Space"

Today's SFI Seminar by Ext Prof @ricard_sole, streaming now — follow this 🧵 for highlights:

Image
"Why #brains? Brains are very costly...it seems like they are not a very good idea to bring complex cognition to a #biosphere that just needs simple replicators."

"I also want to explore the problem of #consciousness, which is around all the time..."

@ricard_sole Image
Two different views of #evolution:

- Stephen Jay Gould and an emphasis on #contingency

- Simon Conway Morris/Pere Albach and an emphasis on #constraints and #convergence

--> "Replay the tape with different results vs. 'the logic of monsters' & 'life's solution'" Image
Read 21 tweets
Nov 9
"Fabricating #Clay: #Machines, #Materials, & #Code"

🧵 Today's SFI Seminar by @leahbuechley (@UNM):
"We build the geometry directly by thinking about the PATH that our 3D printer takes. There's no intermediate slicing software [to render CAD as "2D" layers]."

- @leahbuechley, #3DPrinting clay with a 1.5 mm nozzle
"You can start to think about surface textures - like spikes that you can't do with a traditional slicer. Or...here's a path that's a sine wave, but every other layer is rotated. Or...one creature is following THIS path, and the other is chasing it around."

#clay #3dprinting
Read 11 tweets
Nov 4
"The 21st Century Question: Emergently Engineering the Future"

Follow this 🧵 today and tomorrow for highlights from our 2022 #AppliedComplexity Network and Board of Trustees Symposium:

santafe.edu/events/annual-…

#Web3 #Cities #Polarization #EmergentEngineering #Decentralization
We start with a talk by SFI President David Krakauer:

"Would anyone care to guess why we're so GOOD at building transistors and so CRAP at designing drugs?"

"This thing [points to transistor] lives in a centralized system. This thing [points to cancer drug] lives in US."
"I'm going to pick on economics, because we like to do that at SFI. 'Ooh, look at that cover! So techy. Global, heat maps...' But here's 'Networks' [in the textbook]. THAT'S IT. Here's '#ComplexityEconomics.' NOTHING."
Read 111 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us on Twitter!

:(