To summarize, biological brains do not approximate intractable problems. There's no such thing as being able to approximate intractable problems, this is a fiction that so many machine learning researchers fall into.
I'm not the only that points out the limits of approximation with respect to computational systems.
What can be approximated are the computations that are tractable. We are able to approximate intractable solutions because we constrain ourselves to subsets that do have tractable solutions. Our intuition is a bias to identify these subsets.
• • •
Missing some Tweet in this thread? You can try to
force a refresh
"The brain is a computer" is a damn problematic metaphor. I prefer to say that "the brain is an intuition machine".
The term computer is conventionally understood as to be a digital computer. It's the kind that we program. It's the kind that is designed by minds and manufactured in assembly lines. It's the kind that can't repair itself. It's the kind without any autonomy.
It is a horrible metaphor. The brain is an intuition machine is a better metaphor. It's that kind that learns from experience. It is the kind that develops in an inside-out manner. It is the kind that creates itself. It is the kind that repairs itself. It is autonomous.
"Shut up and calculate" is the affliction we have when we substitute symbols for understanding.
Humans are linguistic bodies. A huge part of our brains has been exapted (verb form of exaptation) for language. Thus it's conceivable that our innate capabilities for understanding have diminished in use.
Simon DeDeo wrote an insightful tweetstorm about explanations that appear intuitive.
Biological cognition is always based on computation subject to time and resource constraints (i.e. fast and frugal). The case where parallel computation cannot be recreated by sequential computation is precisely in this context.
Said differently, smaller brains do not have the same capabilities as larger brains. So when we talk about the notion of cognition or agency for insects or even single-cell life (without brains), we speak of a kind of behavior that is grossly simpler.
However, the computation performed by living things is very different from the computation performed by computers. This is because agency leads to an intentional stance. Intention is what constrains computations towards goals.
So I got myself the Quest 2. I'm not a heavy VR user. Last time I used VR was when the Vive was released. So I skipped one generation of VR headsets. This is the 3rd gen.
The original Vive had 1080×1200 per eye. Quest 2 has 1832×1920 per eye. So the viewing experience is obviously superior. Adoption of VR will all depend on resolution. Seeing the world in low-res is a turn-off.
The 1st gen VR required some setup and was tethered to a computer. Quest 2 in contrast is untethered, the compute is on the headset itself. It's certainly much more convenient to use now that we are spoiled by our mobile devices.
All that is significant emerges from the coupling of slow with fast processes.
The slow side implies a compression or memory process relative to the fast side. That said, there is always change, but the relative difference in speed of the two processes leads to an invariance. It is that emergent invariance that is significant.
The dual of invariance is tension. There are processes that lead to convergence (thus remaining the same) and there are features that are constantly driven towards diversity. All evolution requires push-pull dynamics.