I am so excited that this new paper with @samisaguy is out. We explain how humans perceive number by mathematically deriving an optimal perceptual system with bounded information processing. Four N=100 pre-registered experiments match the predictions.
nature.com/articles/s4156…
People have long hypothesized that there are two systems of number perception: an exact system that lets us perfectly perceive numbers less than about 4, and an approximate one for large numbers.
This observation dates back to William Jevons, who published an experiment in Nature in 1871 where he literally threw beans into a box and saw how accurately he guessed the number. He estimated low numbers perfectly, but could only approximate those above 4 or 5.
But for two systems, nobody explains what is special about low numbers, why high number estimation takes the exact form it does, why and how multiple systems arose in evolution, etc.
The idea of @samisaguy's paper is to "design" an efficient perceptual system for number. What if we got to design a representational system which minimizes error (brain-vs-world) but can only use a fixed amount of information to do so?
Inspired by @markSchurgin and @timothyfbrady super cool paper and Shiny app...
.@samisaguy set up an app to let anyone design a perceptual system, formalized as a distribution of values (what’s in your brain, given what’s in the world). Any distribution can be scored by accuracy and information required. How good can you do?
colala.berkeley.edu:3838/unified_model/
Fortunately, we don't have to design these by hand -- in the paper we mathematically derive the minimum error system subject to the information processing bound. When you do that, you get a system that looks like this:
You get exactness for low numbers, approximately Gaussian high numbers with linearly increasing standard deviation, and people's attested under-estimation bias. Where you switch from exact to approximate changes with your information capacity.
We tested this by having people estimate numbers under varying time demands. Assuming that people accumulate information over time, the model predicts how their error should scale with number and display time. In these plots, left is the model, right is people.
It also makes predictions about how your average guess should scale with number. Left is model, right is people.
And finally it predicts what shape your response distribution should be -- unlike most models, this is not “built in,” but instead derived by solving the optimization problem.
To fit all of these, we essentially have one free parameter (the information capacity), plus individual subject effects. We replicate these effects in three other N=100 experiments which vary other factors.
The overall point is that the difference between small and large numbers, the shape of distribution, underestimation, response to time demands, etc. all can be explained. No need to think about two systems; all you need is one efficient system with limited information capacity.
Share this Scrolly Tale with your friends.
A Scrolly Tale is a new way to read Twitter threads with a more visually immersive experience.
Discover more beautiful Scrolly Tales like this.