I've written the first part of a chapter on the heritability of IQ scores. Focusing on what IQ is attempting to measure. I highlight multiple paradoxical findings demonstrating IQ is not just "one innate thing".
First, a few reasons to write this. 1) The online IQ discourse is completely deranged. 2) IQists regularly invoke molecular heritability as evidence for classic behavioral genetics findings while ignoring the glaring differences (ex: from books by Ritchie and Haier/Colom/Hunt).
Thus, molecular geneticists have been unwittingly drafted into reifying IQ even though we know that every trait is heritable and behavior is highly environmentally confounded. 3) IQ GWAS have focused on crude factor models that perpetuate the "one intelligence" misconception.
So what is an IQ test and the "g" factor? In short, test takes are asked questions related to pattern matching, memory, verbal/numeric reasoning, and general knowledge (some examples below). The weighted average of their scores is then the IQ score.
It turns out people who do poorly on one test tend to do poorly on other tests, which produces weak correlations known as the "positive manifold". These correlations can be summarized with factor analysis, producing a "general" factor (g) that explains 25-45% of total variance.
There's nothing special about the g score: it's just the IQ score computed with a different weighting (the "g loadings", more on these later).
But what could explain these correlations? In fact, many different theories can produce the same exact positive manifold and g patterns:
For example, Thomson's "sampling" theory, where IQ subtests sample from a large number of partially overlapping processes, can produce exactly the pattern of test correlations, leading factor, and factor loadings observed in the UK Biobank. Even though no actual g exists!
Or mutualism theory, where underlying processes interact dynamically over time, together with environmental inputs, to produce an apparent positive manifold. Again, g is not a causal variable, but an emergent statistical byproduct of these mutualistic relationships.
Finally, g/factor theory: where we treat the latent factors as measures of the true causal process itself. One factor doesn't fit the data well, so all sorts of more complex factor models have been proposed. Including a synthesis of sampling + factors in Process Overlap Theory.
I'm stressing the fact that many different theories fit the data in part because a causal/biological "g" is often taken as a given. But also because getting the theory right is critical to understanding what is actually being measured, test bias, and effective tests.
Ok, I promised paradoxical findings so here are five:
1) What are the weights used to compute g? They are highly correlated with how culturally specific a given subtest is (e.g. high for vocab, low for digit memory). So g is just IQ rescaled to emphasize cultural knowledge.
2) Ability + Age differentiation: The *highest* test correlations are among individuals with the *lowest* IQ, yet test correlations also increase with age. So IQ is measuring something different at the low/high ends and is also dynamic through development.
3) There is no "Matthew effect": In study after study, individuals with higher starting IQ do not acquire knowledge/skills faster -- in fact they converge! IQ also can't predict cognitive decline. This means IQ is not a measure of "processing speed" but of baseline knowledge.
4) Socioeconomic status (SES), in contrast, *is* associated with divergence: No SES/IQ differences are observed in kids at 10 months, yet low SES kids were 6 pts behind at age 2, and 15-17 pts (>1SD) behind by age 16. Thus IQ is confounded by SES from the start and throughout.
5) The Flynn Effect: Mean IQ in the population has been increasing over generations, with some studies showing the increase happening on more g/culture loaded subtests and among lower scorers and even within families. Environmental/cultural factors thus reshape IQ over time.
Clearly multiple dynamic phenomena are at play, so how does this fit with theory? In fact, longitudinal data strongly support mutualism, with gains in one cognitive domain translating into gains in other domains. Recently, even shown in RCTs [Stine-Morrow et al. 2024].
The same is observed in cross-sectional analyses, where mutualist/network models consistently fit IQ data better than factor models. [Knyspel + Plomin 2024] even applied networks to twin data and showed reversed relationships between twin "heritability" and g loading!
Finally, while neuroscience suffers from the same construct validity issues as IQ research, the one consistent finding is that there is no "neuro g": g correlates with many different structural/functional patterns, is better explained by network models, and does not replicate.
In short, IQ is indexing a bundle of different and often confounded processes, including individual and cultural shifts. IQ/g scores should be modeled as a dynamic network with environmental interactions, and there's absolutely no reason to treat them as "real" or "innate".
Now that we have a handle on what we are/aren't estimating with IQ, next time I will discuss the heritability and molecular genetic findings. My previous thread on the heritability of Educational Attainment is linked below. /x
Eric Turkheimer has a good piece about a bet he made with Charles Murray regarding the genetic understanding of IQ (or, really, the lack of it). Murray being so wrong in his prediction should make us question his world model, but it's also worth commenting on his response.
Murray has, for some time now, been workshopping the excuse that progress on IQ genetics was blocked by researchers being denied the access to the relevant databases. This is patently untrue!
First, one of the largest genetic analyses to date of *any* trait is of educational attainment, a phenotype Murray himself has used as a proxy for intelligence. Surely a study of 3 million should have been enough to satisfy Murray's prediction.
Murray and most of race twitter has apparently been fooled by this completely fabricated analysis purporting to show African ancestry is associated with IQ. People lie on twitter all the time, but this is both more revealing and more disturbing than usual. A 🧵
Revealing in that it shows how quantitative racism is a just an exercise in manipulating data to fit the preconceived conclusion. Disturbing because this time private data is being used and the results, which cannot be easily verified, are just flatly invented.
What's actually going on? Some guy claims to have an analysis showing that African ancestry differences between siblings are associated with IQ differences in the UK Biobank. Implying an ancestry difference in the within-family influences.
A few thoughts on Herasight, the new embryo selection company. First, the post below and the white paper imply that competitors like Nucleus have been marketing and selling grossly erroneous risk estimates. This is shocking if true! 🧵
I wrote last year about the un-seriousness with which Nucleus approached their IQ product and the damage it could do to genetic prediction and research more broadly (). This appears to have been a broader pattern beyond IQ, extending even to rare disease.theinfinitesimal.substack.com/p/genomic-pred…
People who care about this technology should be furious at Nucleus and their collaborators (as well as Orchid and Genomic Prediction for their own errors). Finding such flaws should not require reverse-engineering by a competitor. These products clearly need independent audits.
Oof. Polygenic scores for IQ lose 75% of their explained variance when adding family controls, even worse than the attenuation for Educational Attainment. These are the scores Silicon Valley is using to select embryos 😬.
The TEDS cohort used here is a very large study with high-quality cognitive assessments collected over multiple time points. It is probably the most impressive twin study of IQ to date. That means very little room for data quality / measurement error issues.
It is important to highlight surprising null results. Just last week we were hypothesizing that large IQ score attenuation could be a study bias or an artifact of the Wilson Effect. Now we see it replicate in an independent study with adults.
@notcomplex_ @krichard1212 The authors fit a non-identifiable Model B, which produces a table full of NA's. Then they try to interpret this model to fix it. That makes no sense. The parameters of this model will be completely arbitrary, so using it to decide what to prune is also statistically invalid.
@notcomplex_ @krichard1212 At various points later on they talk about "Heywood cases", which are out-of-bounds parameters or negative variances, but no such out-of-bounds parameters are actually present in the tables (and, again, you cannot interpret these from the non-identified model).
@notcomplex_ @krichard1212 So none of the decisions make statistical sense and either reflect someone who doesn't know what they're doing or is intentionally trying to find the model fit they like. True to form given they missed a fatal error with model A, misinterpreted AIC comparisons, etc.
Racism twitter has taken to arguing that observed racial differences must be "in part" explained by genetic differences, though they demure on how much. Not only is this claim aggressively misleading, it is completely unsupported by data. A 🧵:
Genetic differences between any two populations can go in *either* direction, matching the phenotypic differences we observe or going against them. Genes also interact with the environment, which makes the whole notion of "explaining" differences intractable.
The mere fact that a trait is heritable within populations tells us nothing about the explanatory factors between populations. See: Lewontin's thought experiment; Freddie de Boer's analogy to a "jumping contest"; or actual derivations (). pubmed.ncbi.nlm.nih.gov/38470926/