Nice! Here we have an interesting paper using genetic ancestry to classify race/ethnicity in modern data and algorithms. Let's take a look at what this paper found: 🧵
First, I don't want to get too hung up on language, but TCB's tweet starts talking about "ethnicity", then shifts to "continental ancestries", and then entirely omits the largest ethnic group in the US: Hispanics. These terms have distinct definitions (). nap.nationalacademies.org/catalog/26902/…
Anyway, how well can this paper actually impute ethnicity from genetic ancestry in a large cancer population ()? ~17% of the time it gets Hispanic classification completely wrong or a no-call! worldscientific.com/doi/10.1142/97…
But even this is an overstatement, because the majority of participants either didn't list race/ethnicity or provided one that didn't fall into an established category. And the ML algorithm is *terrible* at classifying these unlabeled/partially labeled people as no calls.
This creates an interesting paradox where the algorithm can be made to look like it is more accurate over time, but in reality participants are simply drifting to a new unlabeled space in the social construct.
I was also intrigued by the claim that ethnicity is perhaps the least socially constructed variable in social science because an algorithm can classify some of the labels with some accuracy. Is this really true?
Language is a social construct, but AI is able to do a pretty good job at classifying languages.
Religion is a social construct, but AI can do a pretty good job classifying those too, even from a cartoony illustration.
Race is a social construct, but I bet you could easily classify thes-- wait, where was I going with this?
Even more interesting, you can explain a social construct like "money" to an AI and it will figure out the natural divisions within the construct based on visual details.
Can we do the same for race/ethnicity and ancestry? Let's play a game ...
Here's a basic ancestry plot, where each point is a person. Do the green and purple dots reveal two racial groups?
Nope. The green and purple points are sampled from the same population but the purple dots just came from one family in that population.
Okay but that was simulated data. Here's another one, using real data from a large-scale biobank this time. Are these ten different racial groups? Surely the pink-ish groups are a different race from the greens at least?
Nope! These are all Chinese participants of the Kadoorie biobank, color-coded by the cities they were recruited from. Ancestry inference can be extremely sensitive with enough data.
() pubmed.ncbi.nlm.nih.gov/37601966/
Ok, maybe it's unfair to use such closely related populations. Let's look at data from continental groups and use a model-based clustering approach. Surely the two orange/tan clusters here are different races or continents:
Nope! The two groups being distinguished here are Melanesians and ... the rest of the world. Asian, Middle Eastern, European, and African participants all get clustered together because of the sampling of the data. () pmc.ncbi.nlm.nih.gov/articles/PMC60…
I'm making it too hard. Maybe we need more drifted populations and tree-based clustering instead? Look at the deep divergences across these populations, surely *these* must be different races?
Nope. These are all participants from Native American tribes within a single linguistic group. Some of the most diverged populations in the world get lumped together into one socially constructed box.
TLDR: When people say a construct is "the least constructed / best / most replicable in social science", maybe they are telling you more about the quality of the social sciences than the validity of the construct. /x
• • •
Missing some Tweet in this thread? You can try to
force a refresh
Murray and most of race twitter has apparently been fooled by this completely fabricated analysis purporting to show African ancestry is associated with IQ. People lie on twitter all the time, but this is both more revealing and more disturbing than usual. A 🧵
Revealing in that it shows how quantitative racism is a just an exercise in manipulating data to fit the preconceived conclusion. Disturbing because this time private data is being used and the results, which cannot be easily verified, are just flatly invented.
What's actually going on? Some guy claims to have an analysis showing that African ancestry differences between siblings are associated with IQ differences in the UK Biobank. Implying an ancestry difference in the within-family influences.
A few thoughts on Herasight, the new embryo selection company. First, the post below and the white paper imply that competitors like Nucleus have been marketing and selling grossly erroneous risk estimates. This is shocking if true! 🧵
I wrote last year about the un-seriousness with which Nucleus approached their IQ product and the damage it could do to genetic prediction and research more broadly (). This appears to have been a broader pattern beyond IQ, extending even to rare disease.theinfinitesimal.substack.com/p/genomic-pred…
People who care about this technology should be furious at Nucleus and their collaborators (as well as Orchid and Genomic Prediction for their own errors). Finding such flaws should not require reverse-engineering by a competitor. These products clearly need independent audits.
Oof. Polygenic scores for IQ lose 75% of their explained variance when adding family controls, even worse than the attenuation for Educational Attainment. These are the scores Silicon Valley is using to select embryos 😬.
The TEDS cohort used here is a very large study with high-quality cognitive assessments collected over multiple time points. It is probably the most impressive twin study of IQ to date. That means very little room for data quality / measurement error issues.
It is important to highlight surprising null results. Just last week we were hypothesizing that large IQ score attenuation could be a study bias or an artifact of the Wilson Effect. Now we see it replicate in an independent study with adults.
Racism twitter has taken to arguing that observed racial differences must be "in part" explained by genetic differences, though they demure on how much. Not only is this claim aggressively misleading, it is completely unsupported by data. A 🧵:
Genetic differences between any two populations can go in *either* direction, matching the phenotypic differences we observe or going against them. Genes also interact with the environment, which makes the whole notion of "explaining" differences intractable.
The mere fact that a trait is heritable within populations tells us nothing about the explanatory factors between populations. See: Lewontin's thought experiment; Freddie de Boer's analogy to a "jumping contest"; or actual derivations (). pubmed.ncbi.nlm.nih.gov/38470926/
James Lee and @DamienMorris have an interesting perspective paper out describing "some far-reaching conclusions" about the genetics of intelligence. This type of "where are we now" paper is very fun and more people should write them! So, where are we now? 🧵
It's a short paper and it surveys three core findings from the past decade of intelligence genetics. These sections follow a structure that I would cheekily call ... "make a bold claim in the title, then walk it back in the text".
First up, they address the concern that associations with intelligence may actually be mediated by functionally irrelevant traits like physical appearance or pigment. The argument is that IQ GWAS has demonstrated enrichments for CNS/brain structure gene sets. This is true!
The SAT/meritocracy debate has always been a bit odd to me when the test makers themselves have studies showing self-reported high-school GPA is a consistently better predictor of college GPA and always adds on top of SATs.
Clearly SATs are neither the only nor even the best measure we have of college success and "holistic" admissions can be "meritocratic". It's up for debate whether the additional <10% predictive variance SATs give you are worth the high-school testing industrial complex.
A challenge with all of these analyses is they are measured after selection on the predictor variables themselves, which can induce biased estimates through range restriction. The raw correlations are even lower, and it is hard to know whether correcting is appropriate.