The reason machine learning algorithms show bias is that the goal of these algorithms is to learn ALL the patterns in the data including the biases. The "bias" is actually the gap between what the data scientist THINKS is being learned and what's actually being learned. 🧵
An interesting feature of this bias is it's subjective. It depends on what the data scientist INTENDED to learn from the data. For all we know, the data scientist intended to learn all the patterns in the data, racism and all. In which case, there is no bias.
Generally, machine learning does not require us to be specific about what patterns we are trying to learn. It just vaguely picks up all of them. This means we often have no clue what was learned and if it is what we intended to learn.
Traditional statistics isn't like this. In statistics, the first step is specifying what patterns you want to detect. This requires you to have some kind of theory about the structure of the data. Most importantly, this allows you to check if your theory is wrong.
This issue is an huge weakness of the machine learning approach. The vagueness about what is being learned means that we have to do a lot of work after we fit the model to understand the properties of the model itself. In practice, this work is often not done.
The reason we need to do the work is because we can't rely on theory to tell us what the model learned so we must measure it. This means looking at how the model behaves in order to see if it's racist, sexist or has other biases we might care about.
As we see with the many examples of racist algorithms, many of the people using machine learning mistakenly think that they can rely on their intuitions to guess what kinds of patterns are in their dataset and what kind of patterns their algorithms are learning. This is naive.
I think the solution to racism in algorithms (and other biases of this kind) is to be more hands-on about understanding the processes that created the data your model uses and more proactive and explicit about checking that your models have the properties you think they have. 🧵

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with 🔥Kareem Carr | Statistician 🔥

🔥Kareem Carr | Statistician 🔥 Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @kareem_carr

Nov 11
Here is a problem I see with modern liberalism: if you tell a certain kind of liberal that there are two kids drowning and that they can only save one, they would immediately declare that they can save them both, and then act completely surprised when both drown shortly after.
If that same liberal could magically go back in time with all the knowledge of what had happened, that person would do the exact same thing again, and then be just as surprised when both kids drowned for the second time.
It's very hard to say we must sacrifice this one good thing for the sake of this other good thing and remain a liberal in good standing.
Read 4 tweets
Nov 10
People are getting the wrong end of the stick here. Nobody stopped her from experimenting on herself. The hold up is on *publishing* the results.
Here's the quote from the article Image
It's an experiment with exactly one person, no controls, and where multiple procedures were tried.
Read 6 tweets
Nov 9
I honestly get a lot of value out of ChatGPT. It feels built for people like me. I find identifying and correcting its mistakes pretty easy because I'm used to grading student assignments, but I also do things that minimize mistakes like:
I input:
- examples of past solutions to similar problems
- a high-level sketch of the solution to the current problem
- background information if needed
- warnings about any potential complications or pitfalls
For instance, if I want ChatGPT to do a certain kind of computation, I might:
- do a sample calculation by hand on a piece of paper
- get ChatGPT to read the piece of paper and translate it to LaTeX
- tell ChatGPT to study the calculation and extend it to the new situation
Read 10 tweets
Nov 5
This Musk meme speaks to something true which is America is splitting culturally between the college educated vs the non-college educated.

There is however a third group. People who went to college but who think and act like people who didn’t. Image
Basically you have these people who went to elite schools like Harvard or Stanford or Yale, who have law degrees and doctorates in many cases, telling the non-college educated that there’s no point to college because it’s not great job training.
I’m no historian but I don’t think an education was historically about job training. People apprenticed with tradesmen for that. Education was about being acculturated into the superstructure of your civilization. It taught you what humans had done so far and your place in it.
Read 7 tweets
Oct 24
there is clearly a force or mechanism that causes the US electorate to balance at precisely 50% democrat 50% republican no matter what either candidate says or does.

if i was a social scientist, i would be absolutely obsessed with this.
my economics brain says maybe it's that each party is more extreme than the general electorate and is only willing to concede the minimum number of policies necessary to win which is exactly 50% + 1 vote.

this leads to a powerful finetuning mechanism on both sides.
my political science brain says that the fine tuning actually happens at the gerrymandering level. Same drive to stop at roughly 50% + 1 once you've gerrymandered enough seats.
Read 4 tweets
Oct 18
For every ten likes, I'll post a new unhinged mathematician quote. Image
if it's you or logical consistency, you know which one they're picking. Image
technically a physicist quote but i'll allow it Image
Read 6 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Don't want to be a Premium member but still want to support us?

Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal

Or Donate anonymously using crypto!

Ethereum

0xfe58350B80634f60Fa6Dc149a72b4DFbc17D341E copy

Bitcoin

3ATGMxNzCUFzxpMCHL5sWSt4DVtS8UqXpi copy

Thank you for your support!

Follow Us!

:(