I would like to see a thread on the problem of CSAM that doesn’t use automated CSAM reporting tools as the metric to show that there’s a problem.
I’m not denying that there’s a CSAM problem in the sense that there is a certain small population of users who promote this terrible stuff, and that there is awful abuse that drives it. But when we say there’s a “problem”, we’re implying it’s getting rapidly worse.
The actually truth here is that we have no idea how bad the underlying problem is. What we have are increasingly powerful automated tools that detect the stuff. As those tools get better, they generate overwhelming numbers of reports.
The overwhelming number of reports is great for tech companies because it lets them say they’re doing something (and close a few easily replaced accounts.) but it floods law enforcement’s very limited ability to deal with the problem.
It also creates a dynamic where people can say “look at these millions of automated reports, the problem must be out of control and getting worse.” And then tech companies build more automated scanning systems.
At the root of every story on the worsening problem of CSAM, you’ll find someone pointing to increasing numbers of automated detection reports. There is never, ever any data-driven discussion of whether these are cause or effect — or if these reports really help.
After Apple turns on its photo scanning for a billion devices, the number of CSAM reports is going to explode. To staggering numbers. Just based on projecting results on previous platforms to Apple’s customer base.
The CSAM “problem” won’t be any worse than it was the day before Apple ships this update. But it will be apparently worse, in the same way that these graphs from the NYT show it getting apparently worse.
So I would ask people to think hard about whether they have any hard evidence of a worsening problem, before they handwave towards high reporting numbers and move on to the *important* point they want to make.
And moreover, I would ask whether dumping another few tens of millions of potentially noisy reports into the system (Facebook estimates 75% are “non-malicious”) is going to make law enforcement more effective or less.
And as a note, I want to be clear that “X million reports” usually means “X million photos or videos”. It does not mean X million people. The number of people doing this stuff is as small as you would think it is, it isn’t like COVID where it spreads through the population.
The fact that a small number of criminals continue to operate and spread abusive material in increasing amounts is an *indictment* of the effectiveness of CSAM surveillance systems, not an argument for building more of them.

• • •

Missing some Tweet in this thread? You can try to force a refresh
 

Keep Current with Matthew Green

Matthew Green Profile picture

Stay in touch and get notified when new unrolls are available from this author!

Read all threads

This Thread may be Removed Anytime!

PDF

Twitter may remove this content at anytime! Save it as PDF for later use!

Try unrolling a thread yourself!

how to unroll video
  1. Follow @ThreadReaderApp to mention us!

  2. From a Twitter thread mention us with a keyword "unroll"
@threadreaderapp unroll

Practice here first or read more on our help page!

More from @matthew_d_green

10 Aug
Everyone keeps writing these doomed takes about how “the US government is going to force tech companies to comply with surveillance, so they might as well just give in preemptively.” Like it’s inevitable and we should just hope for what scraps of privacy we can.
Even I was pessimistic last week. What I’ve seen in the past week has renewed my faith in my fellow countrymen — or at least made me realize how tired and fed up of invasive tech surveillance they really are.
People are really mad. They know that they used to be able to have private family photo albums and letters, and they could use computers without thinking about who else had their information. And they’re looking for someone to blame for the fact that this has changed.
Read 11 tweets
7 Aug
Someone pointed out that Apple’s Intel Macs probably can’t run their client-side scanning software because they don’t possess a neural engine coprocessor. Real time scanning on Macs is going to require an upgrade to newer M1 hardware (or beyond).
It’s sure a weird thing to pay a ton of money for Apple’s latest hardware, and the first thing they do with it is scan your personal files.
Some other folks have asked whether corporate and enterprise-managed devices will be subject to scanning. What I’ve heard is that enterprise customers are *very* surprised and upset. Apple hasn’t announced if there will be an MDM setting to disable it.
Read 4 tweets
7 Aug
It’s gradually dawning on me how badly Apple screwed up with this content scanning announcement.
If Apple had announced that they were scanning text messages sent through their systems, or photo libraries shared with outside users — well, I wouldn’t have been happy with that. But I think the public would have accepted it.
But they didn’t do that. They announced that they’re going to do real-time scanning of individuals’ *private photo libraries* on their own phones.

That’s… something different. And new. And uncomfortable.
Read 8 tweets
5 Aug
Yesterday we were gradually headed towards a future where less and less of our information had to be under the control and review of anyone but ourselves. For the first time since the 1990s we were taking our privacy back. Today we’re on a different path.
I know the people who did this have good intentions. They think this was inevitable, that we can control it. That it’ll be used only for good, and if it isn’t used for good then that would have happened anyway.
I was alive in the 1990s. I remember we had things like computers that weren’t connected to the Internet, and photo albums that weren’t subject to continuous real-time scanning. Society seemed… stable?
Read 4 tweets
5 Aug
Reading through the analysis. This is not… a security review.
“If we assume there is no adversarial behavior in the security system, then the system will almost never malfunction. Since confidentiality is only broken when this system malfunctions, the system is secure.”
Don’t worry though. There is absolutely no way you can learn which photos the system is scanning for. Why is this good? Doesn’t this mean the system can literally scan for anything with no accountability? Not addressed.
Read 12 tweets
5 Aug
A small update from last night. I described Apple’s matching procedure as a perceptual hash function. Actually it’s a “neural matching function”. I don’t know if that means it will also find *new* content on your device or just known content.
Also, it will use a 2-party process where your phone interacts with Apple’s server (which has the unencrypted database) and will only trigger an alert to Apple if multiple photos match its reporting criteria.
I don’t know anything about Apple’s neural matching system so I’m hopeful it’s just designed to find known content and not new content!

But knowing this uses a neural net raises all kinds of concerns about adversarial ML, concerns that will need to be evaluated.
Read 4 tweets

Did Thread Reader help you today?

Support us! We are indie developers!


This site is made by just two indie developers on a laptop doing marketing, support and development! Read more about the story.

Become a Premium Member ($3/month or $30/year) and get exclusive features!

Become Premium

Too expensive? Make a small donation by buying us coffee ($5) or help with server cost ($10)

Donate via Paypal Become our Patreon

Thank you for your support!

Follow Us on Twitter!

:(