š” My technology > your technology
Thoughts on why the Online Safety Bill is probably a pile of trash, constructed from other trash
Hello there, children of 56kb internet. This week my body decided that I donāt deserve sleep, so naturally that led me to think that YOU donāt deserve this newsletter. But I couldnāt help myself. I read one STUPID story about the STUPID Online Safety Bill and I couldnāt ignore it. So this weekās issue will focus only on that. For a more traditional Horrific/Terrific, read last weekās.
This bill has been floating around like a putrid fart for years now. I remember reading the online harms white paper in 2019 ā and then nearly a year ago, I wrote about the bill in this very newsletter. There are key āchangesā to the bill that the UK news seems to have only just picked up ā but these were already being discussed last March. So maybe thereās been a lot of flip-flopping, but very little has changed:
The language around content which is legal but harmful is still present but is just less explicit
Prison time for tech execs instead of fining the companies was discussed a year ago, and is still being discussed now
The vague rules and even vaguer modes of enforcement still only apply to āthe biggest and most popularā online services who host user-generated content
The vapid career politicians who threw this bill together over what Iām guessing was a four-hour lunch break have, unsurprisingly, a rudimentary understanding of how online social spaces operate. So, there are three really Annoying Things enmeshed in this bill that we should probably take the time to examine. Those are:
Lack of nuance around what constitutes āsocial mediaā
Just likeā¦ literally having no idea what free speech is, and therefore inadvertently restricting it with the power of ignorance
The dangerous assumption that the answer to ābad technologyā is adding more technology
Okay letās gooooooooā¦.
āļø Social media is more than just Instagram and TikTok
This bill manages to be vague enough that you have no idea what itās really going to do, but also narrow enough in scope that whatever it does do, it either wonāt be enough, or it will be too much in one place. Iām a GOOD writer so no, that sentence WASNāT too long, actually.
The companies affected will be ones who host user-generated content, who are also āthe biggest and most popularā. There is no mention of how they will measure what is ābigā and what is āpopularā. Surely with social media ābigā is āpopularā ā because the amount of users on a platform is pretty much the only way to understand its size.
Anyway, even if you disregard all that, youāre still left with a cack-handed interpretation of āsocial mediaā. The phrase āsocial mediaā is becoming less relevant with every passing week. These days online social spaces are extremely varied; some are buried deep within the tightest sub-fandom niche, and others appeal to nearly everyone and extremely accessible. Either way, I donāt think the UK government are aware of anything beyond traditional newsfeed platforms.
What about Minecraft? Apex Legends? Second Life? VR Chat? Or Roblox? Ah yes, good old Roblox ā a gaming platform designed for children, with almost no protection against bullying, or adults who exploit them and steal their money. Just look at this tweet thread where a mother describes the kinds of āgamesā you can find on Roblox, and what happens in them. One of them is called āpublic toilet simulatorā???
Wikipedia also runs on user-generated content ā do the rules apply to them? What about people who run Discord servers? Or Mastodon? If they āfail to protect children from harmful contentā will server-owners be sent to prison?? What happens to the Wikipedia pages that outline school shootings and other horrific things that probably count as ālegal but harmfulā?? Will the regulator just decide? That leads us to the next annoying thingā¦
āļø The bill claims to protect free speech, but will probably degrade it
So, according to the fact sheet, Ofcom (the regulator) will ābe able to make companies change their behaviourā and āhelp companies to comply with the new laws by publishing codes of practice, setting out the steps a company should take to comply with their new dutiesā.
ā” Shocking: this is not a good way to protect free speech. True free speech would not have some central authority (Ofcom) dictating a code of practice to online platforms, or force the platforms to change their behaviour. Iām not even saying these things are a bad idea necessarily ā the rules that Ofcom come up with might be great (lmaoā¦ doubt). So why paint this as a positive step towards preserving free speech? The bill should have probably never mentioned free speech at all. You canāt sit there and demand to radically change the way a website like Wikipedia works āin the name of free speechā. But, of course, I donāt think they even thought about Wikipedia when they sat in the pub while their under-paid assistants came up with these rules, did they?
The bill definitely part of the current trend of new laws that propose to protect the free speech of individuals, but really donāt. They also seem to work hard to actively restrict the free speech of platforms themselves. Any rules that say āplatforms have to do xyz with their contentā limits the platformās freedom of expression ā because it means they donāt even get to create their own community guidelines.
What about places like 4chan or Kiwi Farms? These platforms are maybe too small to be in scope (but who the fuck knows ā they are yet to clarify the scope!). But why should size be the only thing we go by?? 4chan may be āsmallerā than Facebook, but because of the nature of the user base, certainly has potential to inflict serious harm. If they were organised enough to invent Rick-Rolling, they can probably do anythingā¦ thatās all Iām saying.
š¤ Finally, the ultimate lawmakerās fix: if technology is bad, add more technology
Right, Iām just going to say it: the solution to eliminating child sexual abuse materials (CSAM) online is not to spy on people ā itās just not. This is a social issue which requires approaches that sit outside of online platforms, and the tech industry at large. Social media and/or the internet is not the root cause of child p*rn. We all know this, and yet, the proposed solutions are always something like:
āOfcom will be able to require a platform to use highly accurate technology to scan public and private channels for child sexual abuse material. The use of this power will be subject to strict safeguards to protect usersā privacy. Highly accurate automated tools will ensure that legal content is not affectedā
šāāļø Iām justā¦ so tired of reading things like this. Thereās no such thing as āhighly accurate automated toolsā; said tools will probably catch loads of false positives, invade user privacy, and cause loads of general upset. No government should have any sort of hand in our private interactions ā just because they are happening online, it doesnāt mean they should be vulnerable to government scrutiny. This is a classic case of using shoddy technology to counter the effects of other shoddy technology ā when really, technology is not the problem at all.
This bill is as clear and persuasive as my CV was when I was 21. They should probably throw it out and start again. Oh andā¦ get some experience working behind the counter of a high street chemist.
Thank you for reading. I literally threw this together in one afternoon so I hope it made sense.