Recent rioting and unrest in the UK has led to calls for the Online Safety Act to be revisited. Mayor Sadiq Khan has called it “not fit for purpose” and Cabinet Office minister Nick Thomas Symonds suggested that the government could change the law, which was passed under the previous government and includes a raft of measures relevant to the recent riots, including powers to fine social media companies.
Prime Minister Keir Starmer has been less forthcoming about the act and has said only that he would “look more broadly at social media after this disorder”. His spokesperson suggested the act was not under active review.
In practical terms, social media played a key role in the widespread coordination of events in locations across the country. Online platforms have also served as a vehicle through which misinformation and hateful rhetoric has spread.
The act, enforced by the independent media regulator Ofcom, deals with the regulation of online speech and aims to protect users from potential harms including abuse and harassment, fraudulent activity and hate offences.
Specifically, it seeks to place more responsibility on social media companies to ensure their platforms are safe, with fines of up to 10% of their annual revenue being issued to providers whose platforms are deemed unsafe.
In more extreme cases, Ofcom has the power to require advertisers and internet providers to cease working with platforms that do not comply with the regulations. The act passed into law in October 2023, and laws in relation to individual offences are already in effect. For example, it is now an offence to share false information with an intention to cause non-trivial harm.
However, the frustration in the wake of the riots has arisen from the fact that the parts of the act are not due to come into effect until late 2024. These include enforcement powers and other measures that Ofcom could apply to social networking platforms and other platform providers, such as online forums and instant messaging platforms. This raises questions as to what might have been different in the past 14 days had they already been in place.
Algorithm concerns
A key concern has been the way in which algorithms deciding what content is recommended on social networking platforms may have propagated harmful content in relation to the riots – including racist, hateful and violent content.
For example, it was found that people were using TikTok to live-stream content of the riots as they unfolded.
At present, social media platforms such as TikTok, X, Facebook and YouTube are designed to optimise user engagement through their recommendation algorithms, with safety concerns not typically weighted within these systems. X, for example, employs different algorithms for content moderation versus content recommendation.
As a result of this, it is possible that harmful content can be recommended by one algorithm before it is identified as needing to be moderated by another algorithm.
The Online Safety Act aims to address this challenge by requiring platforms to test the safety implications of their recommendation algorithms. That is, when changes are made to their recommendation algorithms, services will be encouraged to collect safety metrics, allowing them to assess whether these algorithm changes are likely to increase individuals’ exposure to illegal content.
By incorporating these safety considerations when designing and refining content recommendation algorithms, it is hoped that fewer individuals will be exposed to harmful content before content moderation teams have had the opportunity to remove it.
Neutral oversight
One of the primary challenges around the regulation of online content is the unwillingness of platform providers to be seen as “arbiters of truth”. For example, X has recently changed the name of its Trust and Safety team to just Safety, as Elon Musk, CEO of X, stated that: “Any organisation that puts ‘Trust’ in their name cannot be trusted as that is obviously a euphemism for censorship.”
Mark Zuckerberg, CEO of Meta, said something similar back in 2016 after the US election, when he stated that Meta “shouldn’t be the arbiter of truth of everything that people say online”.
However, and as recent events have shown, this has not precluded Musk himself from propagating specific narratives in relation to the UK riots and adding fuel to an already inflamed discourse.
The act addresses this challenge by using the independent regulator, Ofcom, to enforce and regulate online content and algorithms. While the law was passed by the UK government, the government does not have powers to determine what content is allowed and what should be disallowed – thus securing political neutrality in the long-term implementation of the act.
Prevailing challenges
At present, the Online Safety Act does not include any legislation about misinformation and disinformation. This appears to be why Khan suggested that in its current form, the act does not go far enough.
The prevailing challenge of misinformation was put in sharp focus by the murders that led to the riots, with content falsely claiming that the Southport attacker was a Muslim migrant trending across several social networking platforms in the aftermath of the incident.
The home secretary Yvette Cooper claimed that social networking platforms “put rocket boosters” under the spread of this content, and there has been much debate as to whether it helped fuel the violence seen on many city streets.
This leaves some observers concerned that, until the act fully comes into force, we are in a legal purgatory around what can and cannot be litigated against online.
However, we won’t really know how effective the Online Safety Act can be until all of it has come into force and it has been tested in another situation like the recent riots.
The authors do not work for, consult, own shares in or receive funding from any company or organisation that would benefit from this article, and have disclosed no relevant affiliations beyond their academic appointment.
This article was originally published on The Conversation. Read the original article.