The Online Safety Act in the UK is, quietly, one of the most important pieces of legislation to have come out of this government. Admittedly, the competition is slim. But as time goes by, and more and more of the act begins to take effect, we’re starting to see how it will reshape the internet.
Social media firms have been told to “tame aggressive algorithms” that recommend harmful content to children, as part of Ofcom’s new safety codes of practice.
The children’s safety codes, introduced as part of the Online Safety Act, let Ofcom set new, tight rules for internet companies and how they can interact with children. It calls on services to make their platforms child-safe by default or implement robust age checks to identify children and give them safer versions of the experience.
The Online Safety Act is much more than just the child-focused aspects, but these are some of the toughest powers handed to Ofcom under the new regulatory regime. Websites will be required to have age-verification tech to know which of its users are children – or, alternatively, to ensure that all its content is safe for children to use.
The content kids do see will need to be kept to a much tighter set of rules than the adult web, with some types of content – including pornography and material relating to suicide, self-harm and eating disorders – strictly banned from young people’s feeds.
Most immediately interesting, though, is the requirement I quoted above. It’s one of the first efforts anywhere in the world to impose a strict requirement on the curation algorithms that underpin most of the biggest social networks, and will see services like TikTok and Instagram required to suppress the spread of “violent, hateful or abusive material, online bullying, and content promoting dangerous challenges” on children’s accounts.
Some fear that Ofcom is trying to have its cake and eat it. The easiest way to suppress such content, after all, is to block, something that doesn’t require faffing about with recommendation algorithms. Anything less, and there’s an inherent gamble: is it worth risking a hefty fine from Ofcom if you decide to allow some violent material on to children’s feeds, even if you can argue you’ve suppressed it beyond where it normally would be?
It might seem an easy fear to dismiss. Who’s going to bat for the right to show violence to children? But I’m already counting down the days until a well-meaning government awareness campaign – maybe about safer streets, maybe something related to drug policy – gets suppressed or blocked under these rules, and the pendulum swings back the other way. Jim Killock, the head of Open Rights Group, an internet policy thinktank, said he was “concerned that educational and help material, especially where it relates to sexuality, gender identity, drugs and other sensitive topics may be denied to young people by moderation systems”.
Of course, there is opposition from the other direction, too. The Online Safety Act was engineered to sit squarely in the Goldilocks zone of policy, after all:
The Goldilocks theory of policy is simple enough. If Mummy Bear says your latest government bill is too hot, and Daddy Bear says your latest government bill is too cold, then you can tuck in knowing that the actual temperature is just right.
Unfortunately, the Goldilocks theory sometimes fails. You learn that what you actually have in front of you is less a perfectly heated bowl of porridge and more a roast chicken you popped in the oven still frozen: frosty on the inside, burnt on the outside, and harmful to your health if you try to eat it.
And so, where Killock is concerned about the chilling effect, others worry this act hasn’t gone far enough. Beeban Kidron, a cross-bench peer and one of the leading proponents for children’s online safety rules, worries that the whole thing is too broad-brush to be helpful. She wrote in the FT (£):
The code is weak on design features, however. While the research shows livestreaming and direct messaging are high risk, there are few mandatory mitigations included to tackle them. Similarly, the requirement for measures to have an existing evidence base fails to incentivise new approaches to safety … How can you provide evidence that something does not work if you don’t try it?
As we celebrate the arrival of the draft code, we should already be demanding that the holes in it are fixed, the exceptions readdressed, the lobbyists contained.
The code is out for consultation, but my sense is that it’s a formality; everyone involved seems to expect the rules as written to be largely unchanged by the time they become binding, later this year. But the fight over what a child-safe internet means is only just beginning.
AI think therefore AI am
One of the reasons I still find the AI sector fascinating – even though I know many readers have rather made up their mind on the whole thing – is that we’re still learning quite fundamental things about how artificial intelligence works.
Take step-by-step reasoning. One of the most useful discoveries in the field of “prompt engineering” was that LLMs such as GPT do much better at answering complex questions if you ask them to explain their thinking step-by-step before giving the answer.
There’s two possible reasons for this, that you can anthropomorphise to “memory” and “thinking”. The first is that LLMs have no ability to reason silently. All they do is generate the next word [technically, the next “token”, sometimes just a fragment of a word] in the sentence, meaning that, unless they’re actively generating new tokens, their ability to handle complex thoughts is constrained. By asking them to “think step-by-step”, you’re allowing the system to write down each part of its answer, and use those intermediate steps to come to its final conclusion.
The other possibility is that step-by-step reasoning literally lets the system do more thinking. Each time an LLM prints a token, it does one pass through its neural network. No matter how difficult the next token is, it can’t do more or less thinking about what it should be (this is wrong, but it’s wrong in the same way that everything you learned about atoms in school is wrong.) Step-by-step thinking might help change that: letting the system spend more passes to answer a question gives it more thinking time. If that’s the case, step-by-step thinking is less like a scratch pad, and more like stalling for time while you answer a hard question.
So which is it? A new paper suggests the latter:
Chain-of-thought responses from language models improve performance across most benchmarks. However, it remains unclear to what extent these performance gains can be attributed to human-like task decomposition or simply the greater computation that additional tokens allow. We show that transformers can use meaningless filler tokens (eg, ‘......’) in place of a chain of thought to solve two hard algorithmic tasks they could not solve when responding without intermediate tokens.
In other words, if you teach a chatbot to just print a dot each time it wants to think, it gets better at thinking. That is, the researchers warn, easier said than done. But the discovery has important ramifications for how we use LLMs, in part because it suggests that what the systems write when they show their working might not be all that relevant to the final answer. If your reasoning can be replaced with a lot of periods, you were probably doing the real work in your head anyway.
If you want to read the complete version of the newsletter please subscribe to receive TechScape in your inbox every Tuesday