Menu

Reactions to the New Online Safety Bill

Click The Arrow For The Table Of Contents
Westminster Bridge at sunny day

The Government’s decision to amend the Online Safety Bill to allow ‘legal but harmful’ content has received a mixed reaction, with criticism coming from The Samaritans. 

 

What Is The Online Safety Bill?

 

The UK government’s Online Safety Bill, initially proposed by former PM Teresa May, is (draft) legislation designed to place a ‘duty of care’ on internet companies which host user-generated content to limit the spread of illegal content on these services. 

 

The idea of the Bill is to prevent the spread of illegal content and activity (e.g. images of child abuse, terror material, and hate crimes), as well as to protect children from harmful material. Until the recent amendment, it was designed to protect adults from legal but harmful content. 

 

Where?

 

The Online Safety Bill applies to social media platforms, video-sharing platforms, search engines, plus other tech services. He requires them to implement systems and processes to remove illegal content as soon as they become aware. The Bill also requires these services to take additional proactive measures about the most harmful ‘priority’ forms of illegal online content.

The Amendment – Legal But Harmful Material OK?

 

Following five months of delays, plus a return to Parliament, Culture Secretary Michelle Donelan confirmed an amendment to the Bill which shifts the emphasis away from general safety towards child safety. The amendment means that social media sites will no longer need to remove material designated “legal but harmful.” 

 

Michelle Donelan said: “Any incentives for social media firms to over-remove people’s legal online content will be taken out of the Online Safety Bill. Firms will still need to protect children and remove content that is illegal or prohibited in their terms of service; however, the Bill will no longer define specific types of legal content that companies must address.” She added: “This removes any future influence governments could have on what private companies do about legal speech on their sites or any risk that companies are motivated to take down legitimate posts to avoid sanctions.” 

 

Why? 

 

The Government says that the amendment has been made over fears that the Bill would stifle free speech and create a quasi-legal category between illegal and legal where sizable platforms and companies would have to remove both illegal content plus any content that had been named as legal but potentially harmful. This echoes concerns by free speech campaigners that the Bill could allow the government or tech platforms to censor content. 

Replaced With What?

 

The Government says the legal but harmful measures will be replaced with a “triple-shield”, which it says will “strike the right balance with its protections for free speech.” This will take the form of “new duties which strengthen Bill’s free speech requirements on major online platforms”, which it says will make them more accountable for their policies. The new duties will prohibit online platforms from removing or restricting user-generated content or suspending or banning users where the content doesn’t go against the platform’s terms of service or the law. 

Concerns

 

The amendment has been met with mixed reactions. Although there is broad agreement that Bill’s emphasis on protecting children is a good thing, concern has been expressed by some groups. For example, problem has been described by the Samaritans in an open letter online, which says: “We have been pleased to see a continued commitment from the Government to protecting vulnerable children as it considers modifying the Bill. But susceptibility to harm from suicide and self-harm content does not end when people reach the age of 18. Anyone, including young adults aged 18-24, can be just as vulnerable to harm from this type of content”. 

What Does This Mean For Your Business?

 

In practical terms, businesses like social media platforms will now need to do less assessing, restricting, and removing content. They will need to relate judgement of content to whether it’s legal or prohibited in their terms of service. Other work may need to be done by platforms too, e.g. looking more closely at their terms of service, which is required to show how they enforce their user age limits, answering to a toughened regulator, being more transparent about how their algorithms work and publishing details of when the regulator Ofcom has taken action against them. Although the amendment still provides protection for children, there is an argument, as made by the Samaritans, that young people over 18 need protection too from certain content.

 

That said, the change should please free speech campaigners, plus there are other aspects to the Bill that should provide a wide range of projections in law, particularly for young people from many types of harmful content and online behaviours.