Britain sets first codes of practice for tech firms in online safety regime


LONDON (Reuters) – Britain’s online safety regime came into force on Monday, requiring social media companies like Meta’s Facebook and ByteDance’s TikTok to take action to tackle criminal activity on their platforms and make them safer by design.
LONDON (Reuters) – Britain’s online safety regime came into force on Monday, requiring social media companies like Meta’s Facebook and ByteDance’s TikTok to take action to tackle criminal activity on their platforms and make them safer by design.
Media regulator Ofcom said it had published its first codes of practice on tackling illegal harms such as child sexual abuse and assisting or encouraging suicide.
Sites and apps have until March 16, 2025, to assess the risks illegal content poses to children and adults on their platforms, Ofcom said.
After the deadline, they will have to start implementing measures to mitigate those risks, such as better moderation, easier reporting and built-in safety tests, Ofcom said.
Ofcom Chief Executive Melanie Dawes said the safety spotlight was now firmly on tech companies.
“We’ll be watching the industry closely to ensure firms match up to the strict safety standards set for them under our first codes and guidance, with further requirements to follow swiftly in the first half of next year,” she said.
The Online Safety Act, which became law last year, sets tougher standards for platforms such as Facebook, YouTube and TikTok, with an emphasis on child protection and the removal of illegal content.
Under the new code, reporting and complaint functions will have to be easier to find and use. High-risk providers will be required to use automated tools called hash-matching and URL detection to detect child sexual abuse material, Ofcom said.
The regulator will be able to issue fines of up to 18 million pounds ($22.3 million) or 10% of a company’s annual global turnover if they fail to comply.
Britain’s Technology Secretary Peter Kyle said the new codes were a “material step change in online safety”.
“If platforms fail to step up the regulator has my backing to use its full powers, including issuing fines and asking the courts to block access to sites,” he said.
(Reporting by Paul Sandle; Editing by Emelia Sithole-Matarise)
Online safety refers to the practice of protecting users from harmful content and interactions on the internet, including measures to prevent cyberbullying, identity theft, and exposure to inappropriate material.
A media regulator is an organization responsible for overseeing and enforcing laws and regulations related to media content and practices, ensuring compliance and protecting the public interest.
Child sexual abuse material refers to any visual depiction of sexually explicit conduct involving a minor, which is illegal and subject to strict penalties under law.
Hash-matching technology is a method used to identify and block illegal content, such as child sexual abuse material, by comparing digital fingerprints of files against known databases.
Automated tools in online safety are software solutions that help detect and manage harmful content on platforms, improving moderation and reporting processes to enhance user safety.
Explore more articles in the Top Stories category











