Stay up to date with notifications from The Independent

Notifications can be managed in browser preferences.

YouTube ups video removals as concerns over child safety grow

Youtube's vice president says the video service has stopped running adverts on over 3.5 million videos since June after concerns about children's exposure to extreme content

Thursday 23 November 2017 11:47 GMT
Comments
YouTube’s Ms Wright cited 'a growing trend around content on YouTube that attempts to pass as family-friendly, but is clearly not'
YouTube’s Ms Wright cited 'a growing trend around content on YouTube that attempts to pass as family-friendly, but is clearly not' (Reuters)

YouTube stepped up enforcement of its guidelines for videos aimed at children, the unit of Alphabet Google said on Wednesday, responding to criticism that it has failed to protect children from adult content.

The streaming video service removed more than 50 user channels in the last week and stopped running ads on over 3.5 million videos since June, YouTube vice president Johanna Wright wrote in a blog post.

“Across the board we have scaled up resources to ensure that thousands of people are working around the clock to monitor, review and make the right decisions across our ads and content policies,” Ms Wright said.

“These latest enforcement changes will take shape over the weeks and months ahead as we work to tackle this evolving challenge.”

YouTube has become one of Google’s fastest-growing operations in terms of sales by simplifying the process of distributing video online but putting in place few limits on content.

Parents, regulators, advertisers and law enforcement have become increasingly concerned about the open nature of the service.

They have contended that Google must do more to banish and restrict access to inappropriate videos, whether it be propaganda from religious extremists and Russia or comedy skits that appear to show children being forcibly drowned.

Concerns about children’s videos gained new force in the last two weeks after reports in BuzzFeed and the New York Times and an online essay by British writer James Bridle pointed out questionable clips.

A forum on the Reddit internet platform dubbed ElsaGate, based on the Walt Disney princess, also became a repository of problematic videos.

Several forum posts Wednesday showed support for YouTube’s actions while noting that vetting must expand even further.

Common Sense Media, an organisation that monitors children’s content online, did not immediately respond to a request to comment about YouTube’s announcement.

YouTube’s Ms Wright cited “a growing trend around content on YouTube that attempts to pass as family-friendly, but is clearly not” for the new efforts “to remove them from YouTube.”

The company relies on review requests from users, a panel of experts and an automated computer programme to help its moderators identify material possibly worth removing.

Moderators now are instructed to delete videos “featuring minors that may be endangering a child, even if that was not the uploader’s intent,” Ms Wright said. Videos with popular characters “but containing mature themes or adult humor” will be restricted to adults, she said.

In addition, commenting functionality will be disabled on any videos where comments refer to children in a “sexual or predatory” manner.

Reuters

Join our commenting forum

Join thought-provoking conversations, follow other Independent readers and see their replies

Comments

Thank you for registering

Please refresh the page or navigate to another page on the site to be automatically logged inPlease refresh your browser to be logged in