Daily BulletinDaily Bulletin

The Conversation

  • Written by Jennifer Beckett, Lecturer in Media and Communications, University of Melbourne

Facebook has recently come fire for not doing enough to keep disturbing content out of our newsfeeds. It hopes a hiring spree will fix the problem.

In a Facebook post Wednesday, company founder and CEO Mark Zuckerberg announced plans to add an additional 3,000 moderators to its community operations team. These new employees will help review posts that are flagged as troubling by the community and “improve the process for doing it quickly”.

Given the recent spate of news stories about Facebook Live being used to stream everything from rapes to murders in real time, it’s fair to say the social media giant has work to do.

Algorithms fall short

So why are people necessary when algorithms can do the job?

Facebook uses software to filter some types of explicit or illegal content, but there are limits to its capabilities. That’s where humans come in.

At the 2015 SWARM conference for community managers, Mia Garlick, Facebook’s director of policy for Australia and New Zealand, noted the social media company uses Microsoft’s PhotoDNA software to weed out known images of child pornography posted onto the platform. And therein lies the rub – the program must already “know” the images to delete them.

Algorithms are only as good as we teach them to be. In the case of something like PhotoDNA, someone – a human – has to find and add additional images to the software’s database.

These automated systems can also be blunt tools. They see the world as black and white, so they’re not always good at deciding when something falls into a grey area – say, making decisions on whether that nude photo is a piece of art or just a nude photo.

image Mark Zuckerburg hopes new content moderators will help prevent violent content getting into the newsfeed. John Adams/flickr, CC BY

This issue also arises when cultural context or local law is at odds with a Facebook post.

In a 2015 blog post explaining Facebook’s community guidelines, Monika Bickert, Facebook’s head of global policy management, and Chris Sonderby, deputy general counsel, touched on the vexed notion of blasphemy to demonstrate how hard it can be to moderate language and imagery in different countries.

The update followed a news story about Facebook blocking a page in Turkey after a court there ruled it was insulting the Prophet Mohammed.

“While blasphemy is not a violation of the Community Standards, we will still evaluate the reported content and restrict it in that country if we conclude it violates local law,” Bickert and Sonderby explained in the post.

In other words, we need humans to catch anything that falls between the cracks. That’s why Facebook relies on its human community to report content and flagged posts are still screened by its team of human moderators.

Moderators are people, too

Good moderation is at the heart of building a safe space online, and Zuckerburg’s stated commitment to building a better Facebook community is an important first step.

Still, one thing is often forgotten in this discussion: the people whose labour ensures the rest of us have a pleasant experience online.

Take Microsoft, creators of the PhotoDNA software used by Facebook, and its online safety team. In December 2016, two former employees of Microsoft’s team sued the technology company after allegedly developing post traumatic stress disorder.

According to the suit, the pair were responsible for reviewing imagery of “horribly violent acts against children”, among other content.

In particular, the lawsuit alleges the two employees did not receive sufficient support from Microsoft despite the traumatic nature of their work.

If the lawsuit is successful, it could have serious ramifications for any company employing content moderators. That could be anything from banks to the media and, of course, Facebook and its 3,000 new moderators.

These companies have a moral duty to help employees who may be routinely exposed to rape and torture imagery, though recent news reports suggest that’s not always the case.

In December, Germany’s Süddeutsche Zeitung newspaper found Facebook’s Berlin moderators facing dismaying conditions. The journalists reported employees were viewing violent and explicit material daily with little support while making “just slightly above the legal minimum wage”. Facebook declined to comment to the publication.

Whether Facebook’s new influx of moderators will create the “global community that works for everyone” Zuckerburg outlined in his recent manifesto remains to be seen. With billions of pieces of content being posted to Facebook everyday, it’s an almost impossible task.

Authors: Jennifer Beckett, Lecturer in Media and Communications, University of Melbourne

Read more http://theconversation.com/facebook-turns-to-real-people-to-fix-its-violent-video-problem-77156

The top villa communities in Dubai


Buying a home in Dubai: Package deals explained


8 Undeniable Signs You Need a Knee Brace


The Conversation


$1.8 billion boost for local government

The Federal Liberal and Nationals Government will deliver a $1.8 billion boost for road and community projects through local governments across Australia.   The package of support will help lo...

Scott Morrison - avatar Scott Morrison

Scott Morrison press conference

PRIME MINISTER: This is a tough day for Australia, a very tough day. Almost 600,000 jobs have been lost, every one of them devastating for those Australians, for their families, for their commun...

Scott Morrison - avatar Scott Morrison


Local economic recovery plans will help towns and regions hit by bushfires get back on their feet as part of a new $650 million package of support from the Morrison Government.   As part of th...

Scott Morrison - avatar Scott Morrison

Business News

Hotdesking might not be ‘dead’ after all

According to Christian Pistauer, Workplace Strategy director of Meta5 Group in Australia, COVID will dramatically change the commercial real estate sector in Australia for many years to come. ”...

Tess Sanders Lazarus - avatar Tess Sanders Lazarus

Office expert: Don't bring your staff back to work until you have done these things

With lockdown restrictions gradually being eased across the country, Australian workplaces are looking at the types of changes needed in order to meet new health and wellness requirements post-l...

Tess Sanders Lazarus - avatar Tess Sanders Lazarus

Major health and wellness brands sign-on to open at Yamanto Central

While COVID restrictions start to ease across the country, plans for Queensland’s newest shopping centre, Yamanto Central, ramp up. Due for completion in the first half of 2021, Yamanto Cent...

Tess Sanders Lazarus - avatar Tess Sanders Lazarus

News Company Media Core

Content & Technology Connecting Global Audiences

More Information - Less Opinion