Facebook Bug Made It Ask Users if Posts They Saw Contained Hate Speech

Share on Facebook Tweet Share Reddit Comment
Facebook Bug Made It Ask Users if Posts They Saw Contained Hate Speech

Facebook has struggled for years to figure out what is and isn't hate speech on its platform. On Tuesday, a "bug " revealed that Facebook might be thinking about crowd-sourcing the question.

At the bottom of each post on my News Feed on Tuesday morning was the same question: "Does this post contain hate speech?" It appeared on everything from news articles to personal updates to a picture of my cat.

The question appeared - and disappeared - on Tuesday morning, as Facebook users began commenting with amusement and alarm at the variety of posts that Facebook wanted to know about.

"This was an internal test we were working on to understand different types of speech, including speech we thought would not be hate. A bug caused it to launch publicly. It's been disabled," a Facebook spokesperson said on Tuesday.

As for why Facebook might be testing a feature like this, a little bit of context is needed.

Just last week, Facebook finally released the guidelines it uses internally to enforce its own rules on the platform. That's important, in part, because Facebook has long struggled to moderate content consistently, or account for the context of a post.

For instance: Facebook once removed Nick Ut's iconic Vietnam War photograph, claiming it violated the site's policy on nudity. The platform defended, and then reversed, the decision, after the photograph's removal outraged basically the entire country of Norway. And minority groups have long said that they believe Facebook's rules and enforcement unfairly punish those who try to call out hate speech and racism.

Here is how Facebook currently defines hate speech, in case you were curious:

"A direct attack on people based on what we call protected characteristics - race, ethnicity, national origin, religious affiliation, sexual orientation, sex, gender, gender identity, and serious disability or disease. We also provide some protections for immigration status. We define attack as violent or dehumanizing speech, statements of inferiority, or calls for exclusion or segregation."

According to the newly-published guidelines, moderators should allow content calling out hateful speech that would otherwise violate their prohibition against it, "but we expect people to clearly indicate their intent" when doing so. "Where the intention is unclear, we may remove the content."

© The Washington Post 2018

Comments

For the latest tech news and reviews, follow Gadgets 360 on Twitter, Facebook, and subscribe to our YouTube channel.

Fujifilm-Xerox Merger: Xerox CEO to Resign in Settlement With Top Shareholders
Snapchat Now Has 191 Million Daily Active Users as Redesign Weighs on Results
 
 

Advertisement

 

Advertisement