"We're testing a feature for you to report Tweets that
seem misleading - as you see them," the social network said from its
safety and security account.
Starting Tuesday, a button would be visible to some users
from the United States, South Korea and Australia to choose "it's
misleading" after clicking "report tweet."
Users can then be more specific, flagging the misleading
tweet as potentially containing misinformation about "health,"
"politics" and "other."
"We're assessing if this is an effective approach so
we're starting small," the San Francisco-based company said.
"We may not take action on and cannot respond to each
report in the experiment, but your input will help us identify trends so that
we can improve the speed and scale of our broader misinformation work."
Twitter, like Facebook and YouTube, regularly comes under
fire from critics who say it does not do enough to fight the spread of
misinformation.
But the platform does not have the resources of its Silicon
Valley neighbors, and so often relies on experimental techniques that are less
expensive than recruiting armies of moderators.
Such efforts have ramped up as Twitter toughened its
misinformation rules during the Covid-19 pandemic and during the US
presidential election between Donald Trump and Joe Biden.
For example, Twitter began blocking users in March who have
been warned five times about spreading false information about vaccines.
And the network began flagging tweets from Trump with a
banner warning of their misleading content during his 2020 re-election
campaign, before the then-president was finally banned from the website for
posting incitements to violence and messages discrediting the election results.
Moderators are ultimately responsible for determining which
content actually violates Twitter's terms of use, but the network has said it
hopes to eventually use a system that relies on both human and automated
analysis to detect suspicious posts.
Concern around Covid-19 vaccine misinformation has become so
rampant that in July Biden said Facebook and other platforms were responsible
for "killing" people in allowing false info around the shots to
spread.
He walked back the remarks the clarify that the false
information itself is what could harm or even kill those who believe it.
0 comments:
Post a Comment