SAN FRANCISCO (KRON) – Nearly a year after COVID-19 vaccines became available in the U.S., YouTube is expanding its crackdown on perceived vaccine misinformation.
As of Wednesday, the platform had already kicked off a number of high-profile creators, including Robert F. Kennedy Jr.
The Google-owned video website updated its guidelines on Wednesday to add specifics for those spreading information contradictory to “currently administered vaccines that are approved and confirmed to be safe and effective by local health authorities and the WHO.”
Here’s what moderators are looking out for:
- Content alleging that vaccines cause chronic side effects, outside of rare side effects that are recognized by health authorities
- Content claiming that vaccines do not reduce transmission or contraction of disease
- Content misrepresenting the substances contained in vaccines
YouTube said it has already removed over 130,00 videos for violating COVID-19 vaccine policies since last year.
There are exceptions to the new guidelines. If a channel doesn’t display a pattern of promoting vaccine hesitancy, YouTube will allow it to upload personal testimonials about the vaccine. The platform will also allow content about vaccine policies, new vaccine trials, and historical vaccine successes or failures.
COVID-19 vaccine ‘misinformation’ has been enough of a concern to health officials that the U.S. Surgeon General issued an advisory in July to warn Americans about the threat.
“Health misinformation is an urgent threat to public health. It can cause confusion, sow mistrust, and undermine public health efforts, including our ongoing work to end the COVID-19 pandemic,” said U.S. Surgeon General Dr. Vivek Murthy.
The surgeon general cited a study performed in the UK and the U.S. which showed “scientific-sounding misinformation is more strongly associated with declines in vaccination intent.”
In the advisory, Murthy called for social media and tech companies to address ‘misinformation’ on their platforms.
Facebook and Twitter have also taken steps to do so.
An analysis by the Center for Countering Digital Hate (CCDH) found that just 12 people accounted for up to 73% of ‘anti-vaccine’ content on Facebook platforms, when looking at posts between Feb. 1, 2021 and March 16, 2021.
Facebook responded to the analysis last month, saying it removed over three dozen pages, groups and Facebook or Instagram accounts linked to these 12 people. They also contend they have removed over 20 million pieces of content since the beginning of the pandemic for spreading COVID-19 and vaccine ‘misinformation’.
The “digital dozen,” as the CCDH called it, also accounted for up to 17% of harmful posts on Twitter.
Since December 2020, Twitter has shared its approach to removing COVID-19 vaccine ‘misinformation’, which includes removing tweets, placing warning labels on tweets that “advance unsubstantiated rumors, disputed claims, as well as incomplete or out-of-context information about vaccines,” and allowing users to report specific tweets for ‘misinformation’.