SAN FRANCISCO —
YouTube said Wednesday it would take down content which contradicts expert consensus about COVID-19 vaccines, updating its policies on misinformation about the pandemic.
The move is the latest by online platforms struggling to contain the spread of hoaxes and false information about the coronavirus and treatments.
YouTube, the Google-owned video-sharing service, said it was expanding its medical misinformation policy “to remove claims about COVID-19 vaccinations that contradict expert consensus from local health authorities or the World Health Organization.”
YouTube said it was acting in anticipation of the release of one or more vaccines, and skepticism among many people about their usefulness.
It said content to be removed would include claims that a vaccine could kill people or cause infertility, or that microchips will be implanted in people who receive a vaccination.
YouTube said it has removed more than 200,000 videos with “dangerous or misleading” COVID-19 information since February, including unverified claims about transmission or unsubstantiated treatments.
In a related action, Facebook on Tuesday announced a ban on ads that discourage people from getting vaccinated, as part of its efforts to contain misinformation.
The tech giants have regularly been accused of allowing anti-vaccine movements to flourish.