Policy & guidelines

YouTube misinformation policy — what's enforced

YouTube's misinformation policy is narrow but strict. Here's what falls under it and what doesn't.

Last updated: Tue May 12 2026 00:00:00 GMT+0000 (Coordinated Universal Time)

YouTube's misinformation policy is narrower than most creators assume. It covers a specific list of topics where misinformation can cause real-world harm. It does not cover "anything that's wrong" — being mistaken is not a policy violation.

What's covered

  1. Health misinformation — content contradicting WHO or local health authority guidance on prevention, transmission, or treatment of conditions like COVID-19, cancer, HIV, etc.
  2. Election misinformation — claims of widespread fraud or that election outcomes were rigged (without official verification), advancing claims that incorrectly identify winners, content meant to suppress voting.
  3. Other harmful misinformation — content that misrepresents medical procedures, encourages dangerous remedies, denies major historical atrocities, etc.

What's NOT covered

How enforcement works

Misinformation videos are typically:

  1. Demonetized first — limited ads or no ads
  2. Search-suppressed — lowered in recommendations and search results
  3. Removed — for severe or repeated violations, with a community guidelines strike

Removal triggers a strike. The strike escalation tier applies (3 in 90 days = termination).

Health misinformation specifically

This is the most active enforcement area. YouTube updates its prohibited claims list based on WHO and CDC guidance. Topics covered in 2026:

If you create health content, link to the WHO or relevant authority in your description and add a disclaimer that your content is informational, not medical advice.

Election misinformation

YouTube applies this policy in election periods (typically 3 months before and after a national election in a covered country). Outside these windows, enforcement is lighter.

Best practices