A man in a QAnon vest waves a US flag as demonstrators gather on October 5, 2020 to protest mandatory flu vaccines for children in Boston, Massachusetts

YouTube said Thursday it was tightening rules for propagation of conspiracy theories, notably targeting the QAnon movement already limited on Twitter and Facebook.

The Google-owned video-sharing service said it was expanding its policies on hate and harassment "to prohibit content that targets an individual or group with that have been used to justify real-world violence."

This could mean removing videos threatening or harassing people by suggesting they are complicit in a like Pizzagate—about a supposed child sex trafficking ring with links to former Democratic White House candidate Hillary Clinton that operated from a Washington pizzeria.

QAnon has grown sharply during the pandemic because it acted as a binding force—mixing its core tenet of anti-Semitic and white supremacist tropes with long-running conspiracy theories about vaccines and 5G , as well as far-right and libertarian politics.

YouTube said it had previously removed "tens of thousands of QAnon videos" and terminated some channels used by the movement, notably those that explicitly threaten violence or deny the existence of major violent events.

Earlier this month Facebook banned QAnon-linked accounts on its core social network and on Instagram. Twitter began a crackdown on QAnon earlier this year.

The latest move by YouTube comes amid heightened tensions over misinformation spreading on social media, while some conservatives have accused platforms of bias in taking down content.