Social media platform Facebook is facing criticism over its handling of posts that allegedly celebrated the deadly Bondi Beach attack in Sydney, which killed 15 people during a Hanukkah gathering on December 14. An anti-hate group has accused the platform of being slow to remove content praising the attackers and promoting extremist ideology.
What the Anti-Hate Group Says
The Community Security Trust (CST), a UK-based organization that works to protect Jewish communities from antisemitism and terrorism, said it identified a significant number of Facebook accounts sharing Islamic State (IS)-supporting content related to the attack.
According to CST, several posts praising the violence remained online for at least two days after the attack, continuing to receive likes, shares, and comments. Some of the accounts were reportedly based in the United Kingdom and were referred to UK counter-terrorism police.
Some Examples of the Content
CST highlighted posts that included:
- A video showing the aftermath of the Bondi Beach attack, accompanied by religious slogans praising the violence
- Images of one of the gunmen along with text glorifying IS leadership
- Posts that received hundreds of likes and multiple shares, indicating continued engagement even after the attack
The attackers—allegedly a father and son inspired by IS ideology—were not believed to have been directly controlled by the group, but authorities said the attack showed signs of planning and ideological influence.
What Is Meta’s Response
Meta, which owns Facebook, said it removed the content after being contacted by journalists and confirmed that the posts violated its policies related to dangerous organizations and individuals. The company declined to answer detailed questions about how long the content remained online.
What Is Government Reaction
UK media regulator Ofcom said social media platforms are legally required to assess reported content and remove illegal material swiftly if it violates UK law. The regulator acknowledged receiving evidence suggesting that terrorist content and illegal hate speech continue to appear on major platforms.
The UK Home Office said content promoting terrorism or violence against communities is unacceptable and emphasized that platforms have a legal duty to prevent the spread of such material.
What Are Broader Security Concerns
The incident comes amid heightened concern over terrorist threats targeting Jewish communities in Western countries. In the UK, two men were recently convicted for plotting an armed attack against Jewish targets in northwest England.
Security officials noted that while such plots may not always be directly directed by extremist organizations, they often involve online radicalization, overseas contacts, and sophisticated planning, including attempts to bypass community security measures.
Conclusion By Factcheck India
The case highlights ongoing challenges faced by social media companies in moderating extremist content in real time. While platforms have policies in place, critics argue that delays in enforcement allow harmful material to spread during critical moments—potentially increasing risk and amplifying extremist propaganda. Authorities continue to urge the public to report suspicious online content, stressing that counter-terrorism efforts rely on cooperation between platforms, regulators, law enforcement, and users.

