The EU’s digital rights chief warned the owner of X, formerly Twitter, that war misinformation flooding the site was a potential breach of EU law.


The social media platform X, formerly known as Twitter, says it is struggling with a flood of posts sharing graphic media, violent speech, and hateful conduct about the Israel-Hamas war. 

But it has received a broadside of criticism, including from a top European Union official, questioning the adequacy of the response and giving its billionaire owner Elon Musk a 24-hour ultimatum to comply with the bloc’s Digital Services Act.

Outside watchdog groups said misinformation about the war abounds on the platform, whose workforce – including its content moderation team – was gutted by Musk after he bought it last year.

Fake and manipulated imagery circulating on X include “repurposed old images of unrelated armed conflicts or military footage that actually originated from video games,” said a Tuesday letter to Musk from European Commissioner Thierry Breton. 

“This appears to be manifestly false or misleading information”.

Breton, the EU’s digital rights chief, also warned Musk in the letter, which was posted on X, that authorities have been flagging “potentially illegal content” that could violate EU laws and “you must be timely, diligent and objective” in removing it when warranted.

San Francisco-based X didn’t immediately respond to a request for comment about Breton’s letter.

A post late on Monday from X’s safety team claimed it is treating the crisis with utmost effort.

“In the past couple of days, we’ve seen an increase in daily active users on @X in the conflict area, plus there have been more than 50 million posts globally focusing on the weekend’s terrorist attack on Israel by Hamas. As the events continue to unfold rapidly, a cross-company leadership group has assessed this moment as a crisis requiring the highest level of response,” it read.

That includes continuing a policy frequently championed by Musk of letting users help rate what might be misinformation, which causes those posts to include a note of context but not disappear from the platform.

Sharing ‘absolutely poisonous’ news sources

The struggle to identify reliable sources for news about the war was exacerbated over the weekend by Musk, who on Sunday posted the names of two accounts he said were “good” for “following the war in real-time”. 

Analyst Emerson Brooking of the Atlantic Council called one of those accounts “absolutely poisonous”. 

Journalists and X users also pointed out that both accounts had previously shared a fake AI-generated image of an explosion at the Pentagon, and that one of them had posted numerous antisemitic comments in recent months. 

Musk later deleted his post.

Brooking posted on X that Musk had enabled fake war reporting by abandoning the blue check verification system for trusted accounts and allowing anyone to buy a blue check.

Brooking said Tuesday that it is “significantly harder to determine ground truth in this conflict as compared to Russia’s invasion of Ukraine” last year and “Elon Musk bears personal responsibility for this”.

‘Musk has made it worse’

He said Musk’s changes to the X platform have made it impossible to quickly assess the credibility of accounts while his “introduction of view monetisation has created perverse incentives for war-focused accounts to post as many times as possible, even unverified rumors, and to make the most salacious claims possible”.

“War is always a cauldron of tragedy and disinformation; Musk has made it worse,” he added. 


Further, Brooking said via email “Musk has repeatedly and purposefully denigrated the idea of an objective media, and he made platform design decisions that undermine such reporting. We now see the result”.

Part of Musk’s drastic changes over the past year included removing many of the people responsible for moderating toxic content and harmful misinformation.

One former member of Twitter’s public policy team said the company is having a harder time taking action on posts that violate its policies because there aren’t enough people to do that work.

“The layoffs are undermining the capacity of Twitter’s trust and safety team, and associated teams like public policy, to provide needed support during a critical time of crisis,” said Theodora Skeadas, one of the thousands of employees who lost their jobs in the months after Musk bought the company.

X policy change over sensitive content

X says it changed one policy over the weekend to enable people to more easily choose whether or not to see sensitive media without the company actually taking down those posts.


“X believes that, while difficult, it’s in the public’s interest to understand what’s happening in real-time,” its statement said.

The company said it is also removing newly created Hamas-affiliated accounts and working with other tech companies to try to prevent “terrorist content” from being distributed online. 

It said it is “also continuing to proactively monitor for antisemitic speech as part of all our efforts. Plus we’ve taken action to remove several hundred accounts attempting to manipulate trending topics”.

Linda Yaccarino, whom Musk named in May as the top executive at X, withdrew from an upcoming three-day tech conference where she was scheduled to speak, citing the need to focus on how the platform was handling the war.

“With the global crisis unfolding, Linda and her team must remain fully focused on X platform safety,” X told the organizers of the WSJ Tech Live conference being held next week in Laguna Beach, California, the US.


Please enter your comment!
Please enter your name here