TikTok and Meta have been told to provide the EU with information that may have caused disinformation on their platforms related to the Israel-Gaza conflict.
Previously, they were given 24 hours to respond to stakeholder concerns.
But this claim was not legally enforced; The last claim is correct.
Both companies have one week to respond. Under its new technical rules, the EU can open a formal investigation if it is not satisfied with their responses.
Following the Hamas attack on Israel, the EU is concerned about the possible spread of terrorism and violent content.
"Next week we will publish our first transparency report under [the new law], which will include more information about what we do to keep our European community safe," a TikTok spokesperson said.
A Meta spokesperson said: "Our teams are coordinating with external fact-checkers in the region to keep our platforms safe, take action against content that violates our policies or local laws, and limit the spread of misinformation. We are happy to provide details of this work beyond what we have already shared, and we will respond to the European Commission."
The EU's latest request comes a week after X, formerly known as Twitter, raised the same concerns.
X said it had removed hundreds of Hamas-affiliated accounts from the platform.
Social media companies have seen a rise in disinformation about the conflict between Israel and Hamas, including doctored images and branded videos.
Meta, TikTok X and Google's CEOs each received a letter from EU Commissioner Thierry Breton in early October, giving them 24 hours to respond.
But these messages are legal under new EU technology laws governing what content is allowed online. Not legally binding claims.
Now, under the Digital Services Act (DSA), companies must respond by a set deadline.
Failure to comply with the DSA can result in fines of up to 6% of a company's global turnover; or the platform may be suspended.
Double the deadline
At this legal stage under the DSA; The commission has set two deadlines for Meta and TikTok.
First, The companies have until October 25 to provide requested information about the "crisis response," and they must answer questions about protecting election integrity by November 8.
TikTok has been tasked with telling the European Commission how it protects minors online by November.
When social media companies asked for more information in the past. Meta is timely; Mr Breton said he had to prove he had acted diligently and objectively.
TikTok later said it "has a special responsibility to protect children and teenagers from violent content and terrorist propaganda."