Meta has launched a series of new platform safety measures in March 2026, targeting three interconnected problems: creator impersonation on Facebook, AI-generated scam content, and the broader flood of low-quality unoriginal posts degrading feed quality. The updates span Facebook, Instagram, WhatsApp, and Messenger, and include both new reporting tools for creators and AI-powered detection systems to identify scammers, fake accounts, and deceptive advertisers.
These announcements come as Meta faces sustained criticism over the quality of content on its platforms, with Facebook in particular being described as an “AI slop hellscape” by widespread user complaints. Meta’s response combines enforcement data, revised content definitions, and new user-facing tools designed to give creators and regular users more control over their experience while reducing exposure to scams and impersonators.
Meta’s 2026 Platform Safety Push: Impersonation Tools, Anti-Scam AI, and Content Quality Updates
New Tools for Creators to Report Impersonation on Facebook
Facebook is testing enhanced content protection tools that allow creators to take action when their reels appear on Meta’s platforms after being posted by impersonators. From a single centralized dashboard, creators can identify and flag such content. The upcoming update is designed to consolidate the reporting process so creators can submit all reports in one place rather than navigating multiple flows.
Meta reports that its earlier enforcement work resulted in a 33% drop in impersonation reports targeting large creators, with 20 million accounts removed in 2025 alone.
How the Dashboard Works
The content protection dashboard shows creators a breakdown of matches, protection issues, and profiles to review. It displays the number of matches detected in a given period, the platforms where those matches were found (Facebook and Instagram), and a list of profiles with posts that match the creator’s protected content, each with a “Take action” option.
It is worth noting that the current tool focuses on matching duplicate content — not detecting unauthorized use of a creator’s likeness — which Meta has acknowledged as an area still needing attention.
Updated Definition of “Original Content” on Facebook
Alongside the new reporting tools, Meta is updating Facebook’s content guidelines to more precisely define what counts as original content. The revised definition includes content that is filmed or produced directly by a creator, as well as reels that remix other content or use overlays to add something new — such as analysis, discussion, or additional information.
Content that will be deemed unoriginal and deprioritized includes re-uploads, minor edits to a creator’s own work, and low-value modifications such as adding borders or captions without substantive new content.
Impact of Previous Efforts
| Metric | Result |
|---|---|
| Views and watch time for original content | Nearly doubled in H2 2025 vs. H2 2024 |
| Accounts removed for impersonation/spam | 20 million in 2025 |
| Drop in impersonation reports (large creators) | 33% reduction |
| Scam ads removed globally in 2025 | Over 159 million |
| Ad content banned in India in 2025 | 12.1 million pieces (93%+ proactively) |
| Accounts taken down linked to scam centers | 10.9 million on Facebook and Instagram |
Meta’s New Anti-Scam Tools Across WhatsApp, Facebook, and Messenger
On March 11, 2026, Meta separately announced a set of new anti-scam tools rolling out across its apps, powered by advanced AI designed to detect sophisticated scam patterns at scale.
WhatsApp Device Linking Warning
Scammers often attempt to trick users into linking their WhatsApp account to a scammer’s device — either by sharing a device linking code or by scanning a falsified QR code. WhatsApp will now alert users when behavioral signals suggest a linking request is suspicious, showing where the request is coming from and flagging the potential risk before the user acts on it.
Facebook Alerts for Suspicious Friend Requests
Meta is testing new warnings on Facebook that trigger when users send or receive a friend request from an account showing signs of suspicious activity — such as few mutual friends or a profile location in a different country. Users will see a contextual alert with safety tips and the option to block or reject the request before accepting.
Advanced Scam Detection on Messenger
Meta is expanding AI-powered scam detection on Messenger to more countries. When a chat with a new contact displays patterns associated with common scams — such as suspicious job offers or requests for money transfers — users are warned and offered the option to send recent messages for an AI scam review. If a scam is detected, Messenger provides information on common scam types and recommends blocking or reporting the account.
AI Used to Detect Celebrity Impersonation and Deceptive Links
Meta has built advanced AI systems that analyze multiple signals — including text, images, and surrounding context — to identify a broader range of sophisticated scam patterns. Specific capabilities include detecting fake fan sentiment, misleading account bios, and false associations with public figures or brands. A separate system proactively detects content that redirects users to websites designed to mimic legitimate ones, protecting thousands of brands from domain impersonation.
Expanded Advertiser Verification
Meta is also expanding its advertiser verification program with the goal of ensuring verified advertisers account for 90% of its ad revenue by end of 2026, up from 70% at the time of the announcement. The verification process is focused on the highest-risk advertising categories and is intended to prevent bad actors from misrepresenting their identity to run scam ads.
Enforcement Against Criminal Scam Networks
Beyond platform tools, Meta participated in a major international law enforcement operation that resulted in the disabling of over 150,000 accounts linked to scam center networks in Southeast Asia. These networks were associated with a range of fraudulent activities including impersonating law enforcement officials to stage fake “digital arrests” and promoting fraudulent cryptocurrency investment schemes.
In India specifically, Meta partnered with the Indian Cyber Crime Coordination Centre (I4C) and the Securities Exchange Board of India (SEBI) to launch the third edition of its “Scam se Bacho” (Beware of Scams) awareness campaign, featuring actor Neena Gupta and prominent digital creators.