Meta Expands Anti-Scam Protections Across WhatsApp, Facebook, and Messenger
Meta introduced new anti-scam protections across WhatsApp, Facebook, and Messenger to counter fraud campaigns that rely on social engineering, impersonation, and malicious links. The updates include WhatsApp warnings when device-linking requests show scam-related behavioral signals, such as attempts to trick users into sharing linking codes or QR codes, and Facebook alerts for suspicious friend requests from accounts with indicators like recent creation or no mutual connections. Messenger is also adding AI-driven scam detection to identify patterns associated with impersonation and spoofed links in chats.
The changes are part of a broader anti-fraud push in which Meta said it worked with international law enforcement to disable more than 150,000 scam-linked accounts and support the arrest of 21 individuals. A separate report on a new cross-industry anti-scam accord involving Meta, Google, Microsoft, Amazon, OpenAI, and others describes a wider effort to share threat intelligence, improve fraud reporting, strengthen transaction verification, and coordinate defenses against scam operations that move across multiple online platforms. A report on Operation Atlantic focuses instead on cryptocurrency approval-phishing enforcement by U.S., U.K., and Canadian authorities and is a different story from Meta's platform-specific product rollout.
Related Entities
Organizations
Sources
Related Stories

Meta Expands AI-Driven Anti-Scam Protections Across Facebook, Messenger, and WhatsApp
Meta announced expanded anti-scam measures across **Facebook**, **Messenger**, and **WhatsApp**, emphasizing AI-driven detection of impersonation, deceptive links, and other fraud patterns, alongside new user-facing warnings intended to interrupt scams earlier in the interaction. Updates include Facebook alerts for suspicious friend requests, WhatsApp warnings for potentially fraudulent device-linking attempts (e.g., QR-code based linking), and Messenger prompts that can offer an AI scam review of recent chat content; Meta also said it is expanding advertiser verification to reduce identity misrepresentation in ads. Separately, Meta described enforcement at scale, reporting the removal of **159 million scam ads in 2025** and **10.9 million** Facebook/Instagram accounts tied to criminal scam centers, amid ongoing scrutiny from US lawmakers and reporting that has questioned the platform’s financial incentives to police fraudulent advertising. Meta also highlighted collaboration with law enforcement targeting “industrialized” scam operations, including actions tied to Southeast Asian scam compounds that resulted in **21 arrests** and the disabling of **150,000+ accounts**, as well as broader efforts to counter “pig-butchering”-style investment fraud. Complementing these initiatives, Meta detailed a privacy-preserving *Messenger* capability—**Advanced Browsing Protection (ABP)**—that warns users about potentially malicious websites opened from encrypted chats by using cryptographic/private-information-retrieval techniques to check links against large blocklists without revealing message contents to Meta. In parallel reporting on the scam ecosystem, researchers described a large network of **paid Meta ad** campaigns using fake media brands and impersonated public figures to push investment scams across dozens of countries, underscoring the continued role of malvertising and disinformation-for-profit tactics in driving victim acquisition.
4 days ago
Social Platforms Face Scrutiny Over Scam Advertising and Messaging-Facilitated APP Fraud
**Meta** said it is pursuing legal action against advertisers it alleges ran **celeb-bait scam ads** on its platforms, including defendants based in **Brazil, China, and Vietnam**. Meta reported suspending the advertisers’ payment methods, disabling related accounts, and blocking domains used in the campaigns, which allegedly misused celebrity images/voices (including synthetic media) to drive victims to scam sites that harvest sensitive data or solicit money via fraudulent products and investment schemes; Meta also said it sent cease-and-desist letters to marketing consultants promoting services to bypass ad-policy enforcement (e.g., “un-ban” services and renting access to trusted accounts). Separately, a **Revolut** report on **Authorised Push Payment (APP) fraud** found scam origination increasingly tied to encrypted/direct messaging channels, with **Telegram** exceeding **20%** of reported scam origination and showing significant growth versus 2024, while **Meta-owned platforms** collectively remained the largest source at **44%**. The report highlighted category concentration by platform, including Telegram and WhatsApp accounting for major shares of **investment** and **job scams** (with Telegram cited as over **58%** of job scams), reinforcing that a large portion of financial fraud is initiated on social and messaging platforms and prompting calls for stronger platform accountability.
2 weeks ago
Meta Expands Safety and Enforcement Measures Across Facebook and Instagram
Meta disclosed a set of new **platform safety and enforcement actions** aimed at reducing harm and abuse on its services. The company filed multiple lawsuits against alleged scam-ad operators in **Brazil, China, Vietnam** and elsewhere, describing tactics including **deepfakes/celebrity impersonation**, “celeb-bait” investment lures, and **cloaking** used to evade ad review; Meta said it also took technical steps such as disabling accounts, suspending scam-linked payment methods, and blocking associated domains, and shared information with industry partners to help them block the same actors. Separately, Meta announced new **Instagram parental-supervision alerts** that notify parents when a teen repeatedly searches for **self-harm or suicide-related terms** within a short time window (initially for supervised accounts in the **U.S., U.K., Australia, and Canada**), and said it is developing similar notifications for teens’ **AI-related conversations** about self-harm. In parallel regulatory developments, EU lawmakers advanced a non-binding opinion supporting **privacy-friendly age verification** and proposing restrictions that would require **parental consent for under-16s** and bar access for children under 13, positioning these measures for potential inclusion in a future **Digital Fairness Act** focused on child protection online, targeted advertising, and addictive design patterns.
2 weeks ago