Skip to main content
Mallory
Mallory

Social Platforms Face Scrutiny Over Scam Advertising and Messaging-Facilitated APP Fraud

app fraudscam advertisinginvestment scamsfraudulent productsjob scamsdirect messagingaccount takedownsencrypted messagingad-policy evasionauthorised push paymentwhatsapptelegramaccount rentingpayment method suspensioncease-and-desist
Updated February 27, 2026 at 10:07 AM2 sources
Social Platforms Face Scrutiny Over Scam Advertising and Messaging-Facilitated APP Fraud

Get Ahead of Threats Like This

Know if you're exposed — before adversaries strike.

Meta said it is pursuing legal action against advertisers it alleges ran celeb-bait scam ads on its platforms, including defendants based in Brazil, China, and Vietnam. Meta reported suspending the advertisers’ payment methods, disabling related accounts, and blocking domains used in the campaigns, which allegedly misused celebrity images/voices (including synthetic media) to drive victims to scam sites that harvest sensitive data or solicit money via fraudulent products and investment schemes; Meta also said it sent cease-and-desist letters to marketing consultants promoting services to bypass ad-policy enforcement (e.g., “un-ban” services and renting access to trusted accounts).

Separately, a Revolut report on Authorised Push Payment (APP) fraud found scam origination increasingly tied to encrypted/direct messaging channels, with Telegram exceeding 20% of reported scam origination and showing significant growth versus 2024, while Meta-owned platforms collectively remained the largest source at 44%. The report highlighted category concentration by platform, including Telegram and WhatsApp accounting for major shares of investment and job scams (with Telegram cited as over 58% of job scams), reinforcing that a large portion of financial fraud is initiated on social and messaging platforms and prompting calls for stronger platform accountability.

Related Entities

Related Stories

Meta Expands AI-Driven Anti-Scam Protections Across Facebook, Messenger, and WhatsApp

Meta Expands AI-Driven Anti-Scam Protections Across Facebook, Messenger, and WhatsApp

Meta announced expanded anti-scam measures across **Facebook**, **Messenger**, and **WhatsApp**, emphasizing AI-driven detection of impersonation, deceptive links, and other fraud patterns, alongside new user-facing warnings intended to interrupt scams earlier in the interaction. Updates include Facebook alerts for suspicious friend requests, WhatsApp warnings for potentially fraudulent device-linking attempts (e.g., QR-code based linking), and Messenger prompts that can offer an AI scam review of recent chat content; Meta also said it is expanding advertiser verification to reduce identity misrepresentation in ads. Separately, Meta described enforcement at scale, reporting the removal of **159 million scam ads in 2025** and **10.9 million** Facebook/Instagram accounts tied to criminal scam centers, amid ongoing scrutiny from US lawmakers and reporting that has questioned the platform’s financial incentives to police fraudulent advertising. Meta also highlighted collaboration with law enforcement targeting “industrialized” scam operations, including actions tied to Southeast Asian scam compounds that resulted in **21 arrests** and the disabling of **150,000+ accounts**, as well as broader efforts to counter “pig-butchering”-style investment fraud. Complementing these initiatives, Meta detailed a privacy-preserving *Messenger* capability—**Advanced Browsing Protection (ABP)**—that warns users about potentially malicious websites opened from encrypted chats by using cryptographic/private-information-retrieval techniques to check links against large blocklists without revealing message contents to Meta. In parallel reporting on the scam ecosystem, researchers described a large network of **paid Meta ad** campaigns using fake media brands and impersonated public figures to push investment scams across dozens of countries, underscoring the continued role of malvertising and disinformation-for-profit tactics in driving victim acquisition.

4 days ago

Meta's Profits from Scam and High-Risk Advertisements Exposed

Reuters uncovered internal Meta documents revealing that the company projected $16 billion in 2024 revenue from advertisements linked to scams and banned goods, accounting for approximately 10% of its total revenue. Meta's safety staff estimated that its platforms were involved in a third of all successful scams in the United States, though some of this involvement may be due to the use of WhatsApp for communication rather than direct ad placement. The documents also showed that Meta only bans advertisers if automated systems are 95% certain of fraud; otherwise, the company imposes higher ad rates as a penalty, potentially incentivizing the acceptance of high-risk ads. Meta's management reportedly weighed the financial benefits of scam ads against potential regulatory costs, with $3.5 billion in revenue every six months coming from ads deemed to have "higher legal risk," such as those impersonating brands or celebrities. The company was willing to forgo only a small fraction of its revenue—about $135 million—to clamp down on suspicious advertisers, suggesting a calculated approach to balancing profit and compliance risk. These revelations have raised significant concerns about Meta's role in facilitating online scams and its internal decision-making regarding fraudulent advertising.

4 months ago
Meta Expands Safety and Enforcement Measures Across Facebook and Instagram

Meta Expands Safety and Enforcement Measures Across Facebook and Instagram

Meta disclosed a set of new **platform safety and enforcement actions** aimed at reducing harm and abuse on its services. The company filed multiple lawsuits against alleged scam-ad operators in **Brazil, China, Vietnam** and elsewhere, describing tactics including **deepfakes/celebrity impersonation**, “celeb-bait” investment lures, and **cloaking** used to evade ad review; Meta said it also took technical steps such as disabling accounts, suspending scam-linked payment methods, and blocking associated domains, and shared information with industry partners to help them block the same actors. Separately, Meta announced new **Instagram parental-supervision alerts** that notify parents when a teen repeatedly searches for **self-harm or suicide-related terms** within a short time window (initially for supervised accounts in the **U.S., U.K., Australia, and Canada**), and said it is developing similar notifications for teens’ **AI-related conversations** about self-harm. In parallel regulatory developments, EU lawmakers advanced a non-binding opinion supporting **privacy-friendly age verification** and proposing restrictions that would require **parental consent for under-16s** and bar access for children under 13, positioning these measures for potential inclusion in a future **Digital Fairness Act** focused on child protection online, targeted advertising, and addictive design patterns.

2 weeks ago

Get Ahead of Threats Like This

Mallory continuously monitors global threat intelligence and correlates it with your attack surface. Know if you're exposed — before adversaries strike.