Live
OpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling SoraOpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling Sora
Policy

Perplexity AI Faces Lawsuit Over Alleged Data Sharing With Meta and Google

A lawsuit filed against Perplexity AI alleges the company shared user data — including search queries and behavioral signals — with Meta and Google without sufficient disclosure. The case targets one of AI search's fastest-growing platforms at a moment when the sector's data practices are under increasing legal and regulatory scrutiny.

D.O.T.S AI Newsroom

D.O.T.S AI Newsroom

AI News Desk

3 min read
Perplexity AI Faces Lawsuit Over Alleged Data Sharing With Meta and Google

Perplexity AI is facing a lawsuit that alleges the company shared user data with Meta and Google without adequate disclosure to its users. The complaint, reported by The Decoder, centers on the data practices underlying Perplexity's AI search product — specifically the claim that user search queries, behavioral signals, and engagement data flowed to two of the world's largest advertising platforms without users' meaningful informed consent.

What the Lawsuit Alleges

The core allegation is that Perplexity integrated tracking and analytics infrastructure from Meta and Google into its platform in ways that transmitted user data to those companies. If accurate, this means that users who turned to Perplexity as an AI-native alternative to Google Search may have had their query behavior logged and transmitted to the very platform they were seeking an alternative to.

The legal theory mirrors cases that have succeeded against other consumer technology companies — most notably the wave of litigation targeting Facebook's tracking pixels, which allowed Facebook to receive data about user behavior on third-party sites without those users' knowledge. Applied to an AI search product, the same mechanism raises additional concerns: search queries to an AI assistant are often more revealing than traditional search terms, as users increasingly ask AI systems personal health questions, financial queries, and other high-sensitivity requests that they might phrase more guardedly on a conventional search engine.

Perplexity's Position

Perplexity has not yet issued a formal public response to the specific allegations. The company has previously emphasized its commitment to privacy as a differentiator from traditional search, pointing to its lack of advertising as evidence that it doesn't have the same incentives to monetize user data that Google does. The lawsuit complicates that positioning, regardless of its ultimate legal outcome.

The Broader Sector Context

The case arrives at a consequential moment for AI search. Perplexity has been one of the fastest-growing consumer AI products of the past two years, positioning itself as a more trustworthy and cited alternative to Google for research queries. It has also been the target of criticism from publishers who allege the platform reproduces copyrighted content without adequate attribution or licensing — a set of disputes that remains unresolved.

The data-sharing lawsuit adds a third dimension of legal exposure. Together, the copyright and privacy questions raise a structural issue about whether AI search products can sustain their growth trajectories while the legal frameworks governing their core practices are still being litigated. The outcome of the Perplexity case will be watched closely by every company building AI-native search and discovery products.

Back to Home

Related Stories

Musk Updates His OpenAI Lawsuit to Route Any $150 Billion Damages Award to the Nonprofit Foundation
Policy

Musk Updates His OpenAI Lawsuit to Route Any $150 Billion Damages Award to the Nonprofit Foundation

Elon Musk has amended his lawsuit against OpenAI with a strategic addition: any damages recovered — potentially up to $150 billion — should be redirected to OpenAI's nonprofit foundation rather than awarded to Musk personally. The update reframes the litigation from a personal grievance into a structural argument about OpenAI's obligations to its original charitable mission.

D.O.T.S AI Newsroom
OpenAI's Child Safety Blueprint Confronts AI's Role in the Surge of Child Sexual Exploitation
Policy

OpenAI's Child Safety Blueprint Confronts AI's Role in the Surge of Child Sexual Exploitation

OpenAI has released a Child Safety Blueprint outlining its approach to detecting, preventing, and reporting AI-generated child sexual abuse material. The document arrives as law enforcement agencies globally report a sharp increase in CSAM volume, with AI tools enabling the production of synthetic material at scale. It is the company's most detailed public statement on the problem it helped create.

D.O.T.S AI Newsroom
Anthropic's Claude Mythos Found Thousands of Zero-Days — So They're Not Releasing It
Policy

Anthropic's Claude Mythos Found Thousands of Zero-Days — So They're Not Releasing It

Anthropic has quietly restricted its most capable new model, Claude Mythos, after the system autonomously discovered thousands of critical vulnerabilities in major operating systems and browsers — including a 27-year-old OpenBSD bug and a 16-year-old FFmpeg flaw. The model is being deployed exclusively through Project Glasswing with 11 vetted security partners. It is the most concrete case yet of an AI lab withholding a model because of genuinely demonstrated risk.

D.O.T.S AI Newsroom