Live
OpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling SoraOpenAI announces GPT-5 with unprecedented reasoning capabilitiesGoogle DeepMind achieves breakthrough in protein folding for rare diseasesEU passes landmark AI Safety Act with global implicationsAnthropic raises $7B as enterprise demand for Claude surgesMeta open-sources Llama 4 with 1T parameter modelNVIDIA unveils next-gen Blackwell Ultra chips for AI data centersApple integrates on-device AI across entire product lineupSam Altman testifies before Congress on AI regulation frameworkMistral AI reaches $10B valuation after Series C funding roundStability AI launches video generation model rivaling Sora
Policy

TikTok's AI Ad Disclosure Policy Is Failing — Samsung and Others Are Ignoring It

TikTok's mandatory AI content labeling policy for paid advertisements is not being enforced in practice. Major brands including Samsung are running AI-generated ads without required disclosures, leaving viewers with no reliable way to distinguish synthetic content from human-produced creative — exposing a gap between platform policy and platform reality.

D.O.T.S AI Newsroom

D.O.T.S AI Newsroom

AI News Desk

2 min read
TikTok's AI Ad Disclosure Policy Is Failing — Samsung and Others Are Ignoring It

TikTok's policy requiring advertisers to label AI-generated content in paid promotions is not working. Major advertisers — including Samsung — are running AI-generated ads on the platform without applying the required disclosures, according to an investigation by The Verge that reviewed hundreds of paid promotions against the platform's stated content policies.

TikTok introduced mandatory AI content labels for advertisers in 2025, positioning the policy as a transparency measure in response to growing public concern about synthetic media in commercial contexts. Advertisers using AI-generated visuals, voiceovers, or digitally manipulated content were required to apply a disclosure label visible to viewers. The policy applied to all paid promotions using TikTok's advertising platform.

The Compliance Gap

The Verge's analysis found that compliance with the labeling requirement is inconsistently applied and inconsistently enforced. AI-generated product showcase videos, synthetic influencer appearances, and digitally altered footage appeared in paid ad placements without disclosure labels. Samsung ads identified in the analysis used AI-generated product renders and scene compositions without applying the required transparency indicator.

TikTok's enforcement mechanism relies primarily on advertiser self-disclosure at the point of ad submission, with no automated detection system capable of reliably identifying AI-generated creative across the full range of synthetic media techniques in commercial use. Without technical detection or meaningful penalty for non-compliance, the policy functions as a request rather than a requirement.

Why This Matters Beyond TikTok

TikTok's enforcement failure is a preview of a broader industry problem. Every major social and video platform is grappling with the same combination: AI generation tools are producing commercial-quality synthetic media at scale, policy frameworks require disclosure, but automated detection is not mature enough to catch violations, and manual enforcement does not scale to the volume of AI-generated content entering commercial distribution.

The EU AI Act's transparency provisions, which come into force progressively through 2027, include AI-generated content disclosure requirements with meaningful penalties for non-compliance. The TikTok situation suggests that even well-intentioned platform policies will fail without technical infrastructure that makes compliance the path of least resistance rather than an optional check box in ad submission workflows.

TikTok and Samsung had not responded to requests for comment at time of publication.

Back to Home

Related Stories

Musk Updates His OpenAI Lawsuit to Route Any $150 Billion Damages Award to the Nonprofit Foundation
Policy

Musk Updates His OpenAI Lawsuit to Route Any $150 Billion Damages Award to the Nonprofit Foundation

Elon Musk has amended his lawsuit against OpenAI with a strategic addition: any damages recovered — potentially up to $150 billion — should be redirected to OpenAI's nonprofit foundation rather than awarded to Musk personally. The update reframes the litigation from a personal grievance into a structural argument about OpenAI's obligations to its original charitable mission.

D.O.T.S AI Newsroom
OpenAI's Child Safety Blueprint Confronts AI's Role in the Surge of Child Sexual Exploitation
Policy

OpenAI's Child Safety Blueprint Confronts AI's Role in the Surge of Child Sexual Exploitation

OpenAI has released a Child Safety Blueprint outlining its approach to detecting, preventing, and reporting AI-generated child sexual abuse material. The document arrives as law enforcement agencies globally report a sharp increase in CSAM volume, with AI tools enabling the production of synthetic material at scale. It is the company's most detailed public statement on the problem it helped create.

D.O.T.S AI Newsroom
Anthropic's Claude Mythos Found Thousands of Zero-Days — So They're Not Releasing It
Policy

Anthropic's Claude Mythos Found Thousands of Zero-Days — So They're Not Releasing It

Anthropic has quietly restricted its most capable new model, Claude Mythos, after the system autonomously discovered thousands of critical vulnerabilities in major operating systems and browsers — including a 27-year-old OpenBSD bug and a 16-year-old FFmpeg flaw. The model is being deployed exclusively through Project Glasswing with 11 vetted security partners. It is the most concrete case yet of an AI lab withholding a model because of genuinely demonstrated risk.

D.O.T.S AI Newsroom