Which major AI companies have policies for reporting CSAM disclosures by users to law enforcement?

Checked on December 11, 2025
Disclaimer: Factually can make mistakes. Please verify important information or breaking news. Learn more.

This fact-check may be outdated. Consider refreshing it to get the most current information.

Executive summary

Several major AI companies publicly commit to detecting and reporting child sexual abuse material (CSAM) to authorities or intermediaries such as the National Center for Missing & Exploited Children (NCMEC). Stability AI explicitly states it reports CSAM to NCMEC via CyberTipline; Amazon (AWS/Bedrock) reports matched CSAM hashes and blocks instances as violations of its terms and integrates detection into its services [1] [2]. Industry-wide pledges and coalition templates also document a trend toward reporting and referral practices [3] [4].

1. Who says they will report — and how they describe it

Public statements and transparency reports from AI platform operators show concrete promises: Stability AI’s 2025 integrity report says the company “report[s] any Child Sexual Exploitation Material (CSAM) to the National Center for Missing and Exploited Children (NCMEC) via their CyberTipline” and that detected CSAM through APIs is promptly referred to NCMEC, which then forwards matters to law enforcement [1]. Amazon’s EU DSA transparency material explains that Amazon Bedrock runs automated CSAM detection, blocks instances, and notes matches “with hashes matched against NCMEC’s verified CSAM database,” indicating operational reporting and blocking processes [2]. Forbes and industry pledges from 2024 state that several tech firms agreed to remove confirmed CSAM from training data and report confirmed CSAM to relevant authorities [3].

2. Industry coordination and templates for reporting

Beyond individual company statements, multi-stakeholder groups are creating reporting norms. The Tech Coalition, with input from NCMEC, developed a reporting template for members to use when referring AI-related child exploitation reports to NCMEC — an explicit signal that reporting to NCMEC is the expected channel for U.S.-focused referrals [4]. Legal and policy advisories likewise suggest companies should train moderation systems on NCMEC or other CSAM hash databases to “quickly identify, remove, and report CSAM” [5].

3. Legal and regulatory pressure shaping reporting practices

Regulatory developments and national laws are pushing companies toward formal reporting and preservation practices. U.S. legislative and state-level actions — including laws addressing AI-generated CSAM — and international frameworks like the EU’s discussions about CSAM detection and the DSA’s obligations create incentives for platforms to detect and report illegal material [6] [7]. At the same time, EU Council deliberations and interim regimes have altered which classes of services are subject to scanning or reporting duties, creating legal complexity for global providers [8] [9].

4. Detection methods and limits companies acknowledge

Companies and analysts highlight reliance on hash-based matching and automated classifiers; Amazon and Stability AI reference hash matching with NCMEC’s database and automated classifiers for blocking and reporting [2] [1]. Legal advisories recommend training moderation systems on known-hash databases and embedding reporting obligations in vendor contracts [5]. However, sources also stress technical limits: AI can generate novel synthetic CSAM that hash-matching alone won’t catch, and policymakers and researchers urge red-teaming and other mitigations to identify novel AI-generated abuse [10] [11].

5. Competing perspectives: voluntary pledges versus mandated duties

Industry pledges (reported by Forbes) show private commitments to report confirmed CSAM [3] while regulatory proposals vary: some EU-level changes removed scanning mandates for certain providers, shifting responsibility and creating debate about whether self-regulation suffices [8]. Legal commentators argue that narrowly scoped safe-harbor rules and DOJ comfort letters could enable more effective testing and reporting practices, indicating that voluntary reporting may be insufficient without clearer legal protections and obligations [10].

6. What the available sources do not fully answer

Available sources do not provide a comprehensive, up-to-date list of every “major AI company” and their precise reporting pipelines to law enforcement beyond the examples above; they do not enumerate, for instance, policies from Google, Microsoft, OpenAI, Meta, or other large model providers in these excerpts. For those companies, available sources do not mention their specific CSAM-reporting commitments in the documents supplied here (not found in current reporting).

7. Bottom line for policymakers and the public

Major AI service operators communicate operational reporting to NCMEC or equivalent authorities and use hash-matching and automated classifiers to detect and block CSAM [1] [2]. Industry coalitions and legal advisors promote standardized reporting templates and integration with NCMEC [4] [5]. But evolving laws, limits of hash-based detection for synthetic or novel AI-generated CSAM, and gaps in which services are legally compelled to scan or report create real-world friction between effective enforcement and privacy/encryption concerns [8] [10]. Reported company commitments are a baseline; oversight, clearer legal obligations, and improved technical detection will determine whether those commitments translate into consistent referrals to law enforcement.

Want to dive deeper?
Which AI companies publicly disclose their procedure for handling user CSAM reports?
How do major AI firms balance user privacy and mandatory CSAM reporting obligations?
What laws require AI companies to report CSAM to law enforcement in the US and EU?
Which AI platforms use automated detection versus human review for CSAM disclosures?
Have any AI companies faced legal action or scrutiny over CSAM reporting practices?