Back
Politicians Rushed Through an Online Speech 'Solution' — Victims Deserve Better (EFF on TAKE IT DOWN Act)
webThis EFF analysis of the TAKE IT DOWN Act is relevant to AI safety governance as it addresses deepfake/synthetic media regulation, content moderation, and the risks of overbroad AI-generated content laws that could undermine encryption and free expression — key concerns in AI deployment policy.
Metadata
Importance: 42/100opinion piececommentary
Summary
The EFF critiques the TAKE IT DOWN Act, a 2025 U.S. law creating a notice-and-takedown system for non-consensual intimate imagery (NCII) including AI-generated deepfakes. While acknowledging the problem, EFF argues the law's overbroad definitions, 48-hour removal deadlines, lack of bad-faith protections, and threats to end-to-end encryption make it a censorship risk. EFF pledges to monitor overreach and push for platform transparency.
Key Points
- •TAKE IT DOWN Act creates a notice-and-takedown system for NCII/deepfakes but applies to a broader category of content than existing NCII laws, risking removal of lawful speech.
- •The 48-hour removal deadline pressures platforms to over-comply without verifying claims, enabling bad-faith censorship of satire, journalism, and political speech.
- •No legal protections exist for providers who receive frivolous or bad-faith takedown requests, creating a one-way censorship ratchet.
- •The law may force platforms to abandon end-to-end encryption to monitor content, turning private communications into surveilled spaces.
- •EFF and coalition partners sought amendments for safeguards; Congress passed the bill unchanged, with main provisions taking effect in 2026.
Cited by 1 page
| Page | Type | Quality |
|---|---|---|
| TAKE IT DOWN Act | Policy | -- |
Cached Content Preview
HTTP 200Fetched Apr 14, 20264 KB
Earlier this year, both chambers of Congress passed the TAKE IT DOWN Act . This bill, while well-intentioned, gives powerful people a new legal tool to force online platforms to remove lawful speech that they simply don't like.
The bill, sponsored by Senate Commerce Chair Ted Cruz (R-TX) and Rep. Maria Salazar (R-FL), sought to speed up the removal of troubling online content: non-consensual intimate imagery (NCII). The spread of NCII is a serious problem, as is digitally altered NCII, sometimes called “deepfakes.” That’s why 48 states have specific laws criminalizing the distribution of NCII, in addition to the long-existing defamation, harassment, and extortion statutes—all of which can be brought to bear against those who abuse NCII. Congress can and should protect victims of NCII by enforcing and improving these laws.
Unfortunately, TAKE IT DOWN takes another approach: it creates an unneeded notice-and-takedown system that threatens free expression, user privacy, and due process, without meaningfully addressing the problem it seeks to solve.
While Congress was still debating the bill, EFF, along with the Center for Democracy & Technology (CDT), Authors Guild, Demand Progress Action, Fight for the Future, Freedom of the Press Foundation, New America’s Open Technology Institute, Public Knowledge, Restore The Fourth, SIECUS: Sex Ed for Social Change, TechFreedom, and Woodhull Freedom Foundation, sent a letter to the Senate outlining our concerns with the proposal.
First, TAKE IT DOWN’s removal provision applies to a much broader category of content—potentially any images involving intimate or sexual content—than the narrower NCII definitions found elsewhere in the law. We worry that bad-faith actors will use the law’s expansive definition to remove lawful speech that is not NCII and may not even contain sexual content.
Worse, the law contains no protections against frivolous or bad-faith takedown requests. Lawful content—including satire, journalism, and political speech—could be wrongly censored. The law requires that apps and websites remove content within 48 hours or face significant legal risks. That ultra-tight deadline means that small apps or websites will have to comply so quickly to avoid legal risk, that they won’t be able to investigate or verify claims.
Finally, there are no legal protections for providers when they believe a takedown request was sent in bad faith to target lawful speech. TAKE IT DOWN is a one-way censorship ratchet, and its fast timeline discourages providers from standing up for their users’ free speech rights.
This new law could lead to the use of automated filters that tend to flag legal content, from commentary to news reporting. Communications providers that offer users end-to-end encrypted messaging, meanwhile, may be served with notices they simply cannot comply with, given the fact that these providers can’t view the contents of messages on their platforms. Platforms could respond
... (truncated, 4 KB total)Resource ID:
3fdbdf3d04dbb295 | Stable ID: sid_VJirrznd95