On May 19, 2025, the federal government enacted a sweeping new law aimed at curbing the spread of nonconsensual intimate images (NCII)—including those generated by artificial intelligence. Dubbed the “Take It Down Act”, this legislation introduces criminal penalties and rapid takedown requirements for platforms hosting such content.
If you operate a digital platform that allows users to upload images, videos, or other media—even in the form of comments or community posts—this law could have significant implications for your business.
The new law makes it a federal crime to knowingly post, share, or distribute nonconsensual explicit images or AI-generated deepfakes of individuals without their permission. The penalties include up to three years in prison and monetary fines.
But the reach of the law doesn’t end with individuals. Social media platforms, content hosting websites, and apps that enable user-generated content (UGC) are now obligated to:
This places a significant onus on platforms to proactively manage and moderate sensitive content—even when it originates from users.
If your platform allows for user-submitted content—photos, videos, audio clips, or even comments—you are now part of the enforcement chain. That includes:
Startups in particular often move fast and build without fully developed content moderation systems. But under the Take It Down Act, failure to act quickly on NCII reports could expose your company to legal risk—and serious reputational damage.
This is especially true for platforms dealing with AI-generated deepfakes, where content may not appear obviously explicit or malicious at first glance. The law doesn’t carve out exceptions for machine-generated media. In fact, it calls it out directly.
The Take It Down Act is a clear signal from lawmakers: platforms must take responsibility for how their users share sensitive or explicit content.
Startups and tech companies need to treat this law as a catalyst to revisit content moderation strategies, terms of service, and legal protections. With the rise of AI-generated media, the line between satire and serious harm is getting harder to detect. Now’s the time to take proactive steps.
Contact The Social Media Law Firm today to get support crafting policies, content procedures, and legal safeguards that scale with your brand.
For more legal tips, give us a follow on Instagram, TikTok, Linkedin, or check out our YouTube Channel.
Subscribe to The Social Media Lawcast on Spotify Podcasts.