Civil rights & AI safety
The statutes, regulations, and survivor-aligned resources our advocacy reports cite. Reducing the barrier to legal recourse for targeted activists.
Aegis never stores the harmful media itself. Cryptographic hashes anchored on Solana create proof of existence that survivors can hand to law enforcement, civil-rights attorneys, or courtrooms — without forcing them to relive the abuse.
Mandates platform takedown of non-consensual synthetic media within 48 hours, with civil penalties for repeat non-compliance.
Pub. L. 119-204 §301–§312
Criminalizes distribution of synthetic intimate imagery without consent. Provides survivors a private right of action.
Cal. Penal Code §647(j)(4)
Section 230 carve-out removing platform immunity for NCII and certain synthetic abuse content.
47 U.S.C. §230(e)(5)
Notice-and-action mechanism requiring platforms to provide accessible, survivor-friendly takedown procedures.
Regulation (EU) 2022/2065
Applies to coordinated targeting of protected classes, supporting hate-crime classification of pattern harassment.
42 U.S.C. §2000a
Criminalizes the sharing or threatening to share synthetic intimate images. Protects survivors regardless of intent.
Online Safety Act 2023