- cross-posted to:
- [email protected]
- [email protected]
- cross-posted to:
- [email protected]
- [email protected]
cross-posted from: https://lemmy.world/post/20664372
X (formerly Twitter) claims that non-consensual nudity is not tolerated on its platform. But a recent study shows that X is more likely to quickly remove this harmful content—sometimes known as revenge porn or non-consensual intimate imagery (NCII)—if victims flag content through a Digital Millennium Copyright Act (DMCA) takedown rather than using X’s mechanism for reporting NCII.
The likelihood of being sued is the issue. The DMCA has tons of case law behind it making it a powerful tool for the media companies. Not so much for NCII. Also, individuals usually don’t have legal teams.
So as with much fuckery in the modern world, we can use the rules as well as they can. If someone’s image is shared as NCII, they only need flag it as a DMCA violation. Problem being, who is going to know that trick?