Apple Faces Federal Lawsuit Alleging It Knowingly Stored Child Sexual Abuse Material on iCloud Servers
Summary
Apple Inc. is facing a class-action lawsuit accusing it of knowingly storing child sexual abuse material (CSAM) on its iCloud servers and profiting from user storage fees. The lawsuit centers on Apple’s development and subsequent abandonment of a CSAM detection tool, NeuralHash, in 2022, following backlash from privacy advocates who feared potential government overreach. Plaintiffs argue that by developing and testing NeuralHash, Apple gained actual knowledge of CSAM on its servers and violated federal law by failing to report it to the National Center for Missing & Exploited Children (NCMEC).
Apple’s reporting rate of suspected CSAM is significantly lower than competitors like Google and Meta, leading to criticism that the company prioritizes privacy over child safety. Apple maintains its commitment to end-to-end encryption prevents effective scanning, but plaintiffs contend this claim is invalid given the company’s prior development of a detection system. The case arrives amidst growing global pressure on tech companies to combat online child exploitation, with debates surrounding regulations like the EU’s “Chat Control” and the UK’s Online Safety Act.
The lawsuit seeks monetary damages, restitution of iCloud storage fees, and a court order requiring Apple to implement CSAM detection measures. Legal experts anticipate Apple will raise defenses based on Section 230 of the Communications Decency Act and argue its decision to halt NeuralHash was a legitimate business choice. The outcome could significantly impact the tech industry, potentially influencing companies’ willingness to develop safety tools and clarifying the legal obligations of platforms regarding illegal content.
(Source:Webpronews)