A new lawsuit accuses Apple of enabling the storage of child sexual abuse material (CSAM) on iCloud, claiming the company abandoned critical detection tools.
Child sexual abuse material RSS
The EU's "Chat Control" proposal to scan encrypted messages for CSAM is back, with new revisions and growing support.
Turkey has officially banned access to the messaging platform Discord due to concerns over illegal content sharing.
The White House announced that leading AI developers, including OpenAI and Microsoft, have committed to reducing the generation of sexual abuse images through responsible data sourcing and more.
TikTok is reportedly being investigated by the U.S. government for not moderating child sexual abuse material (CSAM) on the social media platform and being a hunting ground for predators.
Reddit user u/AsuharietYgvar is confident that they have uncovered Apple's NeuralHash algorithm, which will combat child sexual abuse, deep in iOS' source code. A GitHub repo contains their findings.
Apple has addressed privacy concerns regarding its sex abuse scanning by clarifying that the new feature would only flag accounts with at least 30 iCloud photos matching Child Sexual Abuse Material.
Apple has provided more details about its child safety photo scanning technologies that have been drawing some fire from critics. It has also described the end-to-end flow of its review process.
An open letter demanding that Apple halt the rollout of its photo scanning tech and issue a statement to reaffirm its commitment to privacy now has signatures from over 5,000 individuals and firms.
In an internal memo, Apple's Software VP has acknowledged that people are worried about the company scanning iCloud Photos for child sex abuse material, but says that this is due to misunderstandings.