Reddit user u/AsuharietYgvar is confident that they have uncovered Apple's NeuralHash algorithm, which will combat child sexual abuse, deep in iOS' source code. A GitHub repo contains their findings.
Child sexual abuse material RSS
Apple has addressed privacy concerns regarding its sex abuse scanning by clarifying that the new feature would only flag accounts with at least 30 iCloud photos matching Child Sexual Abuse Material.
Apple has provided more details about its child safety photo scanning technologies that have been drawing some fire from critics. It has also described the end-to-end flow of its review process.
An open letter demanding that Apple halt the rollout of its photo scanning tech and issue a statement to reaffirm its commitment to privacy now has signatures from over 5,000 individuals and firms.
In an internal memo, Apple's Software VP has acknowledged that people are worried about the company scanning iCloud Photos for child sex abuse material, but says that this is due to misunderstandings.