TikTok is reportedly being investigated by the U.S. government for not moderating child sexual abuse material (CSAM) on the social media platform and being a hunting ground for predators.
All our coverage of the Windows 11 2022 update which is rolling out worldwide
windows 11 2022 update
Windows 11 Beta Channel includes fix for Dual SIM calling, and more
windows 11 insider preview
Child sexual abuse material RSS
Reddit user u/AsuharietYgvar is confident that they have uncovered Apple's NeuralHash algorithm, which will combat child sexual abuse, deep in iOS' source code. A GitHub repo contains their findings.
Apple has addressed privacy concerns regarding its sex abuse scanning by clarifying that the new feature would only flag accounts with at least 30 iCloud photos matching Child Sexual Abuse Material.
Apple has provided more details about its child safety photo scanning technologies that have been drawing some fire from critics. It has also described the end-to-end flow of its review process.
An open letter demanding that Apple halt the rollout of its photo scanning tech and issue a statement to reaffirm its commitment to privacy now has signatures from over 5,000 individuals and firms.
In an internal memo, Apple's Software VP has acknowledged that people are worried about the company scanning iCloud Photos for child sex abuse material, but says that this is due to misunderstandings.