Apples says its child abuse prevention system will only scan already flagged photos

Tips & Techniques

Apple has faced harsh criticism after announcing that it will scan iCloud and iPhones for photos of child sexual abuse. The company has now responded to it and clarified that it will only scan the photos that have already been flagged internationally.

One of the features allows Apple to scan stored iCloud images in search of known CSAM (Child Sexual Abuse Material) images. If any of them are found, Apple will report it to the National Center for Missing and Exploited Children (NCMEC). According to Reuters, Apple said on Friday that it will “hunt only for pictures that have been flagged by clearinghouses in multiple countries.”

This feature uses an automatic scanning system to go through your photos. After they reach a certain threshold, Apple gets an alert and a human is allowed to review the content on your iCloud. Apple initially refused to say what the threshold is, but has now confirmed that it’s 30 images. The company added that this number would eventually be reduced in the future as the system improves. Also, the “surveillance,” as this feature was dubbed by some, will only be focused on the US for now.

When asked whether criticism influenced the decisions, Apple declined to respond. The company said that “the project was still in development and changes were to be expected.”

[via Engadget]

Products You May Like

Articles You May Like

Surf Photographer Questions Beach Drones: ‘They Make a lot of Noise’
The Sword-Billed Hummingbird: A Knight of the Andes
Disney Turns Photographers’ Diverse Princess Portraits into Dolls
Watch Deepfake Tech Flawlessly Change an Actor’s Spoken Language
NFL Photographer Shares Dizzying View from Stadium Catwalk

Leave a Reply

Your email address will not be published. Required fields are marked *