Apples says its child abuse prevention system will only scan already flagged photos

Tips & Techniques

Apple has faced harsh criticism after announcing that it will scan iCloud and iPhones for photos of child sexual abuse. The company has now responded to it and clarified that it will only scan the photos that have already been flagged internationally.

One of the features allows Apple to scan stored iCloud images in search of known CSAM (Child Sexual Abuse Material) images. If any of them are found, Apple will report it to the National Center for Missing and Exploited Children (NCMEC). According to Reuters, Apple said on Friday that it will “hunt only for pictures that have been flagged by clearinghouses in multiple countries.”

This feature uses an automatic scanning system to go through your photos. After they reach a certain threshold, Apple gets an alert and a human is allowed to review the content on your iCloud. Apple initially refused to say what the threshold is, but has now confirmed that it’s 30 images. The company added that this number would eventually be reduced in the future as the system improves. Also, the “surveillance,” as this feature was dubbed by some, will only be focused on the US for now.

When asked whether criticism influenced the decisions, Apple declined to respond. The company said that “the project was still in development and changes were to be expected.”

[via Engadget]

Products You May Like

Articles You May Like

Opening Multiple Photographs into a Single Photoshop File
ACDSee Photo Studio Ultimate Now Lets You Edit Faces and Skies with AI
This Canon 5D Mark II has shot over 2.2 million photos
What is Purple Fringing in Photography?
Sony’s CFexpress Type A Card is the New Capacity King, But It’ll Cost You

Leave a Reply

Your email address will not be published.