WASHINGTON — Apple will add a series of new child-safety features to its next big operating system updates for iPhones and iPads.
As a part of iOS 15 and iPadOS 15 updates later this year, the company will implement a feature to detect photos stored in iCloud Photos that depict sexually explicit activities involving children.
“This will enable Apple to report these instances to the National Center for Missing and Exploited Children,” the company said in a notice on its website.
“NCMEC acts as a reporting center for child sexual abuse material and works in collaboration with law enforcement agencies across the US.”
Its method of detecting known child sexual abuse material is “designed with user privacy in mind,” as per Apple.
The company said it is not directly accessing customers’ photos but instead is using a device-local, hash-based matching system to detect child abuse images.
The hashing technology, called NeuralHash, analyzes an image and converts it to a unique number specific to that image, as per the notice.
Only another image that appears nearly identical can produce the same number; for example, images that differ in size or trans-coded quality will still have the same NeuralHash value.
Using another technology called threshold secret sharing, the system ensures that Apple cannot interpret the contents of the safety vouchers unless the iCloud Photos account crosses a threshold of available child sexual abuse content, the notice read.
The cryptographic technology allows Apple to interpret the contents of the safety vouchers associated with the matching child sexual abuse material images only when the threshold is exceeded.
Apple also said it can’t actually see user photos or the results of such scans unless there’s a hit.
“Suppose there’s a match between a user’s photos and the child sexual abuse material database. In that case, Apple then manually reviews each report to confirm sexually explicit images of children, then disable the user’s account and send a report to the National Center for Missing and Exploited Children,” said the tech giant.
If a user feels their account has been mistakenly flagged, as per Apple, “they can file an appeal to have, their account reinstated.”
The system provides a high level of accuracy that ensures less than a one-in-1-trillion chance per year of incorrectly flagging a given account, as per the tech giant.
In addition, with Apple’s iOS 15 update, the iPhone’s Messages app will add new tools to warn children and their parents if they are receiving or sending sexually explicit photos.
“When receiving this type of content, the photo will be blurred, and the child will be warned, presented with helpful resources, and reassured it is okay if they do not want to view this photo,” Apple said.
“As an additional precaution, the child can also be told that to make sure they are safe, and their parents will get a message if they do view it. Similar protections are available if a child attempts to send sexually explicit photos.”
Apple’s iOS 15 also will provide updates to Siri and Search to “provide parents and children expanded information and help if they encounter unsafe situations.”
Siri and Search will intervene when users try to search for child sexual abuse material, displaying prompts that will “explain to users that interest in this topic is harmful and problematic, and provide resources from partners to get help with this issue.”
The iOS 15 update is slated to be available in the fall of 2021, available for iPhone 6s and later models.
(With inputs from ANI)
Edited by Amrita Das and Pallavi Mehra
The post Apple Will Scan IPhones, IPads For Images Of Child Sex Abuse appeared first on Zenger News.