- Apple’s child-safety features have been “misunderstood,” an exec told The Wall Street Journal.
- Earlier this month, the company announced two features that would scan iPhone and iCloud images.
- “I think in no way is this a backdoor,” SVP Craig Federighi told the Journal.
- See more stories on Insider’s business page.
Apple’s Craig Federighi, senior vice president of software engineering, said the company’s plan to scan iPhone users’ photos for child sexual abuse material has been “widely misunderstood.”
“We wish that this had come out a little more clearly for everyone, because we feel very positive and strongly about what we’re doing, and we can see that it’s been widely misunderstood,” Federighi said in a video interview with The Wall Street Journal‘s Joanna Stern published on Friday.
Apple earlier this month announced a feature that would create digital hashes of images as they’re uploaded from iPhones to iCloud accounts. Those hashes would be compared to databases of known child sexual abuse material held by anti-abuse organizations, Apple said.
“Before an image is stored in iCloud Photos, an on-device matching process is performed for that image against the known CSAM hashes,” the summary said.
Critics of the plan said it was a misstep for a company that’s long made privacy a selling point. The Electronic Frontier Foundation (EFF) last Thursday called the hashing-and-matching update a privacy “backdoor,” which could be expanded or exploited. Some Apple employees were worried the feature could be exploited by governments, Reuters reported.
“I think in no way is this a backdoor,” Federighi said in the Journal interview. “I don’t understand – I really don’t understand that characterization.”
The CSAM feature was one of two photo-scanning updates the company announced. The other would scan children’s incoming iMessage photos for nudity, alerting parents when children under 12 years old viewed pornographic content.
“I do believe the sound-bite that got out early was, ‘Oh my god, Apple is scanning my phone for images.’ This is not what is happening,” Federighi told the Journal via video. “This is about images stored in the cloud.”
Eva Galperin, EFF’s director of cybersecurity, said via Twitter that the rollout had not been misunderstood by privacy experts.
“I’d like to take this moment to make it clear to poor Craig that no, I don’t misunderstand Apple’s plans to check photos in iCloud against NCMEC’s database of CSAM,” Galperin said.
“It’s well-meaning but it’s also creating a mechanism that Apple will be forced to use for other things,” she added.