Apple is planning to scan US iPhones for images of child abuse, drawing applause from child protection groups but raising concern among some security researchers that the system could be misused by governments looking to surveil their citizens.
Apple said its messaging app will use on-device machine learning to warn about sensitive content without making private communications readable by the company.
The tool Apple calls “neuralMatch” will detect known images of child sexual abuse without decrypting people’s messages.
If it finds a match, the image will be reviewed by a human who can notify law enforcement if necessary.
But researchers say the tool could be put to other purposes such as government surveillance of dissidents or protesters.
Matthew Green, a security professor at Johns Hopkins University who earlier posted his concerns on Twitter, told the Financial Times that Apple’s move will “break the dam – governments will demand it from everyone”.
Tech companies including Microsoft, Google, Facebook and others have for years been sharing “hash lists” of known images of child sexual abuse.
Apple has also been scanning iCloud, which unlike its messages is not end-to-end encrypted, for such images.
The company has been under pressure from governments and law enforcement to allow for surveillance of encrypted data.
“Apple’s expanded protection for children is a game changer,” John Clark, President & CEO, National Center for Missing & Exploited Children, said in a statement.
“With so many people using Apple products, these new safety measures have lifesaving potential for children who are being enticed online and whose horrific images are being circulated in child sexual abuse material,”
Julia Cordua, the CEO of Thorn, said that Apple’s technology balances “the need for privacy with digital safety for children”.
Thorn, a nonprofit founded by Demi Moore and Ashton Kutcher, uses technology to help protect children from sexual abuse by identifying victims and working with tech platforms.