Customize Consent Preferences

We use cookies to help you navigate efficiently and perform certain functions. You will find detailed information about all cookies under each consent category below.

The cookies that are categorized as "Necessary" are stored on your browser as they are essential for enabling the basic functionalities of the site. ... 

Always Active

Necessary cookies are required to enable the basic features of this site, such as providing secure log-in or adjusting your consent preferences. These cookies do not store any personally identifiable data.

No cookies to display.

Functional cookies help perform certain functionalities like sharing the content of the website on social media platforms, collecting feedback, and other third-party features.

No cookies to display.

Analytical cookies are used to understand how visitors interact with the website. These cookies help provide information on metrics such as the number of visitors, bounce rate, traffic source, etc.

No cookies to display.

Performance cookies are used to understand and analyze the key performance indexes of the website which helps in delivering a better user experience for the visitors.

No cookies to display.

Advertisement cookies are used to provide visitors with customized advertisements based on the pages you visited previously and to analyze the effectiveness of the ad campaigns.

No cookies to display.

admin
Pinned August 23, 2021

<> Embed

@  Email

Report

Uploaded by user
Apple announces new iPhone features to detect child sex abuse
<> Embed @  Email Report

Apple announces new iPhone features to detect child sex abuse

Apple reportedly plans to begin scanning iPhones in the US for child abuse images (updated)

The ‘neuralMatch’ system would ‘continuously scan’ US iPhones for illegal images.

Karissa Bell
K. Bell
August 5th, 2021
Apple announces new iPhone features to detect child sex abuse | DeviceDaily.com
Mike Segar / reuters

Apple is reportedly planning an update that would allow it to scan iPhones for images of child sexual abuse. According to the Financial Times, the company has been briefing security researchers on the “neuralMatch” system, which would “continuously scan photos that are stored on a US user’s iPhone and have also been uploaded to its iCloud back-up system.”

The system would “proactively alert a team of human reviewers if it believes illegal imagery is detected” and human reviewers would alert law enforcement if the images were verified. The neuralMatch system, which was trained using a database from the National Center for Missing and Exploited Children, will be limited to iPhones in the United States to start, the report says.

The move would be somewhat of an about face for Apple, which has previously stood up to law enforcement to defend users’ privacy. The company famously clashed with the FBI in 2016 after it refused to unlock an iPhone belonging to the man behind the San Bernardino terror attack. CEO Tim Cook said at the time that the government’s request was “chilling” and would have far-reaching consequences that could effectively create a backdoor for more government surveillance. (The FBI ultimately turned to an outside security firm to unlock the phone.)

Now, security researchers are raising similar concerns. Though there’s broad support for increasing efforts to fight child abuse, researchers who spoke to the FT said that it could open the door for authoritarian regimes to spy on their citizens, since a system designed to detect one type of imagery could be expanded to other types of content, like terrorism or other content perceived as “anti-government.”

At the same time, Apple and other companies have faced mounting pressure to find ways to cooperate with law enforcement. As the report points out, social media platforms and cloud storage providers like iCloud already have systems to detect child sexual abuse imagery, but extending such efforts to images on a device would be a significant shift for the company.

Apple declined to comment to FT, but the company could release more details about its plans “as soon as this week.”

Update 8/5 4pm ET: Apple confirmed plans to start testing a system that would be able to detect images of child sexual abuse stored in iCloud Photos in the United States. “Apple’s method of detecting known CSAM is designed with user privacy in mind. Instead of scanning images in the cloud, the system performs on-device matching using a database of known CSAM image hashes provided by NCMEC and other child safety organizations,” the company wrote in a statement. “Apple further transforms this database into an unreadable set of hashes that is securely stored on users’ devices.”

The update will be rolling out at a later date, along with several other child safety features, including new parental controls that can detect explicit photos in children’s Messages. 

 

Engadget is a web magazine with obsessive daily coverage of everything new in gadgets and consumer electronics 

(39)