The features, which include scanning users’ iCloud Photos libraries for Child Sexual Abuse Material (CSAM), will be delayed so that Apple can take additional time to make improvements.
Apple confirmed that the feedback from customers, various non-profit and advocacy groups, security researchers, and others, prompted the delay.
Apple has issued the following statement about its decision:
Last month we announced plans for features intended to help protect children from predators who use communication tools to recruit and exploit them, and limit the spread of Child Sexual Abuse Material. Based on feedback from customers, advocacy groups, researchers and others, we have decided to take additional time over the coming months to collect input and make improvements before releasing these critically important child safety features.
Apple’s Child Safety Features attracted criticism from a wide range of individuals, including security researchers, non-profit organizations, and a vocal number of Apple customers, particularly with regards to the scanning of users’ photos libraries.
More to follow…
This article, “Apple Delays Rollout of Controversial Child Safety Features” first appeared on MacRumors.com
Discuss this article in our forums