Siri ‘regularly’ records sex encounters, sends ‘countless’ private moments to Apple contractors
Apple’s Siri AI assistant sends audio of sexual encounters, embarrassing medical information, drug deals, and other private moments recorded without users’ knowledge to human ‘graders’ for evaluation, a whistleblower has revealed.
Recordings from Apple’s Siri voice assistant are fed to human contractors around the world, who grade the AI based on the quality of its response and whether its activation was deliberate, according to an anonymous contractor who spoke to the Guardian. They claimed accidental activations are much more frequent than Apple lets on, especially with Apple Watch users – and wants the company to own up to the problem.
“There have been countless instances of recordings featuring private discussions between doctors and patients, business deals, seemingly criminal dealings, sexual encounters and so on. These recordings are accompanied by user data showing location, contact details, and app data,” the whistleblower revealed.
“Apple is subcontracting out, there’s a high turnover. It’s not like people are being encouraged to have consideration for people’s privacy, or even consider it,” the whistleblower said, explaining that they are concerned these recordings, produced when Siri thinks it hears its “wake word,” could be used against the people who (accidentally) made them – especially given the “broad” amount of user data they claim contractors are “free to look through.” In what sounds like a sick joke on the part of some programmer, the sound of a zipper unzipping often triggers Siri to wake up.
https://www.rt.com/news/465181-apple-siri-human-contractors/