Apple is globally suspending the use of Siri voice recordings after reports emerged that contractors hired by the company listen to ‘private clips’ to improve the voice assistant.
TechCrunch, which broke the news, said Apple plans to issue a software update to seek users’ consent before letting them participate in the product improvement programme.
“We are committed to delivering a great Siri experience while protecting user privacy,” the company said to the outlet. “While we conduct a thorough review, we are suspending Siri grading globally. Additionally, as part of a future software update, users will have the ability to choose to participate in grading.”
In a similar development, Google said it will voluntarily halt listening-in and transcribing Google Assistant recordings for three months in the EU, according to German data regulators.
“The use of automatic speech assistants from providers such as Google, Apple and Amazon is proving to be highly risky for the privacy of those affected,” the German data protection authority said. “This applies not only to people who run a speech assistant, but to all those who come into contact with it, for example if they live in a household in which devices such as Google Assistant are installed are used.”
Last week, The Guardian reported how Siri voice snippets containing “medical information, drug deals, and recordings of couples having sex” are being heard by contractors working for the company around the world.
The objective, like with the case of Amazon and Google, is to listen to them and grade the assistant’s response, including to check if it was invoked by mistake and if the response was appropriate. The snippets themselves are not associated with your Apple ID.
Although Apple said less than one percent of daily requests are being analysed to improve Siri and dictation, the contractor said, “they were motivated to go public about their job because of their fears that such information could be misused.”
Apple’s own terms of service document says pseudonymized Siri requests could be used for quality control. But it stops short of explicitly stating that the work is actually ‘undertaken by humans’.
The iPhone maker is far from the only company to employ human oversight of its voice assistants. Audio recording requests made to Amazon’s Alexa and Google Assistant are also reviewed in a similar fashion.
Google, mid-July, acknowledged voice snippets from the Assistant — leaked to the Belgian news outlet VRT News — revealed sensitive information such as medical conditions and customer addresses.
It’s an industry-wide practice to collect voice recordings to improve speech recognition algorithms, as it’s very much an evolving technology.
But seeking informed consent from users goes a long way towards addressing privacy concerns, not to mention ensuring compliance with EU GDPR requirements.