How to Stop Alexa Gossiping About You
Should we be worried about this invasion of privacy?
Tuesday 5th June 2018
One of Amazon’s voice-activated devices has been caught allegedly emailing a private conversation to a random contact in the user’s address book.
It seems improbable, but Alexa -- the software inside the Echo -- heard the wrong command four times in a row.
According to the user, named Danielle, it woke and recorded a conversation between her and her husband -- mishearing the command. It then asked who to send it to, detected a person’s name, then matched the sound with a person in the user’s contact list.
Then, it misheard the user’s confirmation to send them a recording of their conversation, which was -- fortunately -- fairly mundane. This sequence of events appears to have been confirmed by Amazon in a statement.
The family had Echo devices all over the house as part of their smart home system, but now say they’ve unplugged them. The chances of this happening randomly are undoubtedly low, but should we be worried about this invasion of privacy?
Smart Home or Surveillance?
Devices like Amazon’s Echo, Apple’s Siri, and Google’s Assistant (which runs on Google Home devices) are all voice-activated. They constantly listen out for their ‘wake word’, and then respond to the words after it. But they do make mistakes.
Even if you don’t ask the device to record you, it will record whenever it hears its wake word. The recording is then sent to the cloud for processing, training the company’s algorithm, or improving its recognition of your voice in the future.
If you have any voice-activated devices in your home, you’ve probably activated them by accident, or heard them spring to life in response to something on the TV. There’s also a chance that any device could be hacked, which is rarer but theoretically riskier.
These companies are not out to get you, but the chance that mistakes can happen should make you wary.
How to Delete Voice Recordings
For those of us who already use voice-activated devices, there is a way to clean up old recordings to ensure that they aren’t going to be hacked in the future.
If you use a Google device, you’ll find your records in the Voice & Audio section of your Activity History. Amazon provides similar controls in its Alexa Privacy section. There are instructions for Cortana and Siri users here.
But that just deletes your history. If you’re very privacy-conscious, you probably shouldn’t have any voice-activated devices in your home. At the very least, it’s a good idea not to place these kinds of devices that listen in places where you might have sensitive conversations -- like your home office, and definitely not in your bedroom.
If you have a voice-activated assistant on your phone, it may be switched on by default. You can disable it in the device settings so that it doesn’t pick up your voice accidentally when it’s in your pocket. If you aren’t using the feature, it’s better to be safe than sorry.
Award winning antivirus protection from TotalAV. Stay 100% safe from malware and online threats.