How to Stop Alexa Gossiping About You

Should we be worried about this invasion of privacy?

Tuesday 5th June 2018

One of Amazon’s voice-activated devices has been caught allegedly emailing a private conversation to a random contact in the user’s address book.

It seems improbable, but Alexa -- the software inside the Echo -- heard the wrong command four times in a row.

According to the user, named Danielle, it woke and recorded a conversation between her and her husband -- mishearing the command. It then asked who to send it to, detected a person’s name, then matched the sound with a person in the user’s contact list.

Then, it misheard the user’s confirmation to send them a recording of their conversation, which was -- fortunately -- fairly mundane. This sequence of events appears to have been confirmed by Amazon in a statement.

The family had Echo devices all over the house as part of their smart home system, but now say they’ve unplugged them. The chances of this happening randomly are undoubtedly low, but should we be worried about this invasion of privacy?

Smart Home or Surveillance?

Devices like Amazon’s Echo, Apple’s Siri, and Google’s Assistant (which runs on Google Home devices) are all voice-activated. They constantly listen out for their ‘wake word’, and then respond to the words after it. But they do make mistakes.

Even if you don’t ask the device to record you, it will record whenever it hears its wake word. The recording is then sent to the cloud for processing, training the company’s algorithm, or improving its recognition of your voice in the future.

If you have any voice-activated devices in your home, you’ve probably activated them by accident, or heard them spring to life in response to something on the TV. There’s also a chance that any device could be hacked, which is rarer but theoretically riskier.

These companies are not out to get you, but the chance that mistakes can happen should make you wary.

How to Delete Voice Recordings

For those of us who already use voice-activated devices, there is a way to clean up old recordings to ensure that they aren’t going to be hacked in the future.

If you use a Google device, you’ll find your records in the Voice & Audio section of your Activity History. Amazon provides similar controls in its Alexa Privacy section. There are instructions for Cortana and Siri users here.

But that just deletes your history. If you’re very privacy-conscious, you probably shouldn’t have any voice-activated devices in your home. At the very least, it’s a good idea not to place these kinds of devices that listen in places where you might have sensitive conversations -- like your home office, and definitely not in your bedroom.


If you have a voice-activated assistant on your phone, it may be switched on by default. You can disable it in the device settings so that it doesn’t pick up your voice accidentally when it’s in your pocket. If you aren’t using the feature, it’s better to be safe than sorry.

Award winning antivirus protection from TotalAV. Stay 100% safe from malware and online threats.

Security & Privacy

Intel Chip Bug Could Enable Mass Data Theft by Hackers
A newly uncovered flaw within Intel hardware enables cyber criminals to intercept vitally sensitive data inside a user’s computer
17 April 2019

Security & Privacy

The GUI who Tracked Me
How is your cell being tracked? And by whom?
04 April 2019

Security & Privacy

Apple iOS 12.2 Update Fixes 50+ Vital Security Flaws
With the release of Apple iOS 12.2, the tech juggernaut has released a major update which includes patches for over 50 critical security bugs
04 April 2019