Source |
Network World |
Identifiant |
302096 |
Date de publication |
2017-01-30 04:00:00 (vue: 2017-01-30 04:00:00) |
Titre |
IDG Contributor Network: Hackers could use hidden mal-audio to attack Google Now |
Texte |
There's a fabulous story about a slew of Amazon Echo devices that took it upon themselves to order expensive doll houses from the ecommerce retailer all because a news show host uttered the phrase “Alexa ordered me a dollhouse†on air. The machines heard it from the TV switched on in the room.Researchers say it's not an unlikely scenario. They say not only can attackers issue mal-audio voice commands to any AI listening device that is in audible range, but they can also do it using hidden voice commands. Those are commands that might not even be noticed by the user.To read this article in full or to leave a comment, please click here |
Envoyé |
Oui |
Condensat |
about air all also amazon any are article attack attackers audible audio because but can click commands comment contributor could device devices doll dollhouse†echo ecommerce even expensive fabulous from full google hackers heard here hidden host houses idg issue leave listening machines mal might network: news not noticed now only order phrase please range read researchers retailer room say scenario show slew story switched themselves there those took unlikely upon use user using uttered voice “alexaâ ordered |
Tags |
|
Stories |
|
Notes |
|
Move |
|