Smart speakers have long been linked to lasers. A team based in Tokyo and at the University of Michigan said they could take over , Google Assistant and Apple's Siri by hitting the devices' microphones with beams of light.and , but researchers discovered an unexpected vulnerability -- their susceptibility to
The team's paper, published Monday and previously reported by The New York Times, revealed that researchers tricked a Google Home into opening a garage door from 230 to 350 feet away, by focusing lasers with a telephoto lens.
"Microphones convert sound into electrical signals. The main discovery behind light commands is that in addition to sound, microphones also react to light aimed directly at them," the researchers wrote.
The lasers basically trick the microphones into making electrical signals as if they're hearing someone's voice. A hacker could seemingly use this method to buy stuff online, control smart home switches and remotely unlock and start a car that's linked to the speaker.
In addition to the smart speakers, the researchers tested this method on the, and . They had to be much closer to take control of these devices.
The researchers have let Amazon, Apple, Google, Ford and Tesla know about this potential security issue, the Times noted.
"We are closely reviewing this research paper. Protecting our users is paramount, and we're always looking at ways to improve the security of our devices," a Google spokesperson said in an emailed statement.
Amazon echoed this sentiment, noting that it's engaging with researchers to better understand their work.
"Customer trust is our top priority and we take customer security and the security of our products seriously," a spokesperson for the e-commerce giant said in an email.
Apple didn't offer comment, while Ford and Tesla didn't respond to requests for comment.
First published Nov. 5, 3:59 a.m. PT.
Update, 8:37 a.m. PT: Adds Amazon comment.