This technology supports Amazon and the numerous supporters of the Voice Interoperability Initiative (VII) announced in September and allows brands to get a jumpstart on the development of products that feature multiple simultaneous voice services.
“We are lucky to have worked closely with most of the leaders in voice services to help create products with their wake words embedded. Because of this, we already have all of the data in place to combine multiple models,” explained Todd Mozer, CEO of Sensory. “However, everyone in the industry has found that combining two or more wake word models into a product poses a significant challenge in maintaining optimal performance. The new capabilities we added to TrulyHandsfree overcomes some of the accuracy challenges without significantly boosting MIPS and memory requirements.”
Sensory has already collected the data and built deep learned models for “Hey Cortana”, “Alexa”, “OK Google”, “Hey Siri”, “Hi Bixby”, “Xiao-du Xiao-du” and over 2 dozen other assistants. These wake words have been approved for market and are already shipping in real products with proven performance and reliability in ultra-low power applications like wearables and mobile phones, to high-performance applications like vehicles, smart speakers and home appliances.
TrulyHandsfree employs a new technique to enable multiple wake words without linearly increasing the error rates, which has been the main challenge associated with running multiple wake words on one device. With traditional approaches, adding a new wake word meant the sum of the error rates of each individual wake word, but with TrulyHandsfree the accuracy doesn’t degrade as fast, offering support for more wake words with less false accepts and false rejects.
“Also, on our roadmap for a 2020 release is multi-wake word adaptation that will allow devices to know who is talking to them and to adapt the voice model to better work with each individual user,” Mozer added.
Sensory – www.sensory.com