Posts Tagged ‘trulyhandsfree’
March 11, 2020
For many years Sensory has been considered the de-facto standard for embedded wake words and voice control. In fact, our low power, small footprint TrulyHandsfree has shipped in billions of devices and apps around the globe. Sensory’s best in class accuracy has enabled us to be the only company offering wake word solutions approved by Amazon, Apple, Google, Microsoft, Samsung and LG, as well as international partners like Alibaba, Baidu and Tencent. With global smart speaker sales breaking records year over year, people are definitely getting more comfortable speaking to technology and this creates new opportunities. Savvy voice-first users now want the next wave of use cases that are more compelling than the typical smart speaker trivia, weather and music. While smart speakers are handy and for introducing the convenience afforded by a voice user interface, they aren’t good for actually getting things done. To really accomplish meaningful tasks, the entire Sensory team is excited to charge into the next phase of the voice user interface. A charge that is led by custom branded wake words and domain specific voice assistants.
Your Brand, Your Wake Word
Most voice industry insiders are familiar with the Sonos/Alexa story. Mom brings home a new Sonos One speaker and sets it up with Alexa for voice control. Every time the kids use it to check the weather, play some music, or hear the news, the conversation starts with “Alexa.” After a few weeks Mom asks, “how do you like that new Sonos speaker?” The kids answer in unison, “what Sonos speaker? Oh, you mean the Alexa speaker!” Enabling Alexa or Google to highjack your brand might seem like a good idea to sell a few more units, but it is not a good long-term strategy for building a brand. Unfortunately, Sonos learned the hard way, but your company doesn’t don’t have to. Sensory’s TrulyHandsfree provides companies with the means to create custom branded wake words that deliver accuracy and performance that is equal to or better than anything offered by the digital giants. Very soon you will hear about several high-profile brands, rolling out their custom wake words. Stay tuned for more on that!
Another business case to consider, what if Amazon and Google don’t want to support your specific use case? For example, a recent article in the Boston Globe explains how Amazon decided not to support LifePod and their quest to create a smart speaker tailored to the elderly. At first glance this may have seemed like a showstopper for LifePod, but actually it enabled them to create their own smart speaker solution, and they turned to Sensory to create the “Hello LifePod” branded wake word.
Using Voice to Get It Done
More than just a custom wake word, Sensory is also supporting clients to create domain specific voice assistants. How does this compare to the typical smart speaker? Consider the smart speaker as a generalist. It can do many, many things, but doesn’t excel at any particular task. A domain specific assistant focuses on one task or a collection of tasks and is specifically designed to accomplish these tasks. To drive the point home, a smart speaker is a mile wide and an inch deep. A domain specific voice assistant is a mile deep and an inch wide. For example, when Midea MCA, a global leader in microwave appliances, wanted a voice-controlled microwave, they didn’t need it to play music or talk about the weather. The Sensory team leveraged our TrulyNatural embedded speech recognition and created a custom language model to support microwave cooking tasks. The end product is a specialist. A voice-controlled microwave oven that knows how to cook potatoes, pop popcorn, reheat coffee, melt butter, soften ice cream, defrost vegetables and much more. All by voice. By being a specialist, this voice enabled microwave provides users with advanced features that are usually hidden behind confusing menus and semi-secret button press sequences.
“We firmly believe in ‘consumer-first technology’ and strive to maintain a technological advantage through teamwork and innovation. By teaming up with Sensory we are able to not only modernize the consumer experience offered by our products but do so in a way that addresses the increasingly alarming privacy issue that concerns many consumers worldwide,” said Dr. Scott Sun, Deputy Director of Midea MCA.
Does Your Brand Have A Voice?
Sensory’s embedded AI enables companies to own and control their voice experiences. Pairing a custom branded wake word with domain specific voice assistant creates a voice user experience that is unique to your brand. Every interaction reinforces the brand-customer bond while also effectively completing the task at hand. Products such as smart appliances, remote controls, wireless headsets, hearables, and wearables can all benefit from this next wave of voice user interface.
If you would like to learn more about Sensory, then please contact us here. We would be happy to discuss how we can best support your company’s custom voice strategy.
January 6, 2020
Combined AI Technologies Can Support Local Commands in Dozens of Languages; Demos Held at CES 2020
IRVINE, Calif., January 6, 2020 – Syntiant, the award-winning AI semiconductor start-up providing intelligent voice solutions at the edge, today announced a partnership with Sensory Inc., AI and speech recognition pioneer, to deliver a deep learning, multi-lingual voice interface for battery-powered devices.
The combined solution merges Sensory’s larger, higher-performance TrulyHandsfree™ wake word engine and voice control with Syntiant’s microwatt-power Neural Decision Processors™ (NDPs), bringing low-latency, real-time inference to edge devices, enabling consumers to seamlessly “voice” commands in dozens of languages. Working together, the two technologies could also support additional features, such as voice-based user identification.
“A busy mom in Korea setting a house alarm or teenager in Barcelona raising the volume on his smart speaker, voice commands are becoming more ubiquitous driven by worldwide consumer demand,” said Kurt Busch, CEO of Syntiant. “Collaborating with Sensory allows us to combine their AI with our silicon technology, providing customers a large multi-language library of local commands for just about any application.”
Syntiant and Sensory are accelerating the delivery of a fast, efficient, cloud-free multi-language interface in devices, such as earbuds, smart speakers and smartphones, at a power level orders of magnitude lower than typical MCU offerings.
Custom built to run neural workloads, the Syntiant® NDP100™ and NDP101™ NDPs can support dozens of local voice commands and consume less than 140 microwatts while performing local processing of audio events, increasing privacy, reliability and responsiveness. In addition to voice triggers, other device capabilities include audio event and environment classification, as well as sensor analytics.
“Sensory’s embedded AI voice software combined with Syntiant’s neural network technology opens up a new market for both companies, bringing ultra-low-power, multi-language, local command support to the edge,” said Todd Mozer, Sensory’s CEO. “Our joint solution provides added security because commands are performed at the edge, free from a cloud connection, while devices and applications come alive without pushing a button.”
Sensory is focused on improving user experiences through embedded machine learning technologies such as voice, vision and natural language processing. The company pioneered the use of neural network approaches for embedded speech recognition for consumer electronics with a well-engineered and patented codebase that has shipped in over 2 billion consumer products.
Syntiant will demo the combined technologies at the Bosch Global booth (LVCC, Central Hall – 12401) during CES 2020 in Las Vegas. Sensory is available by appointment at CES.
Founded in 2017 and headquartered in Irvine, Calif., Syntiant Corp. is moving artificial intelligence and machine learning from the cloud to edge devices. Syntiant’s advanced chip solutions merge deep learning with semiconductor design to produce ultra-low-power, high performance, deep neural network processors for always-on applications in battery-powered devices, ranging from hearing aids to smart speakers and mobile phones. The company is backed by some of the world’s strongest strategic investors, including Intel Capital, Microsoft M12, Bosch Ventures and the Amazon Alexa Fund. More information on the company can be found by visiting www.syntiant.com or by following Syntiant on Twitter @Syntiantcorp.
Sensory Inc. creates a safer and superior UX through vision and voice technologies. Sensory’s technologies are widely deployed in consumer electronics applications including mobile phones, automotive, wearables, toys, IoT and various home electronics. With its TrulyHandsfree™ voice control, Sensory has set the standard for mobile handset platforms’ ultra-low power “always listening” touchless control. To date, Sensory’s technologies have shipped in over a billion units of leading consumer products. For more information about this announcement, Sensory or its technologies, visit https://www.sensory.com/, contact firstname.lastname@example.org or for press inquiries contact email@example.com.
For more information, contact:
Michael A. Farino
New Era Communications Services
June 8, 2017
Since the beginning, Sensory has been a pioneer in advancing AI technologies for consumer electronics. Not only did Sensory implement the first commercially successful speech recognition chip, but we also were first to bring biometrics to low cost chips, and speech recognition to Bluetooth devices. Perhaps what I am most proud of though, more than a decade ago Sensory introduced its TrulyHandsfree technology and showed the world that wakeup words could really work in real devices, getting around the false accept and false reject, and power consumption issues that had plagued the industry. No longer did speech recognition devices require button presses…and it caught on quickly!
Let me go on boasting because I think Sensory has a few more claims to fame… Do you think Apple developed the first “Hey Siri” wake word? Did Google develop the first “OK Google” wake word? What about “Hey Cortana”? I believe Sensory developed these initial wake words, some as demos and some shipped in real products (like the Motorola MotoX smartphone and certain glasses). Even third-party Alexa and Cortana products today are running Sensory technology to wake up the Alexa cloud service.
Sensory’s roots are in neural nets and machine learning. I know everyone does that today, but it was quite out of favor when Sensory used machine learning to create a neural net speech recognition system in the 1990’s and 2000’s. Today everyone and their brother is doing deep learning (yeah that’s tongue in cheek because my brother is doing it too! (http://www.cs.colorado.edu/~mozer/index.php). And a lot of these deep learning companies are huge multi-billion-dollar business or extremely well-funded startups.
So, can Sensory stay ahead now and continuing pioneering innovation in AI now that everyone is using machine learning and doing AI? Of course, the answer is yes!
Sensory is now doing computer vision with convolutional neural nets. We are coming out with deep learning noise models to improve speech recognition performance and accuracy, and are working on small TTS systems using deep learning approaches that help them sound lifelike. And of course, we have efforts in biometrics and natural language that also use deep learning.
We are starting to combine a lot of technologies together to show that embedded systems can be quite powerful. And because we have been around longer and thought through most of these implementations years before others, we have a nice portfolio of over 3 dozen patents covering these embedded AI implementations. Hand in hand with Sensory’s improvements in AI software, companies like ARM, NVidia, Intel, Qualcomm and others are investing and improving upon neural net chips that can perform parallel processing for specialized AI functions, so the world will continue seeing better and better AI offerings on “the edge”.
Curious about the kind of on-device AI we can create when combining a bunch of our technologies together? So were we! That’s why we created this demo that showcases Sensory’s natural language speech recognition, chatbots, text-to-speech, avatar lip-sync and animation technologies. It’s our goal to integrate biometrics and computer vision into this demo in the months ahead:
Let me know what you think of that! If you are a potential customer and we sign an NDA, we would be happy to send you an APK of this demo so you can try it yourself! For more information about this exciting demo, please check out the formal announcement we made: http://www.prnewswire.com/news-releases/sensory-brings-chatbot-and-avatar-technology-to-consumer-devices-and-apps-300470592.html
January 5, 2017
Virtual handsfree assistants that you can talk to and that talk back have rapidly gained popularity. First, they arrived in mobile phones with Motorola’s MotoX that had an ‘always listening’ Moto Voice powered by Sensory’s TrulyHandsfree technology. The approach quickly spread across mobile phones and PCs to include Hey Siri, OK Google, and Hey Cortana.
Then Amazon took things to a whole new level with the Echo using Alexa. A true voice interface emerged, initially for music but quickly expanding domain coverage to include weather, Q&A, recipes, and the most common queries. On top of that, Amazon took a unique approach by enabling 3rd parties to develop “skills” that now number over 6000! These skills allow Amazon’s Echo line (with Tap, Dot) and 3rd Party Alexa equipped products (like Nucleus and Triby) to be used to control various functions, from reading heartrates on Fitbits to ordering Pizzas and controlling lights.
Until recently, handsfree assistants required a certain minimum power capability to really be always on and listening. Additionally, the hearable market segment including fitness headsets, hearing aids, stereo headsets and other Bluetooth devices needed to use touch control because of their power limitations. Also, Amazons Alexa had required WIFI communications so you could sit on your couch talking to your Echo and query Fitbit information, but you couldn’t go out on a run and ask Alexa what your heartrate was.
All this is changing now with Sensory’s VoiceGenie!
The VoiceGenie runs an embedded recognizer in a low power mode. Initially this is on a Qualcomm/CSR Bluetooth chip, but could be expanded to other platforms. Sensory has taken an SBC music decoder and intertwined a speech recognition system, so that the Bluetooth device can recognize speech while music is playing.
The VoiceGenie is on and listening for 2 keywords:
For example, a Bluetooth headset’s volume, pairing, battery strength, or connection status can only be controlled by the device itself, so VoiceGenie handles those controls without touching required. VoiceGenie can also read incoming callers’ names and ask the user if they want to answer or ignore. VoiceGenie can call up the phone’s assistant, like Google Assistant or Siri or Cortana, to ask by voice for a call to be made or a song to be played.
Some of the important facts behind the new VoiceGenie include:
This third point is perhaps the least understood, yet the most important. People want a personalized assistant that knows them, keeps their secrets safe, and helps them in their daily lives. This help can be accessing information or controlling your environment. It’s very difficult to accomplish this for privacy and power reasons in a cloud powered environment. There needs to be embedded intelligence. It needs to be low power. VoiceGenie is that low powered voice assistant.
October 6, 2016
It’s always nice when Sensory wins an award. 2016 has been a special year for Sensory because we won more awards than any other year in our 23 year history!!
Check it out:
Sensory Earns Multiple Coveted Awards in 2016
Sensory Inc., a Silicon Valley company that pioneered the hands-free voice wakeup word approach, today, announced it has won over half a dozen awards in 2016 across its product-line, including awards for products, technologies, and people, covering deep learning, biometric authentication and voice recognition.
The awards presented to Sensory include the following:
CTIA Super Mobility 2016™, the largest wireless event in America, announced more than 70 finalists for its 10th annual CTIA Emerging Technology (E-Tech) Awards. Sensory was nominated in the category of Mobile Security and Privacy for its TrulySecure™ technology, along with Nokia, Samsung, SAP, and others. Sensory was presented with the First Place award for the category in a ceremony on September 2016 at the CTIA Las Vegas event.
Speech Technology magazine, the leading provider of speech technology news and analysis, had its 10th annual Speech Industry Awards to recognize the creativity and notable achievements of key influencers (Luminaries), major innovators (Star Performers), and impressive deployments (Implementation Awards). The editors of Speech Technology magazine selected 2016 award winners based on their industry contributions during the past 12 months. Sensory’s CEO, Todd Mozer, was awarded with a Luminary Award, making it his second time winning the prestigious award. Sensory as a company was awarded the Star Performer award along with IBM, Amazon and others.
Two well-known industry analyst firms issued reports highlighting Sensory’s industry contributions for its TrulyHandsfree product and customer leadership, offering awards for innovations, customer deployment, and strategic leadership.
“Sensory has an incredibly talented team of speech recognition and biometrics experts dedicated to advancing the state-of-the-art of each respective field. We are pleased that our TrulyHandsfree, TrulySecure and TrulyNatural product lines are being recognized in so many categories, across the various industries in which we do business,” said Todd Mozer, CEO of Sensory. “I am also thrilled that Sensory’s research and innovations in the deep learning space has been noticed, generating our company prestigious accolades and management recognition.”
May 6, 2016
Rich Nass and Barbara Quinlan from Open Systems Media visited Sensory on their “IoT Roadshow”.
IoT is a very interesting area. About 10 years ago we saw voice controlled IoT on the way, and we started calling the market SCIDs – Speech Controlled Internet Devices. I like IoT better, it’s certainly a more popular name for the segment! ;-)
I started our meeting off by talking about Sensory’s three products – TrulyHandsfree Voice Control, TrulySecure Authentication, and TrulyNatural large vocabulary embedded speech recognition.
Although TrulyHandsfree is best known for its “always on” capabilities, ideal for listening for key phrases (like OK Google, Hey Cortana, and Alexa), it can be used a ton of other ways. One of them is for hands-free photo taking, so no selfie stick is required. To demonstrate, I put my camera on the table and took pictures of Barbara and Rich. (Normally I might have joined the pictures, but their healthy hair, naturally good looks, and formal attire was too outclassing for my participation).
There’s a lot of hype about IoT and Wearables and I’m a big believer in both. That said, I think Amazon’s Echo is the perfect example of a revolutionary product that showcases the use of speech recognition in the IoT space and am looking forward to some innovative uses of speech in Wearables!
Here’s the article they wrote on their visit to Sensory and an impromptu video showing TrulyNatural performing on-device navigation, as well as a demo of TrulySecure via our AppLock Face/Voice Recognition app.
Rich Nass, Embedded Computing Brand Director
If you’re an IoT device that requires hands-free operation, check out Sensory, just like I did while I was OpenSystems Media’s IoT Roadshow. Sensory’s technology worked flawlessly running through the demo, as you can see in the video. We ran through two different products, one for input and one for security.
August 6, 2015
We first came out with TrulyHandsfree about five years ago. I remember talking to speech tech executives at MobileVoice as well as other industry tradeshows, and when talking about always-on hands-free voice control, everybody said it couldn’t be done. Many had attempted it, but their offerings suffered from too many false fires, or not working in noise, or consuming too much power to be always listening. Seems that everyone thought a button was necessary to be usable!
In fact, I remember the irony of being on an automotive panel, and giving a presentation about how we’ve eliminated the need for a trigger button, while the guy from Microsoft presented on the same panel the importance of where to put the trigger button in the car.
Now, five years later, voice activation is the norm… we see it all over the place with OK Google, Hey Siri, Hey Cortana, Alexa, Hey Jibo, and of course if you’ve been watching Sensory’s demos over the years, Hello BlueGenie!
Sensory pioneered the button free, touch free, always-on voice trigger approach with TrulyHandsfree 1.0 using a unique, patented keyword spotting technology we developed in-house– and from its inception, it was highly robust to noise and it was ultra-low power. Over the years we have ported it to dozens of platforms, Including DSP/MCU IP cores from ARM, Cadence, CEVA, NXP CoolFlux, Synopsys and Verisilicon, as well as for integrated circuits from Audience, Avnera, Cirrus Logic, Conexant, DSPG, Fortemedia, Intel, Invensense, NXP, Qualcomm, QuickLogic, Realtek, STMicroelectronics, TI and Yamaha.
This vast platform compatibility has allowed us to work with numerous OEMs to ship TrulyHandsfree in over a billion products!
Sensory didn’t just innovate a novel keyword spotting approach, we’ve continually improved it by adding features like speaker verification and user defined triggers. Working with partners, we lowered the draw on the battery to less than 1mA, and Sensory introduced hardware and software IP to enable ultra-low-power voice wakeup of TrulyHandsfree. All the while, our accuracy has remained the best in the industry for voice wakeup.
We believe the bigger, more capable companies trying to make voice triggers have been forced to use deep learning speech techniques to try and catch up with Sensory in the accuracy department. They have yet to catch up, but they have grown their products to a very usable accuracy level, through deep learning, but lost much of the advantages of small footprint and low power in the process.
Sensory has been architecting solutions for neural nets in consumer electronics since we opened the doors more than 20 years ago. With TrulyHandsfree 4.0 we are applying deep learning to improve accuracy even further, pushing the technology even more ahead of all other approaches, yet enabling an architecture that has the ability to remain small and ultra-low power. We are enabling new feature extraction approaches, as well as improved training in reverb and echo. The end result is a 60-80% boost in what was already considered industry-leading accuracy.
I can’t wait for TrulyHandsfree 5.0…we have been working on it in parallel with 4.0, and although it’s still a long ways off, I am confident we will make the same massive improvements in speaker verification with 5.0 that we are doing for speech recognition in 4.0! Once again further advancing the state of the art in embedded speech technologies!
June 11, 2015
Guest post by: Michael Farino
Sensory’s CEO, Todd Mozer joined Alan Taylor, host of Popular Science Radio, in a fun discussion about artificial intelligence, Sensory’s involvement with the Jibo robot development team, and also gave the show’s listeners a look into the past 20 years of speech recognition. Todd and Alan additionally discussed some of the latest advancements in speech technology, and Todd provided an update on Sensory’s most recent achievements in the field of speech recognition as well as a brief look into what the future holds.
Listen to the full radio show at the link below:
Big Bang Theory, Science, and Robots | FULL EPISODE | Popular Science Radio #269
May 4, 2015
I was at the Mobile Voice Conference last week and was on a keynote panel with Adam Cheyer (Siri, Viv, etc.) and Phil Gray (Interactions) with Bill Meisel moderating. One of Bills questions was about the best speech products, and of course there was a lot of banter about Siri, Cortana, and Voice Actions (or GoogleNow as it’s often referred to). When it was my turn to chime in I spoke about Amazon’s Echo, and heaped lots of praise on it. I had done a bit of testing on it before the conference but I didn’t own one. I decided to buy one from Ebay since Amazon didn’t seem to ever get around to selling me one. It arrived yesterday.
Here are some miscellaneous thoughts:
OK, Amazon… here’s my free advice (admittedly self-serving but nevertheless accurate):
March 3, 2015
It feels like I had a whole week’s worth of the trade show wrapped into one day! By the time mid week hits, I’ll surely be ready to head home! Here are some of the highlights from the first day of Mobile World Congress 2015: