editor's blog
Subscribe Now

Sensory Listens to Your Voice

Today we’ve put up a piece on designing audio subsystems, but there’s more news than that in the audio world. If you read our earlier piece on QuickLogic’s EOS device, and if you were paying attention to details, you might recall a quick mention of a company called Sensory that had partnered with QuickLogic for audio algorithms. Sensory subsequently released a product called TrulyHandsFree, and I connected with them to find out more about who they are and what they do.

In fact, they’ve been at this game for 21 years, so they’re not newcomers. They even sold (and still sell) neural-network-based chips with their algorithms, but their current focus is the algorithms themselves, sold as IP. In fact, they have both software and hardware IP (the latter of which featured on the QuickLogic part).

One of their important applications is biometric authentication: using voice as a security mechanism. It’s mostly for verification – given examples of authorized personnel, confirming by your voice that you’re who you say you are. They can also do some limited identification – that is, listening to your voice and coming up with who you are without your giving them any hints as to who you are. If they have, like, 10 people or so to choose from, they can do this. If they have to identify someone amongst thousands, though, they’re not there. (Yet, anyway.)

They’ve got three levels of product:

  • TrulyHandsFree: this is for low-end consumer products, requiring the least resources to get the job done. Low power, small footprint, always on. Small vocabulary, used for command and control. This is what was incorporated into the QuickLogic part.
  • TrulyNatural: This includes state-of-the-art algorithms for higher-end consumer devices like phones. Can handle a large vocabulary and continuous speech.
  • TrulySecure: this combines audio with video for authentication.

In general, authentication happens through a passphrase (ignoring the video in the last product). It can be a fixed passphrase, but that runs the risk that someone records the authorized person saying the passphrase and then replays it to fool the authentication. It’s better if the system issues random passphrases for the supplicant to utter. Then no one knows ahead of time exactly what will be required to pass.

Of course, with anything like this, you have to deal with false accepts (unauthorized person gets through) and false rejects (authorized person can’t get through). They actually have a dial that lets them set these rates, and the best balance will depend on the application, weighing the risk of unauthorized entry to the inconvenience (or worse) of not being able to get into your own system. There are no testing standards for this. They always assume that the user has done a reasonable training job, and they then look across a variety of noise and environmental conditions that might affect how the sound is perceived by the algorithms.

Of course, with small devices, the challenge is power, since you need this system always to be on. They say that, on average, TrulyHandsFree uses about 1 mA of current. Sound detection requires less than 1 MIPS and runs a couple hundred microamps or less. Once triggered, the recognition part runs 1.5 – 2.5 mA. Processing is staged, with each level ramping up as the prior level directs.

Speech_Graphic_6-3-15_600dpi_red.jpg

(Image courtesy Sensory)

They do as much processing locally as possible – for example, having a wearable work with a phone to do this if there’s not enough oomph in the wearable. That keeps things working even when there’s no connection, and it’s better for privacy. They can escalate to the cloud for more horsepower if necessary, which works particularly well if the thing being requested requires cloud access anyway.

Their latest announcement has them adding deep learning capabilities to their TrulyHandsFree product. They say that this increases their word accuracy by up to 80% while shrinking the size of their acoustic models by a factor of 10. This also lowers their power consumption to the levels discussed above. You can read more in their announcement.

Leave a Reply

featured blogs
Oct 22, 2020
WARNING: If you read this blog and visit the featured site, Max'€™s Cool Beans will accept no responsibility for the countless hours you may fritter away....
Oct 22, 2020
Cadence ® Spectre ® AMS Designer is a high-performance mixed-signal simulation system. The ability to use multiple engines and drive from a variety of platforms enables you to "rev... [[ Click on the title to access the full blog on the Cadence Community site....
Oct 20, 2020
In 2020, mobile traffic has skyrocketed everywhere as our planet battles a pandemic. Samtec.com saw nearly double the mobile traffic in the first two quarters than it normally sees. While these levels have dropped off from their peaks in the spring, they have not returned to ...
Oct 16, 2020
[From the last episode: We put together many of the ideas we'€™ve been describing to show the basics of how in-memory compute works.] I'€™m going to take a sec for some commentary before we continue with the last few steps of in-memory compute. The whole point of this web...

featured video

Demo: Inuitive NU4000 SoC with ARC EV Processor Running SLAM and CNN

Sponsored by Synopsys

Autonomous vehicles, robotics, augmented and virtual reality all require simultaneous localization and mapping (SLAM) to build a map of the surroundings. Combining SLAM with a neural network engine adds intelligence, allowing the system to identify objects and make decisions. In this demo, Synopsys ARC EV processor’s vision engine (VPU) accelerates KudanSLAM algorithms by up to 40% while running object detection on its CNN engine.

Click here for more information about DesignWare ARC EV Processors for Embedded Vision

featured paper

An engineer’s guide to autonomous and collaborative industrial robots

Sponsored by Texas Instruments

As robots are becoming more commonplace in factories, it is important that they become more intelligent, autonomous, safer and efficient. All of this is enabled with precise motor control, advanced sensing technologies and processing at the edge, all with robust real-time communication. In our e-book, an engineer’s guide to industrial robots, we take an in-depth look at the key technologies used in various robotic applications.

Click here to download the e-book

Featured Chalk Talk

Mom, I Have a Digital Twin? Now You Tell Me?

Sponsored by Cadence Design Systems

Today, one engineer’s “system” is another engineer’s “component.” The complexity of system-level design has skyrocketed with the new wave of intelligent systems. In this world, optimizing electronic system designs requires digital twins, shifting left, virtual platforms, and emulation to sort everything out. In this episode of Chalk Talk, Amelia Dalton chats with Frank Schirrmeister of Cadence Design Systems about system-level optimization.

Click here for more information