editor's blog
Subscribe Now

Locus of (Con)Fusion

At the MEPTEC MEMS conference a couple weeks ago, one sensor fusion question kept coming up over and over: Who’s in charge of sensor fusion?

On the one hand, IMU makers in general are giving away sensor fusion packages that help integrate the data from the individual sensors in their combo units. Then there are guys like Movea that don’t make sensors themselves, but integrate across a wide variety of sensors for both high- and low-level motion artifacts (motion in their case, but the concept extends to anything).

So who’s job is it?

I happened to have a conversation with Movea’s Dave Rothenberg that same day, and I brought the topic up.

His first comment was that what most IMU makers refer to as sensor fusion is simply the software required to establish orientation, which is a relatively low-level characteristic. He said that this correlated to Movea’s Foundation series, which they’ve actually de-emphasized a bit since it is hard to sell against free software, even if they do think they do a better job.

The sensor guys say they’re the right place to do it because they know their sensors better than anyone else. That actually covers two separate things: the physical characteristics of the sensors and how they operate, and the low-level data details – formats etc. Dave mentioned that it is work for them to adapt their software to different sensors, since they don’t all look or speak alike. (Area for future possible standardization? Future topic…) But they have to get it right in order for the other pieces that lay over it to work properly: errors at the bottom level will compound as further algorithms manipulate them.

(This also ties into the question of loose vs tight coupling, since a sensor maker is in a better position to do things tightly.)

Of course, it’s unlikely that the sensor vendors will want to take on the higher-level algorithms since those, almost by definition, will, at some point, involve sensors that they don’t make. So it looks like things may go the way of the embedded world, where critical low-level drivers and other bits of firmware are provided by (or in close partnership with) the processor maker, with other companies layering higher-value stuff on top. That seems to be how the sensor world is shaping up, which leaves room both for the sensor guys and for the third-party folks.

Leave a Reply

featured blogs
Oct 3, 2024
Someone with too much time on his hands managed to get Linux to boot on an Intel 4004 in only 4.76 days...

featured paper

A game-changer for IP designers: design-stage verification

Sponsored by Siemens Digital Industries Software

In this new technical paper, you’ll gain valuable insights into how, by moving physical verification earlier in the IP design flow, you can locate and correct design errors sooner, reducing costs and getting complex designs to market faster. Dive into the challenges of hard, soft and custom IP creation, and learn how to run targeted, real-time or on-demand physical verification with precision, earlier in the layout process.

Read more

featured chalk talk

Versatile S32G3 Processors for Automotive and Beyond
In this episode of Chalk Talk, Amelia Dalton and Brian Carlson from NXP investigate NXP’s S32G3 vehicle network processors that combine ASIL D safety, hardware security, high-performance real-time and application processing and network acceleration. They explore how these processors support many vehicle needs simultaneously, the specific benefits they bring to autonomous drive and ADAS applications, and how you can get started developing with these processors today.
Jul 24, 2024
66,728 views