editor's blog
Subscribe Now

Sensing the Squish

We’re used to touch being about locating one or more fingers or items on a surface. This is inherently a 2D process. Although much more richness is being explored for the long-term, one third dimension that seems closer in is pressure: how hard are we pushing down, and can we use that to, for instance, grab an object for dragging?

At the 2011 Touch Gesture Motion conference, one company that got a fair bit of attention was Flatfrog, who uses a light-based approach, with LEDs and sensors around the screen to triangulate positions. At the 2012 Touch Gesture Motion conference, when 2D seemed so 2011, pressure was a more frequent topic of conversation. But clearly a visual technology like Flatfrog’s wouldn’t be amenable to measuring pressure since there is nothing to sense the pressure.

Unless…

If you have a squishy object like a finger, then you can use what I’ll call the squish factor to infer pressure. This is what Flatfrog does: when a finger (for example) touches down, they normalize the width of the item, and then they track as that width widens due to the squishing of the finger (or whatever). Which means that this works with materials that squish. Metal? Not so much.

You might wonder how they can resolve such small movements using an array of LEDs that are millimeters apart. For a single LED and an array of sensors, for example, the resolution might indeed be insufficient. But because they have so many LEDs, the combined measurements from all of them allow them to resolve small micro-structures.

There is a cost to this, of course, in processing: it adds about 100 million instructions per second to the processing. “Ouch!” you say? Actually, it’s not that bad: their basic processing budget without pressure is about 2 billion instructions per second, so this is about a 5% adder.

More information at their website

Leave a Reply

featured blogs
Apr 4, 2025
Gravitrams usually employ a chain or screw lift to hoist their balls from the bottom to the top, but why not use a robot?...

featured chalk talk

Using NXP’s FRDM Ecosystem to Break Down ML Complexity
In this episode of Chalk Talk, Michael Pontikes from NXP and Amelia Dalton explore the details of the FRDM ecosystem from NXP. They explore the scalability component of this ecosystem, the details of the FRDM i.MX 93 Development Board and how the machine learning software and tools of this ecosystem will streamline and simplify your next machine learning enhanced design.
Apr 17, 2025
375 views