editor's blog
Subscribe Now

Gesture Progress

At the recent Interactive Technology Summit (erstwhile Touch Gesture Motion), gesture was featured more on the day I was checking out the TSensors summit. But I did get a chance to talk both to PointGrab and eyeSight to see what has transpired over the last year.

These two companies both aim at similar spaces, gunning for supremacy in laptops, phones, and other household electronics (HVAC, white goods, etc.). Part of the game right now is design wins, and frankly, their design win reports sound very similar. So there seems to be plenty of business to go around – even to the point that it seems that in some cases, a given company is using them both. I don’t know if that’s to check them both out over time or to make them both happy or to use them as negotiation fodder against each other. To hear them tell it, business is good for everyone.

Development continues apace as well. One key change that’s happened in the last year is a move away from using gestures simply to control a mouse. Using the mouse model, for example, if you want to shut off your Windows laptop, then you gesture the mouse to go down to the Start button and do the required clicks to shut down the machine*. The new model is simply to have a “shut down” gesture – the mouse is irrelevant.

PointGrab has already released this; eyeSight has it in the wings.

I discussed the issue of universal gestures with PointGrab. There is an ongoing challenge of developing gestures that are intuitive across cultures (there aren’t many – some say one, some say two…). PointGrab doesn’t actually see this as a big issue; there’s room for everyone to acquire a simple, well-thought out gesture “lexicon” even if it means acquiring some new gestures that weren’t already used in that culture. Their bigger worry is that different companies will use different lexicons, rather than everyone settling on one set of gestures.

PointGrab has also announced what they call Hybrid Action Recognition. This is a way of making gesture recognition smarter, and it consists of three elements (not to be confused with three sequential steps):

  • Watching for movement that suggests that a gesture is coming
  • Looking for specific shapes, like a finger in front of the face
  • Disambiguating look-alike objects

This almost feels to me a bit like yet another form of context awareness: these three tasks establish a context that says, “Hey, this is a gesture; that last thing wasn’t.” At present, this is a static system; in the future, they will be able to make it learn in real time.

 

Meanwhile, eyeSight noted that, in the future, you may have several devices in a given room that are gesture-enabled. Perhaps a laptop, a TV, and a thermostat. If you gesture, which one are you talking to? Well, as humans, our primary indicator is by looking at the person we’re talking to. EyeSight is looking at providing this capability as well: a device would react to a gesture only if you’re looking at it.

They’re also looking farther down the road at more holistic approaches, including gaze, face recognition, and even speech. (As humans, we can talk to someone we’re not looking at, but we use speech to alert them that they’re who we’re talking to.) But this is a ways out…

As an aside, it was noted in a presentation that gaze in particular is good for broad-level use, but doesn’t work well for fine tracking since our eyes actually flit around at high speeds (saccadic movement) – activity that our brain smooths out so that we don’t notice it. A computer could tell that we’re looking at the computer easily enough, but it would have to do a similar smoothing thing in order to be able to identify, for example, which word we’re reading on the screen.

This whole gesture space seems to be moving extraordinarily quickly; there has been significant change in only one year. This is but one reason that it’s all done in software instead of hardware; updates can be anything but minor. The other reason, of course, is that this capability is going onto mainstream consumer devices. Requiring specific hardware would introduce a much higher barrier to inclusion.

This tension between hardware and software is actually going to be playing out in related spaces, but that’s a topic for another time.

 

 

*Unless, heaven help you, you’re on the original Windows 8, in which case you’ll gesture to move the mouse all over the place in a vain attempt to find where to shut things down; then you’ll give up and gesture to bring up your favorite browser to search for “How the #@$(&* do I shut down my @(#$&(# Windows 8 machine???” and find that you go in to Settings (???????) and a few more mouse clicks (really??) done by gestures and Bingo! In only 15 minutes, you’ve managed to shut it off, with only a 50 point rise in your blood pressure! I think that, by this whole Windows 8 fiasco, Microsoft is earning itself its own specific gesture. One that I won’t repeat here, this being a family newspaper and all.

Leave a Reply

featured blogs
Sep 22, 2021
3753 Cruithne is a Q-type, Aten asteroid in orbit around the Sun in 1:1 orbital resonance with the Earth, thereby making it a co-orbital object....
Sep 21, 2021
Placing component leads accurately as per the datasheet is an important task while creating a package footprint symbol. As the pin pitch goes down, the size and location of the component lead play a... [[ Click on the title to access the full blog on the Cadence Community si...
Sep 21, 2021
Learn how our high-performance FPGA prototyping tools enable RTL debug for chip validation teams, eliminating simulation/emulation during hardware debugging. The post High Debug Productivity Is the FPGA Prototyping Game Changer: Part 1 appeared first on From Silicon To Softw...
Aug 5, 2021
Megh Computing's Video Analytics Solution (VAS) portfolio implements a flexible and scalable video analytics pipeline consisting of the following elements: Video Ingestion Video Transformation Object Detection and Inference Video Analytics Visualization   Because Megh's ...

featured video

ARC® Processor Virtual Summit 2021

Sponsored by Synopsys

Designing an embedded SoC? Attend the ARC Processor Virtual Summit on Sept 21-22 to get in-depth information from industry leaders on the latest ARC processor IP and related hardware and software technologies that enable you to achieve differentiation in your chip or system design.

Click to read more

featured paper

Designing an Accurate, Multifunction Lithium-Ion Battery-Testing Solution

Sponsored by Texas Instruments

This paper highlights the benefits of a discrete solution over an integrated solution in order to meet current and future battery testing challenges. It also includes an example of a highly flexible battery testing design.

Click to read more

featured chalk talk

Traveo II Microcontrollers for Automotive Solutions

Sponsored by Mouser Electronics and Infineon

Today’s automotive designs are more complicated than ever, with a slew of safety requirements, internal memory considerations, and complicated power issues to consider. In this episode of Chalk Talk, Amelia Dalton chats with Marcelo Williams Silva from Infineon about the Traveo™ II Microcontrollers that deal with all of these automotive-related challenges with ease. Amelia and Marcelo take a closer look at how the power efficiency, smart IO signal paths, and over the air firmware updates included with this new MCU family will make all the time-saving difference in your next automotive design.

Click here for more information about Cypress Semiconductor Traveo™ II 32-bit Arm Automotive MCUs