feature article
Subscribe Now

Automotive Vision

When you head out onto the road, your car is managed by a sophisticated parallel processor: your brain. That computing engine is able to do an amazing number of things all at the same time, so much so that we’re still not sure how it works.

And many things have to be done in parallel when you drive a car. Obviously, you have to be able to actuate the various controls – steering and working clutch and accelerator together. You need to observe all manner of threats around you – that pothole, the soccer ball that’s rolling into the street, those pedestrians waiting to cross (you do stop for them, right?).

In fact, the number of things we really should be looking for at all times is quite enormous. Each trip is an independent event, and the probability of something going wrong on any given trip is not changed by the fact that nothing went wrong yesterday. But our cranial parallel processor seems to have this risk diminution algorithm, particularly for routes we travel often. The more nothing happens, the less we pay attention.

And what do we do with that attention? OK, let’s not get into the phone debate – for some reason, we have created the scapephone in order to give ourselves permission to do all kinds of other distracting things. Listening to music… listening to the news… here’s a really bad one: listening to This American Life or books on tape. All of these things are specifically intended to put you cognitively somewhere –anywhere – besides on this boring road that you’re tired of seeing twice a day.

But take away all of that distraction, and what happens? Your parallel processor gets bored and starts to entertain itself with your own internal songs or political musings or that old standby – sexual fantasy. And there you are again, off somewhere else when you’re supposed to be driving. Our brains are self-distracting in the absence of other distractions. You hear advocates righteously declare that you should be 100% focused on driving all the time – yet we know that that’s not really possible; it just makes for good political grandstanding.

That doesn’t change the fact, however, that we’re barreling down the road in a big ol’ hunk o’ steel (or composite), and we can do a lot of damage to ourselves and others with it. True 100%-conscious driving would involve constantly checking our speed, noting every speed limit sign (you do see them all, right?), checking around on all sides of the car many times per minute to make sure that no idiot is lurking in your blind spot, estimating the distance to the car in front of you, monitoring how wet the road is, timing when to dim your brights due to an oncoming car, and, well, the list goes on.

It’s actually pretty remarkable how good a job we do, but our failings are frequent and obvious, whether in the dents in the cars around us, the newspaper headlines, or those red sections of the road as seen on an online map with a traffic view. That’s very often because, despite all the things we should be doing in parallel while driving, we really do very few of them. It’s just that the risk is low enough that we get away with it most of the time.

But here’s where things are changing, and changing fast. (At least, in automotive terms.) All these tasks that we need to do in parallel – which we don’t really do religiously – can increasingly be done by machines. In fact, machines can probably do a much better job than we can, but from a legal and cultural standpoint, the issue of control remains a sticking point. We’re used to controlling – and being responsible for –the operation of our vehicles, and it may be tough to change that.

So technology is chipping away at this by degrees. Ford had a big display at CES, and I had a conversation with them illuminating some of the things they’re working on to increase safety – some of which are already in place, even in modest cars like the Focus and the Fiesta.

Just as happens when we do a conscientious job driving, driver assistance technologies (officially called Advanced Driver Assistance Systems, or ADAS) operate by observing artifacts and then, when needed, taking some action. We use our eyes and ears (and occasionally touch, when that rattle can be felt in the steering wheel) to tell if something’s wrong, and we use our hands and feet to take action.

Machines are inferior to the extent that any single machine can’t do nearly as much as our bodies can, but, with multiple machines, we can make a combined unit that does a far better job.

For example, when looking forward, we have a number of different machine options: visual cameras, infrared, radar, and LIDAR are examples. And each has strengths and weaknesses. Visual cameras obviously suffer in low-light conditions; radar can mistake large metallic stationary objects for other cars; and LIDAR is best for close-in detection, meaning we need something else to get an earlier warning. Each can be used for the job to which it’s best suited, however, and the inputs can be fused to draw overall conclusions about safety threats and for making other operational decisions.

Then there’s the question of what to do with the information. For example, a car with auto-braking can apply the brakes itself when it sees the need. As implemented today, this happens as a hard brake; in the future, it will become more subtle, with intermediate force being used until the end, when a hard stop can be forced if necessary.

By contrast, in a car whose driver remains more in control, the driver may be informed of a looming problem, with the brakes being automatically applied only if the car decides that the driver isn’t braking fast enough. And if you’re wondering how it knows how hard you’re braking, it’s not just about calculating closing speed – there’s a sensor gauging how fast you are pushing down on the brake pedal. If you’re being too gentle, you may get some help.

Based on some of the embedded vision technology stories we’ve heard, this type of visually-guided control would appear to be a work in progress. But it turns out that some of these vision-processing algorithms are surprisingly well-established – so much so that they’ve been cast into dedicated silicon since 2007, provided by a company called Mobileye.

Mobileye was founded specifically for the purposes of addressing automotive safety. They worked with ST Microelectronics to build a custom ASIC; they call it the EyeQ chip. The idea was to put in place enough horsepower to run certain specific algorithms in parallel, relying only on data from a single monocular camera – no fancy 3D stuff. The first version had four dedicated accelerators that they call “vision computation engines,” or VCEs, working along with two ARM946E cores. The specific tasks of this system included:

  • Forward collision warning
  • Automatic high-beam control
  • Lane departure
  • Detecting the closing distance to the car in front of you

Alternatively, you could use the engines for pedestrian detection.

While this chip was largely a proof of concept, it was used in various BMW, Cadillac, and Volvo models.

The EyeQ2 was a second version; they claim that it has six times the processing power of the original EyeQ. This has five VCEs, 3 “vector microcode processors” (VMPs), and two MIPS34K processors. This bulked up the number of things that could be done in parallel. (Note the switch from ARM to MIPS, a decision that was based on old-school benchmarking.) This chip has seen more commercial uptake, shipping in the millions of units.

And a third generation, the EyeQ3, is in development now. It adds another VMP and two more MIPS cores, along with faster clocks and memory and all of the support circuitry, resulting in another six-fold increase in processing power. This version will ship in 2014.

The main point here is that, while many vision algorithms are still finding their way via soft implementation in order to facilitate frequent changes and updates, Mobileye is on their third generation of dedicated hard silicon. (Obviously parts of it are highly programmable, but key algorithms have been hardened.) Note that they didn’t go down the custom route just to be stubborn; they couldn’t (and claim they still can’t) find off-the-shelf technology (even silicon IP) that would do what they needed with the performance they wanted. At such time as they find something that works for them, they say they’re happy to stop doing a custom solution.

So while Ford puts the entire car together, tying the sensing technology to the braking response (for example), they’ve been working with Mobileye to handle the vision portion. Such alliances seem likely to proliferate as specialist companies develop increasingly arcane algorithms and as auto manufacturers turn more and more into systems integrators.

We can also start refining some of the cruder operations we have. Ford is talking about “partial high beams” –shutting down the high beams only on the left side for the sake of an oncoming driver, leaving the right side bright. Or bit-mapping the mirrors in a way that allows images to be projected on them. This is more than just having the car automatically do more of what we do now; it’s changing what we do.

And at some point, the cars are likely to drive us. We know it can be done – heck, some states have already specifically OKed the concept. And we know Google is doing it (they’re probably looking to cut costs for generating street views – driverless cars don’t ask for raises or stop for pizza). Of course, we have to resolve who’s responsible if things get banged up.

I have to say, given the amount of driving I’ve done over the last 12 months, with several roadtrips of significant distance, that I’d be OK with having the car take over the driving (cruise control seems so crude by comparison). Except that such a car would probably hew religiously to the speed limit. Not that I ever exceed it myself… just sayin…

So it’s pretty clear that, given enough time, our biological parallel processors are going to get a break, with numerous other parallel systems taking over. These new widgets will pay strict attention all the time, never getting sleepy, becoming bored, or fantasizing about the donuts in the shop you just passed. And they will remain completely immune to both Ira Glass’s low-key vocalizations and that super-cool groove going on in the background.

One thought on “Automotive Vision”

Leave a Reply

featured blogs
May 30, 2023
Explore our 2022 environmental, social, and governance (ESG) report to explore our sustainable business practices and our progress in building a smart future. The post Synopsys 2022 ESG Report: Building a Smart Future Together appeared first on New Horizons for Chip Design....
May 25, 2023
Register only once to get access to all Cadence on-demand webinars. Unstructured meshing can be automated for much of the mesh generation process, saving significant engineering time and cost. However, controlling numerical errors resulting from the discrete mesh requires ada...
May 8, 2023
If you are planning on traveling to Turkey in the not-so-distant future, then I have a favor to ask....

featured video

Synopsys 224G & 112G Ethernet PHY IP Demos at OFC 2023

Sponsored by Synopsys

Watch this video of the Synopsys 224G & 112G Ethernet PHY IP demonstrating excellent performance and IP successful ecosystem interoperability demonstrations at OIF.

Learn More

featured contest

Join the AI Generated Open-Source Silicon Design Challenge

Sponsored by Efabless

Get your AI-generated design manufactured ($9,750 value)! Enter the E-fabless open-source silicon design challenge. Use generative AI to create Verilog from natural language prompts, then implement your design using the Efabless chipIgnite platform - including an SoC template (Caravel) providing rapid chip-level integration, and an open-source RTL-to-GDS digital design flow (OpenLane). The winner gets their design manufactured by eFabless. Hurry, though - deadline is June 2!

Click here to enter!

featured chalk talk

DIN Rail Power Solutions' Usage and Advantages
Sponsored by Mouser Electronics and MEAN WELL
DIN rail power supplies with their clean installation and quiet fanless design can be a great solution to solve the common power supply concerns that come along with industrial and building automation applications. In this episode of Chalk Talk, Kai Li from MEAN WELL and Amelia Dalton discuss the variety of benefits that DIN rail power supplies can bring to your next design. They examine how these power supplies can help with power buffering, power distribution, redundancy and more.
Nov 28, 2022