feature article
Subscribe Now

0.6V, and Still the Memory Persisted

Near-Threshold, Low-Voltage, Low-Power SRAM for IoT SoC Design

A big chunk of the vast, imagined IoT (Internet of Things) includes the industrial IoT (IIoT), which is responsible for controlling and monitoring energy-consuming processes based on large motors, pumps, fans, generators, heavy-duty actuators, heaters, and other high-power equipment. Such IIoT applications aren’t as concerned with low-power design compared to the medical and consumer side of the IoT—wearables and the like. Medical and consumer IoT designs are more sensitive to high-volume unit costs and low-power considerations, which makes them good targets for ASIC implementations.

The semiconductor foundries have developed low-power and low-leakage processes including FDSOI variants that suit these low-power IoT applications, as long as the applications demand the sort of high manufacturing volumes that would support a full-on ASIC development project. Microprocessor IP vendors have developed low-power processor cores that can be used in such applications. The modular RISC-V and Arm Cortex-M series processor cores are but two examples. But the ever-present SRAM IP has been left out of this equation, or so claims sureCore CEO Paul Wells.

“SRAM IP has not changed in 20 years,” Wells told me at the recent DAC held in San Francisco last June. There’s a simple reason for that, he explained. For the last 20 years, simply waiting for the next semiconductor process node to come on line would automatically give SoC designers faster memory that operated at lower power levels. Because there were plenty of other fires to extinguish and mountains to climb, on-chip SRAM design stagnated.

That was before Dennard Scaling and Moore’s Law went to their graves. (If you read EEJournal regularly, the deaths of these two industry stalwarts should not come as a surprise. For example, see “Fifty (or Sixty) Years of Processor Development…for This? ‘Dennard Scaling and Moore’s Law are dead, Now what?’ – a 4-act play by Dr. David Patterson” and my more recent blog titled “Monty Python, Dead Parrots, Moore’s Law, and the ITRS.”)

The IoT has the potential to require stratospheric, mobile-phone production volumes that make electronics vendors’ mouths water. But without the relatively large lithium batteries that come along with mobile phones and certainly without the ready juice fix of a daily (or nightly) recharge, these IoT designs need to sip power. Today, everything IoT seemingly needs to run on small, coin-cell batteries or locally harvested power and needs to run essentially forever. That’s a tall order.

The coincident deaths of Moore’s Law and Dennard Scaling along with a sudden strong interest in millions of low-power “things” that connect to the Internet have opened a market window for new, on-chip memory IP, which prompted sureCore to develop 6T SRAM blocks that could operate at very low voltage, reducing dynamic power levels by as much as 80%. The resulting SRAM design for TSMC’s 40 ULV process technology can operate at supply voltages as low as 0.6V across process corners and over the full industrial temperature range of -40° to 125°C.

Achieving that feat required the sureCore designers to toss out a lot of excess design margin, which in turn required a significant amount of “high-sigma” statistical simulation and analysis to ensure that there was still sufficient margin for the SRAM to operate over process variations, without process modifications, and even down to 0.6V. Monte Carlo simulation and extensive physical verification including IR drop, electromagnetic interference, and crosstalk checks replaced typical, fast-and-loose, she’ll-be-right-mate engineering margins. The result is sureCore’s EverOn low-power SRAM IP core.

Figure 1 below shows a graph that plots energy consumption versus supply voltage for the EverOn SRAM core built in TSMC’s 40 ULV process technology versus factory SRAM IP. The graph’s taken from sureCore’s EverOn data sheet:

Figure 1: The sureCore EverOn SRAM consumes less power at the same supply voltage and operates at much lower power at low supply voltages compared to foundry SRAM.

Three key things to note from this graph:

  1. The sureCore EverOn SRAM always consumes significantly less power than standard foundry SRAM at any supply voltage.
  2. The sureCore SRAM continues to operate down to 0.6V, much lower than the characterization curve for foundry SRAM, which stops at 0.9V.
  3. In the most extreme comparison, the EverOn SRAM core operating at 0.6V consumes less than 20% of the energy per MHz compared to the foundry SRAM core operating at 1.2V.

In addition, sureCore’s EverOn SRAM design divides each SRAM core into four independently controlled banks. Each EverOn SRAM bank can be active, asleep (with data retained), or off, as shown in Figure 2 below:

Figure 2: All EverOn SRAMs are divided into four independent memory banks and the operational/power state of each bank is under software control, which can result in even lower power operation.

The EverOn core’s control circuitry allows software to control the state of each of the four banks and the SRAM periphery circuity, making even lower power operation under active software management possible. You power up only the RAM you need, when you need it.

So what’s new here? Hasn’t near-threshold ASIC logic been around for a while? Yes, said Wells, but foundries’ ASIC SRAM bit-cell designs have not followed. Because SRAM blocks are stepped and repeated and because they consume relatively large amounts of on-chip area, standard-cell SRAM IP is designed for optimal packing density, to save silicon area and not to save power. Remember, saving SRAM power was once a simple result of Moore’s Law, Dennard Scaling, and process-node migration, so SRAM IP didn’t get redesigned when there were so many other, more important things to do.

With Moore’s Law and Dennard Scaling dead and on-chip process variability on the rise in the land of nanometer process geometries, the possibility of “weak” bit cells in an SRAM block increases. This weakness materializes operationally in the form of multiple problems that produce random read and write errors when the SRAM block operates at near-threshold supply voltages. That’s why the factory SRAM curve in Figure 1 above stops at 0.9V.

Engineers know you rarely get something for nothing, and that’s the case here too. First, EverOn SRAM blocks are 5-to-10% larger than big standard foundry SRAM blocks. Second, there’s about a 10% performance loss if you lower the supply voltage all the way down to 0.6V. However, you get full speed if you wind the supply voltage up to 1.2V.

Now the EverOn SRAM’s area penalty might seem like it’s an issue, but there are a couple of factors to consider. First, sureCore’s EverOn SRAMs are intended for less memory-intensive applications (meaning they’re NOT intended to store images, audio, video, or large blocks of code). These are relatively small memories that max out at 8K by 72 bits or 16K by 36 bits. That’s 576 Kbits, if you do the math. These EverOn SRAMs are designed to hold small blocks of code and tiny data sets—which are more appropriate for wearable IoT devices or smart dust sensors.

The EverOn memories must be small to effectively manage process variation, and sureCore has run test chips in TSMC’s 40 ULV process to prove the design. Standard 125°C, 1000-hour, HTOL (High Temperature Operating Life) tests on these chips surfaced no PVT failures in the SRAM design. In other words, there’s still sufficient design margin to achieve reliable operation as proven by all that time at Monte Carlo.

Second, the conventional approach to building low-power, low-voltage bit cells for SRAM IP is to use a 10T cell design, which is a 67% increase in area over a 6T bit cell. By comparison, the EverOn SRAM core’s 10% area penalty really doesn’t seem like all that much.

That left one potential fly in the ointment for me to investigate. I asked Wells about the adoption of TSMC’s 40 ULV process technology, and he admitted that the uptake at that process node wasn’t as great as expected. He then added, “That’s why sureCore has migrated the EverOn design to TSMC’s 28 HPC+ process technology.” Now 28nm process technology is the current sweet spot for ASIC density/cost and it’s destined to stay that way for quite a while. FinFET technology and EUV lithography increase the cost of ASIC design and manufacturing, so 28nm processes like TSMC’s 28 HPC+ are going to be in the Goldilocks zone for IoT devices for the foreseeable future.

Leave a Reply

featured blogs
Oct 23, 2020
The Covid-19 pandemic continues to impact our lives in both expected and unexpected ways. Unfortunately, one of the expected ways is a drop in charitable donations. Analysts predict anywhere from a 6% decrease '€“ with many planning for a bigger decline than that. Also, mor...
Oct 23, 2020
[From the last episode: We noted that some inventions, like in-memory compute, aren'€™t intuitive, being driven instead by the math.] We have one more addition to add to our in-memory compute system. Remember that, when we use a regular memory, what goes in is an address '...
Oct 23, 2020
Any suggestions for a 4x4 keypad in which the keys aren'€™t wobbly and you don'€™t have to strike a key dead center for it to make contact?...
Oct 23, 2020
At 11:10am Korean time this morning, Cadence's Elias Fallon delivered one of the keynotes at ISOCC (International System On Chip Conference). It was titled EDA and Machine Learning: The Next Leap... [[ Click on the title to access the full blog on the Cadence Community ...

featured video

Demo: Inuitive NU4000 SoC with ARC EV Processor Running SLAM and CNN

Sponsored by Synopsys

Autonomous vehicles, robotics, augmented and virtual reality all require simultaneous localization and mapping (SLAM) to build a map of the surroundings. Combining SLAM with a neural network engine adds intelligence, allowing the system to identify objects and make decisions. In this demo, Synopsys ARC EV processor’s vision engine (VPU) accelerates KudanSLAM algorithms by up to 40% while running object detection on its CNN engine.

Click here for more information about DesignWare ARC EV Processors for Embedded Vision

featured paper

An engineer’s guide to autonomous and collaborative industrial robots

Sponsored by Texas Instruments

As robots are becoming more commonplace in factories, it is important that they become more intelligent, autonomous, safer and efficient. All of this is enabled with precise motor control, advanced sensing technologies and processing at the edge, all with robust real-time communication. In our e-book, an engineer’s guide to industrial robots, we take an in-depth look at the key technologies used in various robotic applications.

Click here to download the e-book

Featured Chalk Talk

Protecting Circuitry with eFuse IC

Sponsored by Mouser Electronics and Toshiba

Conventional fuses are rapidly becoming dinosaurs in our electronic systems. Finally, there is circuit protection technology that doesn’t rely on disposable parts and molten metal. In this episode of Chalk Talk, Amelia Dalton chats with Jake Canon of Toshiba about eFuse - a smart solution that will get rid of those old-school fuses once and for all.

Click here for more information about Toshiba efuses