feature article
Subscribe Now

Flash Flood

We Never Thought...

Flash was great for storing small amounts of data.  It was even pretty cool when we could pop a small card into our early digital cameras and save our low-res images with reasonable speed and security.  It was expensive, but we were willing to pay the big premium because it could uniquely solve the non-volatile storage problem for certain classes of mobile devices.  Flash seemed like a novelty, though, and those of us that knew all the evil quirks and kluges under the hood held a healthy respect for the potential of the technology to bite us.  We knew that flash could “wear out,” that flash wasn’t truly random access, and that flash was very temperamental to bring up on a new process node – usually lagging far behind standard CMOS processes.  Special techniques had to be applied to even get flash to appear to behave well, and we expected the curtain to fall aside on that magic at literally any moment.  REAL non-volatile storage was done with hard disk drive (HDD) technology – always had been, always would be.

Uh, until now.

A few years ago, the analyst line on flash was that it was great for the niche it occupied, but that it wasn’t headed for more mainstream storage applications.  When NAND flash started to grab market share, we became even more skeptical.  From an architecture point of view, it seemed like a step backward – even though performance and cost could potentially be much better.  Progress marched ahead, however, and the rhetorical question became: “Will flash someday challenge hard drives for some specialized applications like laptops?”

“Of course not!” we all laughed.  Flash has a 100X disadvantage on cost per megabyte, and it isn’t gaining any on HDD as that technology moves forward.  Nonetheless, we designed flash into mobile devices because we had to – HDD wasn’t up to the rigors of pocket pounding that most mobile devices dished out, and we wanted lots of storage in our mobile companions.  We had photos and music and video and – wait a minute! We’re suddenly dealing with tens of gigabytes of flash in our mobile devices. 

The “billions served” effect drove down prices even farther, and investment in flash technology increased.  Mobile systems had a huge appetite for cheap, fast, reliable non-volatile storage – and those products were selling in the millions of units. 

I’ve been on a panel at the annual Flash Memory Summit for a couple of years now.  The topics of discussion have generally centered around mobile-device related issues.  This year, it was pretty much the same – NAND flash getting cheaper and faster, better management technology alleviating concerns about wear and masking the lack of true random access capability, data centers adopting flash-based SSDs… Wait – rewind… Data Centers?!?

You may already be involved in the debate about SSDs replacing (or more accurately, augmenting) HDDs in enterprise data centers, but the argument goes like this:  Data centers typically use large numbers of high-speed, moderate-capacity, high-cost disk drives to deliver the required capacity and access rate.  However, if you introduce SSD into the mix, you can create a net improvement in every important metric – cost per gigabyte, cost per IOPS, power per gigabyte, and power per IOPS.  Flash-based SSDs are better at speed and power-efficiency, and HDDs are better at cost-per-gigabyte, so mixing the two in the correct proportions can yield better overall results.  Furthermore, since the SSD now carries the burden of the “speed” part of the operation, HDDs can be downgraded to less expensive, lower-speed models to further optimize cost per storage.

While we in the embedded and mobile world have been long since converted, the current revolution in flash cost and performance may spell the beginning of the end of the venerable hard drive – even in the last bastions of HDD supremacy like data centers.  HDDs have become a major bottleneck in system performance, reliability, and power consumption.  In every dimension except cost-per-storage, SSDs have already far surpassed their performance.  In the area of reliability, while we mulled over the tendency of flash to “wear out,” we accepted as fact that the moving, spinning, grinding parts of hard drives would always fail over time (where time is a single-digit number of years), and that the industry was treading water with reliability (which doesn’t sell hardware to consumers) while pushing ahead with cost and density progress (which most certainly does).

NAND flash is now perhaps the most aggressively pushed and downsized technology driving down the Moore’s Law freeway, with no major service stops required until we hit about 20nm technology.  Even at that process node, additional work, like multi-level cell technology (storing of multiple bits per memory cell), promise to increase effective density even farther.  As flash races ahead, other technologies like ferroelectric RAM (FeRAM), magnetoresistive RAM (MRAM), programmable metallization cells (PMC), and phase change memory (PCM) all are nipping at its heels, with advantages in areas like speed, power consumption, and reliability.

For those of us designing embedded systems, this means simply that massive quantities of reliable, power-efficient, fast mass storage are just around the corner.  With hardware foundations including ubiquitous wireless connectivity, incredible processing power, and now practically unlimited storage capacity all available to us in creating our systems, we have to broaden our thinking in architecting our embedded and mobile devices. 

Also, as technologies like flash that cut their teeth in the embedded market make their way into areas like desktop and enterprise computing, we see the continuation of the emerging trend for mobile and embedded systems to lead rather than follow the trends of their less mobile brethren. For us as embedded designers, that’s a big responsibility to bear.

Leave a Reply

featured blogs
Sep 29, 2023
Our ultra-low-power SiWx917 Wi-Fi SoC with an integrated AI/ML accelerator simplifies Edge AI for IoT device makers. Accelerate your AIoT development....
Sep 29, 2023
Cadence has become a contributor-level member of the Automotive Working Group in the Universal Chiplet Interconnect Express (UCIe) Consortium. Last year, the Consortium ratified the UCIe specification, which was established to standardize a die-to-die interconnect for chiplet...
Sep 28, 2023
See how we set (and meet) our GHG emission reduction goals with the help of the Science Based Targets initiative (SBTi) as we expand our sustainable energy use.The post Synopsys Smart Future: Our Climate Actions to Reduce Greenhouse Gas Emissions appeared first on Chip Des...
Sep 27, 2023
On-device generative AI brings many exciting advantages, including cost, privacy, performance and personalization '“ offering significant enhancements in utility, productivity and entertainment with use cases across industries, from the commonplace to the creative....
Sep 21, 2023
Not knowing all the stuff I don't know didn't come easy. I've had to read a lot of books to get where I am....

Featured Video

Chiplet Architecture Accelerates Delivery of Industry-Leading Intel® FPGA Features and Capabilities

Sponsored by Intel

With each generation, packing millions of transistors onto shrinking dies gets more challenging. But we are continuing to change the game with advanced, targeted FPGAs for your needs. In this video, you’ll discover how Intel®’s chiplet-based approach to FPGAs delivers the latest capabilities faster than ever. Find out how we deliver on the promise of Moore’s law and push the boundaries with future innovations such as pathfinding options for chip-to-chip optical communication, exploring new ways to deliver better AI, and adopting UCIe standards in our next-generation FPGAs.

To learn more about chiplet architecture in Intel FPGA devices visit https://intel.ly/45B65Ij

featured paper

Intel's Chiplet Leadership Delivers Industry-Leading Capabilities at an Accelerated Pace

Sponsored by Intel

We're proud of our long history of rapid innovation in #FPGA development. With the help of Intel's Embedded Multi-Die Interconnect Bridge (EMIB), we’ve been able to advance our FPGAs at breakneck speed. In this blog, Intel’s Deepali Trehan charts the incredible history of our chiplet technology advancement from 2011 to today, and the many advantages of Intel's programmable logic devices, including the flexibility to combine a variety of IP from different process nodes and foundries, quicker time-to-market for new technologies and the ability to build higher-capacity semiconductors

To learn more about chiplet architecture in Intel FPGA devices visit: https://intel.ly/47JKL5h

featured chalk talk

Secure Authentication ICs for Disposable and Accessory Ecosystems
Sponsored by Mouser Electronics and Microchip
Secure authentication for disposable and accessory ecosystems is a critical element for many embedded systems today. In this episode of Chalk Talk, Amelia Dalton and Xavier Bignalet from Microchip discuss the benefits of Microchip’s Trust Platform design suite and how it can provide the security you need for your next embedded design. They investigate the value of symmetric authentication and asymmetric authentication and the roles that parasitic power and package size play in these kinds of designs.
Jul 21, 2023
8,949 views