feature article
Subscribe Now

Something’s Coming but I Can’t Say What!

I just received an unusual briefing, and now I’m sitting here at the start of this column looking at a blank page thinking to myself, “So, how do I go about explaining this one?” What we’re talking about here is a next-generation processing chip that’s based on a new processing architecture. Known as a hierarchical learning processor (HLP), this technology is intended to be a game-changer for tasks like artificial intelligence (AI) training, high-performance computing (HPC), and metaverse processing (don’t ask, I’ll tell you later).

Generally speaking, when I receive a briefing on something like this, its creators thrust eye-wateringly complicated architectural block diagrams at me with reckless abandon. It’s also common for me to be taken on a deep dive through the inner workings of the beast, only occasionally managing to raise my nose above the surface of the PowerPoint slide deck to gasp for air. Not this time. Although I have been given a hint of a sniff of a whiff of some of the high-level functional units of which this device is composed, I’ve not even been presented with an image of a black box.

Is this bodacious beauty available now? Sadly not. Test silicon and performance signoff are scheduled for this year (2022), with customer samples following in 2023 and full-on production in 2024.

During the course of my career, I’ve been regaled with promises of delight, I’ve been told tall tales of wonders that are to come, and I’ve seen vaporware versions of hardware and software you wouldn’t believe (it’s the audacity of the vaporware you wouldn’t believe; the hardware and software didn’t exist in any of the dimensions with which I’m familiar). I know you’ll find this difficult to believe but — on occasion — many of the promised products didn’t turn out as their proponents originally proposed. (This might be a good time for you to imagine me standing tout seul in the middle of an unforgiving city on a cold blustery night with a tumbleweed rolling down the street in the background, my lower lip quivering, and a little tear rolling down my cheek.)

Based on these past experiences, in the normal scheme of things, this is the point where I would typically have bailed out of the briefing, muttering platitudes while sidling towards the nearest exit. However, this wasn’t the normal scheme of things because I was being briefed by industry veteran Dr. Venkat Mattela, who is Founder and CEO of Ceremorphic.

Venkat has an interesting way of going about things. First, he looks for a problem that needs solving. Next, he spends the time and effort to build the underlying technology required to address the problem. It’s only then that he starts working on game-changing products based on his new technology. This is the approach Venkat took at Redpine Signals, which was the previous company of which he was the Founder and CEO. Under his leadership, Redpine Signals delivered breakthrough innovations and industry-first products leading to the development of an ultra-low power wireless solution that outperformed products from wireless industry giants by as much as 26 times in terms of energy efficiency.

Silicon Labs acquired Redpine Signals’ wireless assets for $308 million in March 2020. If this had been yours truly, I would have taken some time to run for the hills and bask on the beach (I never metaphor I didn’t like). By comparison, only one month later, in April 2020, Venkat founded Ceremorphic. 

So, let me tell you what I do know. There are already other players in this area, so — as Venkat says — if HLP offers only 50% improvement over existing solutions, no one will be interested. By comparison, if HLP can outperform the competition by 50X in terms of performance and power efficiency, then this will cause people to sit up and pay attention.

It turns out that the underlying architectural concepts for HLP have been years in the making. Furthermore, a physical realization of HLP technology has been in development since 2018. Initial development employed TSMC’s 7nm process node, switching to the newly available 5nm technology node when Ceremorphic was launched in 2020.

With the exception of things like GPUs and FPGAs, the vast majority of companies creating processor chips for AI are targeting ultra-low-power devices for inferencing at the edge. These devices are typically not intended to scale up to data centers and the cloud. By comparison, HLP technology is primarily targeted at high-end deployments and applications like data centers, AI training, automotive advanced driver assistance systems (ADAS) and autonomous driving (AD) systems, robotics, life sciences, and metaverse processing.

“Remind me again; what’s metaverse processing?” I hear you cry. I’m pleased to see that you’re paying attention. As the Ceremorphic website tells us: “The convergence of extended Reality (XR), augmented reality (AR), mixed reality (MR), and virtual reality (VR) is an exciting new capability that enables the unprecedented creation of virtual worlds. Metaverse processing creates a greater overlap of digital and physical spaces to make productivity and entertainment applications abundant.” Well, I’m glad we’ve cleared that up.

One interesting aspect to all of this is that — although it wouldn’t initially be cost-effective to use the latest and greatest technology node to create processing engines targeted only at cheap-and-cheerful IoT devices — HLP technology is based on chiplets that allows it to start off targeting high-end applications and be scaled down to mid-range and low-end applications over time.

Another point that caught my attention was the use of intellectual property (IP). When it comes to functions like PCIe 6.0 / CXL 3.0 connectivity interfaces, most companies will purchase this IP from a third-party. By comparison, while Venkat has no problem purchasing commodity IP blocks, he prefers to develop his own low-energy incarnations of high-end IPs like PCIe 6.0 / CXL 3.0 because this allows him to further differentiate his products. (Although it’s not a big focus for them, the folks at Ceremorphic won’t be averse to licensing this IP to other companies working in different markets.)

The limited details that are available regarding the QS 1, which will be the first member of the HLP family, include the following key features:

  • Custom machine learning processor (MLP) running at 2GHz.
  • Custom floating-point unit (FPU) running at 2GHz.
  • Patented multi-thread processing macro-architecture, ThreadArch-based RISC-V processor for proxy processing running at 1GHz.
  • Custom video engines for metaverse processing running at 1GHz along with an Arm Cortex-M55 core (the M55 core is the first to feature Arm Helium vector processing technology for enhanced, energy-efficient digital signal processing (DSP) and machine learning (ML) performance).
  • Quantum-resistant security microarchitecture.
  • Custom designed X16 PCIe 6.0 / CXL 3.0 connectivity interface.
  • Open AI framework software support with optimized compiler and application libraries.
  • Soft error rate: (100,000)^-1

Ceremorphic is backed by more than 100 patents in core technologies, it’s led by proven members of the management team, board of directors, and technical advisors, and it currently has 150 full-time employees with plans to ramp up to 250 this year.

So, at the end of the day I have to say that I think this briefing was a success because — if nothing else — it’s left me longing for more on the nitty-gritty details front. How about you? Do you have any thoughts you’d care to share? 

4 thoughts on “Something’s Coming but I Can’t Say What!”

Leave a Reply

featured blogs
Jan 27, 2023
Wow, it's already the last Friday in January, so time for one of my monthly update posts where I cover anything that doesn't justify its own full post or which is an update to something I wrote about earlier. Automotive Security I have written about automotive secur...
Jan 26, 2023
By Slava Zhuchenya Software migration can be a dreaded endeavor, especially for electronic design automation (EDA) tools that design companies… ...
Jan 24, 2023
We explain embedded magnetoresistive random access memory (eMRAM) and its low-power SoC design applications as a non-volatile memory alternative to SRAM & Flash. The post Why Embedded MRAMs Are the Future for Advanced-Node SoCs appeared first on From Silicon To Software...
Jan 19, 2023
Are you having problems adjusting your watch strap or swapping out your watch battery? If so, I am the bearer of glad tidings....

featured video

Synopsys 224G & 112G Ethernet PHY IP OIF Interop at ECOC 2022

Sponsored by Synopsys

This Featured Video shows four demonstrations of the Synopsys 224G and 112G Ethernet PHY IP long and medium reach performance, interoperating with third-party channels and SerDes.

Learn More

featured chalk talk

"Scalable Power Delivery" for High-Performance ASICs, SoCs, and xPUs

Sponsored by Infineon

Today’s AI and Networking applications are driving an exponential increase in compute power. When it comes to scaling power for these kinds of applications with next generation chipsets, we need to keep in mind package size constraints, dynamic current balancing, and output capacitance. In this episode of Chalk Talk, Mark Rodrigues from Infineon joins Amelia Dalton to discuss the system design challenges with increasing power density for next generation chipsets, the benefits that phase paralleling brings to the table, and why Infineon’s best in class transient performance with XDP architecture and Trans Inductor Voltage Regulator can help power  your next high performance ASIC, SoC or xPU design.

Click here for more information about computing and data storage from Infineon