editor's blog
Subscribe Now

Shootout at the FinFET Corral

It’s high noon at IEDM. Both Intel and IBM have “late-breaking news” with their 14-nm FinFET numbers. The giant room is filled to bursting capacity. I’m lucky enough to have some space along the side wall, far from the screen. So far, in fact, that much of what’s on the screen is completely illegible.

Oh, and did I mention photography is not allowed? So… you can’t see the information, you can’t record it even if you saw it… you could busily write what little you can see but then you’re not listening… Oh well, the paper is in the proceedings and I should be able to get the slides after the fact. Right?

Nope. IBM politely declined. Intel didn’t respond at all. (Good thing the proceedings have contact information…) So if the paper is the only record of what happened, then why bother with the presentation? Except for those in the center of the room…

Yeah, I was frustrated, since, in these presentations, you can get a better sense of context and perspective, but only if you have a photographic memory. And I don’t. (And getting less so with each day.) There were definitely points that were made in the presentations that are not in the paper… so I can’t report them.

The whole deal here is Intel’s 14-nm bulk-silicon process vs. IBM’s 14-nm SOI process. And here’s the major takeaway: cost and performance have improved. Moore’s Law, reported as dead at the leading nodes, has taken a few more breaths. It’s just like the good old days, where area shrunk enough to make up for increased costs, and performance gained substantially.

I was going to compare some numbers here, but it’s too spotty to find numbers that they both reported in their papers. For instance, IBM reports a 35% performance improvement over 22 nm; as far as I can tell, Intel reported a performance improvement in the presentation, but didn’t put it in the paper. (I assume that’s intentional.)

Some notable process points:

  • IBM
    • Has a dual-work-function process that allows optimizing both low- and high-VT devices without resorting to doping. No details provided on that process.
    • 15 layers of copper
    • Includes deep-trench embedded DRAM.
  • Intel
    • Uses sub-fin doping.
    • Fin is now much more rectangular than their last edition.
    • 13 interconnect layers
    • They use air-gapped interconnects: pockets of air between lines on select metal layers that reduce capacitance by 17%. They were not willing to discuss how they do the air-gapping, just that they do.
    • Their random variation for VT, which grew from node to node for many nodes, is almost down to where it was at the 90-nm node.

Select images data follow…

[Suggestion to IEDM: require that presentations be made available. They shouldn’t be presenting material if they don’t have the cojones to stand behind it after the presentation…]

Cross-sections:

IBM:

IBM_photo.png

Intel:

Intel_photo.png

 

Pitches:

IBM:

 IBM_table.png

 Intel:

Intel_table.png


Transistor performance:

IBM:

IBM_xstor.png

Intel:

Intel_xstor.png

All images courtesy IEDM.

Leave a Reply

featured blogs
Nov 30, 2023
No one wants to waste unnecessary time in the model creation phase when using a modeling software. Rather than expect users to spend time trawling for published data and tediously model equipment items one by one from scratch, modeling software tends to include pre-configured...
Nov 27, 2023
See how we're harnessing generative AI throughout our suite of EDA tools with Synopsys.AI Copilot, the world's first GenAI capability for chip design.The post Meet Synopsys.ai Copilot, Industry's First GenAI Capability for Chip Design appeared first on Chip Design....
Nov 6, 2023
Suffice it to say that everyone and everything in these images was shot in-camera underwater, and that the results truly are haunting....

featured video

Dramatically Improve PPA and Productivity with Generative AI

Sponsored by Cadence Design Systems

Discover how you can quickly optimize flows for many blocks concurrently and use that knowledge for your next design. The Cadence Cerebrus Intelligent Chip Explorer is a revolutionary, AI-driven, automated approach to chip design flow optimization. Block engineers specify the design goals, and generative AI features within Cadence Cerebrus Explorer will intelligently optimize the design to meet the power, performance, and area (PPA) goals in a completely automated way.

Click here for more information

featured webinar

Rapid Learning: Purpose-Built MCU Software Tools for Data-Driven Embedded IoT Systems

Sponsored by ITTIA

Are you developing an MCU application that captures data of all kinds (metrics, events, logs, traces, etc.)? Are you ready to reduce the difficulties and complications involved in developing an event- and data-centric embedded system? This webinar will quickly introduce you to excellent MCU-specific software options for developing your next-generation data-driven IoT systems. You will also learn how to recognize and overcome data management obstacles. Register today as seats are limited!

Register Now!

featured chalk talk

Energy Storage Systems
Increasing electric vehicle sales, decreasing battery sales, and a shift in energy consumption has made energy storage systems more important than ever before. In this episode of Chalk Talk, Amelia Dalton chats with Gijs Werner from Amphenol FCI Basics about the functions and components involved in commercial energy storage systems, residential energy storage systems and EV charging stations. They investigate the qualifications needed for connectors in energy storage systems and what kind of connectors Amphenol FCI Basics offers for your next energy storage system design.
Apr 3, 2023
28,748 views