industry news
Subscribe Now

Flex Logix Improves Deep Learning Performance by 10X with New EFLX4K AI eFPGA Core

Available on Any Process in 6-8 Months: Enables Reconfigurable AI Matrix Multipliers of Any Size

MOUNTAIN VIEW, Calif.– June 25, 2018 – Flex LogixÒTechnologies, Inc., the leading supplier of embedded FPGA (eFPGA) IP, architecture and software, announced today a new member of the EFLXÒ4K eFPGA Core Family: the EFLX4K Artificial Intelligence (AI) eFPGA core. This new core has been specifically designed to enhance the performance of deep learning by 10X and enable more neural network processing per square millimeter.

Many companies are using FPGA to implement AI and more specificllay machine learning, deep learning and neural networks as approaches to achieve AI. The key function needed for AI are matrix multipliers, which consist of arrays of MACs (multiplier accumulators). In existing FPGA and eFPGAs, the MACs are optimized for DSPs with larger multipliers, pre-adders and other logic which are overkill for AI. For AI applications, smaller multipliers such as 16 bits or 8 bits, with the ability to support both modes with accumulators, allow more neural network processing per square millimeter.

“We’ve had significant customer interest in using eFPGA for AI applications because of the performance advantages it can deliver in these chip design,” said Geoff Tate, CEO and co-founder of Flex Logix. “In fact, one of the first customers Flex Logix announced was Harvard who chose our eFPGA for their deep learning design. Not only will they be presenting this chip at Hot Chips next month, but they are already working on a follow-on 16nm AI chip that will make more extensive use of the Flex Logix EFLX eFPGA.”

AI customers want more MACs/second and more MACs/square millimeter, but they also want the flexibility of eFPGA to reconfigure designs as AI algorithms are changing rapidly. They require the ability to switch between 8 and 16 bit modes as needed and to implement matrix multipliers of varying sizes to meet their applications’ performance and cost constraints.

The EFLX4K AI eFPGA core leverages Flex Logix’s patented XFLX™ interconnect for an area-efficient, reconfigurable AI solution. It uses a new AI-MAC architecture capable of implementing 8-bit MACs or 16-bit MACs reconfigurably (as well as 16×8 and 8×16).  A single EFLX4K AI core in 16nm for example will be about 1.2 square millimeters with 441 8-bit MACs running at 1GHz for a throughput of 441 GigaMACs/second at worst case silicon conditions.  The EFLX4K AI core can be arrayed up to at least a 7×7, enabling performance of ~22 TeraMacs/second in worst case silicon conditons.

The EFLX4K AI is footprint compatible with currently available, silicon proven EFLX4K cores to allow architects and designers the ultimate flexibility in RTL reconfigurability.

The EFLX4K AI eFPGA core is fully supported by Flex Logix’s existing software flow using the EFLX Compiler.

As with any EFLX core, the EFLX4K AI eFPGA core can be implemented on any process in six to eight months.  Flex Logix’s unique eFPGA architecture (XFLX high density interconnect & ArrayLinx™ interconnect for tiling arrays of any size) enables much faster architectural innovation and iteration than traditional FPGA design approaches.  A EFLX4K AI eFPGA core is ~1.2x the area of an EFLX4K Logic core (same width, ~1.2x height).

See us at DAC Booth #2318 this week to learn more or contact us at info@flex-logix.com.

Leave a Reply

featured blogs
Apr 24, 2026
A thought experiment in curiosity, confusion, and cosmic consequences....

featured paper

Quickly and accurately identify inter-domain leakage issues in IC designs

Sponsored by Siemens Digital Industries Software

Power domain leakage is a major IC reliability issue, often missed by traditional tools. This white paper describes challenges of identifying leakage, types of false results, and presents Siemens EDA’s Insight Analyzer. The tool proactively finds true leakage paths, filters out false positives, and helps circuit designers quickly fix risks—enabling more robust, reliable chip designs. With detailed, context-aware analysis, designers save time and improve silicon quality.

Click to read more

featured chalk talk

Connecting the World Through Space
Sponsored by Mouser Electronics and Qorvo
In this episode of Chalk Talk, Ryan Jennings from Qorvo and Amelia Dalton explore the critical components and design challenges inherent in LEO satellite infrastructure and how Qorvo’s solutions are enabling the next generation of space-based connectivity. 
Mar 30, 2026
29,704 views