feature article
Subscribe Now

Intel Acquires Omnitek

FPGA AI Race Heats Up

Following on the recent announcement of the new Agilex FPGA family, Intel announced they are acquiring Omnitek – a developer of video and vision acceleration IP for FPGAs. As the market for FPGA-powered acceleration heats up, one of the killer applications is video and machine vision. FPGAs are extremely well suited to video analytics, as the combination of acceleration capabilities and flexible handling of high bit-rate data streams, plus the adaptability to various application-specific parameters is a near ideal fit.

Omnitek is a UK-based company with about 40 employees that has been in business for twenty years. They focus on IP and services for the design of video and vision systems, based on FPGAs and SoCs. They specialize in high-performance video/vision and AI/machine-learning for markets including medical, broadcast, professional AV, automotive, government homeland security, aerospace/defense, industrial/scientific, consumer, and test & measurement.  The company also manufactures video test & measurement equipment to complement their IP and services offerings.

Omnitek is being rolled into the programmable systems group (PSG, which is the former Altera team) at Intel. The combination makes sense, as video and vision acceleration with machine learning is clearly one of the key battlegrounds for Intel PSG in their ongoing rivalry with Xilinx. Xilinx has made strides in repositioning itself to go head-to-head with Intel in the data center acceleration market with their “ACAP” FPGA offering, their Alveo accelerator cards, and a rich portfolio of IP. Xilinx also bolstered their offering with the acquisition of DeePhi Tech, a  Beijing-based start-up with capabilities in machine learning built around Xilinx devices.

For Intel, the Omnitek acquisition is a strategic move, strengthening the company’s own machine-learning portfolio. while effectively blocking Xilinx from taking full advantage of Omnitek’s offering. Omnitek has developed over 220 FPGA IP cores, as well as accompanying software for video-related applications. In the battle to capture the rapidly growing intelligent video acceleration market, FPGA companies will need to engage customers who do not have experienced teams of FPGA designers, so ready-baked optimized IP, reference designs, software, and services will make the advantages of FPGAs available to a much larger set of customers and applications.

At almost the same time, Omnitek announced the availability a new IP offering for convolutional neural networks (CNNs), “delivering world-leading performance per watt at full FP32 accuracy” with the Intel Arria 10 GX FPGA. The Omnitek Deep Learning Processing Unit (DPU) achieves 135 GOPS/W at 32-bit floating point accuracy when running the VGG-16 CNN in an Arria 10 GX 1150. The design “employs a novel mathematical framework combining low-precision fixed point maths with floating point maths to achieve this very high compute density with zero loss of accuracy.”

The DPU is scalable across both Intel’s Arria 10 GX and Stratix 10 GX devices, and it can be tuned for either low cost or high performance for either embedded or data center use. Omnitek says the DPU is “fully software programmable in C/C++ or Python using standard frameworks such as TensorFlow, enabling it to be configured for a wide range of standard CNN models including GoogLeNet, ResNet-50 and VGG-16 as well as custom models.”

With this acquisition, Intel gets a strong boost in IP, but, more importantly a big influx of talent – engineers already expert in creating applications based on Intel FPGAs. Intel’s strategy in FPGA appears to be to deploy as many ready-made applications as possible into high-growth markets, and to have as many of those solutions as possible already bundled into Intel hardware solutions. This takes advantage of the company’s incredible breadth of technology as well as capitalizing on their dominance in the data center. This approach creates a formidable barrier to companies such as NVidia and Xilinx, who aim to make a living selling into what amounts to an Intel-owned ecosystem. The fortification of Intel’s FPGA offering – in hardware, IP, and software – raises the bar for what those competitors will have to do in order to get a foothold in the acceleration part of those systems.

Looking at just the AI acceleration market in the data center, NVidia has created a decent business selling GPUs as acceleration engines for both training and inference use in data centers. The problem with GPU-based solutions is that they do well in the acceleration department but don’t provide much benefit in the performance-per-watt department. FPGA-based accelerators are enormously more power efficient, but they have traditionally had a steep development curve to get them programmed optimally.

It appears that Intel is taking a multi-pronged approach to squeeze NVidia out of this space. First, Intel has significantly improved the performance of their Xeon processors for AI inferencing tasks. With something like a 30x inferencing upgrade announced in their recent release, the baseline for applications that would even need acceleration is raised significantly. After all, if you’ve already got Xeon-based servers sitting there and they can handle your particular AI task, why go for acceleration at all?

Then, for those who still need acceleration, there are many competitors in the market, including NVidia with their GPUs, companies such as Xilinx and Achronix with third-party FPGA solutions, and Intel, themselves, with their own range of accelerators including FPGAs from PSG. In many situations, Intel and their OEMs are packaging FPGAs into servers, cards, and even the same packages as Xeon processors. With Intel’s FPGA hardware already sitting in the system, the argument for bringing in third-party accelerators is made even more difficult.

Regardless of the competitive landscape, several things are clear. First, FPGAs will play a much larger role than ever in the enormous opportunities emerging in and out of the data center with the explosion of data generated by the latest generations of cameras, sensors, and assorted IoT devices. The challenge of processing and moving all that data is a perfect match for FPGAs, but FPGA companies must be creative in their approach to reducing the learning curve for engineering teams wanting to take advantage of their capabilities. By pre-engineering IP, reference designs, software, and entire applications – in addition to radically upgrading development tool suites – FPGA suppliers can enormously expand their customer base and the number of systems in which they are deployed. We are clearly at the doorstep of the biggest opportunity for FPGA market growth in history, but that growth will not occur without some serious innovation in the ecosystem. It will be interesting to watch.

Leave a Reply

featured blogs
Dec 1, 2020
If you'€™d asked me at the beginning of 2020 as to the chances of my replicating an 1820 Welsh dresser, I would have said '€œzero,'€ which just goes to show how little I know....
Dec 1, 2020
More package designers these days, with the increasing component counts and more complicated electrical constraints, are shifting to using a front-end schematic capture tool. As with IC and PCB... [[ Click on the title to access the full blog on the Cadence Community site. ]...
Dec 1, 2020
UCLA’s Maxx Tepper gives us a brief overview of the Ocean High-Throughput processor to be used in the upgrade of the real-time event selection system of the CMS experiment at the CERN LHC (Large Hadron Collider). The board incorporates Samtec FireFly'„¢ optical cable ...
Nov 25, 2020
[From the last episode: We looked at what it takes to generate data that can be used to train machine-learning .] We take a break from learning how IoT technology works for one of our occasional posts on how IoT technology is used. In this case, we look at trucking fleet mana...

featured video

Improve SoC-Level Verification Efficiency by Up to 10X

Sponsored by Cadence Design Systems

Chip-level testbench creation, multi-IP and CPU traffic generation, performance bottleneck identification, and data and cache-coherency verification all lack automation. The effort required to complete these tasks is error prone and time consuming. Discover how the Cadence® System VIP tool suite works seamlessly with its simulation, emulation, and prototyping engines to automate chip-level verification and improve efficiency by ten times over existing manual processes.

Click here for more information about System VIP

featured paper

Simplify your isolated current & voltage sensing designs

Sponsored by Texas Instruments

Learn how the latest isolated amplifiers and isolated ADCs can operate with a single supply on the low side, and why this offers substantial benefits over traditional solutions.

Click here to download the whitepaper

Featured Chalk Talk

Benefits of FPGAs & eFPGA IP in Futureproofing Compute Acceleration

Sponsored by Achronix

In the quest to accelerate and optimize today’s computing challenges such as AI inference, our system designs have to be flexible above all else. At the confluence of speed and flexibility are today’s new FPGAs and e-FPGA IP. In this episode of Chalk Talk, Amelia Dalton chats with Mike Fitton from Achronix about how to design systems to be both fast and future-proof using FPGA and e-FPGA technology.

Click here for more information about the Achronix Speedster7 FPGAs