feature article
Subscribe Now

Xilinx Unveils U55C Data Center Card

More Power for Less Power

With Supercomputing 2021 underway this week, all eyes are focused on high-performance computing (HPC) and the incredible advances we are seeing in the world’s fastest computers. (OK, not really ALL eyes, some are still focused on Tik Tok and… well, you know).

The landscape in HPC is changing rapidly, with supercomputers playing a much larger role in solving the world’s most critical problems. With direct impact on crises such as climate change and the global pandemic, and the rise of new and challenging workloads such as AI, supercomputing is spreading its wings and having a tangible impact on the everyday life of everyone on Earth. 

To mis-paraphrase Peter Parker, handling this great responsibility requires great power. And, that means great computing power as well as copious quantities of electrical energy. Since the death of Dennard Scaling for processors, our ability to increase compute performance has come primarily from parallelism – piling more processors into each package, rack, and room. And, since parallelism scales more-or-less infinitely, the real-world limiting factor for what we can achieve in computing performance is energy. 

Moving forward, energy efficiency is likely to be the most critical consideration for HPC. 

Our old buddy the Von Neumann processor carries the load for the vast majority of computers in the world today. But, for all his great attributes, Von Neumann is not particularly adept at power efficiency. Compounding the problem, the massive amount of data being created today causes crushing problems with memory size and bandwidth on applications that need to consume, store, and process that data. And, all that memory activity translates into more energy consumption.

For decades, FPGAs have had the potential to accelerate broad classes of applications with substantially better energy efficiency than conventional processors. But, taking advantage of the parallel processing capabilities of FPGAs required significant engineering investment, and required rare talent in FPGA design. And, in order to deploy FPGAs in HPC systems, FPGA design skill was required in both the hardware and software realms.

A few years ago, Xilinx declared that they were going “data center first” in their pursuit of emerging markets for FPGAs and programmable logic technology. That meant they needed to overcome both the hardware and software barriers in order to get their silicon into systems where it could shine. And, they were working at a disadvantage as Intel, their primary rival in the FPGA business (formerly Altera), had a death grip on the data center market. That meant Xilinx was swimming upstream trying to get their chips into data centers and supercomputers.

Two of the key elements of Xilinx’s strategy were their Alveo pre-designed accelerator cards, and their Vitis unified software platform, addressing the hardware and software deployment challenges. Alveo forced Xilinx from their usual comfort zone – selling components – to selling systems/solutions. And, Vitis took them away from their historic user – the RTL-savvy digital designer – to the much broader audience of software developers. Both of these required rethinking the way the company did business. 

Now, three years or so into the data center endeavor, Xilinx is announcing a new, much more capable Alveo card – the U55C, along with a standards-based API-driven clustering solution that allows them to be deployed at a massive scale – upwards of a thousand FPGAs in a system. The card is a single-slot full height, half length (FHHL) form factor with 150W max power. It doubles the amount of ultra-high-bandwidth HBM2 per FPGA to 16GB (compared with the previous, dual-slot Alveo U280). It offers superior compute density. HBM2 is key in many HPC applications which are memory bandwidth limited.  U55C also packs increased compute density in a smaller form factor, with higher power efficiency. Xilinx says U55C is “built for high-density streaming data, high IO math, and big compute problems that require scale out like big data analytics and AI applications.”

Xilinx designed Alveo to work with existing data center infrastructure and network, giving them the vehicle they need to cross the Intel moat into the well-fortified data center and HPC markets. The new RoCE v2-based clustering solution enables customers to build large FPGA-based HPC clusters on top of their existing infrastructure – without having to hire a team of FPGA experts. The API-driven clustering solution takes advantage of RoCE v2 standards and data center bridging coupled with 200 Gbps bandwidth. This enables an Alveo network to compete with InfiniBand networks in performance and latency, without the vendor lock-in. 

Of course, once your fancy hardware is in the system, you still have to program it. 

That’s where the Vitis unified software platform comes in. HPC developers can develop using normal, high-level languages and AI frameworks. scaling out data pipelining with shared workloads and memory across hundreds of Alveo cards, independent of what server platform and network they use.

Xilinx has come a long way with the Vitis development platform over the last few years, making FPGA-accelerated computing more readily available to to software developers and data scientists without FPGA or hardware expertise. Vitis is somewhat more focused than Intel’s oneAPI framework, but shares the common goal of simplifying the task of deploying applications on complex, heterogeneous computing hardware. 

Vitis supports major AI frameworks such as Pytorch and Tensorflow, as well as high-level programming languages like C, C++ and Python. It abstracts away the usual FPGA challenges such as RTL design, synthesis, layout, and tining closure. Xilinx also provides a growing library of pre-optimized IP, giving developers a jump start on deploying many typical HPC applications in existing data centers.

The Alveo U55C card is currently available  from Xilinx and distributors. The company also provides evaluation via public cloud-based FPGA-as-a-Service providers, as well as select colocation data centers for private previews. Clustering is available now for private previews, with general availability expected in the second quarter of next year.

Obviously, the Alveo strategy plays nicely with the impending acquisition of Xilinx by AMD, and will be a key weapon as AMD continues to do battle with Intel over the fertile ground of data center and HPC computing. It will be interesting to watch.

Leave a Reply

featured blogs
Jan 31, 2023
At CadenceLIVE Europe last year, Ludwig Nordstrom of AWS presented Scaling to 1 Million+ Core to Reduce Time to Results, with up to 90% Discount on Compute Costs . I think that there are currently two trends in EDA infrastructure that cut across almost all design tools. They ...
Jan 30, 2023
By Hossam Sarhan Work smarter, not harder. Isn't that what everyone is always telling you? Of course, it's excellent advice,… ...
Jan 24, 2023
We explain embedded magnetoresistive random access memory (eMRAM) and its low-power SoC design applications as a non-volatile memory alternative to SRAM & Flash. The post Why Embedded MRAMs Are the Future for Advanced-Node SoCs appeared first on From Silicon To Software...
Jan 19, 2023
Are you having problems adjusting your watch strap or swapping out your watch battery? If so, I am the bearer of glad tidings....

featured video

Synopsys 224G & 112G Ethernet PHY IP OIF Interop at ECOC 2022

Sponsored by Synopsys

This Featured Video shows four demonstrations of the Synopsys 224G and 112G Ethernet PHY IP long and medium reach performance, interoperating with third-party channels and SerDes.

Learn More

featured chalk talk

TE APL: Flexibility for Any Use

Sponsored by Mouser Electronics and TE Connectivity

Connectors can make a big difference when it comes to reducing system complexity and ease of use but did you know they can also help with automation and sustainability as well? In this episode of Chalk Talk, Amelia Dalton and Anita Costamagna from TE discuss TE’s APL Connectivity solutions. They dig into the details of these connector solutions and how you can get started using these connector solutions in your next design.

Click here for more information about TE Connectivity Appliance Solutions