industry news
Subscribe Now

CacheQ’s New Compiler Supports Multi-Threading CPU Acceleration, Enables Software Developers to “Write Once, Accelerate Anywhere”

486% Speedup Over Single-Thread Execution on x86 with 12 Cores; 400% Faster than Single-Threaded GCC on Apple M1 with Eight Cores

LOS GATOS, Calif., May 25, 2021 (GLOBE NEWSWIRE) — CacheQ Systems, Inc., a leader in compiler tools for heterogeneous compute systems, today announced support for multi-threading acceleration for CPUs with multiple physical cores through its new compiler, offering software developers the ability to “write once, accelerate anywhere.”

A feature of the CacheQ Compiler Collection, the compiler eliminates manual code rewriting and the use of threading libraries or complex parallel execution APIs such as OpenMP or MPI. It takes single-threaded C code and generates executables that can run on CPUs, leveraging many physical x86 cores with or without hyperthreading, as well as Arm and RISC-V cores.

Using CacheQ’s flexible compiler, users generate code for multicore processors on the same or different architectures and benchmark usage with runtime variables. They can add to the hardware for ultimate performance and power usage or reduce the number of cores and allocate other processes to achieve more optimal performance per watt of power consumed.

The result is a speedup of more than 486% over single-thread execution on X86 processors with 12 logical cores, based on benchmarks from the Black Scholes financial algorithm that simulates human behavior in stock trading. An Apple M1 processor with eight Arm cores is 400% faster than the single-threaded GNU Compiler Collection (GCC).

“This is a game changer for software developers to take full advantage of parallel processing power without spending years learning to code with OpenMP or MPI,” remarks Clay Johnson, CEO of CacheQ Systems. “They can accelerate a single-thread algorithm with our tools to quickly compile and target any CPU with two or more cores.”

The CacheQ Compiler Collection
CacheQ enables software developers to easily develop and deploy custom hardware accelerators for heterogeneous compute systems including FPGAs, CPUs and GPUs. Its CacheQ Compiler Collection is modelled after the GCC tool suite, including a user interface similar to common open-source compilers. Because it requires limited code modification, development time is shortened and system quality improved.

The tool suite enables compilation, linting and error detection, performance prediction, profiling, debug and visualization of the generated virtual engine. It supports target hardware including single and multicore processors, as well as heterogeneous compute systems with FPGA accelerators connected to x86 and Arm processors.

The CacheQ Compiler Collection supports C code and C++ through “hybrid” access of an exported function call.

Additional Benchmarks
Additional benchmarks of the CacheQ Compiler Collection highlight its ability to span high-end servers and consumer electronic devices. Execution of the M1 processor with two cores outperformed the x86 chip with 11 cores, demonstrating a cost-per-watt advantage. The Apple M1 processor with four cores performed 210% faster than the x86 with 12 cores. Overall, it performed approximately 1,476% faster than the single-threaded GCC running on x86 using the CacheQ Compiler Collection.

All simulations were performed on the same code compiled for different targets. Benchmarks were performed on an Intel i7-8700k x86 CPU running at 3.7GHz with six physical cores and hyperthreading for 12 logical cores available running Ubuntu 18.04. Apple M1 benchmarks were captured with a Parallels VM running native Arm Ubuntu 20.04 image.

Availability and Pricing
The CacheQ compiler tools are shipping now through a limited access program. Pricing is available on request.

Visit the CacheQ website for additional information or to view a brief video on the new multicore compiler technology. Requests for more information should be sent to

About CacheQ Systems
CacheQ Systems, headquartered in Los Gatos, Calif., with a development center in Longmont, Colo., was founded in 2018 to develop compiler technology that brings acceleration technology to developers with little to no hardware-specific modifications or pragmas.

Leave a Reply

featured blogs
Aug 12, 2022
Relive Design Automation Conference (DAC) 2022 with intern Samir Banerjee, including the latest on cloud EDA tools, climate action, and diversity in engineering. The post Reflecting on the 59th Design Automation Conference (DAC) appeared first on From Silicon To Software....
Aug 11, 2022
Speed increase requirements keep on flowing by in all the domains surrounding us . The s ame applies to memory storage too . Earlier mobile devices used eMMC based flash storage, which was a significantly slower technology. With increased SoC processing speed, pairing it with...
Jul 27, 2022
It's easy to envisage a not-so-distant future when sophisticated brain-computer interfaces become available for general-purpose infotainment use....

featured video

Synopsys End-to-End Solution for Energy-Efficient SoCs

Sponsored by Synopsys

This video shows a holistic approach to energy-efficient System-on-Chip (SoC) design with Synopsys’ end-to-end solution for software-driven power analysis and optimization, from architecture to signoff.

Learn more about Synopsys’ Energy-Efficient SoCs Solutions:

featured chalk talk

Introducing Vivado ML Editions

Sponsored by Xilinx

There are many ways that machine learning can help improve our IC designs, but when it comes to quality of results and design iteration - it’s a game changer. In this episode of Chalk Talk, Amelia Dalton chats with Nick Ni from Xilinx about the benefits of machine learning design optimization, what hierarchical module-based compilation brings to the table, and why extending a module design into an end-to-end flow can make all the difference in your next IC design.

Click here for more information about Vivado ML