industry news
Subscribe Now

Deci Works With Qualcomm to Make Generative AI Accessible for Wide Range of Applications

Deci’s groundbreaking new models streamline the deployment of advanced Generative Artificial Intelligence on the Qualcomm Cloud AI 100 solution, unlocking cost effective, real-time AI processing

TEL AVIV, Israel, January 17th, 2024 — Deci, the deep learning company harnessing artificial intelligence (AI) to build AI, announced today it is collaborating with Qualcomm Technologies, Inc. to introduce advanced Generative Artificial Intelligence (AI) models tailored for the Qualcomm® Cloud AI 100, Qualcomm Technologies’ performance and cost-optimized AI inference solution designed for Generative AI and large language models (LLMs). This working relationship between the two companies is designed to make AI accessible for a wider range of AI-powered applications, resulting in the democratization of Generative AI’s transformative power for developers everywhere.

“Together with Qualcomm Technologies we are pushing the boundaries of what’s possible in AI efficiency and performance” said Yonatan Geifman, CEO and co-founder of Deci. “Our joint efforts streamline the deployment of advanced AI models on Qualcomm Technologies’ hardware, making AI more accessible and cost-effective, and economically viable for a wider range of applications. Our work together is a testament to our vision of making the transformational power of generative AI available to all.”

Through the relationship, Deci will work with Qualcomm Technologies to launch two groundbreaking models. The first model is DeciCoder-6B, a 6 billion parameter model for code generation engineered with a focus on performance at scale. Supporting eight programming languages (C, C#, C++, GO, RAST, Python, Java, JavaScript), it outperforms established models such as CodeGen2.5-7B, StarCoder-7B, and CodeLlama-7B. In fact, in Python, DeciCoder achieves a 3-point lead over models more than twice its size, such as StarCoderBase 15.5B. The model also stands out for its remarkable memory and computational efficiency, boasting 19x higher throughput compared to similar models when running on Qualcomm’s Cloud AI 100.

The second model, DeciDiffusion 2.0, is a 732 million parameter text-to-image diffusion model that sets new standards by outperforming Stable Diffusion v1.5, operating at 2.6 times the speed with on-par image quality. Both models are meticulously optimized to leverage the full potential of the Qualcomm Cloud AI 100 solution. These models are designed to enable users across various industries to experience exceptional performance from the outset at a more competitive price point.

Both DeciCoder-6B and DeciDiffusion 2.0 were developed using Deci’s Neural Architecture Search Technology, AutoNAC™, its proprietary, hardware-aware technology that democratizes the use of Neural Architecture Search for enterprises of all sizes. The distinctive architecture of both models ensures efficient scaling of batching while maintaining minimal memory usage and avoiding any increase in latency. Additionally, the models were designed to handle large batches, enabling maximal utilization of the computational power of the Qualcomm’s Cloud AI 100 cores. DeciCoder-6B and DeciDiffusion have been released under Apache-2.0 and CreativeML Open RAIL++-M Licenses, respectively.

About Deci

Deci enables deep learning to live up to its true potential by using AI to build better AI. With the company’s deep learning development platform, AI developers can build, optimize, and deploy faster and more accurate models for any environment including cloud, edge, and mobile, allowing them to revolutionize industries with innovative products. The platform is powered by Deci’s proprietary automated Neural Architecture Construction technology (AutoNAC), which generates state-of-the-art, efficient deep learning models’ architecture and allows teams to accelerate inference performance, shorten development cycles and reduce computing costs. Founded in 2019, Deci’s team of deep learning engineers and scientists are dedicated to eliminating production-related bottlenecks across the AI lifecycle.

Leave a Reply

featured blogs
Jul 25, 2025
Manufacturers cover themselves by saying 'Contents may settle' in fine print on the package, to which I reply, 'Pull the other one'”it's got bells on it!'...

featured paper

Agilex™ 3 vs. Certus-N2 Devices: Head-to-Head Benchmarking on 10 OpenCores Designs

Sponsored by Altera

Explore how Agilex™ 3 FPGAs deliver up to 2.4× higher performance and 30% lower power than comparable low-cost FPGAs in embedded applications. This white paper benchmarks real workloads, highlights key architectural advantages, and shows how Agilex 3 enables efficient AI, vision, and control systems with headroom to scale.

Click to read more

featured chalk talk

Infineon Automotive MOSFETs Expertise and the OptiMOS™ 7 Advantage
Sponsored by Mouser Electronics and Infineon
In this episode of Chalk Talk, Joseph Sara and Amelia Dalton explore the mega trends motivating MOSFET innovation and the details of Infineon’s cutting-edge MOSFET solutions. They also explore the benefits of Infineon’s OptiMOS 7 platform and how you can take advantage of Infineon MOSFETs for your next design.
Jul 21, 2025
8,925 views