chalk talk
Subscribe Now

Scaling Embedded Deep Learning Inference Performance with Dedicated Neural Network DSP

 

Neural networks are taking over a broad range of exciting applications these days. But, the amount of computation required for neural network inferencing can be daunting. In this episode of Chalk Talk, Amelia Dalton chats with Pulin Desai of Cadence Design Systems about some new processor IP designed specifically for neural network inferencing.

Click here for more information about Tensilica Vision DSPs for Imaging, Computer Vision, and Neural Networks

Leave a Reply

featured blogs
Jul 19, 2019
I am sure you can't fail to have noticed that tomorrow is the 50th anniversary of the first landing on the moon. It has been everywhere for the last few weeks. I was a teenager. I remember... [[ Click on the title to access the full blog on the Cadence Community site. ]...
Jan 25, 2019
Let'€™s face it: We'€™re addicted to SRAM. It'€™s big, it'€™s power-hungry, but it'€™s fast. And no matter how much we complain about it, we still use it. Because we don'€™t have anything better in the mainstream yet. We'€™ve looked at attempts to improve conven...