Home » Markets » AI

AI processor intellectual property aimed at TinyML

Ceva has announced a neural network processing core aimed at SoCs needing to run TinyML models.

Ceva NeuPro-Nano block

There are two versions:

  • NPN32 with 32 int8 MACs
  • NPN64 with 64 int8 MACs

“Both of which benefit from Ceva-NetSqueeze for direct processing of compressed model weights,” according to the company. “NPN32 is optimised for most TinyML workloads targeting voice, audio, object detection and anomaly detection. The NPN64 provides 2x performance acceleration using weight sparsity, greater memory bandwidth, more MACs and support for 4bit weights for more complex on-device AI use such as object classification, face detection, speech recognition and health monitoring.”


The single cores include sufficient hardware, said Ceva, for neural network compute, feature extraction, running control code and running DSP code without an external microcontroller.


NeuPro Studio is the associated AI stack, which works with AI inference frameworks including TFLM and µTVM.

A pre-trained model zoo has been created with TinyML models for voice, vision and sensing use.

Find NPN32 and NPN64 on this web page


Leave a Reply

Your email address will not be published. Required fields are marked *

*