CEVA, a developer of signal processing platforms and artificial intelligence processors, has unveiled NeuPro, a specialized Artificial Intelligence (AI) processor family for deep learning inference. The NeuPro family of processors is designed for smart and connected edge device vendors looking for a streamlined way to quickly take advantage of the significant possibilities that deep neural network technologies offer.
NeuPro builds on CEVA’s position and experience in deep neural networks for computer vision applications. Dozens of customers are already deploying the CEVA-XM4 and CEVA-XM6 vision platforms along with the CDNN neural network software framework in consumer, surveillance and ADAS products. This new family of dedicated AI processors offers a considerable step-up in performance, ranging from 2 Tera Ops Per Second (TOPS) for the entry-level processor to 12.5 TOPS for the most advanced configuration.
The NeuPro processor line extends the use of AI beyond machine vision to new edge-based applications including natural language processing, real-time translation, authentication, workflow management, and many other learning-based applications that make devices smarter and reduce human involvement.
Ilan Yona, vice president and general manager of the Vision Business Unit at CEVA, commented: “It’s abundantly clear that AI applications are trending toward processing at the edge, rather than relying on services from the cloud. The computational power required along with the low power constraints for edge processing, calls for specialized processors rather than using CPUs, GPUs or DSPs. We designed the NeuPro processors to reduce the high barriers-to-entry into the AI space in terms of both architecture and software. Our customers now have an optimized and cost-effective standard AI platform that can be utilized for a multitude of AI-based workloads and applications.”
The NeuPro architecture is composed of a combination of hardware-based and software-based engines coupled for a complete, scalable and expandable AI solution. Optimizations for power, performance, and area (PPA) are achieved using a precise mix of hardware, software and configurable performance options for each application tier.
The NeuPro family comprises four AI processors offering different levels of parallel processing:
Each processor consists of the NeuPro engine and the NeuPro VPU. The NeuPro engine includes the hardwired implementation of neural network layers among which are convolutional, fully-connected, pooling, and activation. The NeuPro VPU is a cost-efficient programmable vector DSP, which handles the CDNN software and provides software-based support for new advances in AI workloads. NeuPro supports both 8-bit and 16-bit neural networks, with an optimized decision made in real time in order to deliver the best tradeoff between precision and performance. The MAC units achieve better than 90% utilization when running, ensuring highly optimized neural network performance. The overall processor design reduces DDR bandwidth substantially, improving power consumption levels for any AI application.
The NeuPro family, coupled with CDNN, CEVA’s neural network software framework, provides a deep learning solution for developers to easily and efficiently generate and port their proprietary neural networks to the processor. CDNN supports the full gamut of layer types and network topologies, enabling fastest time-to-market.
In conjunction with the NeuPro processor line, CEVA will also offer the NeuPro hardware engine as a Convolutional Neural Network (CNN) accelerator. When combined with the CEVA-XM4 or CEVA-XM6 vision platforms, this provides a flexible option for customers seeking a single unified platform for imaging, computer vision and neural network workloads.