Cadence Accelerates On-Device and Edge AI Performance, Efficiency with New Neo NPU IP and NeuroWeave SDK for Silicon Design
September 14, 2023 | Cadence Design Systems, Inc.Estimated reading time: 4 minutes

Cadence Design Systems, Inc. unveiled its next-generation AI IP and software tools to address the escalating demand for on-device and edge AI processing. The new highly scalable Cadence Neo Neural Processing Units (NPUs) deliver a wide range of AI performance in a low-energy footprint, bringing new levels of performance and efficiency to AI SoCs. Delivering up to 80 TOPS performance in a single core, the Neo NPUs support both classic and new generative AI models and can offload AI/ML execution from any host processor—including application processors, general-purpose microcontrollers and DSPs—with a simple and scalable AMBA® AXI interconnect. Complementing the AI hardware, the new NeuroWeave™ Software Development Kit (SDK) provides developers with a “one-tool” AI software solution across Cadence AI and Tensilica® IP products for no-code AI development.
The new highly scalable Cadence® Neo™ Neural Processing Units (NPUs) deliver a wide range of AI performance in a low-energy footprint, bringing new levels of performance and efficiency to AI SoCs. Complementing the AI hardware, the new NeuroWeave™ Software Development Kit (SDK) provides a "one-tool" AI software solution across Cadence AI and Tensilica IP products for no-code AI development. (Photo: Business Wire)
The new highly scalable Cadence® Neo™ Neural Processing Units (NPUs) deliver a wide range of AI performance in a low-energy footprint, bringing new levels of performance and efficiency to AI SoCs. Complementing the AI hardware, the new NeuroWeave™ Software Development Kit (SDK) provides a "one-tool" AI software solution across Cadence AI and Tensilica IP products for no-code AI development. (Photo: Business Wire)
“While most of the recent attention on AI has been cloud-focused, there are an incredible range of new possibilities that both classic and generative AI can enable on the edge and within devices,” said Bob O’Donnell, president and chief analyst at TECHnalysis Research. “From consumer to mobile and automotive to enterprise, we’re embarking on a new era of naturally intuitive intelligent devices. For these to come to fruition, both chip designers and device makers need a flexible, scalable combination of hardware and software solutions that allow them to bring the magic of AI to a wide range of power requirements and compute performance, all while leveraging familiar tools. New chip architectures that are optimized to accelerate ML models and software tools with seamless links to popular AI development frameworks are going to be incredibly important parts of this process.”
The flexible Neo NPUs are well suited for ultra-power-sensitive devices as well as high-performance systems with a configurable architecture, enabling SoC architects to integrate an optimal AI inferencing solution in a broad range of products, including intelligent sensors, IoT and mobile devices, cameras, hearables/wearables, PCs, AR/VR headsets and advanced driver-assistance systems (ADAS). New hardware and performance enhancements and key features/capabilities include:
- Scalability: Single-core solution is scalable from 8 GOPS to 80 TOPS, with further extension to hundreds of TOPS with multicore
- Broad configuration range: supports 256 to 32K MACs per cycle, allowing SoC architects to optimize their embedded AI solution to meet power, performance and area (PPA) tradeoffs
- Integrated support for a myriad of network topologies and operators: enables efficient offloading of inferencing tasks from any host processor—including DSPs, general-purpose microcontrollers or application processors—significantly improving system performance and power
- Ease of deployment: shortens the time to market to meet rapidly evolving next-generation vision, audio, radar, natural language processing (NLP) and generative AI pipelines
- Flexibility: Support for Int4, Int8, Int16, and FP16 data types across a wide set of operations that form the basis of CNN, RNN and transformer-based networks allows flexibility in neural network performance and accuracy tradeoffs
- High performance and efficiency: Up to 20X higher performance than the first-generation Cadence AI IP, with 2-5X the inferences per second per area (IPS/mm2) and 5-10X the inferences per second per Watt (IPS/W)
Since software is a critical part of any AI solution, Cadence also upgraded its common software toolchain with the introduction of the NeuroWeave SDK. Providing customers with a uniform, scalable and configurable software stack across Tensilica DSPs, controllers and Neo NPUs to address all target applications, the NeuroWeave SDK streamlines product development and enables an easy migration as design requirements evolve. It supports many industry-standard domain-specific ML frameworks, including TensorFlow, ONNX, PyTorch, Caffe2, TensorFlow Lite, MXNet, JAX and others for automated end-to-end code generation; Android Neural Network Compiler; TF Lite Delegates for real-time execution; and TensorFlow Lite Micro for microcontroller-class devices.
“For two decades and with more than 60 billion processors shipped, industry-leading SoC customers have relied on Cadence processor IP for their edge and on-device SoCs. Our Neo NPUs capitalize on this expertise, delivering a leap forward in AI processing and performance,” said David Glasco, vice president of research and development for Tensilica IP at Cadence. “In today’s rapidly evolving landscape, it’s critical that our customers are able to design and deliver AI solutions based on their unique requirements and KPIs without concern about whether future neural networks are supported. Toward this end, we’ve made significant investments in our new AI hardware platform and software toolchain to enable AI at every performance, power and cost point and to drive the rapid deployment of AI-enabled systems.”
“At Labforge, we use a cluster of Cadence Tensilica DSPs in our Bottlenose smart camera product line to enable best-in-class AI processing for power-sensitive edge applications,” said Yassir Rizwan, CEO of Labforge, Inc. “Cadence’s AI software is an integral part of our embedded low power AI solution, and we’re looking forward to leveraging the new capabilities and higher performance offered by Cadence’s new NeuroWeave SDK. With an end-to-end compiler toolchain flow, we can better solve challenging AI problems in automation and robotics—accelerating our time to market to capitalize on generative AI-based application demand and opening new market streams that may not have been possible otherwise.”
The Neo NPUs and the NeuroWeave SDK support Cadence’s Intelligent System Design™ strategy by enabling pervasive intelligence through SoC design excellence.
Testimonial
"Our marketing partnership with I-Connect007 is already delivering. Just a day after our press release went live, we received a direct inquiry about our updated products!"
Rachael Temple - AlltematedSuggested Items
FuriosaAI Closes $125M Investment Round to Scale Production of Next-Gen AI Inference Chip
07/31/2025 | BUSINESS WIREFuriosaAI, a semiconductor company building a new foundation for AI compute, today announced it has completed a $125 million Series C bridge funding round. The investment continues a period of significant momentum for Furiosa as global demand for high-performance, efficient AI infrastructure soars.
Siemens, PTC, and Dassault Systèmes Named Leaders in ABI Research's PLM Assessment for Large Discrete Manufacturers
07/28/2025 | PRNewswireThe Product Lifecycle Management (PLM) market has witnessed significant developments over the past year with the rise of Software-as-a-Service (SaaS), digital twins, and Generative AI (Gen AI) becoming integral for large discrete manufacturers.
Elementary, Mr. Watson: Rein in Your Design Constraints
07/10/2025 | John Watson -- Column: Elementary, Mr. WatsonI remember the long hours spent at the light table, carefully laying down black tape to shape each trace, cutting and aligning pads with surgical precision on sheets of Mylar. I often went home with nicks on my fingers from the X-Acto knives and bits of tape all over me. It was as much an art form as it was an engineering task—tactile and methodical, requiring the patience of a sculptor. A lot has changed in PCB design over the years.
Stop Using Spreadsheets—and You Can Quote Me on That
07/03/2025 | Nolan Johnson, SMT007 MagazineMeeting changing business needs and a thriving market sometimes means process efficiencies. Not all those efficiency improvements take place on the shop floor, however. For EMS companies, growth can also mean changing out the business operations software to have one more chance to close a business deal that better aligns with the company’s sweet spot. One of the pressing issues for EMS companies is the great deal of attention on sales and quoting software solutions and how to make them perform better. Chintan Sutaria, the founder and former CEO of CalcuQuote now working on other projects at OpenJar, explains.
IPC-CFX, 2.0: How to Use the QPL Effectively
07/02/2025 | Chris Jorgensen, Global Electronics AssociationIn part one of this series, we discussed the new features in CFX Version 2.0 and their implications for improved inter-machine communication. But what about bringing this new functionality to the shop floor? The IPC-CFX-2591 QPL is a powerful technical resource for manufacturers seeking CFX-enabled equipment. The Qualified Product List (QPL) helps streamline equipment selection by listing models verified for CFX compliance through a robust third-party virtual qualification process.