Expedera Launches Origin Evolution NPU IP for Generative AI in Edge Devices

Expedera's Launch of Origin Evolution NPU IP



On May 20, 2025, Expedera Inc., a prominent innovator in scalable Neural Processing Unit (NPU) semiconductor intellectual property (IP), introduced its latest breakthrough, the Origin Evolution NPU IP. This cutting-edge technology is tailored to advance Generative Artificial Intelligence (GenAI) performance specifically for edge devices. The unique architecture enables efficient processing, addressing the challenges associated with large language models (LLMs) and traditional neural networks in resource-constrained environments.

As the demand for AI capabilities rises in everyday devices, Expedera's new product aims to match the computational demands of modern applications without compromising performance. Origin Evolution incorporates a unique packet-based architecture that ensures exceptional NPU efficiency. It significantly enhances the ability to run inference for LLMs on edge hardware, a crucial advancement that notably decreases latency and dodges the security issues inherent in cloud-based solutions.

Unique Challenges in Edge Computing


Running LLMs and neural networks on limited-resource devices entails a series of challenges due to their extensive computational requirements and large model sizes. Consequently, there needs to be a specialized hardware setup for edge applications that can balance power, performance, area (PPA), and memory limitations. As noted by Siyad Ma, CEO and co-founder of Expedera, "Origin Evolution is a radical advancement providing an AI inference engine with out-of-the-box compatibility with popular LLM and CNN networks."

This operational flexibility allows it to serve diverse applications across smartphones, automobiles, and data centers.

Performance and Scalability


One of the hallmark features of Origin Evolution is its impressive scalability, reaching 128 TFLOPS for a single core and potentially exceeding petaflops with multiple cores deployed in unison. This adaptability allows various configuration choices to optimize performance according to the specific needs of applications. Furthermore, the NPU reduces memory and system power consumption significantly while boosting processor utilization rates. Notably, the packet-based processing reduces external memory movements by over 75% when processing complex models like Llama 3.2 1B and Qwen2 1.5 B.

In memory-constrained scenarios, Origin Evolution is capable of delivering thousands of effective TFLOPS and processing multiple tokens per square millimeter of silicon. The architecture is designed to support both custom 'black box' layers and acknowledged networks such as Llama3, ChatGLM, MobileNet, and Yolo, providing broad compatibility with contemporary computing demands.

Implementing Existing Models


The innovation does not stop with scalability. Users are enabled to deploy pre-existing trained models without any loss in accuracy or need for retraining, providing a seamless transition to enhanced efficiency and performance. Unlike prior models, Origin Evolution's packet-based infrastructure mitigates large memory movements, making it adept at handling the considerable demands posed by LLMs. The efficient routing of network packets through specialized processing blocks allows the system to adaptively manage various operations and data types, driving performance across multiple workloads efficiently.

Origin Evolution boasts a high-speed external memory streaming interface that complies with cutting-edge DRAM and HBM standards. Coupled with an advanced software stack, it offers full support for numerous network representations and compatibility with popular frameworks, making it a versatile choice.

Conclusion


In summary, Expedera's Origin Evolution NPU IP marks a significant shift in how generative AI can be amplified in edge devices. By combining technical prowess and a commitment to usability, Expedera addresses the growing demands of AI-powered applications in various industries. As Expedera continues to pave the way for future innovations in AI hardware, its latest product promises to empower manufacturers and developers alike to better harness the capabilities of AI at the edge.

For additional details on the Origin Evolution NPU IP, visit Expedera's blog.

About Expedera


With its headquarters in Santa Clara, California, Expedera is committed to delivering customizable NPU semiconductor solutions. The company’s IP has achieved recognition through impressive deployments across millions of devices, underlining its influence in transforming AI inference applications within edge computing and data centers.

Topics Consumer Technology)

【About Using Articles】

You can freely use the title and article content by linking to the page where the article is posted.
※ Images cannot be used.

【About Links】

Links are free to use.