Expedera provides scalable Neural Processing Unit (NPU) semiconductor intellectual property (IP) for edge AI inference applications. The company's flagship product line, Origin, offers neural engine IP solutions designed to improve performance, power efficiency, and latency while reducing costs and complexity in AI processing. Expedera's NPU architecture utilizes a unique packet-based approach, which enables parallel execution across multiple layers and eliminates the need for hardware-specific optimizations. This allows customers to run trained neural networks without reducing model accuracy.
Expedera's Origin NPUs support generative AI on edge devices, including native support for large language models (LLMs) and stable diffusion. The Origin product family scales from three GOPS to 128 TOPS per core, with options for up to PetaOps performance. Expedera claims its NPUs can achieve up to 18 TOPS/W and TOPS/mm², with processor utilization averaging 80% compared to the 20-40% industry norm. The company's solutions have been deployed in over 10 million consumer devices.
Expedera's NPU IP is designed for applications including smartphones, automotive advanced driver assistance systems (ADAS), virtual/augmented reality devices, smart home technology, industrial automation, and entertainment systems. The company offers multiple product lines tailored for different performance needs, from the low-power E1 series for always-on applications to the high-performance E8 series for automotive and data center use cases.
By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.