Founded in Israel, NeuReality specializes in providing hardware infrastructure to streamline computation processes to accelerate the development of AI applications. Its technology assists AI inferencing deployment by offering a comprehensive system solution that seamlessly supports a vast array of deep learning models.
NeuRealitys’ infrastructure includes an architecture called the Network Addressable Processing Unit (NAPU) to maximize the capabilities of Deep Learning Accelerators (DLAs). This design facilitates inference through purpose-built NAPUs that are specifically crafted for AI inference tasks, enhancing performance and efficiency.
The NR1 AI Inference Solution by NeuReality is a comprehensive system comprising architecture, hardware, and software platforms. Launched in November 2023, it intends to simplify the installation, utilization, and management of AI, making it accessible and manageable for users.
The hardware offerings of the solution include,
1. NR1 Network Addressable Processing Unit : A network addressable inference Server-on-a-Chip with an embedded Neural Network Engine, which is a workflow-optimized hardware device with specialized processing units, native network capabilities, and virtualization capabilities
2. NR1-M™ AI Inference Module : A full-height, double-wide PCIe card containing one NR1 Network Addressable Processing Unit (NAPU) system-on-chip and a network addressable Inference Server that can connect to an external Deep Learning Accelerator (DLA)
3. NR1-S AI Inference Appliance : An optimized design for an inference server that contains NR1-M modules with the NR1 NAPU, which enables disaggregated AI service in a scalable and efficient architecture
The software component of the solution is designed to be compatible with any trained model from any development environment. It includes tools that handle the entire AI pipeline and enable seamless connection of AI workflows to any environment.
NeuReality differs from traditional AI chip connectivity in replacing CPU-centric support with an AI-centric platform. This allows components to communicate directly with AI accelerators, eliminating the need for commands to pass through a central processing unit. Additionally, the technology enables AI accelerators to connect with each other, enhancing efficiency and performance in AI applications. It also claims its AI-centric approach reduces complexity, cost, and power usage. The firm claims it can reduce inference costs by 90%.
Key customers and partnerships
In November 2021, NeuReality teamed up with IBM to develop advanced AI inference technology by enhancing its NR1 chip and system for use in IBM's Hybrid Cloud.
In September 2021, the company partnered with Xilinx to develop AI-centric platforms that improve real-world AI applications. This collaboration was intended to focus on NeuReality's AI-centric inference platform NR1-P and address existing issues with AI deployment including cost and complexity.
By using this site, you agree to allow SPEEDA Edge and our partners to use cookies for analytics and personalization. Visit our privacy policy for more information about our data collection practices.