Embedded engineering services handle MCU hardware and firmware design, helping customers deliver the world’s first Network Addressable Processing Unit to decrease cost and energy in AI data centres
Arrow Electronics supported the development of the world’s first 7nm Network Addressable Processing Unit (NR1 NAPU™) housed in the complete NR1-S™ AI Inference Appliance from NeuReality – delivering competitive advantages in cost and power savings versus traditional CPU-centric architecture.
The NR1-S, when paired with AI accelerators in an AI inference server, reduces data centre costs by up to 90 per cent and increases energy efficiency by up to 15 times while delivering linear scalability without performance drop-offs or lags as additional AI accelerators are added, according to NeuReality.
Bringing extensive embedded design skills to the project, Arrow’s in-house experts provided firmware and hardware design guidance, developed and validated power management firmware. Arrow also handled debugging of the microcontroller (MCU) and platform power flows to support the successful NAPU bring-up, NR1-S and integrated NeuReality software – all performed in record time.
The Arrow team also helped select the most suitable MCU to provide the interface crosslink between system components of the PCIe card and server.
The NR1 NAPU™ is a custom server-on-a-chip that provides the full performance of each dedicated AI accelerator from approximately 30 per cent today to 100 per cent full utilisation – boosting total output and reducing silicon waste. The NAPU not only migrates services including network termination, quality of service, and AI data pre-and post-processing, but also improves data flow for the high volume and variety of AI pipelines.
The NeuReality system architecture eliminates the performance bottleneck caused by traditional CPU-centric system architecture relied upon today by all AI Inference systems and hardware manufacturers. As a result, the NR1-S increases cost savings and energy efficiency of running high-volume, high-variety AI data pipelines – a top financial concern in the deployment of today’s power-hungry conventional and generative AI applications.
“Our NAPU addresses the major bottlenecks that restrict performance in today’s AI accelerators, such as power management and transferring data from the network into the AI accelerator, typically a GPU, FPGA or ASIC,” said Eli Bar-Lev, director of hardware at NeuReality. “Arrow’s support with the hardware and firmware for power management and thermal engineering allowed us to focus resources on a complete silicon-to-software AI inference solution which will reduce the AI market barriers for governments and businesses around the world.”
“This exciting project can potentially make cloud and on-premise enterprise AI inferencing more affordable and faster, thereby increasing access to valuable services in healthcare and medical imaging, banking and insurance, and AI-driven customer call centres and virtual assistants,” said Vitali Damasevich, director engineering Eastern Europe and engineering solutions centre EMEA.
Credit:NeuReality
Lenovo has redefined laptop innovation with the introduction of the Yoga Slim 9i, the world’s…
TDK Corporation (TSE 6762) announces the introduction of multiple option configurations for the TDK-Lambda brand…
Introducing MECC: A seamless, scalable solution to unlock global IoT connectivity using nRF9151 module Nordic…
TOLL package 650V GaN HEMTs contribute to improving power supply efficiency Willich-Münchheide, Germany, March 05,…
Development tool offers powerful debugging at a more affordable price for professional engineers, students and…
Infineon Technologies AG (FSE: IFX / OTCQX: IFNNY) today announced the addition of P-channel power MOSFETs…