$20 Million Investment in NeuReality to Advance Carbon-Neutral AI Data Centres
NeuReality, a company with a mission of redefining AI inference and data centre infrastructure for the AI era, announced it secured $20 million in fresh funding.
This financial injection stemmed from contributions by the European Innovation Council (EIC) Fund, Varana Capital, Cleveland Avenue, XT Hi-Tech, and OurCrowd. The influx of capital is poised to expedite the proliferation of NeuReality's NR1 AI Inference Solution across a broader customer and user base. This strategic move aims to transition NeuReality swiftly from the initial deployment phase to expansion across various markets, regions, and sectors within generative AI.
The recent funding round elevated NeuReality's total capital to $70 million, signalling a robust endorsement following the successful rollout of its 7nm AI inference server-on-a-chip, the NR1 NAPU (Network Addressable Processing Unit), produced by TSMC last year. As the cornerstone of the comprehensive NR1 AI Inference Solution, the NR1 NAPU has been instrumental in NeuReality's delivery of an AI-focused system architecture. This architecture empowers organisations to deploy generative AI applications and large language models (LLMs) efficiently, without the need for extensive investment in limited and often underutilised GPUs.
“In order to mitigate GPU scarcity, optimisation at the system and datacenter level are key,” said Naveen Rao, VP of Generative AI at Databricks. Rao, a NeuReality Board member and early investor in the startup added: “To enable greater access to compute for generative AI, we must remove market barriers to entry with a far greater sense of urgency. NeuReality's innovative system – engineered at the data centre architecture level – represents that tipping point.”
Enterprises face other big challenges in deploying trained AI models and apps, known as the AI Inference process. Running live AI data to solve a task can be complex and costly, with a record of poor scalability from AI accelerators and system bottlenecks caused by CPUs.
“Our disruptive AI Inference technology is unbound by conventional CPUs, GPUs, and NICs. We didn’t try to just improve an already flawed system. Instead, we unpacked and redefined the ideal AI Inference system from top to bottom and end to end, to deliver breakthrough performance, cost savings, and energy efficiency,” said NeuReality’s CEO Moshe Tanach, pointing to the paltry 30-40% utilisation rate of AI accelerators.
“Investing in more and more DLAs, GPUs, LPUs, TPUs…won’t address your core issue of system inefficiency,” remarked Tanach. “It's akin to installing a faster engine in your car to navigate through traffic congestion and dead ends — it simply won't get you to your destination any faster. NeuReality, on the other hand, provides an express lane for large AI pipelines, seamlessly routing tasks to purpose-built AI devices and swiftly delivering responses to your customers, while conserving both resources and capital.”
NeuReality's NR1-M and NR1-S systems, which are full-height PCIe cards that integrate easily into server racks, drive 100% AI accelerator utilisation. Each system houses internal NAPUs that run on any AI accelerator and operate independently from the CPU, eliminating a CPU requirement altogether. By connecting directly to Ethernet, NR1 efficiently manages AI queries from vast data pipelines originating from millions of users and billions of devices. Compatible server configurations were demonstrated by AMD, IBM, Lenovo, Qualcomm, and Supermicro at NeuReality's product launch at the SC23 international conference in Denver last November.
Since its Series A funding in 2022, NeuReality has taken delivery of its NR1 NAPUs. The firm is driving early AI deployments with select cloud service customers and enterprise customers in financial services, business services, and government, focusing on current natural language processing, automated speech recognition, recommendation systems, and computer vision. The additional $20 million propels NeuReality for broader AI Inference deployment as both conventional and generative AI applications surge in demand.
The investment of the EIC Fund, the venture arm of the European Commission’s EIC Accelerator program, shows its support for the firm’s solution which not only brings optimised performance, but also improved energy efficiency in AI deployment to the marketplace. The EIC investment further addresses two important industries for Europe – advanced semiconductors and AI, both of which are expected to be major drivers of economic growth in the coming years.
“We recognise the increasing importance of the European Union’s leadership in AI and blazing a different path of high efficiency versus high spending,” said Svetoslava Georgieva, Chair of the EIC Fund Board. “NeuReality’s vision and disruptive technology align with our commitment to fortifying the region’s deep tech and AI investments for a sustainable, AI-powered future.”
NeuReality had already secured a substantial grant from the EIC Accelerator program last year to support various development steps of the firm’s innovation aimed at solving the cost and complexity of AI Inference at its core architectural level.
“We appreciate this tremendous vote of confidence from the EU and recognise that the backing of influential investors and the European Union will drive us forward as we pioneer next-gen semiconductor advancements in AI Inference,” Tanach said. “It's a significant step towards a greener, more democratised AI future.”