Nvidia has reportedly spent more than $900 million in cash and stock to hire Enfabrica CEO Rochan Sankar and key team members, and to license the company’s networking technology. The deal underscores how central interconnects and memory fabrics have become to scaling AI data centers.

09/17/2025
Nvidia has entered into a talent and technology deal worth over $900 million with Enfabrica, the AI infrastructure startup behind a new class of networking and memory fabric chips. The transaction, reported by CNBC and Reuters, combines cash and stock, and brings Enfabrica’s CEO Rochan Sankar and other leaders into Nvidia while licensing the company’s core technology.
Rather than a traditional acquisition, this is closer to an “acquihire plus license” that locks in Enfabrica’s expertise in connecting huge fleets of GPUs as a single, efficient system.
As AI models grow, the bottleneck is shifting from raw compute to how fast data can move between GPUs, CPUs, memory, and storage. If the network cannot keep up, extremely expensive accelerators sit idle waiting for data. Enfabrica has focused on that exact pain point, building an accelerated compute fabric that can link tens of thousands of AI chips with high bandwidth and low latency.
By investing heavily in Enfabrica’s technology and team, Nvidia is signaling that solving interconnect and memory scaling is as important as shipping the next generation of GPUs. It is a bet that the winners in AI will be the ones who treat the data center as a single, composable system, not a pile of isolated servers.
Enfabrica’s architecture brings networking, PCIe/CXL connectivity, and pooled memory into a unified fabric designed for modern AI workloads. Their ACF-based systems and EMFASYS memory fabric aim to:
Connect very large GPU clusters over Ethernet while keeping latency low
Pool tens of terabytes of DDR5 memory so models are not constrained by on-card HBM alone
Improve GPU utilization and reduce total cost of ownership by avoiding stranded resources
In practical terms, that means AI data centers can serve larger context windows, more concurrent users, and more complex agentic workloads without linearly adding more GPUs.
Long before this deal, Enfabrica asked ANML to help translate a deep, technical story into a clear, confident brand and digital presence. We partnered with their team to design a future-facing website and campaign system that:
Positions Enfabrica as a leader in next-generation AI networking and memory fabrics
Explains complex concepts like elastic memory fabrics and accelerated compute fabrics in approachable language
Gives their sales and partnerships teams a platform that feels as advanced as the technology behind it
The goal was simple: make it obvious, in a few seconds, why Enfabrica matters to anyone trying to run AI at scale.
Seeing Nvidia make a nine-figure commitment to the team and technology is a strong validation of that story and of Enfabrica’s role in the AI infrastructure stack.
For CMOs, product leaders, and CTOs building AI platforms, this deal is another sign that infrastructure differentiation is shifting toward:
Interconnect and memory fabrics, not just faster chips
Open, composable architectures that can work across vendors and generations
Clear narratives that help non-specialists understand why these layers matter
Companies that sit in critical parts of the AI value chain now have an opportunity and a challenge. The opportunity is to define a clear category story around the specific bottleneck they solve. The challenge is to explain it in a way that resonates with buyers who are not deep protocol experts.
That is where brand, product storytelling, and UX need to keep pace with the technology. When your platform is solving problems at rack or fleet scale, the way you communicate that story can be as important as the silicon itself.
Nvidia reportedly committed more than $900 million in cash and stock to hire Enfabrica CEO Rochan Sankar and other key employees, and to license the company’s networking technology. The arrangement is structured as a talent and technology deal rather than a full corporate acquisition.
Traditional data center networks treat servers, accelerators, and memory as separate islands connected through multiple layers of switches and adapters. Enfabrica’s approach collapses many of those layers into a single fabric that can pool memory and move data with much lower latency, enabling larger, more efficient AI systems built on standard Ethernet.