Nvidia, EPRI, Prologis, and InfraPartners Team Up to Build Smaller Data Centers

This post was originally published on Data Center Knowledge

Nvidia continued its partnership streak on Tuesday, announcing plans to team up with EPRI, Prologis, and InfraPartners to study smaller-scale data centers for distributed inference, which would be used for real-time processing across various sectors. 

The partners will examine the deployment of so-called micro data centers – ranging from 5 to 20 MW – located at or near utility substations to enable quicker setup. The goal is to deliver inference capabilities near the point of data creation while optimizing underutilized infrastructure and easing strain on congested transmission systems, the companies said.

With five pilot sites planned across the US by year-end 2026, the collaboration aims to establish a replica model for rapid, scalable deployment.

“AI is transforming every industry, and the energy system will need to continue to evolve to meet increasing demand,” EPRI President and CEO Arshad Mansoor said in a release. “Using existing grid capacity to bring inference compute closer to where it’s needed – quickly and reliably – is a win for all.”

Related:Nvidia, Dassault Systèmes Partner for Digital Twin AI Platform

Inference, the process of using trained AI models to generate output, has different infrastructure requirements than training. Demand for AI inference is exploding as workloads shift

Read the rest of this post, which was originally published on Data Center Knowledge.

Previous Post

Community-First Data Centers: Engagement That Works

Next Post

Passing the Torch: Building a Workforce for the Next Generation of Data Centers