AI is rewriting the rules of data center design, but most data centers werenโt built with AI in mind.
Across the industry, operators are facing the same challenge: existing facilities designed for 5-15 kW per rack are now expected to support AI workloads pushing 50 kW, 100 kW, and beyond. The instinctive response is often drastic: major mechanical rebuilds, new data halls, or entirely new sites. But there is a way to plan a retrofit that isnโt so interruptive to everyday operations.
High-density AI capacity does not require rebuilding your data center from scratch. It requires rethinking how cooling is delivered.
Will Most of the Data Center Growth Be Greenfield?
Despite the attention on greenfield hyperscale campuses, the majority of near-term AI capacity will come from retrofitting existing infrastructure:
- Colocation facilities upgrading halls for AI tenants
- Hyperscalers densifying live campuses faster than utilities can expand
- Enterprises repurposing legacy data halls for AI and HPC
These sites share common constraints:
- Limited white space
- Fixed floor layouts
- Existing power and mechanical systems
- Tight deployment timelines
Tearing everything out isnโt just expensive, itโs often impossible.
Why Traditional Cooling Retrofits Fall Short
When operators attempt to retrofit for AI using conventional approaches, they quickly hit limits.
In-row and In-Rack CDUs
Originally designed to extend air cooling, these solutions:
- Consume valuable white space
- Scale unevenly as loads increase
- Add operational complexity inside the data hall
- Become difficult to rebalance as AI workloads shift
Nautilus offers an in-row CDU, the RCD, because some designs and tenants still require row-level cooling. However, when considering efficiency and scale, facility-wide cooling is often the better approach.
The Retrofit Breakthrough: Facility-Scale Liquid Cooling
The most effective AI retrofits donโt start inside the rack, they start at the facility level.
Facility-scale Cooling Distribution Units (CDUs) enable operators to introduce high-density liquid cooling without reworking the data hall itself.
Key characteristics that make this possible:
- Off whitespace floor, skidded deployment
- Multi-megawatt scale capacity per unit
- Designed for simple parallel scalability
- Support for hybrid environments (air + liquid)
- Compatibility with multiple cooling methods, including direct-to-chip, rear-door, immersion, and hybrid approaches
Instead of threading complexity through the data hall, facility-scale CDUs centralize it where it belongs.
Retrofitting Without Disruption
One of the biggest fears in retrofit projects is operational disruption. Facility-scale liquid cooling directly addresses this concern.
Because cooling capacity is added outside the white space:
- Live environments remain operational
- Construction inside the data hall is minimized
- AI capacity can be phased in alongside existing workloads
This allows operators to selectively convert portions of a facility to high-density AI while maintaining legacy systems elsewhere.
Speed Delivers Operational Agility
AI demand is moving faster than traditional infrastructure timelines.
Retrofit-friendly cooling solutions must be:
- Prefabricated, not custom-built
- Deployable in weeks or months, not years
- Designed to adapt as workloads evolve
Facility-scale CDUs succeed here because they behave like infrastructure products, not bespoke mechanical projects.
What This Looks Like in Practice
In real-world retrofit deployments, operators are increasingly turning to facility-scale CDUs, like Nautilusโ EcoCore FCD, to add AI-ready liquid cooling capacity without rebuilding their data halls.
EcoCore FCD was designed specifically for high-density and retrofit-constrained environments:
- Delivers megawatt-scale cooling per unit
- Deploys off-floor, preserving white space
- Integrates with existing cooling sources and legacy infrastructure
- Scales incrementally as AI demand grows
Its design is informed by hundreds of thousands of hours of real-world liquid cooling operation, where pressure management, reliability, and ease of integration matter more than theoretical models.
Designing to Future-Proof Your Workload
A successful retrofit isnโt just about meeting todayโs rack density, itโs about planning for future scale.
Facility-scale cooling enables:
- Higher density ceilings without re-architecting
- Cleaner upgrades as GPU generations evolve
- Long-term flexibility in cooling allocation
The result is an AI-ready facility that can evolve without repeated mechanical interventions.
Retrofitting for AI Successfully
AI doesnโt require operators to abandon their existing data centers; it requires them to stop forcing legacy cooling models to do jobs they were never designed for.
Facility-scale liquid cooling provides a pragmatic retrofit path:
- High-density capability
- Faster deployment
- Minimal disruption
- Infrastructure-level scalability
Retrofitting for AI isnโt about rebuilding everything. Itโs about putting the right cooling infrastructure in the right place and letting the data hall do what it does best.
Nautilus pioneered liquid cooling over 5 years ago. Our technologies are operator-built and proven with over 500K+ unit hours of runtime globally. Connect with one of our experts to support your next project.