ai/ml data center projects analysis
on this page
overview
ai/ml workloads are driving the most significant transformation in data center infrastructure history. 140 ai/ml projects represent 23.2% of the us data center database but account for disproportionate investment and power demand due to fundamentally different infrastructure requirements.
key statistics
Metric | Value |
Total AI/ML Projects | 140 (23.2% of database) |
Total Investment | $538.2B disclosed |
Average Investment | $6.9B per project |
Total Power Capacity | 65.2 GW disclosed |
Average Power | 717 MW per project |
Projects with Data | 78 investment / 91 power |
why ai/ml is different
traditional data centers operate at 15-20 kw per rack. ai/ml infrastructure requires:
- power density: 100-140 kw per rack (5-10x increase)
- cooling: liquid cooling mandatory (direct-to-chip or immersion)
- networking: 400-800 gb/s gpu interconnects (nvidia quantum-2 infiniband, spectrum-x ethernet)
- scale: gigawatt-scale projects increasingly common
- speed: compressed construction timelines (18-24 months typical)
leading ai deployments
these projects represent the cutting edge of ai infrastructure, demonstrating the scale and technical requirements driving the industry:
Project | Location | GPUs | Power (MW) | Investment | Status |
xAI Colossus | Memphis, TN | 230,000 (150K H100, 50K H200, 30K GB200) | 300 | Undisclosed | Operational (2024) |
CoreWeave Portfolio | 33 facilities (US & Europe) | 250,000+ (H100/H200/GB200/GB300) | 420 active / 2,200 contracted | $6B+ (Lancaster PA alone) | Expanding |
Meta Prometheus | New Albany, OH | 500,000+ (Blackwell/MI300/MTIA) | 1,020 | Part of Hyperion program | 2026 launch |
Project Jupiter/Stargate | Santa Teresa, NM | TBD | TBD | $165B | Announced (OpenAI) |
Stargate Abilene Campus | Abilene, TX | 100,000 per building capacity | 1,200 | $40B | Operational (Oracle/Crusoe) |
Meta 24K GPU Clusters | Undisclosed | 49,152 H100 (2x 24,576 clusters) | Undisclosed | Undisclosed | Operational (2024) |
Applied Digital Ellendale | Ellendale, ND | 50,000 H100 SXM capacity | 180 (400 campus / 1,000 pipeline) | CoreWeave $7B lease | Energized (2024) |
Crusoe Abilene Phase 1 | Abilene, TX | 100,000 per building capacity | 200 (1,200 full campus) | $15B joint venture | H1 2025 occupancy |
technology requirements
gpu specifications
GPU Model | Memory | Power (TDP) | Bandwidth | Deployment Status |
NVIDIA A100 | 40/80GB HBM2e | 250-400W | 1.6 TB/s | Widespread (previous gen) |
NVIDIA H100 | 80GB HBM3 | 350-700W (PCIe/SXM5) | 3 TB/s | Current generation |
NVIDIA H200 | 141GB HBM3e | 700W | 4.8 TB/s (+43% vs H100) | Deploying now |
NVIDIA B200 | 180-192GB HBM3e | 1,000W (600W typical) | 7.7 TB/s | Next generation |
NVIDIA B300 (Blackwell Ultra) | 288GB HBM3e | 1,400W | Enhanced | Cutting edge |
NVIDIA GB200 NVL72 | 13.5TB total (72 GPUs) | 120kW per rack | 1.44 exaflops | Rack-scale system |
NVIDIA GB300 NVL72 | 21TB total (72 GPUs) | 140kW per rack | 1.1 exaflops FP4 | CoreWeave first deployment |
AMD MI300X | 192GB HBM3 | Varies | 5.3 TB/s | Crusoe $400M order |
power density evolution
Infrastructure Type | Rack Power | Cooling Method | Deployment |
Traditional Enterprise | 5-10 kW | Air cooling | Legacy |
Traditional Hyperscale | 15-20 kW | Air cooling | Still common |
AI Training (H100/H200) | 100-130 kW | Direct liquid cooling | Current standard |
AI Training (GB200 NVL72) | 120 kW | Liquid-cooled rack | Deploying 2025 |
AI Training (GB300 NVL72) | 140 kW | Liquid-cooled rack | Cutting edge 2025 |
High-Density AI | 300+ kW | Immersion cooling | CyrusOne Intelliscale |
cooling technology
direct-to-chip liquid cooling (dominant):
- cold plates attached directly to gpu/cpu
- 70-98% heat capture at source
- vendors: vertiv (cdu 70-2300kw), supermicro (dlc-2), lenovo (neptune), dell, hpe cray
- coreweave: all new facilities liquid cooling foundation (130kw+ racks)
- lambda labs: dallas-fort worth facility liquid-cooled for highest-density gpus
immersion cooling (emerging):
- servers submerged in dielectric fluid
- single-phase or two-phase systems
- vendors: grc, liquidstack, asperitas
- best for 80-100kw+ rack densities
- market: 11.10b (2030), 17.91% cagr
networking requirements:
- nvidia quantum-2 infiniband: 400gb/s (ndr), up to 51.2tb/s aggregate
- deployments: coreweave (3.2tbps per vm), lambda labs, meta 24k clusters, azure nd h100 v5
- nvidia spectrum-x ethernet: 800gb/s with rdma
- deployment: xai colossus (100,000 h100 on single rdma fabric)
- nvlink 4.0: 900gb/s per gpu (3.6tb/s bisectional in 8-gpu systems)
geographic distribution
top 10 states for ai/ml projects
Rank | State | AI/ML Projects |
1 | Texas | 10 |
2 | Pennsylvania | 8 |
3 | South Carolina | 8 |
4 | Indiana | 6 |
5 | New York | 6 |
6 | Wisconsin | 6 |
7 | Arizona | 5 |
8 | California | 5 |
9 | Oklahoma | 5 |
10 | Wyoming | 5 |
clustering patterns
texas dominance (10 projects, 11.0 gw total state capacity):
- deregulated power market enables flexibility
- oncor interconnection queue: 186 gw pending
- major projects: stargate abilene (15b, 1.2 gw), lambda plano, coreweave austin/plano
pennsylvania emergence (8 projects, 16.9 gw total state capacity):
- marcellus shale natural gas advantage
- major project: coreweave lancaster ($6b, 100 mw scalable to 300 mw)
- homer city energy campus: 4.5 gw planned
emerging ai hubs:
- north dakota: applied digital ellendale (180 mw, 50k gpu capacity), coreweave lease ($7b)
- tennessee: xai colossus memphis (300 mw, 230k gpus operational)
- ohio: meta prometheus new albany (1,020 mw, 500k+ gpus, 2026)
- arizona: vermaland la osa (3 gw, 20b)
project status distribution
Status | Projects | Percentage |
Under Construction | 53 | 37.9% |
Planned | 42 | 30.0% |
Operational | 24 | 17.1% |
Announced | 10 | 7.1% |
Expansion | 9 | 6.4% |
Canceled | 2 | 1.4% |
key insight: 67.9% of ai/ml projects are under construction or planned, indicating massive near-term capacity additions (2025-2027).
top sponsors
Rank | Sponsor | AI/ML Projects |
1 | Microsoft | 10 |
2 | Google LLC / Google | 13 combined |
3 | Meta / Meta Platforms Inc. | 8 combined |
4 | Amazon Web Services | 4 |
5 | Tract | 4 |
6 | CoreSite | 4 |
7 | Aligned Data Centers | 4 |
8 | PowerHouse Data Centers | 4 |
analysis: why ai drives different infrastructure
training vs inference workloads
training (high compute intensity):
- massive parallel gpu clusters (10k-500k+ gpus)
- high-bandwidth networking critical (400-800gb/s gpu interconnects)
- power density: 100-140kw/rack
- examples: meta prometheus (500k gpus), xai colossus (230k gpus)
- cooling: liquid cooling mandatory
- location: near power sources, less latency-sensitive
inference (lower per-request compute, high throughput):
- distributed across multiple sites
- latency-sensitive (user-facing)
- lower power density possible but still 3-5x traditional
- examples: openai api infrastructure, google tpu v7 ironwood (inference-optimized)
- cooling: direct liquid or advanced air
- location: near users/edge locations
the gpu density challenge
rack-scale systems (nvidia nvl72 example):
- 72 gpus + 36 cpus per rack
- 120-140kw power per rack
- 13.5-21tb gpu memory per rack
- must deploy in multiples of 18 nodes
- requires liquid cooling infrastructure
cluster-scale challenges:
- 100,000 gpu cluster (xai colossus):
- 1,389 racks of 72 gpus each
- ~173 mw for gpu compute alone
- single rdma fabric connectivity
- 122 days construction time
- power delivery at this scale requires utility partnerships or on-site generation
geographic clustering drivers
power availability (primary constraint):
- texas: deregulated market, 186 gw queue
- pennsylvania: marcellus shale gas, 16.9 gw capacity
- ohio: competitive market, ample generation
- north dakota: low-cost power, minimal constraints
alternative power sources:
- on-site natural gas: 9 of 11 gigawatt projects
- nuclear partnerships: 24+ gw smr commitments (microsoft-constellation, amazon-x-energy, google-kairos, switch-oklo)
- renewable + battery: xai colossus (150 mw megapack battery backup)
- hydrogen: lambda terrasite-tx1 (1 gw campus, hydrogen-powered)
construction speed:
- xai colossus: 100,000 gpus in 122 days
- standard timeline: 18-24 months for ai facilities
- traditional hyperscale: 24-36 months
investment patterns
hyperscaler build-out (vertical integration):
- microsoft: 10 ai/ml projects
- google: 13 ai/ml projects
- meta: 8 ai/ml projects (including 1 gw prometheus)
- amazon: 4 ai/ml projects
specialized ai infrastructure operators:
- coreweave: 250k gpu fleet, 2.2 gw contracted capacity
- crusoe: 1.6 gw operational/construction, 10+ gw pipeline
- applied digital: ellendale hpc campus
- lambda labs: gigawatt-scale gpu cloud
financial sponsors entering ai:
- nvidia: $8.13b portfolio (applied digital, coreweave investments)
- blackstone: airtrunk $16b acquisition
- digitalbridge: 5.4 gw portfolio
- blue owl capital: crusoe $15b joint venture
project economics:
- average ai/ml project: $6.9b investment
- average power: 717 mw per project
- contrast traditional: $2-3b average, 100-200 mw typical
- premium pricing: ai colocation commands 2-3x traditional rates
complete ai/ml projects table
mega-projects (≥$10b or ≥1gw)
Project Name | Location | Sponsor | Power (MW) | Investment | Status |
Project Jupiter (Stargate Santa Teresa) | Santa Teresa, NM | BorderPlex Digital Assets / OpenAI | TBD | $165.0B | Announced |
Stargate Abilene Campus | Abilene, TX | OpenAI / Oracle / Crusoe | 1,200 | $40.0B | Operational |
Vermaland La Osa Data Center Park | Eloy, AZ | Vermaland LLC | 3,000 | $33.0B | Planned |
Vantage Data Centers - Frontier Campus | Various | Vantage Data Centers | 1,400 | $25.0B | Under Construction |
Google PJM Data Center Infrastructure | Pennsylvania | 670 | $25.0B | Planned | |
Tract Buckeye Data Center Park | Buckeye, AZ | Tract | 1,800 | $20.0B | Planned |
AWS AI Innovation Campuses | Various | Amazon Web Services | 960 | $20.0B | Under Construction |
Applied Digital Toronto AI Data Center | Toronto (Canada partner) | Applied Digital | 430 | $16.0B | Planned |
Pennsylvania Digital I (PAX) | Pennsylvania | Pennsylvania Data Center Partners | 1,350 | $15.0B | Planned |
Project Mica (Google AI Campus) | Various | Google LLC | 700 | $10.0B | Planned |
Meta Richland Parish (Hyperion Campus) | Louisiana | Meta Platforms, Inc. | 2,000 | $10.0B | Under Construction |
AWS Richmond County Campus | Georgia | Amazon Web Services | TBD | $10.0B | Announced |
Homer City Energy Campus | Pennsylvania | Homer City Redevelopment | 4,500 | $10.0B | Planned |
Data City Texas | Texas | Energy Abundance Development Corp | 5,000 | TBD | Planned |
Delta Gigasite / Fibernet MercuryDelta | Mississippi | Fibernet MercuryDelta LLC | 4,000 | TBD | Planned |
Joule Capital Partners - Millard County | Utah | Joule Capital Partners | 4,000 | TBD | Planned |
TECfusions Keystone Connect | Pennsylvania | TECfusions | 3,000 | TBD | Under Construction |
Shippingport Power Station | Pennsylvania | Frontier Group of Companies | 2,700 | $3.2B | Planned |
PowerHouse - Grand Prairie Campus | Texas | PowerHouse Data Centers | 1,800 | TBD | Under Construction |
Crusoe/Tallgrass AI Data Center | Wyoming | Crusoe Energy Systems | 1,800 | TBD | Announced |
Tract Silver Springs Park | Nevada | Tract | 1,600 | TBD | Planned |
Prometheus Hyperscale - Natrona/Converse | Wyoming | Prometheus Hyperscale | 1,500 | $0.5B | Planned |
data access
full project list: all 140 ai/ml projects available in database
- location:
support/datacenters/analysis/ai_ml_projects_simplified.json
- includes: project name, state, city, sponsor, power capacity, investment, status
- sortable by investment, power, sponsor, status, geography
dimensional research:
- gpu inventory: 1m+ deployed gpus tracked
- technology specs: h100/h200/b200/b300/gb200/gb300
- cooling vendors: vertiv, supermicro, grc, liquidstack, lenovo, dell, hpe
- networking: quantum-2 infiniband, spectrum-x ethernet specifications
- operators: coreweave, lambda, crusoe, applied digital, hyperscalers
future outlook (2025-2030)
projected growth
- ai/ml share: 23.2% (2025) → 40-50% (2030)
- average project size: 10-15b (2030)
- typical power: 717 mw average → 1-2 gw standard
- gpu clusters: 100k-500k gpus → 500k-2m gpus
- cooling: liquid cooling universal → immersion increasing for 300+ kw racks
technology evolution
gpu generations:
- current: h100/h200 (80-141gb, 350-700w)
- 2025-2026: b200/b300 (180-288gb, 600-1,400w)
- 2027+: next-gen blackwell, 2x performance/watt improvements
- rack systems: gb200/gb300 nvl72 (120-140kw) becoming standard
cooling infrastructure:
- direct-to-chip liquid cooling: dominant for 100-140kw racks
- immersion cooling: growing for 300+ kw racks
- hybrid systems: liquid for gpus/cpus, air for ancillary
- vendors scaling: vertiv cdu 2300 (2.3mw capacity), supermicro in-row cdu (1.8mw)
networking bandwidth:
- 2025: 400gb/s quantum-2, 800gb/s spectrum-x
- 2026-2027: 800gb/s, 1.6tb/s standards
- 2028+: 3.2tb/s and beyond
power solutions
grid constraints driving alternatives:
- on-site generation: natural gas increasingly common (9 of 11 gw+ projects)
- nuclear smrs: 24+ gw committed (2027-2035 deployment)
- battery integration: tesla megapack (xai 150mw backup)
- hydrogen: lambda terrasite-tx1 (1 gw hydrogen-powered)
interconnection queue crisis:
- oncor (texas): 186 gw queue
- pjm (mid-atlantic): 270+ gw queue
- multi-year delays standard
- forcing geographic diversification to less constrained markets
market dynamics
consolidation likely:
- specialized ai infrastructure operators (coreweave, crusoe, lambda) gaining share
- hyperscalers building captive capacity (meta prometheus, google project mica)
- traditional colocation adapting (equinix 100 liquid-cooled dcs, cyrusone intelliscale)
- financial sponsors backing buildout (nvidia, blackstone, digitalbridge, kkr)
pricing premium:
- ai colocation: 100-150 traditional)
- gpu cloud: premium for h100/h200 vs commodity compute
- long-term contracts: 3-7 year commitments typical
- pre-leasing common: projects leased before construction
regulatory evolution:
- permitting: ai-specific review processes emerging
- power allocation: utility commissions prioritizing economic development
- environmental: renewable energy requirements increasing
- community impact: noise, water use, thermal pollution concerns
using this analysis
for infrastructure operators
- capacity planning: 717 mw average project size, 100-140kw rack density
- technology roadmap: liquid cooling mandatory, gb200/gb300 deployment timeline
- geographic strategy: cluster analysis shows power availability primary driver
- competitive intelligence: 140 projects tracked with investment, sponsor, status
for investors
- market size: $538b disclosed investment across 140 projects
- growth trajectory: 67.9% projects under construction/planned (near-term capacity)
- key players: hyperscalers (microsoft, google, meta, amazon) + specialists (coreweave, crusoe, lambda)
- deal flow: mega-projects ($10b+) increasingly common
for technology vendors
- cooling demand: liquid cooling mandatory for 140 projects (vertiv, supermicro, grc, liquidstack)
- gpu sales: 1m+ deployed, 500k-2m+ needed by 2030
- networking: 400-800gb/s interconnects (nvidia quantum-2, spectrum-x)
- power infrastructure: smr partnerships, battery storage, on-site generation
for policymakers
- economic development: average $6.9b investment, 717 mw power per project
- infrastructure planning: 65.2 gw disclosed capacity from ai/ml alone
- grid impact: concentrated demand requiring utility coordination
- incentive design: competition for mega-projects intense
related resources
- gpu infrastructure - detailed gpu specifications and deployment tracking
- liquid cooling - cooling technology vendors and requirements
- power infrastructure - nuclear partnerships, grid constraints, alternative solutions
- project jupiter/stargate - $165b new mexico mega-project analysis
- state rankings - geographic distribution and competitive analysis
- mega-projects - projects >$10b or >1gw
ai/ml infrastructure represents the most significant data center buildout in history. the 140 projects tracked here demonstrate fundamentally different requirements: 5-10x power density, mandatory liquid cooling, gigawatt-scale projects, and compressed timelines. success requires solving the power delivery challenge through grid upgrades, on-site generation, or nuclear partnerships. the next 5 years (2025-2030) will determine which regions and operators can deliver the infrastructure needed for ai leadership.