ai/ml data center projects analysis

overview

ai/ml workloads are driving the most significant transformation in data center infrastructure history. 140 ai/ml projects represent 23.2% of the us data center database but account for disproportionate investment and power demand due to fundamentally different infrastructure requirements.

key statistics

MetricValue
Total AI/ML Projects140 (23.2% of database)
Total Investment$538.2B disclosed
Average Investment$6.9B per project
Total Power Capacity65.2 GW disclosed
Average Power717 MW per project
Projects with Data78 investment / 91 power

why ai/ml is different

traditional data centers operate at 15-20 kw per rack. ai/ml infrastructure requires:

  • power density: 100-140 kw per rack (5-10x increase)
  • cooling: liquid cooling mandatory (direct-to-chip or immersion)
  • networking: 400-800 gb/s gpu interconnects (nvidia quantum-2 infiniband, spectrum-x ethernet)
  • scale: gigawatt-scale projects increasingly common
  • speed: compressed construction timelines (18-24 months typical)

leading ai deployments

these projects represent the cutting edge of ai infrastructure, demonstrating the scale and technical requirements driving the industry:

ProjectLocationGPUsPower (MW)InvestmentStatus
xAI ColossusMemphis, TN230,000 (150K H100, 50K H200, 30K GB200)300UndisclosedOperational (2024)
CoreWeave Portfolio33 facilities (US & Europe)250,000+ (H100/H200/GB200/GB300)420 active / 2,200 contracted$6B+ (Lancaster PA alone)Expanding
Meta PrometheusNew Albany, OH500,000+ (Blackwell/MI300/MTIA)1,020Part of Hyperion program2026 launch
Project Jupiter/StargateSanta Teresa, NMTBDTBD$165BAnnounced (OpenAI)
Stargate Abilene CampusAbilene, TX100,000 per building capacity1,200$40BOperational (Oracle/Crusoe)
Meta 24K GPU ClustersUndisclosed49,152 H100 (2x 24,576 clusters)UndisclosedUndisclosedOperational (2024)
Applied Digital EllendaleEllendale, ND50,000 H100 SXM capacity180 (400 campus / 1,000 pipeline)CoreWeave $7B leaseEnergized (2024)
Crusoe Abilene Phase 1Abilene, TX100,000 per building capacity200 (1,200 full campus)$15B joint ventureH1 2025 occupancy

technology requirements

gpu specifications

GPU ModelMemoryPower (TDP)BandwidthDeployment Status
NVIDIA A10040/80GB HBM2e250-400W1.6 TB/sWidespread (previous gen)
NVIDIA H10080GB HBM3350-700W (PCIe/SXM5)3 TB/sCurrent generation
NVIDIA H200141GB HBM3e700W4.8 TB/s (+43% vs H100)Deploying now
NVIDIA B200180-192GB HBM3e1,000W (600W typical)7.7 TB/sNext generation
NVIDIA B300 (Blackwell Ultra)288GB HBM3e1,400WEnhancedCutting edge
NVIDIA GB200 NVL7213.5TB total (72 GPUs)120kW per rack1.44 exaflopsRack-scale system
NVIDIA GB300 NVL7221TB total (72 GPUs)140kW per rack1.1 exaflops FP4CoreWeave first deployment
AMD MI300X192GB HBM3Varies5.3 TB/sCrusoe $400M order

power density evolution

Infrastructure TypeRack PowerCooling MethodDeployment
Traditional Enterprise5-10 kWAir coolingLegacy
Traditional Hyperscale15-20 kWAir coolingStill common
AI Training (H100/H200)100-130 kWDirect liquid coolingCurrent standard
AI Training (GB200 NVL72)120 kWLiquid-cooled rackDeploying 2025
AI Training (GB300 NVL72)140 kWLiquid-cooled rackCutting edge 2025
High-Density AI300+ kWImmersion coolingCyrusOne Intelliscale

cooling technology

direct-to-chip liquid cooling (dominant):

  • cold plates attached directly to gpu/cpu
  • 70-98% heat capture at source
  • vendors: vertiv (cdu 70-2300kw), supermicro (dlc-2), lenovo (neptune), dell, hpe cray
  • coreweave: all new facilities liquid cooling foundation (130kw+ racks)
  • lambda labs: dallas-fort worth facility liquid-cooled for highest-density gpus

immersion cooling (emerging):

  • servers submerged in dielectric fluid
  • single-phase or two-phase systems
  • vendors: grc, liquidstack, asperitas
  • best for 80-100kw+ rack densities
  • market: 4.87b(2025)4.87b (2025) → 11.10b (2030), 17.91% cagr

networking requirements:

  • nvidia quantum-2 infiniband: 400gb/s (ndr), up to 51.2tb/s aggregate
    • deployments: coreweave (3.2tbps per vm), lambda labs, meta 24k clusters, azure nd h100 v5
  • nvidia spectrum-x ethernet: 800gb/s with rdma
    • deployment: xai colossus (100,000 h100 on single rdma fabric)
  • nvlink 4.0: 900gb/s per gpu (3.6tb/s bisectional in 8-gpu systems)

geographic distribution

top 10 states for ai/ml projects

RankStateAI/ML Projects
1Texas10
2Pennsylvania8
3South Carolina8
4Indiana6
5New York6
6Wisconsin6
7Arizona5
8California5
9Oklahoma5
10Wyoming5

clustering patterns

texas dominance (10 projects, 11.0 gw total state capacity):

  • deregulated power market enables flexibility
  • oncor interconnection queue: 186 gw pending
  • major projects: stargate abilene (40b,1.2gw),crusoeabilene(40b, 1.2 gw), crusoe abilene (15b, 1.2 gw), lambda plano, coreweave austin/plano

pennsylvania emergence (8 projects, 16.9 gw total state capacity):

  • marcellus shale natural gas advantage
  • major project: coreweave lancaster ($6b, 100 mw scalable to 300 mw)
  • homer city energy campus: 4.5 gw planned

emerging ai hubs:

  • north dakota: applied digital ellendale (180 mw, 50k gpu capacity), coreweave lease ($7b)
  • tennessee: xai colossus memphis (300 mw, 230k gpus operational)
  • ohio: meta prometheus new albany (1,020 mw, 500k+ gpus, 2026)
  • arizona: vermaland la osa (3 gw, 33b),tractbuckeye(1.8gw,33b), tract buckeye (1.8 gw, 20b)

project status distribution

StatusProjectsPercentage
Under Construction5337.9%
Planned4230.0%
Operational2417.1%
Announced107.1%
Expansion96.4%
Canceled21.4%

key insight: 67.9% of ai/ml projects are under construction or planned, indicating massive near-term capacity additions (2025-2027).

top sponsors

RankSponsorAI/ML Projects
1Microsoft10
2Google LLC / Google13 combined
3Meta / Meta Platforms Inc.8 combined
4Amazon Web Services4
5Tract4
6CoreSite4
7Aligned Data Centers4
8PowerHouse Data Centers4

analysis: why ai drives different infrastructure

training vs inference workloads

training (high compute intensity):

  • massive parallel gpu clusters (10k-500k+ gpus)
  • high-bandwidth networking critical (400-800gb/s gpu interconnects)
  • power density: 100-140kw/rack
  • examples: meta prometheus (500k gpus), xai colossus (230k gpus)
  • cooling: liquid cooling mandatory
  • location: near power sources, less latency-sensitive

inference (lower per-request compute, high throughput):

  • distributed across multiple sites
  • latency-sensitive (user-facing)
  • lower power density possible but still 3-5x traditional
  • examples: openai api infrastructure, google tpu v7 ironwood (inference-optimized)
  • cooling: direct liquid or advanced air
  • location: near users/edge locations

the gpu density challenge

rack-scale systems (nvidia nvl72 example):

  • 72 gpus + 36 cpus per rack
  • 120-140kw power per rack
  • 13.5-21tb gpu memory per rack
  • must deploy in multiples of 18 nodes
  • requires liquid cooling infrastructure

cluster-scale challenges:

  • 100,000 gpu cluster (xai colossus):
    • 1,389 racks of 72 gpus each
    • ~173 mw for gpu compute alone
    • single rdma fabric connectivity
    • 122 days construction time
  • power delivery at this scale requires utility partnerships or on-site generation

geographic clustering drivers

power availability (primary constraint):

  • texas: deregulated market, 186 gw queue
  • pennsylvania: marcellus shale gas, 16.9 gw capacity
  • ohio: competitive market, ample generation
  • north dakota: low-cost power, minimal constraints

alternative power sources:

  • on-site natural gas: 9 of 11 gigawatt projects
  • nuclear partnerships: 24+ gw smr commitments (microsoft-constellation, amazon-x-energy, google-kairos, switch-oklo)
  • renewable + battery: xai colossus (150 mw megapack battery backup)
  • hydrogen: lambda terrasite-tx1 (1 gw campus, hydrogen-powered)

construction speed:

  • xai colossus: 100,000 gpus in 122 days
  • standard timeline: 18-24 months for ai facilities
  • traditional hyperscale: 24-36 months

investment patterns

hyperscaler build-out (vertical integration):

  • microsoft: 10 ai/ml projects
  • google: 13 ai/ml projects
  • meta: 8 ai/ml projects (including 1 gw prometheus)
  • amazon: 4 ai/ml projects

specialized ai infrastructure operators:

  • coreweave: 250k gpu fleet, 2.2 gw contracted capacity
  • crusoe: 1.6 gw operational/construction, 10+ gw pipeline
  • applied digital: ellendale hpc campus
  • lambda labs: gigawatt-scale gpu cloud

financial sponsors entering ai:

  • nvidia: $8.13b portfolio (applied digital, coreweave investments)
  • blackstone: airtrunk $16b acquisition
  • digitalbridge: 5.4 gw portfolio
  • blue owl capital: crusoe $15b joint venture

project economics:

  • average ai/ml project: $6.9b investment
  • average power: 717 mw per project
  • contrast traditional: $2-3b average, 100-200 mw typical
  • premium pricing: ai colocation commands 2-3x traditional rates

complete ai/ml projects table

mega-projects (≥$10b or ≥1gw)

Project NameLocationSponsorPower (MW)InvestmentStatus
Project Jupiter (Stargate Santa Teresa)Santa Teresa, NMBorderPlex Digital Assets / OpenAITBD$165.0BAnnounced
Stargate Abilene CampusAbilene, TXOpenAI / Oracle / Crusoe1,200$40.0BOperational
Vermaland La Osa Data Center ParkEloy, AZVermaland LLC3,000$33.0BPlanned
Vantage Data Centers - Frontier CampusVariousVantage Data Centers1,400$25.0BUnder Construction
Google PJM Data Center InfrastructurePennsylvaniaGoogle670$25.0BPlanned
Tract Buckeye Data Center ParkBuckeye, AZTract1,800$20.0BPlanned
AWS AI Innovation CampusesVariousAmazon Web Services960$20.0BUnder Construction
Applied Digital Toronto AI Data CenterToronto (Canada partner)Applied Digital430$16.0BPlanned
Pennsylvania Digital I (PAX)PennsylvaniaPennsylvania Data Center Partners1,350$15.0BPlanned
Project Mica (Google AI Campus)VariousGoogle LLC700$10.0BPlanned
Meta Richland Parish (Hyperion Campus)LouisianaMeta Platforms, Inc.2,000$10.0BUnder Construction
AWS Richmond County CampusGeorgiaAmazon Web ServicesTBD$10.0BAnnounced
Homer City Energy CampusPennsylvaniaHomer City Redevelopment4,500$10.0BPlanned
Data City TexasTexasEnergy Abundance Development Corp5,000TBDPlanned
Delta Gigasite / Fibernet MercuryDeltaMississippiFibernet MercuryDelta LLC4,000TBDPlanned
Joule Capital Partners - Millard CountyUtahJoule Capital Partners4,000TBDPlanned
TECfusions Keystone ConnectPennsylvaniaTECfusions3,000TBDUnder Construction
Shippingport Power StationPennsylvaniaFrontier Group of Companies2,700$3.2BPlanned
PowerHouse - Grand Prairie CampusTexasPowerHouse Data Centers1,800TBDUnder Construction
Crusoe/Tallgrass AI Data CenterWyomingCrusoe Energy Systems1,800TBDAnnounced
Tract Silver Springs ParkNevadaTract1,600TBDPlanned
Prometheus Hyperscale - Natrona/ConverseWyomingPrometheus Hyperscale1,500$0.5BPlanned

data access

full project list: all 140 ai/ml projects available in database

  • location: support/datacenters/analysis/ai_ml_projects_simplified.json
  • includes: project name, state, city, sponsor, power capacity, investment, status
  • sortable by investment, power, sponsor, status, geography

dimensional research:

  • gpu inventory: 1m+ deployed gpus tracked
  • technology specs: h100/h200/b200/b300/gb200/gb300
  • cooling vendors: vertiv, supermicro, grc, liquidstack, lenovo, dell, hpe
  • networking: quantum-2 infiniband, spectrum-x ethernet specifications
  • operators: coreweave, lambda, crusoe, applied digital, hyperscalers

future outlook (2025-2030)

projected growth

  • ai/ml share: 23.2% (2025) → 40-50% (2030)
  • average project size: 6.9b(2025)6.9b (2025) → 10-15b (2030)
  • typical power: 717 mw average → 1-2 gw standard
  • gpu clusters: 100k-500k gpus → 500k-2m gpus
  • cooling: liquid cooling universal → immersion increasing for 300+ kw racks

technology evolution

gpu generations:

  • current: h100/h200 (80-141gb, 350-700w)
  • 2025-2026: b200/b300 (180-288gb, 600-1,400w)
  • 2027+: next-gen blackwell, 2x performance/watt improvements
  • rack systems: gb200/gb300 nvl72 (120-140kw) becoming standard

cooling infrastructure:

  • direct-to-chip liquid cooling: dominant for 100-140kw racks
  • immersion cooling: growing for 300+ kw racks
  • hybrid systems: liquid for gpus/cpus, air for ancillary
  • vendors scaling: vertiv cdu 2300 (2.3mw capacity), supermicro in-row cdu (1.8mw)

networking bandwidth:

  • 2025: 400gb/s quantum-2, 800gb/s spectrum-x
  • 2026-2027: 800gb/s, 1.6tb/s standards
  • 2028+: 3.2tb/s and beyond

power solutions

grid constraints driving alternatives:

  • on-site generation: natural gas increasingly common (9 of 11 gw+ projects)
  • nuclear smrs: 24+ gw committed (2027-2035 deployment)
  • battery integration: tesla megapack (xai 150mw backup)
  • hydrogen: lambda terrasite-tx1 (1 gw hydrogen-powered)

interconnection queue crisis:

  • oncor (texas): 186 gw queue
  • pjm (mid-atlantic): 270+ gw queue
  • multi-year delays standard
  • forcing geographic diversification to less constrained markets

market dynamics

consolidation likely:

  • specialized ai infrastructure operators (coreweave, crusoe, lambda) gaining share
  • hyperscalers building captive capacity (meta prometheus, google project mica)
  • traditional colocation adapting (equinix 100 liquid-cooled dcs, cyrusone intelliscale)
  • financial sponsors backing buildout (nvidia, blackstone, digitalbridge, kkr)

pricing premium:

  • ai colocation: 200300/kw/month(vs200-300/kw/month (vs 100-150 traditional)
  • gpu cloud: premium for h100/h200 vs commodity compute
  • long-term contracts: 3-7 year commitments typical
  • pre-leasing common: projects leased before construction

regulatory evolution:

  • permitting: ai-specific review processes emerging
  • power allocation: utility commissions prioritizing economic development
  • environmental: renewable energy requirements increasing
  • community impact: noise, water use, thermal pollution concerns

using this analysis

for infrastructure operators

  • capacity planning: 717 mw average project size, 100-140kw rack density
  • technology roadmap: liquid cooling mandatory, gb200/gb300 deployment timeline
  • geographic strategy: cluster analysis shows power availability primary driver
  • competitive intelligence: 140 projects tracked with investment, sponsor, status

for investors

  • market size: $538b disclosed investment across 140 projects
  • growth trajectory: 67.9% projects under construction/planned (near-term capacity)
  • key players: hyperscalers (microsoft, google, meta, amazon) + specialists (coreweave, crusoe, lambda)
  • deal flow: mega-projects ($10b+) increasingly common

for technology vendors

  • cooling demand: liquid cooling mandatory for 140 projects (vertiv, supermicro, grc, liquidstack)
  • gpu sales: 1m+ deployed, 500k-2m+ needed by 2030
  • networking: 400-800gb/s interconnects (nvidia quantum-2, spectrum-x)
  • power infrastructure: smr partnerships, battery storage, on-site generation

for policymakers

  • economic development: average $6.9b investment, 717 mw power per project
  • infrastructure planning: 65.2 gw disclosed capacity from ai/ml alone
  • grid impact: concentrated demand requiring utility coordination
  • incentive design: competition for mega-projects intense

ai/ml infrastructure represents the most significant data center buildout in history. the 140 projects tracked here demonstrate fundamentally different requirements: 5-10x power density, mandatory liquid cooling, gigawatt-scale projects, and compressed timelines. success requires solving the power delivery challenge through grid upgrades, on-site generation, or nuclear partnerships. the next 5 years (2025-2030) will determine which regions and operators can deliver the infrastructure needed for ai leadership.

on this page