ai infrastructure competition
on this page
overview
specialized ai infrastructure providers are challenging traditional hyperscalers with gpu-optimized data centers designed for ai training and inference workloads. this analysis examines 20 projects totaling 7.8 gw of capacity across providers including stargate (openai/oracle), coreweave, crusoe energy, applied digital, and lancium.
key findings
- stargate leads investment: $500 billion openai/oracle initiative
- crusoe leads capacity: 3,000 mw across 3 projects (38.5% market share)
- applied digital pure-play gpu: 100% of portfolio gpu-optimized
- coreweave fastest growth: 7 projects across 7 states
- power innovation: flared gas, nuclear, renewable partnerships
market share by capacity
| Provider | Projects | Capacity (MW) | States | GPU Focus | Share |
| Crusoe Energy | 3 | 3,000 | 3 | 67% | 38.5% |
| Stargate (OpenAI/Oracle) | 5 | 2,200 | 3 | 100% | 28.2% |
| Applied Digital | 4 | 1,346 | 2 | 100% | 17.3% |
| CoreWeave | 7 | 1,272 | 7 | 100% | 16.3% |
| Lancium | 1 | 1,200 | 1 | 100% | N/A |
note: lancium capacity overlaps with stargate abilene campus
stargate initiative (openai/oracle)
competitive position
stargate represents the most ambitious ai infrastructure project globally, with $500 billion planned investment over multiple years. the initiative involves openai, oracle, softbank, nvidia, and mgx.
major projects
| Project | State | Capacity (MW) | Status | Partners |
| Abilene Campus | Texas | 1,200 | Planned | Oracle/Crusoe/Lancium |
| Wisconsin Evaluation | Wisconsin | 1,000 | Evaluation | Oracle/OpenAI |
| Santa Teresa Campus | New Mexico | TBD | Under Construction | Oracle/STACK/OpenAI |
| Milam County Site | Texas | TBD | Planned | Oracle/OpenAI |
| Shackelford County | Texas | TBD | Planned | Oracle/OpenAI |
strategic differentiators
- unprecedented scale: $500 billion total investment target
- vertical integration: openai controls training infrastructure
- nvidia partnership: priority gpu allocation
- oracle cloud integration: oci optimized for llm training
- geographic diversification: texas, new mexico, wisconsin campuses
technology stack
- gpus: nvidia h100, h200, gb200 superchips
- networking: nvidia infiniband for low-latency training
- storage: high-performance parallel file systems
- power: behind-meter generation, utility partnerships
competitive implications
stargate’s scale threatens traditional cloud providers’ ai dominance. openai gains infrastructure independence while oracle accelerates cloud growth.
coreweave
competitive position
coreweave operates 7 projects across 7 states with 1,272 mw capacity. founded as crypto mining operation, pivoted to ai infrastructure in 2020.
major projects
| Project | State | Capacity (MW) | Status | Special Feature |
| Polaris Forge 1 | North Dakota | 530 | Operational | Wind power |
| Related Digital Campus | Wyoming | 302 | Under Construction | Nuclear candidate |
| Digital Crossroads Hammond | Indiana | 200 | Expansion | Midwest power |
| Lancaster County Campus | Pennsylvania | 100 | Planned | Nuclear partnership |
| Core Scientific Muskogee | Oklahoma | 100 | Operational | Crypto conversion |
strategic differentiators
- kubernetes native: cloud-native gpu orchestration
- spot pricing: auction-based gpu marketplace
- crypto legacy: repurposed mining infrastructure
- nvidia partnership: preferred cloud service provider
- inference focus: shifting from training to inference workloads
customer segmentation
- ai startups: 40%+ revenue (anthropic, character.ai, etc.)
- enterprises: 30% (financial services, healthcare)
- rendering/vfx: 15% (legacy business)
- research institutions: 15%
funding and growth
- $642 million series b: led by magnetar capital (2023)
- $7.5 billion debt facility: blackstone/carlyle/pimco (2024)
- $19 billion valuation: ipo planned 2024/2025
- 5-year cagr: 200%+ revenue growth
crusoe energy
competitive position
crusoe operates 3 projects across 3 states with 3,000 mw capacity. company pioneered flared gas utilization for data centers, now expanding to ai infrastructure.
major projects
| Project | State | Capacity (MW) | Status | Power Source |
| Crusoe/Tallgrass AI Campus | Wyoming | 1,800 | Planned | Natural gas/nuclear |
| Stargate Abilene Campus | Texas | 1,200 | Planned | Grid/renewables |
| ExxonMobil Alaska Crypto | Alaska | TBD | Pilot | Flared gas |
strategic differentiators
- flared gas pioneer: monetizes stranded energy
- sustainability narrative: prevents methane emissions
- oil & gas partnerships: exxonmobil, devon energy
- modular deployment: rapid installation near wellheads
- energy arbitrage: utilizes otherwise-wasted power
technology innovation
- mobile data centers: containerized for oil field deployment
- gas turbine integration: direct fuel-to-compute
- carbon credits: methane reduction generates offsets
- grid services: demand response capabilities
pivot to ai
transitioning from crypto mining to ai infrastructure. stargate partnership validates ai business model. wyoming campus represents largest ai deployment.
applied digital
competitive position
applied digital operates 4 projects across 2 states with 1,346 mw capacity. publicly traded (nasdaq: apld) pure-play gpu infrastructure provider.
major projects
| Project | State | Capacity (MW) | Status | Customer |
| Polaris Forge 1 | North Dakota | 530 | Operational | CoreWeave |
| Toronto AI Data Center | South Dakota | 430 | Under Construction | TBD |
| Polaris Forge 2 | North Dakota | 280 | Planned | TBD |
| JMS01 | North Dakota | 106 | Operational | Mixed |
strategic differentiators
- vertical integration: develops, builds, operates facilities
- public markets access: nasdaq listing provides capital
- midwest focus: cheap power, favorable regulations
- hpc heritage: evolved from cryptocurrency mining
- turnkey solutions: builds for customers or operates
business model evolution
- phase 1 (2021-2022): cryptocurrency mining
- phase 2 (2022-2023): transition to hpc/ai
- phase 3 (2023-2024): gpu cloud services
- phase 4 (2024+): hyperscale ai infrastructure
financial performance
- fy2024 revenue: $231 million (up 150% yoy)
- gpu deployment: 20,000+ gpus operational
- pipeline: 1.5+ gw in development
- customers: coreweave (anchor), enterprises, ai startups
lancium
competitive position
lancium operates 1 project with 1,200 mw capacity (stargate abilene campus). company specializes in behind-the-meter power and grid services.
major project
| Project | State | Capacity (MW) | Status | Partners |
| Stargate Abilene Campus | Texas | 1,200 | Planned | Oracle/Crusoe/OpenAI |
strategic differentiators
- interruptible power: sells grid services during peak demand
- ercot expertise: deep texas power market knowledge
- behind-meter generation: natural gas, renewables
- flexible loads: designed for curtailment
- economic arbitrage: profits from power price volatility
business model
lancium provides ai infrastructure that can be curtailed during grid stress, earning capacity payments while reducing compute during high-price periods.
competitive dynamics
gpu capacity race
nvidia allocation: critical bottleneck for all providers
- stargate: priority allocation through partnership
- coreweave: preferred cloud partner status
- applied digital: public market visibility
- crusoe: strategic partnership discussions
power strategies
flared gas (crusoe): pioneered stranded energy utilization renewables (coreweave/applied): wind/solar power purchase agreements nuclear (stargate): investigating smr partnerships behind-meter (lancium): natural gas, grid services
customer segmentation
ai foundation models: stargate (openai), anthropic (coreweave) enterprises: banking, healthcare, legal seeking private gpu startups: character.ai, perplexity, midjourney research: universities, national labs
geographic strategies
north dakota (applied digital): cheap power, cold climate cooling texas (stargate/crusoe/lancium): deregulated ercot market wyoming (crusoe/coreweave): mining legacy, nuclear potential pennsylvania (coreweave): nuclear partnerships, pjm grid
technology differentiation
cooling systems
liquid cooling: 80%+ of new ai facilities deploying rear-door heat exchangers or direct-to-chip air cooling: limited to less than 50 kw/rack deployments immersion cooling: specialized applications, limited adoption free cooling: north dakota climate advantage
power density
traditional: 5-10 kw/rack ai training: 50-100 kw/rack liquid cooled: 100-200 kw/rack future (gb200): 200+ kw/rack
networking
infiniband: nvidia mellanox for training clusters ethernet: microsoft/amd competitors gaining traction optical: 800gbps links standard, 1.6tbps emerging
software stack
coreweave: kubernetes-native orchestration stargate: oracle cloud infrastructure applied digital: proprietary management crusoe: cloud-native platform
market outlook
capacity growth
ai infrastructure providers adding 5-10 gw annually, exceeding traditional hyperscaler growth rates. stargate alone targeting 5+ gw by 2030.
investment trends
- venture capital: $5+ billion raised 2023-2024
- debt facilities: blackstone, pimco providing billions
- strategic investments: nvidia, microsoft taking stakes
- public markets: applied digital ipo, coreweave planned
gpu supply chain
- h100 availability: improving but constrained
- h200 deployment: ramping 2024-2025
- gb200 superchip: 2025 deliveries
- amd mi300x: alternative gaining traction
power challenges
- grid constraints: 2-5 year interconnection timelines
- behind-meter: accelerates deployment
- nuclear options: smrs proposed but years away
- renewable mandates: customers demanding carbon-free
competitive threats
hyperscalers entering market: aws trainium, google tpu, microsoft maia colocation providers: qts, vantage, cyrusone adding gpu capacity vertical integration: openai, anthropic, meta building own international competition: china, middle east infrastructure
customer trends
training vs inference
- training: declining share as models stabilize
- inference: 70%+ of compute hours by 2025
- edge inference: 5g driving distributed deployment
- hybrid: fine-tuning blends training/inference
workload characteristics
llm training: 1,000-10,000 gpus, weeks-months duration inference serving: 10-100 gpus, continuous operation fine-tuning: 100-1,000 gpus, days-weeks duration rlhf: human-in-loop, variable duration
pricing models
reserved capacity: annual contracts, predictable pricing spot pricing: auction-based, cost-optimized on-demand: pay-per-use, premium pricing revenue share: aligned incentives for startups
data sources
analysis based on 604 documented us data center projects. ai infrastructure providers tracked through sec filings, press releases, industry publications, and company disclosures. gpu capacity estimates conservative given allocation constraints.
last updated: october 17, 2025