Appendix II
Estimated Global Compute Distribution and Capacity
This appendix documents the global distribution of computing power as of Q4 2024, with particular focus on AI accelerators and general-purpose computing hardware. All computing power is measured in floating point operations per second (FLOP/s), enabling direct comparison between different types of computing devices.
Methodology
Our analysis tracks two primary categories of computing power:
- Specialized AI Hardware: Data center GPUs (e.g., NVIDIA H100s, AMD MI300X), TPUs, and custom ASICs (e.g., Microsoft MAIA, Meta MTIA)
- General Purpose Computing: Consumer devices including smartphones, personal computers, and gaming consoles
The analysis covers 2023-2024, with several key methodological considerations:
Time Period Normalization
Many source datasets span 2022-2024. For these three-year totals, we estimate the 2023-2024 portion as 80% of the total, reflecting:
- Accelerating deployment rates through the period
- Increased manufacturing capacity in later years
- Reported quarterly growth patterns (Mann 2024)
Infrastructure Accounting
To account for infrastructure deployed before our analysis period that remains in active use, we apply a 1.3× multiplier to new deployments. This multiplier derives from:
- Historical hardware retirement rates (Epoch AI 2024)
- Typical infrastructure lifecycle lengths
- Reported infrastructure retention patterns (Mann 2024)
Consumer Device Calculations
For devices like smartphones, we calculate:
$$\text{Total Capacity} = \text{Quarterly Sales} \times \text{Number of Quarters} \times \text{Device Performance}$$
Key assumptions include:
- 275M average quarterly smartphone sales (Ming 2024)
- 8 quarters (covering 2023-2024)
- 80/20 split between iPhone/Android market share
- Device performance from manufacturer specifications
Results
Hardware Specifications
We begin by documenting the performance characteristics of key computing hardware. Table 6.1 shows peak theoretical performance for major AI accelerators and consumer devices.
Table 6.1: Hardware Specifications (2023-2024)
| Hardware | Type | Peak Performance (FLOP/s) |
Source |
|---|---|---|---|
| Data Center AI Chips | |||
| AMD MI300X OAM | GPU | $2.61 \times 10^{15}$ | Mann 2024 |
| NVIDIA H100 | GPU | $9.89 \times 10^{14}$ | Epoch AI 2024 |
| NVIDIA A100 | GPU | $3.12 \times 10^{14}$ | Epoch AI 2024 |
| Microsoft MAIA | ASIC | $8.00 \times 10^{14}$ | Mann 2024 |
| Meta MTIA | ASIC | $3.54 \times 10^{14}$ | Mann 2024 |
| AWS Tranium | ASIC | $2.50 \times 10^{14}$ | Mann 2024 |
| AWS Inferentia | ASIC | $1.92 \times 10^{14}$ | Mann 2024 |
| Consumer Devices | |||
| iPhone 14 Pro | Mobile SoC | $2.00 \times 10^{12}$ | Ming 2024 |
| Samsung S24 | Mobile SoC | $3.40 \times 10^{12}$ | Ming 2024 |
| Average PC CPU | CPU | $4.08 \times 10^{10}$ | Epoch AI 2024 |
| Average PC GPU | GPU | $4.60 \times 10^{12}$ | Epoch AI 2024 |
| PS5 | Console | $1.03 \times 10^{13}$ | Epoch AI 2024 |
Estimated New Hardware Deployments
Table 6.2 documents NVIDIA’s GPU shipments, which represent the majority of new AI computing capacity deployed during this period.
These shipments are distributed across major cloud providers and AI companies. Table 6.3 details how this computing capacity is allocated, including both NVIDIA GPUs and custom AI accelerators
Table 6.2: NVIDIA GPU Shipments (2023-2024)
| Type | Units | Computing Power (FLOP/s) |
Source/Calculation |
|---|---|---|---|
| A100s (2023) | 2,260,000 | $7.05 \times 10^{20}$ | Units$\times 3.12 \times 10^{14}$ FLOP/s per A100 |
| H100s (2023) | 1,500,000 | $1.48 \times 10^{21}$ | Units$\times 9.89 \times 10^{14}$ FLOP/s per H100 |
| A100s (2024) | 2,000,000 | $6.24 \times 10^{20}$ | Units$\times 3.12 \times 10^{14}$ FLOP/s per A100 |
| H100s (2024) | 2,000,000 | $1.98 \times 10^{21}$ | Units$\times 9.89 \times 10^{14}$ FLOP/s per H100 |
| Total 2022-2024 | 7,760,000 | $4.79 \times 10^{21}$ | Sum of all rows above |
Estimated Base Computing Capacity
To establish total worldwide computing capacity, we first document the base capacity that existed at the start of our analysis period (Q1 2023).
Estimated Global Computing Distribution
Finally, Table 6.6 presents the complete worldwide distribution of computing capacity as of Q4 2024, combining base capacity, new deployments, and applying our temporal adjustment factors.
Analysis and Implications
Distribution of Computing Power
The results reveal several key insights about global computing power distribution:
- Cloud Concentration: The four largest cloud/AI providers (Meta, Microsoft/OpenAI, Google/DeepMind, and Amazon/Anthropic) collectively control 15.9% of global computing power. This represents a significant concentration of high-performance computing resources.
- Consumer Device Significance: Despite lower per-unit performance, consumer devices collectively represent 32.83% of global computing power, with smartphones alone accounting for 23.23%. This highlights the massive scale of deployed consumer hardware.
- Legacy Infrastructure: Pre-2023 and other cloud infrastructure remains significant, representing 51.28% of total capacity. This suggests substantial inertia in computing infrastructure deployment.
Growth Patterns
The data reveals several notable trends in computing power growth:
- Accelerating Deployment: New capacity added in 2023-2024 ($2.82 \times 10^{22}$ FLOP/s) is approximately 7 times larger than the Q1 2023 base capacity ($3.98 \times 10^{21}$ FLOP/s).
Table 6.3: Estimated Major Cloud Provider Deployments (2022-2024)
| Provider | Hardware Type | Computing Power (FLOP/s) |
Source/Calculation |
|---|---|---|---|
| Microsoft/OpenAI | |||
| H100 equivalents | $6.53 \times 10^{20}$ | Epoch AI 2024 | |
| AMD MI300X | $2.51 \times 10^{20}$ | Mann 2024 | |
| MAIA | $1.58 \times 10^{20}$ | Mann 2024 | |
| Total New | $1.06 \times 10^{21}$ | Sum of three rows above | |
| 2023-2024 (80%) | $8.50 \times 10^{20}$ | Total New $\times$ 0.8 (2-year portion of 3-year total) | |
| Total Q4 2024 | $1.38 \times 10^{21}$ | 2023-2024 total $\times$ 1.3 (infrastructure multiplier) | |
| Amazon/Anthropic | |||
| H100/A100 | $2.87 \times 10^{20}$ | Epoch AI 2024 | |
| Inferentia | $1.75 \times 10^{20}$ | Units$\times 1.92 \times 10^{14}$ FLOP/s per chip | |
| Tranium | $9.15 \times 10^{19}$ | Units$\times 2.50 \times 10^{14}$ FLOP/s per chip | |
| Total New | $5.53 \times 10^{20}$ | Sum of three rows above | |
| 2023-2024 (80%) | $4.42 \times 10^{20}$ | Total New $\times$ 0.8 | |
| Total Q4 2024 | $7.19 \times 10^{20}$ | 2023-2024 total $\times$ 1.3 | |
| Google/DeepMind | |||
| TPUs (Q1 2023) | $2.93 \times 10^{19}$ | Epoch AI 2024 | |
| TPUs (Q4 2024) | $9.10 \times 10^{20}$ | Epoch AI 2024 | |
| New TPUs 2023-2024 | $8.81 \times 10^{20}$ | Q4 2024 TPUs - Q1 2023 TPUs | |
| H100/A100 | $3.16 \times 10^{20}$ | Mann 2024 | |
| Total New GPUs | $3.16 \times 10^{20}$ | Same as H100/A100 row | |
| Total New TPUs | $8.81 \times 10^{20}$ | Same as New TPUs row | |
| New GPUs + TPUs | $1.20 \times 10^{21}$ | Sum of Total New GPUs and TPUs | |
| Total Q4 2024 | $1.23 \times 10^{21}$ | (New GPUs + TPUs) $\times$ 1.3 - Q1 2023 TPUs | |
| Meta | |||
| A100 | $6.68 \times 10^{18}$ | Epoch AI 2024 | |
| H100 | $3.46 \times 10^{20}$ | Units $\times$ 9.89 $\times$ $10^{14}$ FLOP/s per H100 | |
| H100/A100 | $3.96 \times 10^{20}$ | Sum of A100 and H100 rows | |
| AMD MI300X | $4.52 \times 10^{20}$ | Units $\times$ 2.61 $\times$ $10^{15}$ FLOP/s per MI300X | |
| MTIA | $5.31 \times 10^{20}$ | Units $\times$ 3.54 $\times$ $10^{14}$ FLOP/s per MTIA | |
| Total New | $1.38 \times 10^{21}$ | Sum of all hardware rows | |
| 2023-2024 (80%) | $1.10 \times 10^{21}$ | Total New $\times$ 0.8 | |
| Total Q4 2024 | $1.79 \times 10^{21}$ | 2023-2024 total $\times$ 1.3 | |
| Other NVIDIA Customers | |||
| H100/A100 | $1.38 \times 10^{21}$ | Ming 2024 | |
| MI300X | $1.52 \times 10^{21}$ | Mann 2024 | |
| Total New | $1.54 \times 10^{20}$ | Sum of H100/A100 and MI300X rows | |
| 2023-2024 (80%) | $1.23 \times 10^{21}$ | Total New $\times$ 0.8 (2-year portion of 3-year total) | |
| Total Q4 2024 | $2.00 \times 10^{21}$ | 2023-2024 total$\times$ 1.3 (infrastructure multiplier) | |
Table 6.4: Base Computing Capacity (Q1 2023)
| Type | Computing Power (FLOP/s) |
Source/Calculation |
|---|---|---|
| TPU | $2.93 \times 10^{19}$ | Epoch AI 2024 |
| GPU | $3.95 \times 10^{21}$ | Epoch AI 2024 |
| Total | $3.98 \times 10^{21}$ | Sum of TPU and GPU rows |
Table 6.5: Estimated New Computing Capacity (Q1 2023 - Q4 2024)
| Category | Computing Power (FLOP/s) |
Source/Calculation |
|---|---|---|
| Specialized AI Hardware | ||
| NVIDIA GPUs (avg of source 1) | $4.82 \times 10^{21}$ | Epoch AI 2024 |
| NVIDIA GPUs (avg of source 2) | $4.79 \times 10^{21}$ | Sum of all NVIDIA shipments from Table 6.2 |
| Avg NVIDIA GPUs | $4.81 \times 10^{21}$ | Average of two NVIDIA GPU sources above |
| New Google TPUs | $8.81 \times 10^{20}$ | Q4 2024 TPUs - Q1 2023 TPUs from Table 6.3 |
| New Cloud Hardware | $2.35 \times 10^{21}$ | Sum of all non-NVIDIA cloud hardware from Table 6.3 |
| Consumer Devices | ||
| Active iPhones | $5.98 \times 10^{21}$ | \( 275\text{M units/quarter} \times 8\ \text{quarters} \times 80\% \times 2.00 \times 10^{12}\ \text{FLOP/s per iPhone} \) |
| Active Androids | $1.50 \times 10^{21}$ | \( 275\text{M units/quarter} \times 8\ \text{quarters} \times 20\% \times 3.40 \times 10^{12}\ \text{FLOP/s per Android} \) |
| PC CPUs | $1.96 \times 10^{19}$ | \( 60\text{M units/quarter} \times 8\ \text{quarters} \times 4.08\ \times 10^{10}\ \text{FLOP/s per CPU} \) |
| PC GPUs | $2.21 \times 10^{21}$ | \( 60\text{M units/quarter} \times 8\ \text{quarters} \times 4.60\ \times 10^{12}\ \text{FLOP/s per GPU} \) |
| Game Consoles | $8.64 \times 10^{20}$ | \( 84\text{M units} \times 1.03\ \times 10^{13}\ \text{FLOP/s per console} \) |
| Total New | $2.82 \times 10^{22}$ | Sum of all rows above |
Table 6.6: Estimated Total Worldwide Computing Capacity (Q4 2024). Note that these calculations exclude cloud CPU compute.
| Category | Computing Power (FLOP/s) |
Share (%) | Source/Calculation |
|---|---|---|---|
| Cloud/AI Providers | |||
| Meta | $1.79 \times 10^{21}$ | 5.57% | From Table 6.3 Total Q4 2024 |
| Microsoft/OpenAI | $1.38 \times 10^{21}$ | 4.29% | From Table 6.3 Total Q4 2024 |
| Google/DeepMind | $1.23 \times 10^{21}$ | 3.81% | From Table 6.3 Total Q4 2024 |
| Amazon/Anthropic | $7.19 \times 10^{20}$ | 2.23% | From Table 6.3 Total Q4 2024 |
| Consumer Computing | |||
| Smartphones | $7.48 \times 10^{21}$ | 23.23% | Sum of Active iPhones and Androids from Table 6.5 |
| PC CPUs/GPUs | $2.23 \times 10^{21}$ | 6.92% | Sum of PC CPUs and GPUs from Table 6.5 |
| Game Consoles | $8.64 \times 10^{20}$ | 2.68% | From Table 6.5 |
| Other Cloud/Pre-2023 | $1.65 \times 10^{22}$ | 51.28% | Base capacity $\times$ 1.3 + remaining new capacity not allocated above |
| Total | $3.22 \times 10^{22}$ | 100.00% | Sum of all rows above |
- Custom Hardware: Major cloud providers are increasingly deploying custom AI accelerators (MAIA, MTIA, Inferentia) alongside NVIDIA GPUs, though these still represent a minority of total capacity.
- Consumer Updates: The rapid replacement cycle of consumer devices (particularly smartphones) means this sector maintains a significant share of total computing power despite lower per-unit performance.
Methodological Limitations
Several factors affect the precision of these estimates:
- Utilization Rates: Our analysis uses peak theoretical performance. Actual
achieved
performance varies significantly based on:
- Workload characteristics
- Cooling and power constraints
- Network and memory bottlenecks
- Temporal Distribution: The 80% factor used for 2023-2024 portion of 2022-2024 deployments is a rough estimate based on the degree to which NVIDIA sales and chip FLOP capacity have greatly increased in the past two years (Macrotrends LLC 2024).
- Consumer Device Usage: Not all consumer devices are actively used for computation at any given time, potentially overestimating their contribution to global computing power.
References
- Ai’s rising tide lifts all chips as amd instinct, cloudy silicon vie for a slice of nvidia’s pie. The Register, Dec.
- Data on machine learning hardware. Updated December 30, 2024.
- Nvidia ai gpu shipments to hit 4m in 2024. SMYG Limited News, Jun.
- Nvidia revenue 2010-2024 — nvda. Accessed January 2024