Appendix II

Estimated Global Compute Distribution and Capacity

This appendix documents the global distribution of computing power as of Q4 2024, with particular focus on AI accelerators and general-purpose computing hardware. All computing power is measured in floating point operations per second (FLOP/s), enabling direct comparison between different types of computing devices.

Methodology

Our analysis tracks two primary categories of computing power:

  1. Specialized AI Hardware: Data center GPUs (e.g., NVIDIA H100s, AMD MI300X), TPUs, and custom ASICs (e.g., Microsoft MAIA, Meta MTIA)
  2. General Purpose Computing: Consumer devices including smartphones, personal computers, and gaming consoles

The analysis covers 2023-2024, with several key methodological considerations:

Time Period Normalization

Many source datasets span 2022-2024. For these three-year totals, we estimate the 2023-2024 portion as 80% of the total, reflecting:

  • Accelerating deployment rates through the period
  • Increased manufacturing capacity in later years
  • Reported quarterly growth patterns (Mann 2024)

Infrastructure Accounting

To account for infrastructure deployed before our analysis period that remains in active use, we apply a 1.3× multiplier to new deployments. This multiplier derives from:

  • Historical hardware retirement rates (Epoch AI 2024)
  • Typical infrastructure lifecycle lengths
  • Reported infrastructure retention patterns (Mann 2024)

Consumer Device Calculations

For devices like smartphones, we calculate:

$$\text{Total Capacity} = \text{Quarterly Sales} \times \text{Number of Quarters} \times \text{Device Performance}$$

Key assumptions include:

  • 275M average quarterly smartphone sales (Ming 2024)
  • 8 quarters (covering 2023-2024)
  • 80/20 split between iPhone/Android market share
  • Device performance from manufacturer specifications

Results

Hardware Specifications

We begin by documenting the performance characteristics of key computing hardware. Table 6.1 shows peak theoretical performance for major AI accelerators and consumer devices.

Table 6.1: Hardware Specifications (2023-2024)

Hardware Type Peak Performance
(FLOP/s)
Source
Data Center AI Chips
AMD MI300X OAM GPU $2.61 \times 10^{15}$ Mann 2024
NVIDIA H100 GPU $9.89 \times 10^{14}$ Epoch AI 2024
NVIDIA A100 GPU $3.12 \times 10^{14}$ Epoch AI 2024
Microsoft MAIA ASIC $8.00 \times 10^{14}$ Mann 2024
Meta MTIA ASIC $3.54 \times 10^{14}$ Mann 2024
AWS Tranium ASIC $2.50 \times 10^{14}$ Mann 2024
AWS Inferentia ASIC $1.92 \times 10^{14}$ Mann 2024
Consumer Devices
iPhone 14 Pro Mobile SoC $2.00 \times 10^{12}$ Ming 2024
Samsung S24 Mobile SoC $3.40 \times 10^{12}$ Ming 2024
Average PC CPU CPU $4.08 \times 10^{10}$ Epoch AI 2024
Average PC GPU GPU $4.60 \times 10^{12}$ Epoch AI 2024
PS5 Console $1.03 \times 10^{13}$ Epoch AI 2024

Estimated New Hardware Deployments

Table 6.2 documents NVIDIA’s GPU shipments, which represent the majority of new AI computing capacity deployed during this period.

These shipments are distributed across major cloud providers and AI companies. Table 6.3 details how this computing capacity is allocated, including both NVIDIA GPUs and custom AI accelerators

Table 6.2: NVIDIA GPU Shipments (2023-2024)

Type Units Computing Power
(FLOP/s)
Source/Calculation
A100s (2023) 2,260,000 $7.05 \times 10^{20}$ Units$\times 3.12 \times 10^{14}$ FLOP/s per A100
H100s (2023) 1,500,000 $1.48 \times 10^{21}$ Units$\times 9.89 \times 10^{14}$ FLOP/s per H100
A100s (2024) 2,000,000 $6.24 \times 10^{20}$ Units$\times 3.12 \times 10^{14}$ FLOP/s per A100
H100s (2024) 2,000,000 $1.98 \times 10^{21}$ Units$\times 9.89 \times 10^{14}$ FLOP/s per H100
Total 2022-2024 7,760,000 $4.79 \times 10^{21}$ Sum of all rows above

Estimated Base Computing Capacity

To establish total worldwide computing capacity, we first document the base capacity that existed at the start of our analysis period (Q1 2023).

Estimated Global Computing Distribution

Finally, Table 6.6 presents the complete worldwide distribution of computing capacity as of Q4 2024, combining base capacity, new deployments, and applying our temporal adjustment factors.

Analysis and Implications

Distribution of Computing Power

The results reveal several key insights about global computing power distribution:

  1. Cloud Concentration: The four largest cloud/AI providers (Meta, Microsoft/OpenAI, Google/DeepMind, and Amazon/Anthropic) collectively control 15.9% of global computing power. This represents a significant concentration of high-performance computing resources.
  2. Consumer Device Significance: Despite lower per-unit performance, consumer devices collectively represent 32.83% of global computing power, with smartphones alone accounting for 23.23%. This highlights the massive scale of deployed consumer hardware.
  3. Legacy Infrastructure: Pre-2023 and other cloud infrastructure remains significant, representing 51.28% of total capacity. This suggests substantial inertia in computing infrastructure deployment.

Growth Patterns

The data reveals several notable trends in computing power growth:

  • Accelerating Deployment: New capacity added in 2023-2024 ($2.82 \times 10^{22}$ FLOP/s) is approximately 7 times larger than the Q1 2023 base capacity ($3.98 \times 10^{21}$ FLOP/s).

Table 6.3: Estimated Major Cloud Provider Deployments (2022-2024)

Provider Hardware Type Computing Power
(FLOP/s)
Source/Calculation
Microsoft/OpenAI
H100 equivalents $6.53 \times 10^{20}$ Epoch AI 2024
AMD MI300X $2.51 \times 10^{20}$ Mann 2024
MAIA $1.58 \times 10^{20}$ Mann 2024
Total New $1.06 \times 10^{21}$ Sum of three rows above
2023-2024 (80%) $8.50 \times 10^{20}$ Total New $\times$ 0.8 (2-year portion of 3-year total)
Total Q4 2024 $1.38 \times 10^{21}$ 2023-2024 total $\times$ 1.3 (infrastructure multiplier)
Amazon/Anthropic
H100/A100 $2.87 \times 10^{20}$ Epoch AI 2024
Inferentia $1.75 \times 10^{20}$ Units$\times 1.92 \times 10^{14}$ FLOP/s per chip
Tranium $9.15 \times 10^{19}$ Units$\times 2.50 \times 10^{14}$ FLOP/s per chip
Total New $5.53 \times 10^{20}$ Sum of three rows above
2023-2024 (80%) $4.42 \times 10^{20}$ Total New $\times$ 0.8
Total Q4 2024 $7.19 \times 10^{20}$ 2023-2024 total $\times$ 1.3
Google/DeepMind
TPUs (Q1 2023) $2.93 \times 10^{19}$ Epoch AI 2024
TPUs (Q4 2024) $9.10 \times 10^{20}$ Epoch AI 2024
New TPUs 2023-2024 $8.81 \times 10^{20}$ Q4 2024 TPUs - Q1 2023 TPUs
H100/A100 $3.16 \times 10^{20}$ Mann 2024
Total New GPUs $3.16 \times 10^{20}$ Same as H100/A100 row
Total New TPUs $8.81 \times 10^{20}$ Same as New TPUs row
New GPUs + TPUs $1.20 \times 10^{21}$ Sum of Total New GPUs and TPUs
Total Q4 2024 $1.23 \times 10^{21}$ (New GPUs + TPUs) $\times$ 1.3 - Q1 2023 TPUs
Meta
A100 $6.68 \times 10^{18}$ Epoch AI 2024
H100 $3.46 \times 10^{20}$ Units $\times$ 9.89 $\times$ $10^{14}$ FLOP/s per H100
H100/A100 $3.96 \times 10^{20}$ Sum of A100 and H100 rows
AMD MI300X $4.52 \times 10^{20}$ Units $\times$ 2.61 $\times$ $10^{15}$ FLOP/s per MI300X
MTIA $5.31 \times 10^{20}$ Units $\times$ 3.54 $\times$ $10^{14}$ FLOP/s per MTIA
Total New $1.38 \times 10^{21}$ Sum of all hardware rows
2023-2024 (80%) $1.10 \times 10^{21}$ Total New $\times$ 0.8
Total Q4 2024 $1.79 \times 10^{21}$ 2023-2024 total $\times$ 1.3
Other NVIDIA Customers
H100/A100 $1.38 \times 10^{21}$ Ming 2024
MI300X $1.52 \times 10^{21}$ Mann 2024
Total New $1.54 \times 10^{20}$ Sum of H100/A100 and MI300X rows
2023-2024 (80%) $1.23 \times 10^{21}$ Total New $\times$ 0.8 (2-year portion of 3-year total)
Total Q4 2024 $2.00 \times 10^{21}$ 2023-2024 total$\times$ 1.3 (infrastructure multiplier)

Table 6.4: Base Computing Capacity (Q1 2023)

Type Computing Power
(FLOP/s)
Source/Calculation
TPU $2.93 \times 10^{19}$ Epoch AI 2024
GPU $3.95 \times 10^{21}$ Epoch AI 2024
Total $3.98 \times 10^{21}$ Sum of TPU and GPU rows

Table 6.5: Estimated New Computing Capacity (Q1 2023 - Q4 2024)

Category Computing Power
(FLOP/s)
Source/Calculation
Specialized AI Hardware
NVIDIA GPUs (avg of source 1) $4.82 \times 10^{21}$ Epoch AI 2024
NVIDIA GPUs (avg of source 2) $4.79 \times 10^{21}$ Sum of all NVIDIA shipments from Table 6.2
Avg NVIDIA GPUs $4.81 \times 10^{21}$ Average of two NVIDIA GPU sources above
New Google TPUs $8.81 \times 10^{20}$ Q4 2024 TPUs - Q1 2023 TPUs from Table 6.3
New Cloud Hardware $2.35 \times 10^{21}$ Sum of all non-NVIDIA cloud hardware from Table 6.3
Consumer Devices
Active iPhones $5.98 \times 10^{21}$ \( 275\text{M units/quarter} \times 8\ \text{quarters} \times 80\% \times 2.00 \times 10^{12}\ \text{FLOP/s per iPhone} \)
Active Androids $1.50 \times 10^{21}$ \( 275\text{M units/quarter} \times 8\ \text{quarters} \times 20\% \times 3.40 \times 10^{12}\ \text{FLOP/s per Android} \)
PC CPUs $1.96 \times 10^{19}$ \( 60\text{M units/quarter} \times 8\ \text{quarters} \times 4.08\ \times 10^{10}\ \text{FLOP/s per CPU} \)
PC GPUs $2.21 \times 10^{21}$ \( 60\text{M units/quarter} \times 8\ \text{quarters} \times 4.60\ \times 10^{12}\ \text{FLOP/s per GPU} \)
Game Consoles $8.64 \times 10^{20}$ \( 84\text{M units} \times 1.03\ \times 10^{13}\ \text{FLOP/s per console} \)
Total New $2.82 \times 10^{22}$ Sum of all rows above

Table 6.6: Estimated Total Worldwide Computing Capacity (Q4 2024). Note that these calculations exclude cloud CPU compute.

Category Computing Power
(FLOP/s)
Share (%) Source/Calculation
Cloud/AI Providers
Meta $1.79 \times 10^{21}$ 5.57% From Table 6.3 Total Q4 2024
Microsoft/OpenAI $1.38 \times 10^{21}$ 4.29% From Table 6.3 Total Q4 2024
Google/DeepMind $1.23 \times 10^{21}$ 3.81% From Table 6.3 Total Q4 2024
Amazon/Anthropic $7.19 \times 10^{20}$ 2.23% From Table 6.3 Total Q4 2024
Consumer Computing
Smartphones $7.48 \times 10^{21}$ 23.23% Sum of Active iPhones and Androids from Table 6.5
PC CPUs/GPUs $2.23 \times 10^{21}$ 6.92% Sum of PC CPUs and GPUs from Table 6.5
Game Consoles $8.64 \times 10^{20}$ 2.68% From Table 6.5
Other Cloud/Pre-2023 $1.65 \times 10^{22}$ 51.28% Base capacity $\times$ 1.3 + remaining new capacity not allocated above
Total $3.22 \times 10^{22}$ 100.00% Sum of all rows above
  • Custom Hardware: Major cloud providers are increasingly deploying custom AI accelerators (MAIA, MTIA, Inferentia) alongside NVIDIA GPUs, though these still represent a minority of total capacity.
  • Consumer Updates: The rapid replacement cycle of consumer devices (particularly smartphones) means this sector maintains a significant share of total computing power despite lower per-unit performance.

Methodological Limitations

Several factors affect the precision of these estimates:

  1. Utilization Rates: Our analysis uses peak theoretical performance. Actual achieved performance varies significantly based on:
    • Workload characteristics
    • Cooling and power constraints
    • Network and memory bottlenecks
  2. Temporal Distribution: The 80% factor used for 2023-2024 portion of 2022-2024 deployments is a rough estimate based on the degree to which NVIDIA sales and chip FLOP capacity have greatly increased in the past two years (Macrotrends LLC 2024).
  3. Consumer Device Usage: Not all consumer devices are actively used for computation at any given time, potentially overestimating their contribution to global computing power.

References

  1. T. Mann. 2024. Ai’s rising tide lifts all chips as amd instinct, cloudy silicon vie for a slice of nvidia’s pie. The Register, Dec.
  2. Epoch AI. 2024. Data on machine learning hardware. Updated December 30, 2024.
  3. Ming. 2024. Nvidia ai gpu shipments to hit 4m in 2024. SMYG Limited News, Jun.
  4. Macrotrends LLC. 2024. Nvidia revenue 2010-2024 — nvda. Accessed January 2024
University of Oxford