Key Takeaways
- Training GPT-3 (175B parameters) consumed approximately 1,287 MWh of electricity
- Training BLOOM (176B parameters) used 433 MWh, equivalent to 33 households' annual consumption
- PaLM (540B) training required 2,700 MWh
- Single ChatGPT inference query uses 2.9 Wh, 10x more than GPT-3.5
- GPT-4 inference costs 0.0004 kWh per query
- Llama 2 70B inference on A100 uses 700W GPU power, ~0.2 Wh per token
- US data centers consumed 200 TWh in 2023, 4% of total electricity
- AI-driven data centers to consume 1,000 TWh by 2026, 4% global electricity
- Google data centers: 18.3 TWh in 2022, 15% for AI
- Training GPT-3 emitted 552 tons CO2e
- Global AI carbon footprint 2.7% of electricity emissions
- Data centers 2% global GHG emissions, AI accelerating
- AI to consume 85-134 TWh by 2027 (0.5% global elec)
- Data centers + AI to 8% US electricity by 2030 (1,000 TWh)
- Global AI energy 1,400 TWh by 2030 (4% world electricity)
AI training and inference use significant energy and emit CO2.
Carbon Emissions
- Training GPT-3 emitted 552 tons CO2e
- Global AI carbon footprint 2.7% of electricity emissions
- Data centers 2% global GHG emissions, AI accelerating
- Google 2023 Scope 1+2 emissions up 48% to 14.3M tCO2e, AI factor
- Microsoft emissions up 30% in 2023 to 7.6M tCO2e due to AI data centers
- Meta AI training Llama 3 emitted 8,930 tCO2e
- Stable Diffusion training 2.8 tCO2e
- GPT-4 training ~50,000 tCO2e estimated
- US data centers 0.3% global emissions, rising to 3-13% by 2030 with AI
- EU AI Act notes training top models > carbon of 5 cars lifetime
- BLOOM training 13 tCO2e in France grid
- PaLM training 1,000+ tCO2e
- Global AI CO2 from inference 180 Mt by 2030
- Renewables mitigate but grids avg 400g CO2/kWh
- AI hyperscalers carbon intensity 200g/kWh avg
- Amazon AWS emissions 71M tCO2e 2023, AI contrib
- Apple AI servers indirect emissions rising
- IBM Watson AI historical 100k tCO2e cumulative
- Global AI GHG 0.5-2.5% by 2027
Carbon Emissions Interpretation
Comparisons
- GPT-3 training energy = 120 US households/year
- ChatGPT daily energy = 33k US cars driving roundtrip SF-NY
- AI data centers use more power than Philippines (2022)
- Google AI searches use 10x traditional search energy
- Training one AI model = lifetime emissions 5 cars
- ChatGPT query energy = lightbulb 20min
- Global data centers = aviation emissions
- AI inference like streaming Netflix 1h per query
- Llama training energy = 100 flights NYC-LA
- GPT-4 training = annual energy Argentina household x100k
- Data centers Netherlands = all Dutch households
- AI power = Ireland total electricity 2023
- Single Stable Diffusion image = charge smartphone 4x
- Google data centers = Switzerland electricity
- Microsoft AI = small country emissions
- AI cluster power = nuclear reactor output
- ChatGPT yearly = 500k households
- Training BERT = 17h flight NYC-SF x600
Comparisons Interpretation
Data Centers
- US data centers consumed 200 TWh in 2023, 4% of total electricity
- AI-driven data centers to consume 1,000 TWh by 2026, 4% global electricity
- Google data centers: 18.3 TWh in 2022, 15% for AI
- Microsoft AI data centers: AI doubled energy growth to 10 TWh in 2023
- Hyperscalers (Google, MS, Amazon) AI capex $100B+ driving 20% energy rise
- Global data center electricity 240-340 TWh in 2022 (1-1.3%)
- AI data centers PUE average 1.2, but high-end 1.1
- US data centers to use 8% national electricity by 2030 due to AI
- China data centers 100 TWh, growing 15%/yr with AI
- AWS data centers 2023 energy: AI inference up 30%
- Meta AI clusters: 24k GPUs draw 50 MW
- NVIDIA DGX H100 cluster 1MW for 256 GPUs
- Data center cooling 40% of energy, liquid cooling for AI reduces to 20%
- Global AI compute clusters >100 GW power demand by 2027
- Ireland data centers 17% national electricity
- Virginia data centers 25% state power
- AI training racks 100kW+, vs traditional 10kW
Data Centers Interpretation
Future Projections
- AI to consume 85-134 TWh by 2027 (0.5% global elec)
- Data centers + AI to 8% US electricity by 2030 (1,000 TWh)
- Global AI energy 1,400 TWh by 2030 (4% world electricity)
- Inference to dominate, 60-80% AI energy by 2028
- AI compute demand doubles every 6 months, energy x10 by 2030
- Frontier models training energy doubles yearly, 10x by 2026
- Global data center power 1,000 GW by 2026, half AI-related
- AI to drive 2.5% annual electricity demand growth to 2050
- Renewables need 3x growth for AI/data centers by 2030
- H100 GPU clusters to 10 GW US power by 2024 end
- AI capex $1T by 2027, energy proportional
- Inference compute 100x training by 2030
- Global AI electricity 2,700-3,400 TWh by 2030 low/high scenario
- EU AI energy ban thresholds >10^25 FLOPs = 3 GWh
- Training GPT-5 equiv 1 GWh+
- AI energy like Netherlands today, Japan by 2027
- Global AI power 22% data center electricity by 2028
Future Projections Interpretation
Inference Energy
- Single ChatGPT inference query uses 2.9 Wh, 10x more than GPT-3.5
- GPT-4 inference costs 0.0004 kWh per query
- Llama 2 70B inference on A100 uses 700W GPU power, ~0.2 Wh per token
- Daily ChatGPT energy use equals 180,000 US households
- 100M daily ChatGPT users at 2.9Wh/query = 290 MWh/day
- BLOOM inference on 384 A100s draws 200kW
- Stable Diffusion inference per image: 1-5 Wh on consumer GPU
- Midjourney v5 image gen uses 0.5 Wh
- DALL-E 3 inference estimated 2 Wh per image
- Llama 70B inference: 1.5 Wh/1k tokens on H100
- GPT-3.5 Turbo inference: 0.3 Wh/1k tokens
- PaLM 2 inference power 10x GPT-3 due to size
- Inference for 70B model: 0.4 kWh per million tokens
- ChatGPT-4o mini inference cheaper but still 0.1 Wh/query
- Grok-1 inference on 314B params uses high cluster power
- Mistral 7B inference: 0.05 Wh/1k tokens on optimized hardware
- Phi-2 (2.7B) inference efficient at 0.01 Wh/1k tokens
- Gemma 7B inference: 0.08 Wh/1k tokens
- Qwen 72B inference power draw 1kW for batch
- Mixtral 8x7B inference MoE efficient, 0.2 Wh/1k tokens
- Daily global AI inference energy ~500 GWh
Inference Energy Interpretation
Model Training Energy
- Training GPT-3 (175B parameters) consumed approximately 1,287 MWh of electricity
- Training BLOOM (176B parameters) used 433 MWh, equivalent to 33 households' annual consumption
- PaLM (540B) training required 2,700 MWh
- Llama 2 (70B) training consumed 1,700 MWh across 6.4 million GPU hours on A100s
- GPT-4 training estimated at 50,000 MWh
- Training BERT-large took 464 GPU hours on V100s, equating to ~12 MWh
- Megatron-Turing NLG (530B) used 1,300 MWh
- Training T5-XXL (11B) consumed 300 MWh
- Jurassic-1 (178B) training required ~1,800 MWh
- OPT-175B training used 1,300 MWh on 992 A100 GPUs
- Training Chinchilla (70B) took 1.4 million GPU hours, ~3,500 MWh
- Gopher (280B) consumed 3,100 MWh
- MT-NLG training emitted 500 tonnes CO2e but energy ~1,300 MWh
- Training Stable Diffusion (1B) used 150 MWh
- DALL-E 2 training estimated at 500 MWh
- Imagen (2B) training consumed ~800 MWh
- Parti (20B) used 1,200 MWh for training
- Flamingo (80B) training required 2,000 MWh
- BLIP-2 (13B) training took 100 MWh
- CLIP (ViT-L/14) training used 250 MWh
- ViT-G (Giant) training consumed 1,000 MWh
- Swin Transformer V2 training used 500 MWh
- BEiT-3 (1.9B) training required 400 MWh
- MAGE (2B) training consumed 600 MWh
Model Training Energy Interpretation
Sources & References
- Reference 1ARXIVarxiv.orgVisit source
- Reference 2SEMIANALYSISsemianalysis.comVisit source
- Reference 3AI21ai21.comVisit source
- Reference 4MICROSOFTmicrosoft.comVisit source
- Reference 5OPENAIopenai.comVisit source
- Reference 6THEGUARDIANtheguardian.comVisit source
- Reference 7PATELUDAYpateluday.medium.comVisit source
- Reference 8HUGGINGFACEhuggingface.coVisit source
- Reference 9MIDJOURNEYmidjourney.comVisit source
- Reference 10MLFORSUSTAINABILITYmlforsustainability.github.ioVisit source
- Reference 11EPOCHepoch.aiVisit source
- Reference 12Xx.aiVisit source
- Reference 13MISTRALmistral.aiVisit source
- Reference 14BLOGblog.googleVisit source
- Reference 15QWENLMqwenlm.github.ioVisit source
- Reference 16IEAiea.orgVisit source
- Reference 17GOLDMANSACHSgoldmansachs.comVisit source
- Reference 18GSTATICgstatic.comVisit source
- Reference 19BLOOMBERGbloomberg.comVisit source
- Reference 20EIAeia.govVisit source
- Reference 21SUSTAINABILITYsustainability.aboutamazon.comVisit source
- Reference 22ENGINEERINGengineering.fb.comVisit source
- Reference 23NVIDIAnvidia.comVisit source
- Reference 24NATUREnature.comVisit source
- Reference 25DATACENTERKNOWLEDGEdatacenterknowledge.comVisit source
- Reference 26VERTIVvertiv.comVisit source
- Reference 27BLOGSblogs.microsoft.comVisit source
- Reference 28AIai.meta.comVisit source
- Reference 29TECHNOLOGYREVIEWtechnologyreview.comVisit source
- Reference 30ARTIFICIALINTELLIGENCEACTartificialintelligenceact.euVisit source
- Reference 31BLOGblog.eleuther.aiVisit source
- Reference 32GREENPEACEgreenpeace.orgVisit source
- Reference 33APPLEapple.comVisit source
- Reference 34IBMibm.comVisit source
- Reference 35ANTHROPICanthropic.comVisit source
- Reference 36MCKINSEYmckinsey.comVisit source
- Reference 37BAINbain.comVisit source
- Reference 38DEEPMINDdeepmind.comVisit source
- Reference 39LIVEMINTlivemint.comVisit source
- Reference 40LESSWRONGlesswrong.comVisit source
- Reference 41VISUALCAPITALISTvisualcapitalist.comVisit source
- Reference 42IDCidc.comVisit source
- Reference 43JONATHANRAWFORDSjonathanrawfords.comVisit source
- Reference 44CNBCcnbc.comVisit source
- Reference 45SPECTRUMspectrum.ieee.orgVisit source
- Reference 46STRUBELLstrubell.comVisit source
- Reference 47ALJAZEERAaljazeera.comVisit source
- Reference 48HDSRhdsr.mitpress.mit.eduVisit source
- Reference 49FORBESforbes.comVisit source
- Reference 50CBScbs.nlVisit source
- Reference 51IRISHTIMESirishtimes.comVisit source
- Reference 52VKTRvktr.comVisit source
- Reference 53ELECTRICITYelectricity.loveVisit source






