How much calories does an LLM burn?
Every request to GPT-4 burns 430 calories, approximately 17% of an average male diet
A clickbait quote from me for the dear reader, to tease you into the topic of AI energy consumption, something many people rarely give any thought to.
You might wonder, how did I come up with this number? Glad you ask, I’d like to do the math with you.
Let’s fuel an LLM
Many people are aware of the fact that large language models consume lots of energy. It is in the name, “large” models. We don’t have the exact tally for GPT-4, as the technical paper does not outline it. (Khowaja et al. 2024) went down the rabbit hole to find out how much energy GPT-3 and Meta’s LLama family of models consumed during training.
In this blog post, we will use the numbers for LLama-combined, which performs head-to-head with GPT4 (depending on the benchmark). For training the combined LLama model, 2638000 kWh were necessary.
Your response to that number will likely be: ok, but I don’t care, this number means nothing to me.
No worries there, I reacted exactly the same way, humans aren’t wired to make sense of millions of kilowatt-hours.
We’ll break down how much energy that actually is:
- one killowat hour multiplies out to an energy value of 860421 calories
- according to the Energy Information administration, an average household consumes 899 kwH per month (“Frequently Asked Questions (FAQs) - U.S. Energy Information Administration (EIA) — Eia.gov”)
- an average female should eat about 1600 to 2000 calories, whereas an average male should consume 2000 to 2500 calories
- an average american household has 3.13 members (“Average Family Size in the U.S. 1960-2022 | Statista — Statista.com”)
From this we can work out the math:
Per citizen, we need 287 kilowatt hours per month, and an additional 0.08 kwH for food (75000 calories), which is 3445.2 kWh per year.
When training mixed LLama we consumed the energy budget of 765 citizens for an entire year.
Equivalently, we could nourish 25 doctorate students, until they are 30 years old, which is the average age for a doctorate (“Age Distribution of Doctorate Recipients U.S. 2021 | Statista — Statista.com”).
Great, armed with the knowledge of 25 doctorate students, we go to work!
The energy consumption of inference
The true energy consumption of inference isn’t exactly known, but the best guestimates provide a number of 0.0005 kWh per request.
This is ~430 calories, roughly one plate of prawn spaghetti (if one can believe the BBC).
Using one of our 25 doctorate students, we could work out how much time they would have to answer the same request. We would assume they have access to a library (no digital devices), and are told the question verbatim.
Energy consumption throughout the day is relatively uniform (except for sport), so 104 calories per hour.
The math works out to:
(((430 calories) / (104 calories / hour)) / 25 people) * 1/60 hour ~ 10 minutes / people
So for every request to GPT-4, we could put our 25 doctorate students to work for 10 minutes to devise an answer. Having 25 doctorate students agree on a response is unlikely, but the equation is pretty mind-boggling.
OpenAI has claimed GPT-5 will reach PhD-level intelligence. Thus, every request could be equated with 25 PhD students working 10 minutes for you. Now that is a number one can let sink in.
On a more serious note
The consumption presented here is a guestimate based on available data and comes backed in with two assumptions:
- our PhD students don’t travel etc - also, they’re American (other countries have much lower energy estimates)
- GPT-4 and GPT-5 will be more energy intensive (but with advances in chips they may not need to be)
However, I think it gives a much better feeling of what exactly is our quality standard for an LLM.
If I put 25 PhD students to work, I expect them not to hallucinate, simply. On the other hand, if we believe the marketing around LLM, if we put the same amount of energy into an LLM, its not as tragic.
That’s why it is important to spell out energy consumption in a currency we understood (such as human labour hours), rather than an abstract quantity (such as kWh).