I’ve been reading articles claiming that training and running AI models use a surprising amount of water for cooling data centers, but the numbers are all over the place. I’m trying to understand the real environmental impact before I decide how heavily to rely on AI tools for my projects. Can anyone break down how water usage is measured for AI, what factors affect it, and how big a problem this really is compared to other tech or industries?
Short version. Yes, AI uses a lot of water. The numbers look wild because people talk about different things and use different scopes.
Here is the practical breakdown.
-
Where the water goes
- Data centers use water for cooling.
- Two main types.
• Direct water cooling on site, like cooling towers that evaporate water.
• Indirect use at power plants that generate the electricity for the data center. - Most articles mix these or only count one, so the numbers jump around.
-
Rough numbers you can use
These are ballpark, based on public research and company disclosures up to 2024.Per model training
- A big training run for something like GPT‑3 size reported in one paper used about 700 thousand liters of water, including power plant water. That is about 180 thousand gallons.
- Some newer estimates for even larger models go into several million liters for a single full training run, depending on where the data center sits and how hot and dry the location is.
- Smaller models or fine‑tunes are much lower, more in the tens of thousands of liters or less.
Per usage, like chats and API calls
- One study estimated about 0.5 liter of water per 10 to 50 prompts for a big model, including power generation water, at typical US data centers. So something in the range of a few hundred milliliters per short chat session.
- Numbers vary a lot by location. Clean grid and dry air, lower water. Coal or gas plants with water cooling, higher water.
Per data center
- Some hyperscale data centers use around 20 to 50 million liters per year on site.
- Add more if local electricity still uses water‑cooled thermal plants. Wind and solar use far less water.
-
Why numbers differ so much
- Scope. Some count only onsite water. Some include the water used at power plants.
- Location. Desert region vs cool, humid region gives huge differences.
- Time of year. Hot summers increase water for cooling.
- Type of workload. A data center serving video streaming has a different profile than one hammered with AI training.
-
How to think about impact
- If you compare one large model training run to personal activities.
• Hundreds of thousands to a few million liters is similar to the yearly indoor water use of dozens to a few hundred average US households. - But training runs happen occasionally. Inference happens all the time and will matter more as usage scales.
- If you compare one large model training run to personal activities.
-
What you as a user influence
- Your individual chat use has a small direct footprint, comparable to other online actions. For example, watching an hour of HD video uses water too, through power generation and sometimes cooling at video data centers.
- The big impact comes from aggregate demand and how fast companies deploy bigger models.
-
What companies can and should do
Things to look for in their sustainability reports if you care and want to compare vendors.- Water usage effectiveness, WUE.
• This is liters of water per kilowatt hour of IT load. Lower is better. - Location choices.
• Cool climates reduce cooling needs.
• Regions with low water stress reduce local impact. - Cooling design.
• Closed‑loop systems that recirculate water instead of evaporating large amounts.
• Air cooling where climate allows it. - Energy mix.
• More wind, solar, nuclear means less water per kWh, especially versus old coal and gas plants.
- Water usage effectiveness, WUE.
-
How to interpret the headlines
- When you see “training this model used X Olympic pools of water”, check.
• Does it include power plant water.
• Is it one training run, or training plus months of usage.
• Is it comparing to realistic baselines, like household water or agriculture in that region. - Agriculture and industry still dwarf AI water use today, although local stress near specific data centers can be serious.
- When you see “training this model used X Olympic pools of water”, check.
-
Practical takeaways for you
- If you want lower impact use.
• Prefer providers that publish WUE and water stress data, and that commit to using less water in stressed regions.
• Use smaller, efficient models when you do not need giant ones. Lots of tasks do not need a frontier model.
• Avoid unnecessary compute heavy use, like pointlessly re‑running big jobs.
- If you want lower impact use.
So, yes, AI is non‑trivial for water. A single big model training can equal dozens of homes for a year. But your single chat is closer to a sip or two. The real issue is how many models get trained, where data centers are built, and how clean and water‑efficient the power and cooling setup is.
Short version: the headlines are both right and misleading at the same time.
@sterrenkijker already laid out a nice structured breakdown, so I’ll just come at it from a slightly different angle and sanity‑check the “is this actually a big deal?” part.
- What the scary numbers actually mean
When you see “training Model X used a million liters of water” you’re usually looking at:
- Onsite cooling water +
- Water used at power plants to make the electricity
People love converting that into “Olympic pools” because it sounds dramatic, but never say what the comparison baseline is. A million liters sounds insane until you realize:
- A single large irrigated farm can burn through that in a day.
- A midsize factory can use that in a week.
So yes, it’s large, but it’s industrial‑scale large, not “end of civilization by chatbot” large.
- Order‑of‑magnitude reality check
Very roughly, with all the caveats:
-
Big frontier model training (GPT‑3-ish or larger):
Think tens to hundreds of US households’ yearly indoor water use, once per major training run. Not nothing, but also not “every question you ask kills a river.” -
Inference (you asking questions):
Depending on grid + datacenter, you’re in the territory of a few to a few tens of mL of water per medium chat, counting power‑plant water. So more like a sip than a bottle. The “0.5 L per 10–50 prompts” type numbers some papers use are fine as rough ballparks, but they’re super location‑dependent and honestly a bit too tidy for my taste.
- Where I slightly disagree with the framing
A lot of analyses, including some like what @sterrenkijker summarized, lean hard on “your chat is tiny, don’t worry.” That’s technically correct but also a bit too comforting. The problem is not your chat; it’s:
- Billions of chats per day
- Multiple companies retraining giant models frequently
- Datacenters built in already water‑stressed regions because land and taxes are cheap
So individual use is small, but aggregate demand can absolutely make a dent, especially locally. Locals don’t care that agriculture uses more water nationally if their town’s aquifer is getting squeezed by a new AI campus.
- The part almost nobody talks about
Two big blind spots in most coverage:
-
Temporal footprint
A model might be trained once, but it runs for years. Over that lifetime, inference can dominate the water footprint, not the flashy “we trained this monster model” moment. News pieces usually focus on the training because the numbers are sexier. -
Water stress, not just liters
A liter of water used in a wet, cold region is not the same moral issue as a liter pulled from a drought‑prone basin. You really want:
“Liters of water per kWh × local water‑stress factor × total energy over model life.”
Almost nobody publishes that in a digestible way, which is why the numbers you see feel all over the place.
- How I’d mentally model it
If you want a quick intuitive model without drowning in PDFs:
- Treat AI like “another heavy digital service,” in the same rough league as:
• HD video streaming
• Big multiplayer games
• Crypto mining (although crypto is often worse on pure energy) - For personal behavior, put “AI chats” mentally in the same bucket as “watching some online video” in terms of per‑use impact. Not zero, not huge.
- The real lever is systemic: what regions, what cooling tech, what energy mix.
- Should you actually care personally?
Honestly:
- If you live in a very water‑stressed place and a hyperscale datacenter is being proposed nearby, yes, this is something to care and yell about.
- If you are just chatting with models a few hours a week, your marginal impact is extremely small compared to, say, diet, flying, lawn irrigation, or consumer junk. Anyone telling you to stop using AI chats “to save water” is kind of missing the scale picture.
- Practical filters for the hype
When you read the next dramatic article, check:
- Are they including power‑plant water or only onsite?
- Is it “one training run” or “training + 1 year of usage”?
- Are they comparing AI to something realistic in that region (agriculture, industry), or just waving around pool metaphors?
If they skip those three, I’d treat the conclusion as more vibes than science.
tl;dr:
Yes, AI uses a lot of water in absolute terms, and it can be a real local problem depending on where datacenters are built. Your individual usage is more like a sip than a shower, but at planetary scale, it’s yet another growing industrial load that needs to be constrained and designed properly.