this post was submitted on 15 Aug 2025
623 points (96.0% liked)

Technology

74114 readers
2883 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

The University of Rhode Island's AI lab estimates that GPT-5 averages just over 18 Wh per query, so putting all of ChatGPT's reported 2.5 billion requests a day through the model could see energy usage as high as 45 GWh.

A daily energy use of 45 GWh is enormous. A typical modern nuclear power plant produces between 1 and 1.6 GW of electricity per reactor per hour, so data centers running OpenAI's GPT-5 at 18 Wh per query could require the power equivalent of two to three nuclear power reactors, an amount that could be enough to power a small country.

you are viewing a single comment's thread
view the rest of the comments
[–] Corkyskog@sh.itjust.works 1 points 21 hours ago* (last edited 21 hours ago) (5 children)

Why not... for the ignorant such as myself?

[–] EncryptKeeper@lemmy.world 3 points 20 hours ago* (last edited 20 hours ago) (4 children)

AI models require a LOT of VRAM to run. Failing that they need some serious CPU power but it’ll be dog slow.

A consumer model that is only a small fraction of the capability of the latest ChatGPT model would require at least a $2,000+ graphics card, if not more than one.

Like I run a local LLM with a etc 5070TI and the best model I can run with that thing is good for like ingesting some text to generate tags and such but not a whole lot else.

[–] Corkyskog@sh.itjust.works 1 points 20 hours ago* (last edited 20 hours ago) (3 children)

How slow?

Loading up a website with flash and GIF in the 90s dialup slow... Or worse?

[–] Evono@lemmy.dbzer0.com 3 points 20 hours ago

Basicly I can run 9b models on my 16gb gpu mostly fine like getting responses of lets say 10 lines in a few seconds.

Bigger models if they don't outright crash take for the same task then like 5x or 10x longer so long it isn't even useful anymore

So very worse.

load more comments (2 replies)
load more comments (2 replies)
load more comments (2 replies)