this post was submitted on 25 Jun 2024
648 points (98.6% liked)
LinkedinLunatics
3563 readers
5 users here now
A place to post ridiculous posts from linkedIn.com
(Full transparency.. a mod for this sub happens to work there.. but that doesn't influence his moderation or laughter at a lot of posts.)
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Fwiw a LLM uses as much power as 10 regular Google searches... So it's almost nothing in the grand scope of things. It might even save some for the people who don't know how to utilize search engines properly.
We also need more data centers, not fewer. And they use almost no water compared to other utilities.
I'm not sure that's even a valid comparison? I'd love to know where you got that data point.
LLMs run until they decide to output an end-of-text token. So the amount of power used will vary massively depending on the prompt.
Search results on the other hand run nearly instantaneously, and can cache huge amounts of data between requests, unlike LLMs where they need to run every request individually.
I'd estimate responding to a typical ChatGPT query uses at least 100x the power of a single Google search, based on my knowledge of databases and running LLMs at home.
https://old.lemmy.world/comment/10803727
E: I've cleaned up the comment for saucing ease in case you looked at it right away but I'll quote it here for ease.
Thanks for the links. I was able to find the original source for that claim, which has actually usage numbers: https://iea.blob.core.windows.net/assets/18f3ed24-4b26-4c83-a3d2-8a1be51c8cc8/Electricity2024-Analysisandforecastto2026.pdf
0.3Wh / request for Google 2.9Wh / request for ChatGPT
That does however reference the same paper as your linked articles, which I can't find without a paywall: https://www.sciencedirect.com/science/article/abs/pii/S2542435123003653?dgcid=author
I'd love to know how they came up with that number for ChatGPT, but it looks like I was a bit off with my estimates regardless. There's probably some scaling efficiencies they're taking advantage of at that size.