Well that’s of the same order of magnitude as the quoted figure. I was suggesting that it sounded vastly larger than it should be.
Well that’s of the same order of magnitude as the quoted figure. I was suggesting that it sounded vastly larger than it should be.
It’s true, I don’t know how large the models are that are being accessed in data centers. Although if the article’s estimate is correct, it’s sad that such excessively-demanding models are always being used for use-cases that could often be handled with much lower power usage.
140Wh seems off.
It’s possible to run an LLM on a moderately-powered gaming PC (even a Steam Deck).
Those consume power in the range of a few hundred watts and they can generate replies in a seconds, or maybe a minute or so. Power use throttles down when not actually working.
That means a home pc could generate dozens of email-sized texts an hour using a few hundred watt-hours.
I think that the article is missing some factor, such as how many parallel users the racks they’re discussing can support.
You thinking of Apple headsets. These are budget things, maybe $300.
He decided that it was unethical to have an AI/LLM impersonate a real person, but set up the “wizard” as an AI assistant for his fake crypto site helpline.
That was covered pretty well already!
Or maybe it’s using Fluidic logic.