this post was submitted on 01 Jul 2025
2119 points (98.4% liked)
Microblog Memes
8456 readers
2559 users here now
A place to share screenshots of Microblog posts, whether from Mastodon, tumblr, ~~Twitter~~ X, KBin, Threads or elsewhere.
Created as an evolution of White People Twitter and other tweet-capture subreddits.
Rules:
- Please put at least one word relevant to the post in the post title.
- Be nice.
- No advertising, brand promotion or guerilla marketing.
- Posters are encouraged to link to the toot or tweet etc in the description of posts.
Related communities:
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I do, because they're not at full load the entire time it's in use
They are, it'd be uneconomical not to use them fully the whole time. Look up how batching works.
I mean I literally run a local LLM, while the model sits in memory it's really not using up a crazy amount of resources, I should hook up something to actually measure exactly how much it's pulling vs just looking at htop/atop and guesstimating based on load TBF.
Vs when I play a game and the fans start blaring and it heats up and you can clearly see the usage increasing across various metrics
One user vs a public service is apples to oranges and it's actually hilarious you're so willing to compare them.
It's literally the same thing, the obvious difference is how much usage it's getting at a time per gpu, but everyone seems to assume all these data centers are running at full load at all times for some reason?
It's explicitly and literally not the same thing.
The highest likelihood is you have literally no idea how any of this works and are just joining the crowd of AI bad because energy usage and have done zero research or hands on knowledge of how these tools actually work.