this post was submitted on 01 Jul 2025
2110 points (98.4% liked)

Microblog Memes

8385 readers
2765 users here now

A place to share screenshots of Microblog posts, whether from Mastodon, tumblr, ~~Twitter~~ X, KBin, Threads or elsewhere.

Created as an evolution of White People Twitter and other tweet-capture subreddits.

Rules:

  1. Please put at least one word relevant to the post in the post title.
  2. Be nice.
  3. No advertising, brand promotion or guerilla marketing.
  4. Posters are encouraged to link to the toot or tweet etc in the description of posts.

Related communities:

founded 2 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] jsomae@lemmy.ml 12 points 2 days ago* (last edited 2 days ago) (16 children)

I know she's exaggerating but this post yet again underscores how nobody understands that it is training AI which is computationally expensive. Deployment of an AI model is a comparable power draw to running a high-end videogame. How can people hope to fight back against things they don't understand?

[–] cantstopthesignal@sh.itjust.works 30 points 2 days ago* (last edited 2 days ago) (1 children)

She's not exaggerating, if anything she's undercounting the number of tits.

[–] MotoAsh@lemmy.world 4 points 2 days ago

Well you asked for six tits but you're getting five. Why? Because the AI is intelligent and can count, obviously.

[–] domdanial@reddthat.com 25 points 2 days ago (2 children)

I mean, continued use of AI encourages the training of new models. If nobody used the image generators, they wouldn't keep trying to make better ones.

[–] jsomae@lemmy.ml 3 points 2 days ago (1 children)

you are correct, and also not in any way disagreeing with me.

[–] domdanial@reddthat.com 2 points 2 days ago
load more comments (1 replies)
[–] FooBarrington@lemmy.world 20 points 2 days ago (22 children)

It's closer to running 8 high-end video games at once. Sure, from a scale perspective it's further removed from training, but it's still fairly expensive.

[–] jsomae@lemmy.ml 2 points 2 days ago (4 children)

really depends. You can locally host an LLM on a typical gaming computer.

[–] FooBarrington@lemmy.world 9 points 2 days ago

You can, but that's not the kind of LLM the meme is talking about. It's about the big LLMs hosted by large companies.

[–] floquant@lemmy.dbzer0.com 5 points 2 days ago* (last edited 2 days ago) (2 children)

True, and that's how everyone who is able should use AI, but OpenAI's models are in the trillion parameter range. That's 2-3 orders of magnitude more than what you can reasonably run yourself

load more comments (2 replies)
[–] Thorry84 5 points 2 days ago* (last edited 2 days ago)

Well that's sort of half right. Yes you can run the smaller models locally, but usually it's the bigger models that we want to use. It would also be very slow on a typical gaming computer and even a high end gaming computer. To make it go faster not only is the hardware used in datacenters more optimised for the task, it's also a lot faster. This is both a speed increase per unit as well as more units being used than you would normally find in a gaming PC.

Now these things aren't magic, the basic technology is the same, so where does the speed come from? The answer is raw power, these things run insane amounts of power through them, with specialised cooling systems to keep them cool. This comes at the cost of efficiency.

So whilst running a model is much cheaper compared to training a model, it is far from free. And whilst you can run a smaller model on your home PC, it isn't directly comparable to how it's used in the datacenter. So the use of AI is still very power hungry, even when not counting the training.

[–] CheeseNoodle@lemmy.world 3 points 2 days ago

Yeh but those local models are usually pretty underpowered compared to the ones that run via online services, and are still more demanding than any game.

[–] brucethemoose@lemmy.world 1 points 2 days ago* (last edited 2 days ago)

Not at all. Not even close.

Image generation is usually batched and takes seconds, so 700W (a single H100 SXM) for a few seconds for a batch of a few images to multiple users. Maybe more for the absolute biggest (but SFW, no porn) models.

LLM generation takes more VRAM, but is MUCH more compute-light. Typically one has banks of 8 GPUs in multiple servers serving many, many users at once. Even my lowly RTX 3090 can serve 8+ users in parallel with TabbyAPI (and modestly sized model) before becoming more compute bound.

So in a nutshell, imagegen (on an 80GB H100) is probably more like 1/4-1/8 of a video game at once (not 8 at once), and only for a few seconds.

Text generation is similarly efficient, if not more. Responses take longer (many seconds, except on special hardware like Cerebras CS-2s), but it parallelized over dozens of users per GPU.


This is excluding more specialized hardware like Google's TPUs, Huawei NPUs, Cerebras CS-2s and so on. These are clocked far more efficiently than Nvidia/AMD GPUs.


...The worst are probably video generation models. These are extremely compute intense and take a long time (at the moment), so you are burning like a few minutes of gaming time per output.

ollama/sd-web-ui are terrible analogs for all this because they are single user, and relatively unoptimized.

load more comments (19 replies)
[–] PeriodicallyPedantic@lemmy.ca 6 points 2 days ago (12 children)

Right, but that's kind of like saying "I don't kill babies" while you use a product made from murdered baby souls. Yes you weren't the one who did it, but your continued use of it caused the babies too be killed.

There is no ethical consumption under capitalism and all that, but I feel like here is a line were crossing. This fruit is hanging so low it's brushing the grass.

[–] Randelung@lemmy.world 3 points 2 days ago

"The plane is flying, anyway."

load more comments (11 replies)
load more comments (12 replies)