174
Deepseek bangers (hexbear.net)
you are viewing a single comment's thread
view the rest of the comments
[-] FloridaBoi@hexbear.net 34 points 1 week ago* (last edited 1 week ago)

Someone told me it’s like 97% more energy efficient or that it consumes 97% less energy. Is that true

Edit: this comment has a summary saying that the model has 93% compression ratios so maybe that’s there efficiency number canes from

[-] Yuritopiaposadism@hexbear.net 31 points 1 week ago

so you are saying ai is destroying the planet for nothing?!

[-] KurtVonnegut@hexbear.net 24 points 1 week ago

Literally the same business model as Bitcoin.

[-] FloridaBoi@hexbear.net 6 points 1 week ago

Funny how it also caused a massive demand spike for gpus

[-] peeonyou@hexbear.net 3 points 1 week ago

why's that funny? it's because of what GPUs do.. namely massively parallel computations

[-] FumpyAer@hexbear.net 5 points 1 week ago

We also got news from Trump of a huge tariff to all chips manufactured outside the US, so that may be panic buying as well.

[-] peeonyou@hexbear.net 4 points 1 week ago

ah! i hadn't heard of that

[-] Pastaguini@hexbear.net 15 points 1 week ago

Not for nothing - for the investment portfolio of energy investors. You know, the highest priority of all.

[-] ragebutt@lemmy.dbzer0.com 28 points 1 week ago

The exact figures aren’t documented but it’s a pretty massive decline in energy usage (though probably not 97%), enough so that stocks related to power consumption took a pretty notable hit

[-] sewer_rat_420@hexbear.net 13 points 1 week ago

It consumes less energy now, but also consumed less energy in its creation. This is directly reflected in the cost to the user - the API is 10-30x cheaper per token than openAI

[-] glimmer_twin@hexbear.net 9 points 1 week ago

Isn’t 97% more efficient still really bad compared to like, a search engine

That's because LLMs aren't supposed to be search engines. They are pretty good at summarizing documents in certain cases, but don't have a big enough context window to effectively plow through massive troughs of data.

[-] Stolen_Stolen_Valor@hexbear.net 7 points 1 week ago

The “AI” is effectively just autocomplete powered by the internet. It could by powered by your 2001 flip phone probably. The whole thing is smoke and mirrors, hype, and snake oil bought by people who don’t understand what’s happening or people only concerned with line go up.

It could by powered by your 2001 flip phone probably

LLMs are fundamentally billion-dimensional logistic regressions that require massive context windows and training sets. It is difficult to create a more computationally expensive system than an LLM for that reason. I have a fairly nice new laptop, and it can barely run Deepseek-r1:14b (14 billion parameter model. Not technically the same model as deepseek-r1:671b as it is a fine-tune of qwen-2.5:14b that uses the deepseek chain reasoning. It can run the 7b model fine, however. There isn't a single piece of consumer-grade hardware capable of running the full 671b model.

this post was submitted on 27 Jan 2025
174 points (99.4% liked)

technology

23539 readers
328 users here now

On the road to fully automated luxury gay space communism.

Spreading Linux propaganda since 2020

Rules:

founded 4 years ago
MODERATORS