Ancestors

Toot

Written by David Gerard on 2025-01-28 at 02:19

DeepSeek slaps OpenAI, tech stocks crash

https://awful.systems/post/3389920

=> More informations about this toot | More toots from dgerard@awful.systems

Descendants

Written by swlabr@awful.systems on 2025-01-28 at 03:33

Working on this take:

Aside from stitching together a bigger and bigger trenchcoat, has open AI done anything else? Just goes to show how vacuous LLMs are if this is what it takes to catch up/outpace them.

=> More informations about this toot | More toots from swlabr@awful.systems

Written by istewart@awful.systems on 2025-01-28 at 04:56

We regret to inform you that Stonks have been replaced by Storks*. Go long long legs, big beaks, and affinity for coastal shorelines and wetlands

=> More informations about this toot | More toots from istewart@awful.systems

Written by FredFig@awful.systems on 2025-01-28 at 05:41

I must’ve read the words “Jevon’s Paradox” a hundred times today, I didn’t realize that many people had their livelihood predicated on NVidia going to the moon forevermore.

=> More informations about this toot | More toots from FredFig@awful.systems

Written by EtnaAtsume@lemmy.world on 2025-01-28 at 06:06

Servers must be having a hell of a time since I can’t get a verification code sent to my email 😆

=> More informations about this toot | More toots from EtnaAtsume@lemmy.world

Written by wndy@awful.systems on 2025-01-28 at 15:12

Like, it’s not even THAT much better. I mean, not so much so that everyone should flood it lmao. The main plus was no restriction on tokens used, but that’s useless when it’s getting overloaded all the time.

I would say it’s just barely noticeably better than the free tier of GPT. Which makes it a little annoying to go back but w/e.

=> More informations about this toot | More toots from wndy@awful.systems

Written by Evinceo@awful.systems on 2025-01-28 at 22:53

Can’t people run it locally supposedly?

=> More informations about this toot | More toots from Evinceo@awful.systems

Written by Architeuthis@awful.systems on 2025-01-29 at 11:35

The 671B model although ‘open sourced’ is a 400+GB download and is definitely not runnable on household hardware.

=> More informations about this toot | More toots from Architeuthis@awful.systems

Written by gerikson@awful.systems on 2025-01-29 at 16:00

What? You don’t have a spare $6,000 to run nonsense generators at home? Peasant…

xcancel.com/carrigmat/status/1884244369907278106?…

=> More informations about this toot | More toots from gerikson@awful.systems

Written by swlabr@awful.systems on 2025-01-29 at 17:05

What? You don’t have a spare $6,000 to run nonsense generators at home?

Really sorry about this but this sounds like the premise to a shitty boomer joke:

“If I wanted to spend another 6000 for a home nonsense generator, I’d get married again!”

etc.

=> More informations about this toot | More toots from swlabr@awful.systems

Written by wndy@awful.systems on 2025-01-29 at 11:42

Not people who can’t afford 100k to spin up their own servers. It’s going to be a game changer for AI startups and such though who won’t have to spend as much as previously thought.

=> More informations about this toot | More toots from wndy@awful.systems

Written by gerikson@awful.systems on 2025-01-28 at 06:36

SoftBank, one of the big backers behind Stargate, is notorious for WeWork-level funding disasters — but five days from floating the idea to the crash might be a new record.

I LOLd

=> More informations about this toot | More toots from gerikson@awful.systems

Written by self@awful.systems on 2025-01-30 at 00:39

Is the R1 model better than all existing models? Well, it benchmarks well. But everyone trains their models to the benchmarks hard. The benchmarks exist to create headlines about model improvements while everyone using the model still sees lying slop machines. No, no, sir, this is much finer slop, with a bouquet from the rotting carcass side of the garbage heap.

[…]

This crash doesn’t mean AI sucks now or that it’s good now. It just means OpenAI, and everyone else whose stock dipped, was just throwing money into a fire. But we knew that.

Slop generators are cheap now, and that’s a sea change — but the output is still terrible slop, just more of it.

this bares repeating. I’ve seen quite a few people declare that DeepSeek fixes all of the issues with LLMs as a technology, but that just isn’t true. a DeepSeek LLM is still an unreliable plagiarism machine with no known use case trained on massive amounts of stolen data, even if OpenAI and other American ghouls were the ones who did the theft in the first place.

there’s a small victory in that Altman and friends were exposed very publicly as lying grifters, and that’s worth celebrating. but it’s very important to not get swept up in a hype wave, especially one crafted by people who are much more competent at managing public opinion than Altman & co. from what I understand: no, this thing isn’t meaningfully open source. no, you can’t run the good version at home. sure, it performs great at the benchmarks we know were designed to be cheated. yeah, DeepSeek LLMs are probably still an environmental disaster for the same reason most supposedly more efficient blockchains are — perverse financial incentives across the entire industry.

but hey, good news for the boy genius Prompt Engineer at your company: he gets to requisition another top end gaming PC, absolutely drowning in RGB, to run the shit version of DeepSeek on. maybe in a couple months he can spin switching from OpenAI’s rentseeking to a DeepSeek LLM startup’s slightly cheaper rentseeking into a mild pay bump.

=> More informations about this toot | More toots from self@awful.systems

Written by David Gerard on 2025-01-30 at 00:47

so you can run the good version at home! this thread tells how to build a workstation for it.

tl;dr 768GB RAM.

with that, you can run the largest deepseek model, or even open a tab in chrome

apparently it’s not very fast, but it does in fact do the stuff

=> More informations about this toot | More toots from dgerard@awful.systems

Written by self@awful.systems on 2025-01-30 at 01:08

ah, I stand corrected! the figures I was looking at previously were for doing it at acceptable speeds in a data center.

can you imagine the intensity of the RGB in the boy genius Prompt Engineer’s new $6000 custom top end gaming PC with server components? maybe they’ll have the LLM slowly plagiarize them a Python script that turns on more RGB when the GPU’s under load.

=> More informations about this toot | More toots from self@awful.systems

Written by blakestacey@awful.systems on 2025-01-30 at 01:38

yeah, DeepSeek LLMs are probably still an environmental disaster for the same reason most supposedly more efficient blockchains are — perverse financial incentives across the entire industry.

the waste generation will expand to fill the available data centers

oops all data centers are full, we need to build more data centers

=> More informations about this toot | More toots from blakestacey@awful.systems

Proxy Information
Original URL
gemini://mastogem.picasoft.net/thread/113903584446516571
Status Code
Success (20)
Meta
text/gemini
Capsule Response Time
297.066112 milliseconds
Gemini-to-HTML Time
3.454952 milliseconds

This content has been proxied by September (3851b).