Tried out the new and popular “Deepseek” LLM with my standard “tell me facts about the author of PCalc” query. At least half were misleading or straight up hallucinations. LLMs are not a suitable technology for looking up facts, and anybody who tells you otherwise is… probably trying to sell you a LLM.
=> View attached media | View attached media | View attached media | View attached media
=> More informations about this toot | More toots from jamesthomson@mastodon.social
I then asked for a list of ten Easter eggs in the app, and every single one was a hallucination, bar the Konami code, which I did actually do.
=> View attached media | View attached media | View attached media | View attached media
=> More informations about this toot | More toots from jamesthomson@mastodon.social
@jamesthomson I wonder if some other calculator had some of these Easter eggs?
It’s very specific.
=> More informations about this toot | More toots from samir@functional.computer
@samir LLMs are great at specific sounding fabrications!
=> More informations about this toot | More toots from jamesthomson@mastodon.social
@jamesthomson It’s really impressive how on-point they are. I’d love to see some of those Easter eggs. (Please don’t actually waste your time though.)
Total nonsense, of course, but I think that’s a feature for a lot of people.
=> More informations about this toot | More toots from samir@functional.computer
@samir @jamesthomson They are plausibility machines. everything they say is plausible - & that's the problem.
As an expert, it takes seconds to debunk the 20% that's obviously shite, 75% is correct, and that last 5% takes you two hours and a load of research, to be sure the plausible lie isn't some obscure fact you didn't know/forgot!
(Not quite the same percentages when it's you asking about your own life, I bet!)
=> More informations about this toot | More toots from Dss@infosec.exchange This content has been proxied by September (3851b).Proxy Information
text/gemini