I asked “What was the moneyline for Thursday's game between the Chargers and Broncos?”
=> More informations about this toot | More toots from gruber@mastodon.social
Here’s Siri's utterly botched result for “What was the Vikings-Bears score last week?”, and ChatGPT's response to the exact same query.
Siri is technically correct that the Vikings and Bears did not play between Dec 8 and last Saturday, Dec. 15, but they did play last week on Sunday. [Edit: Monday] ChatGPT completely nailed it, with the correct score in the first paragraph and more details below, including sources at ESPN and Reuters.
=> View attached media | View attached media
=> More informations about this toot | More toots from gruber@mastodon.social
@gruber I don’t think Siri is using AI, that’s just the old Siri, you’ve been able to always ask for scores haven’t you?
=> More informations about this toot | More toots from monorailtimes@worldkey.io
@monorailtimes @gruber This has always been laughably bad for me.
=> More informations about this toot | More toots from webology@mastodon.social
@webology @gruber The scores part is definitely old Siri, I’ve had success with other questions from Siri that use Apple Intelligence
=> More informations about this toot | More toots from monorailtimes@worldkey.io
@gruber Oddly, just typing “Vikings Bears score” gives me the correct info
=> More informations about this toot | More toots from tahewett@mastodon.social
@gruber
Asking a LLM for accurate answers is misusing it. It simply isn't in its nature to do that.
If you want accurate answers, you need an old fashioned hand coded application. It was designed for it.
LLMs can be good for suggestions, inspiration or first drafts.
By the way, there have been non-LLM applications for that since forever. But they have never had all the money LLMs have behind, so they have always been small tools.
A tailored app will always be better than a LLM for the same feature. The time to use an LLM is when it is the only viable option.
=> More informations about this toot | More toots from jgg@qoto.org
@gruber just asking for the score, it worked correctly. Asking it what the score was for Week 15, also worked. I think it was confused by “last week”.
But, if you’re using ChatGPT integration with Siri, you can say “Ask ChatGPT what the Vikings-Bears score was last week”
=> More informations about this toot | More toots from jmabeebiz@mastodon.social
@jmabeebiz @gruber like it’s always confused by “what’s the weather like tomorrow” after midnight. Technically correct, but socially akward!
=> More informations about this toot | More toots from Luk
@gruber Siri is being pedantic but technically correct (the best kind of correct!) in that Monday’s (not Sunday’s) game is still “this week”
Asking for the most recent score from either team gives the correct answer.
=> More informations about this toot | More toots from sfrazer434@mastodon.social
@sfrazer434 But a “week” in football has a different meaning.
=> More informations about this toot | More toots from gruber@mastodon.social
@gruber Everyone here in Chicagoland thanks you for reminding us of that game.
=> More informations about this toot | More toots from drdrang@fosstodon.org
@gruber forgive me if I am missing something but isn’t chat gpt based Siri improvements supposed to land in March? Like, Siri can do three things and if you step outside those carefully controlled boundaries, you are doomed. I’m assuming that’s going to continue to be true until at least March
=> More informations about this toot | More toots from crazybutable@mastodon.social
@crazybutable I can only judge what we have today, and today, Siri is an embarrassment for very typical and easy questions like this.
=> More informations about this toot | More toots from gruber@mastodon.social
@gruber My self hosted LLM is llama3.2 running on my M1 Max Mac Studio and backed with duckduckgo. That nailed it as well. The links are also correct.
There’s really no excuse for #Siri being worse than my self hosted solution.
=> More informations about this toot | More toots from tcb@infosec.exchange
@gruber Just one more example why the ChatGPT app is assigned to my action button.
=> More informations about this toot | More toots from wetstevs@techhub.social
@gruber There was an ex-Scientologist who was offering a million-dollar prize if any Scientologist could prove they had the abilities promised by Scientologist's "Operating Thetan" level, which has something to do with having cause over all matter, time, and space. And I remember Jason Beghe, another ex-Scientologist, laughing and saying he'd give a million dollars to anyone who could prove they had achieved the most basic level of Scientology, which is Clear—a state in which a person no longer gets sick, etc.
I give that long, rambling anecdote because it reminds me of this ever-more seemingly quixotic and ambitious journey Apple is on.
They are promising the world, but I would settle for what they promised even pre-Siri with Voice Control that came out with the iPhone 3GS.
I am personally not even hitting 50% success rate with calling the right contact in 2024 with Siri. Maybe it's my contacts' names in particular, but the most common problem is that when I call a contact using Siri it tries to call me instead (thinking I'm saying my own name) and then says, there's no work number for (my name), which number would you prefer to call (of the other numbers listed for me)?
I have no idea why this keeps persisting but it has for a year or so. I'm not saying my own name, and I'm not saying work.
So I do not hold Siri to the standard of ChatGPT. I hold it to what was promised in Voice Control that came out with the iPhone 3GS, and it's failing at that. And I'm sure I could call Apple and troubleshoot and maybe there are a million reasons why, but with a better product you don't have to do that. And I honestly dread the thought of calling them and going through the song and dance when we both know it's just problematic. And it's gotten worse over time somehow. I don't recall having all these listening issues with my first iPhone, which was the 5s.
=> More informations about this toot | More toots from swingerofbirch@c.im This content has been proxied by September (ba2dc).Proxy Information
text/gemini