this post was submitted on 29 Aug 2025
686 points (98.6% liked)

Technology

74599 readers
3864 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] finitebanjo@lemmy.world 19 points 14 hours ago (1 children)

Because it costed them money, lol. The suits upstairs gave a quote in the article talking about how they will withdraw AI from all 500 locations they were implemented, and it also talks about how McDonalds did the exact same little dance over a year ago.

[–] Prove_your_argument@piefed.social 6 points 14 hours ago (4 children)

The mcdonalds thing was because the model they implemented was misinterpreting people and incorrectly placing orders. Yeah, obviously the thing wasn't working right so they pulled that. Sounds just like early personal assistants on phones and other devices, hell my wife still struggles with those. They clearly needed more time developing and testing it with a diverse range of customers from all over. I don't know if they trained it using recordings from real drive throughs from all over, but they should have.

The 18000 water example probably didn't cost anyone anything. Regardless of if it was intentional or not, it wouldn't have been fulfilled as part of an order. They mention it "crashing the system" - whatever that means in this context is impossible to know. Did it take down all of taco bell? Did it cause the LLM to stop responding on JUST this one site? All of them? Did it eventually time out and start working right? it's impossible to know because the details just aren't there and we have no insight as to the system architecture. I always assume there is a method to rely on traditional ordering where a person listening in while the chatbot talks to the person can take over and fix the problem. It's not like there aren't drive through workers still there.

[–] _stranger_@lemmy.world 4 points 10 hours ago* (last edited 10 hours ago) (1 children)

A drive through menu shouldn't have crippling security vulnerabilities that are trivial to reproduce just by speaking near it.

McDonald's thing was because "AI" is a scam.l, and the only way to make money off of it is to shut down your AI selling business after pocketing as much VC as possible (unless your Nvidia of course).

Totally agree. Without details we don't have any idea what actually went wrong.

[–] finitebanjo@lemmy.world 4 points 13 hours ago

Even if it's only a receipt for 18,000 waters or it fills up a screen it costs them time and resources.

Every single AI halucinates, always has and always will. It's useless for this.

[–] Prove_your_argument@piefed.social -1 points 14 hours ago (2 children)

Really the only cost here is the impact to consumer attitudes towards taco bell and AI because the video and news of this is circulating. One error is whatever, but public perception doesn't typically involve much critical thinking.

People are still irrationally terrified of all manner of technology even though science backs it up, like vaccines.

[–] chonglibloodsport@lemmy.world 5 points 10 hours ago (1 children)

What do you mean science backs it up? Science is finding massive social problems with technology all the time. Social media and its negative impacts on mental health (especially for teen and preteen girls), for example. Microplastics everywhere, for another. Climate change anyone?

[–] Prove_your_argument@piefed.social 0 points 9 hours ago (1 children)

One person commits suicide from LLMs: OH MY GOD BAN ALL LLMS REQUIRE IDS AND REGULATE THEM TO THE GROUND. (Please ignore all cases of suicide for therapy patients. Therapy is always effective and results in positive outcomes, right?)

One person dies in a car crash with a semi-autonomous L2 car: OH MY GOD BAN ALL SELF DRIVING CARS PEOPLE ARE DYING LEFT AND RIGHT (ignore billions of miles per significant accident for the robot vs hundreds of thousands for humans.)

Just two examples, and odds are you have your own personal opinion about how you absolutely loathe one or another. Maybe you feel like you're losing control with self driving cars, or maybe you feel like chat bots have encroached on your field of work because you're a dev and we've had countless layoffs after over-hiring during covid lockdowns.) Either way, there's studies and there's kneejerk reactions, and in our world the latter is winning right now.

[–] chonglibloodsport@lemmy.world 3 points 8 hours ago

Sorry dude, but cars are technology too, not just self driving cars. Every death due to cars is a technology death. You can’t escape the reality of tradeoffs.

[–] finitebanjo@lemmy.world -1 points 13 hours ago (1 children)

Unlike vaccines, AI has no use case and is always a net negative.

[–] Prove_your_argument@piefed.social 0 points 11 hours ago (1 children)

I just don't agree man. It won't do what most people want it to do, it doesn't at all work like some kind of science fiction "AI" that we classically think of. It's great at organizing patterns and helping create models to do a specific use case, but when you try to do some real convoluted multilevel thing it just doesn't.

We've been using ML for a ton of tools in tech for a long time. Crowdstrike, Darktrace and Abnormal are all very successful in the realm of what they do thanks to ML (aka "AI".)

OCR has been used for so long and has gotten really fucking good, thanks to ML.

I don't think we're gonna replace humans for thinking, but we can definitely replace them for boring repetitive actions.

[–] finitebanjo@lemmy.world 1 points 9 hours ago* (last edited 9 hours ago) (1 children)

We're talking about different things. This article is about Language Models. The discussion is about Language Model.

If you ask a language model via prompt to organize patterns and create models you will get slop that small children would recognize is wrong. It's garbage.

[–] Prove_your_argument@piefed.social 0 points 9 hours ago (1 children)

What's the architecture of taco bell's implementation?

Which LLM are they using?

[–] finitebanjo@lemmy.world 1 points 7 hours ago

Does not matter, all useless.