this post was submitted on 01 May 2026
116 points (92.0% liked)

Technology

84257 readers
3918 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
all 39 comments
sorted by: hot top controversial new old
[–] magnue@lemmy.world 2 points 42 minutes ago

If you supplied humans with the same image and asked for the same estimate I'd be curious to know the difference in results.

[–] FauxLiving@lemmy.world 9 points 2 hours ago

I tried to build a deck with my smartphone, it couldn't drive a single nail.

[–] eager_eagle@lemmy.world 24 points 4 hours ago (1 children)

Waste of energy. It's like asking a person to estimate a non-trivial angle. Either use a model trained for that task, or don't bother.

[–] Corkyskog@sh.itjust.works 19 points 3 hours ago (1 children)

The point is they are advertising that these models can do it.

[–] eager_eagle@lemmy.world 2 points 1 hour ago* (last edited 1 hour ago) (1 children)

You’d expect the same answer each time. It’s the same photo, the same model, the same question. But you won’t get the same answer.

I don't know what ads show that, but anyone who knows the first thing about LLMs knows you don't get the same answer twice.

I'd get this expectation 5 years ago when most people weren't familiar with it, but come on... you don't need to feed it an image 500 times to see that.

[–] Sandbar_Trekker@lemmy.today 1 points 17 minutes ago

Technically, you can get the same answer twice from an LLM, but only when you control the full input. When a model is being run, a random seed/hash is applied to the input. If you run the model locally you could force the seed to always be the same so that you would always get the same answer for a given question.

[–] Buffalox@lemmy.world 47 points 5 hours ago (2 children)

It’s the same photo, the same model, the same question. But you won’t get the same answer. Not even close — and the differences are large enough to cause a hypoglycaemic emergency.

OK I wonder if there's something wrong with the photo.
The photo:

WTF!!??
That's like estimating the carbs in 2 slices of standard sandwich bread! Of course not all bread has the same amount of sugar, but a reasonable range based on an average should be a dead easy answer.

I thought the headline sounded crazy, but try to read the article, and it actually becomes worse. I have said it many times before, these AI chatbots should not be legal, they put lives at risk.

[–] inari@piefed.zip 24 points 5 hours ago (5 children)

To be fair there's no way of knowing what the filling is, so the AI may be guessing based on that too

[–] Grail@multiverse.soulism.net 6 points 1 hour ago

Nope, Claude and Gemini both guessed fewer carbs than are in the bread.

[–] Buffalox@lemmy.world 1 points 1 hour ago* (last edited 57 minutes ago) (1 children)

What in the picture indicates any form of filling?
What you can see is cheese, there is probably butter too, but those 2 have zero carbohydrates, so adding carbohydrates based on filling would be pure speculation.
There are no carbohydrates to see beyond the bread.
There is no evidence of any filling, as there is zero bulge in the bread.
The answer should be based on what can be seen, with a remark to that effect, and that there possibly could be more if it contains filling that isn't visible.

The AI could ask about a possible filling, instead of just making shit up with zero evidence.

[–] jim_v@lemmy.world 1 points 22 minutes ago* (last edited 22 minutes ago)

To your point -

If a friend texted me the same picture and question, I would do exactly what you described. Try to give a calculated guess that wouldn't change.

Unless I was lazy and Googled it.

Google's carbohydrate tool says 8g, then the AI overview goes on to contradict that by saying "A standard cheese sandwich typically contains between 25 and 35g."

[–] Ludicrous0251@piefed.zip 21 points 3 hours ago (1 children)

Friendly reminder that LLMs don't do math, they guess what number should come next, just like words.

It can probably link the image to the words "a photo of a sandwich on a plate", and interpret the question as "how many calories are in a sandwich" but from there it is just guessing at the syntax of an answer, but not at finding any truth.

It knows sandwiches have calories and those tend to be 3-4 digit numbers, but also all numbers kinda look the same, so what's to say it's not 2, 5, or 12 digits?

[–] monkeyslikebananas2@lemmy.world 8 points 2 hours ago

Tool-powered agents can do math though. The issue is the fuzziness of it trying to guess carbs. It doesn’t know weight, ingredients, or anything other than a picture. These tools can be useful but not for this. Maybe one day but not yet.

Whoever claims an AI (LLM or agents) can do that and charging their users is lying and defrauding them.

[–] Carnelian@lemmy.world 34 points 5 hours ago

The apps are advertising that they can do this tho. Many of them are aggressively sponsoring YouTubers who advertise you can basically just wave your phone over the food and it takes away all the “work” from traditional calorie counting apps

[–] PatrickYaa@feddit.org 15 points 5 hours ago (1 children)

But the ai assumes itself infallible, at least it could ask...

[–] inari@piefed.zip 8 points 4 hours ago* (last edited 4 hours ago)

That's true, it should ask follow-up questions, or at least clarify its assumptions

[–] MightEnlightenYou@lemmy.world 2 points 5 hours ago* (last edited 5 hours ago) (3 children)

People should read the top comments on Hackernews instead of anyone here, they're more informed on the topic than Lemmy is

[–] brucethemoose@lemmy.world 2 points 1 hour ago* (last edited 1 hour ago)

Better yet, download Qwen 3.5/3.6, with a "raw" notepad like Mikupad. Try it yourself:

https://huggingface.co/ubergarm/Qwen3.6-27B-GGUF

https://github.com/lmg-anon/mikupad

One might observe:

  • Chat formating, and how janky the "thinking" block is.

  • How words are broken up into tokens, not characters.

  • How particularly funky that gets with numbers.

  • Precisely how sampling "randomizes" the answers by visualizing "all possible answers" with the logprobs display.

  • And, thus, precisely how and why carb counting in ChatGPT fails, yet a measly local LLM on a desktop/phone could get it right with a little tooling or adjustment.

This is exactly what OpenAI/Anthropic don't want you to do. They want users dumb and tethered, like a cloud subscription or social media platform. Not cognizant of how tools they are peddling as magic lamps actually work. And why, and how, they're often stupid.

[–] Oisteink@lemmy.world 8 points 4 hours ago

Yeah - if you’re after AI fanbois you should head over there. They’re not that bright, but if you check show and tell you can see what claude’s been ut to last two days

[–] prole@lemmy.blahaj.zone 2 points 3 hours ago

HN is full of techno fascists

[–] psycho_driver@lemmy.world -1 points 5 hours ago* (last edited 5 hours ago) (1 children)

Bruh a couple of months ago I asked it (Gemini) to check the number of characters, including spaces, in a potential game character name because I was working at the time and couldn't stop to check my in-head count. It told me 21--I had counted 20. I thought I must have gotten distracted and miscounted. Later when I had time to actually focus on the issue it turned out AI had miscounted a 20 character string (maybe counting the null terminating character?).

[–] boonhet@sopuli.xyz 18 points 5 hours ago (1 children)

AI doesn't see individual characters, it sees tokens, with most tokens being a word or part of a word. That's why per-character questions have such a high failure rate.