this post was submitted on 24 Aug 2025
325 points (94.5% liked)

Technology

74361 readers
2931 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
top 50 comments
sorted by: hot top controversial new old
[–] archchan@lemmy.ml 9 points 5 hours ago

Knowing Google, they care more about blurring the lines between AI and reality to confuse and force it onto people than they do about saving a few dollars on storage costs.

[–] klemptor@startrek.website 45 points 11 hours ago

I'm huge into makeup, and I watch a lot of beauty content on YouTube because I want to see how certain makeup looks and performs before I buy it. This AI bullshit defeats the purpose of demonstrating makeup.

[–] FauxLiving@lemmy.world 83 points 13 hours ago (3 children)

“AI”

Sharpening, Denoising and upscaling barely count as machine learning. They don’t require AI neural networks.

[–] CeeBee_Eh@lemmy.world 9 points 7 hours ago (1 children)

They don’t require AI neural networks.

Sharpening and denoising don't. But upscalers worth anything do require neural nets.

Anything that uses a neural network is the definition of AI.

[–] ccunix@sh.itjust.works 3 points 6 hours ago (2 children)

Not true

Company I used to work for had excellent upscalers running on FPGAs that they developed 20+ years ago.

The algorithms have been there for years, just AI gives it bit of marketing sprinkle to something that has been a solved problem for years.

[–] CeeBee_Eh@lemmy.world 1 points 40 minutes ago

Well, the algorithms that make up many neural networks have existed for over 60 years. It's only recently that hardware has been able to make it happen.

AI gives it bit of marketing sprinkle to something that has been a solved problem for years.

Not true and I did say "any upscaler that's worth anything". Upscaling tech has existed at least since digital video was a thing. Pixel interpolation is the simplest and computationally easiest method. But it tends to give a slight hazy appearance.

It's actually far from a solved problem. There's a constant trade-off beyond processing power and quality. And quality can still be improved by a lot.

[–] Probius@sopuli.xyz 4 points 5 hours ago

Depends on what you're trying to upscale.

[–] hushable@lemmy.world 39 points 12 hours ago* (last edited 12 hours ago) (2 children)

Sharpening is a simple convolution, doesn't even count as ML.

I really hate that everything gets the AI label nowadays

[–] FauxLiving@lemmy.world 26 points 12 hours ago (1 children)

The “ai bad” brainrot has everyone thinking that any algorithm is AI and all AI is ChatGPT.

[–] hushable@lemmy.world 7 points 6 hours ago

just today someone told me that Vocaloid was also AI music, they are either too dumb to make some basic fact-checking or true believers trying to hype up AI by any means necessary

load more comments (1 replies)
[–] Evotech@lemmy.world 4 points 9 hours ago

But you can use AI for that

[–] jerkface@lemmy.ca 1 points 5 hours ago (1 children)

This is shitty journalism that massively distorts what actually happened. It's just traditional video filters, and AI panic.

[–] Sp00kyB00k@lemmy.world 5 points 5 hours ago (1 children)

There is no AI panic. There is a distrust against the intention of the companies pushing it. Can you trust Google, Amazon, Microsoft, Meta, Anthropic etc?

[–] jerkface@lemmy.ca 0 points 3 hours ago* (last edited 3 hours ago)

There is an AI panic, just like there was a microprocessor panic 50 years ago. Distrust and panic are different things. There is also AI distrust. There is also an AI revolution, an AI bubble, and a whole new AI epoch. There's lots of AI shit going on right now, and panic is certainly one of them.

This article is AI panic because it's what we would call a hallucination if an LLM wrote it. There is no AI in this story. People in a panic often jump at nothing.

[–] someguy3@lemmy.world 20 points 10 hours ago (1 children)

It's very likely to do with compression codecs to save money.

[–] RedditRefugee69@lemmynsfw.com 8 points 6 hours ago

Ostensibly, yes. Just like the Patriot Act was to fight terrorism.

[–] Kissaki@feddit.org 38 points 14 hours ago

Nice

(linked from the article about a Netflix series upscale)

[–] jjjalljs@ttrpg.network 26 points 13 hours ago (2 children)

Seems like this should be illegal, Google should be broken up, and its leadership imprisoned

[–] KairuByte@lemmy.dbzer0.com 4 points 9 hours ago (2 children)

I’m down for a breakup but I don’t see how we could twist this into illegality.

[–] jjjalljs@ttrpg.network 4 points 8 hours ago (2 children)

You could probably make it illegal to alter people's videos without their explicit consent. But also the Republicans have shown us that laws mean what the people in charge want

[–] CeeBee_Eh@lemmy.world 2 points 7 hours ago (1 children)

without their explicit consent.

By signing up to this service you agree to allow us to alter or modify your content as we require for efficient operation or to increase content engagement

[–] jjjalljs@ttrpg.network 1 points 6 hours ago

You can make that kind of thing illegal. I think "shrink wrap eulas" are dubious. Rule that fine print with a bunch of other stuff doesn't count as explicit. Like there are rules now about cookie acceptance that has changed how the web works, and most sites don't try to hide the cookie thing because that's against the rules.

[–] KairuByte@lemmy.dbzer0.com 3 points 8 hours ago (1 children)

I kinda doubt you’d be able to write a law that would actually have the effect you’re looking for. In the case of what you just wrote, all YouTube would need to do is write into their ToS that by uploading to their platform you’ve given them explicit permission to alter the video for purposes of storage space or increasing/decreasing quality.

[–] jjjalljs@ttrpg.network 1 points 7 hours ago

I think you're under estimating what the law can do, probably because most of the time it's used to bolster rich assholes.

[–] jonesey71@lemmus.org -1 points 6 hours ago (1 children)

It might not be currently illegal but I think there should be a law defining "crimes against society" that only applies to corporations and politicians. It could be vague like "disorderly conduct" but just for corpos and politicians and would include things like lying to the public and could have punishments like corpos losing their business license (death) and banishment to the moon/sun for politicians.

[–] KairuByte@lemmy.dbzer0.com 4 points 5 hours ago

Overly vague laws are never a good thing.

load more comments (1 replies)
[–] Ganbat@lemmy.dbzer0.com 65 points 16 hours ago* (last edited 9 hours ago) (3 children)

From what I've seen so far, the case here seems to be that it's only being done to shorts, and what's happening is that they're being permanently stored at a lower quality and size and are then upscaled on the fly. I mean... it feels kinda fair to me. Theres a good reason YouTube has so little competition, and it's because how hard and expensive maintaining a service like this is. They're always trying to cut costs, and storage is gonna be a big cost. Personally, I'm glad it's just shorts for now. It absolutely shouldn't be happening to people who are paying for the service or making money for it, though.

[–] ObviouslyNotBanana@piefed.world 64 points 15 hours ago (1 children)

I mean yeah, it doesn't seem entirely unreasonable. But if it actually was reasonable, wouldn't they just inform the uploader?

[–] T156@lemmy.world 23 points 15 hours ago (1 children)

Or give an option to toggle. Surely letting people turn it off would save them even more resources, if they don't have to bother with upscaling the video in the first place.

[–] 01189998819991197253@infosec.pub 14 points 14 hours ago (1 children)

It likely costs them less to upscale than it does to store and serve a full sized video, so they're not giving the uploader the choice.

[–] exu@feditown.com 3 points 8 hours ago

Storage is very cheap. This only makes sense if they actually do the upscaling client side

[–] Dragomus@lemmy.world 25 points 13 hours ago (1 children)

It's not so much that they down- and upscale the video of shorts, their algorithm changes the look of people. It warps skin and does a strange sort of sharpening that makes things look quite unreal and almost plastic.

It is a filter that evens the look with images generated by, say, grok or one of the other AI filters.

In a year people will think that "AI-look" is a normal video look, and stuff generated with it is what humans can look like. We will see crazed AI-fashion looks popping up.

[–] boonhet@sopuli.xyz 8 points 13 hours ago

Yeah, upscaling can generate artefacts and such.

[–] Zarxrax@lemmy.world 31 points 15 hours ago (6 children)

It would not make any sense for them to be upscaled on the fly. It's a computationally intensive operation, and storage space is cheap. Is there any evidence of it being done on the fly?

[–] baggins@lemmy.ca 18 points 15 hours ago (5 children)

It would if they can do it on your device.

load more comments (5 replies)
load more comments (5 replies)
[–] NigelFrobisher@aussie.zone 0 points 4 hours ago

Counterpoint: no they couldn’t.

[–] basiclemmon98@lemmy.dbzer0.com 5 points 12 hours ago* (last edited 12 hours ago) (1 children)

Well, youtube is not even intended to host quality content anymore, but besides that, this appears to just be visual tweaks. This title is trying to be vague enough that one could assume it's tweaking the content itself which would be of real concern. It's not doing that (for now). Video graphics seems like an awefully minor thing to be screaming about AI over. Especially when AI has actual reprocussions in the knowledge accuracy sector.

[–] xthexder@l.sw0.com 7 points 10 hours ago

From what I've heard this mostly happens on YT Shorts, and the AI upscaling they're doing is making people look like plastic and uncanny as hell.
I haven't noticed on normal videos, since that's pretty much all I watch.

[–] desmosthenes@lemmy.world 6 points 13 hours ago

yucky, shorts lol

load more comments
view more: next ›