this post was submitted on 10 Feb 2026
208 points (99.5% liked)

Technology

80928 readers
4392 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

LOS ANGELES (AP) — The world's biggest social media companies face several landmark trials this year that seek to hold them responsible for harms to children who use their platforms. Opening statements for the first, in Los Angeles County Superior Court, began on Monday.

Instagram's parent company Meta and Google's YouTube face claims that their platforms deliberately addict and harm children. TikTok and Snap, which were originally named in the lawsuit, settled for undisclosed sums.

Jurors got their first glimpse into what will be a lengthy trial characterized by dueling narratives from the plaintiffs and the two remaining social media companies named as defendants. Opening arguments in the landmark case began Monday at the Spring Street Courthouse in downtown Los Angeles.

Mark Lanier delivered the opening statement for the plaintiffs first, in a lively display where he said the case is as "easy as ABC," which he said stands for "addicting the brains of children." He called Meta and Google "two of the richest corporations in history" who have "engineered addiction in children's brains."

At the core of the Los Angeles case is a 19-year-old identified only by the initials "KGM," whose case could determine how thousands of other, similar lawsuits against social media companies will play out. She and two other plaintiffs have been selected for bellwether trials — essentially test cases for both sides to see how their arguments play out before a jury and what damages, if any, may be awarded, said Clay Calvert, a nonresident senior fellow of technology policy studies at the American Enterprise Institute.

top 13 comments
sorted by: hot top controversial new old
[–] Perspectivist@feddit.uk 2 points 1 hour ago

My unpopular opinion is that social media is simply inherently incompatible with human nature. I don't think it's anyone's fault per se. It's like heroin in the sense that it doesn't matter how you distribute it - it's going to cause harm because hijacking our reward systems is the reason we use it in the first place. If you modify it so all that goes away, then what you're left with is water - and nobody wants that.

I don't know what the solution is, though. I don't think banning it is a solution, but I'm not sure how to square the harmfulness of it. It's not just kids it's bad for - it's everyone. And yeah, there are degrees to it - perhaps Lemmy is objectively better than an algorithm-based message board like Reddit, but something being better doesn't make it good. A non-toxic heroin that you can't OD on is also better than the alternative, but it's still going to be harmful. It's an arbitrary line we collectively just decide to draw somewhere - even though you could argue infinitely about nudging it one way or the other.

[–] ageedizzle@piefed.ca 3 points 2 hours ago

Are there any lawyers in here that can tell us how likely this thing is to succeed?

[–] NateNate60@lemmy.world 20 points 7 hours ago (1 children)

America really has a litigation culture, not because people are particularly fond of lawsuits, but because problems which are generally solved by legislative enactments or actions by regulatory bodies in other countries, aren't in the US, and thus the only way to find out who is right is to go to court.

[–] tyler@programming.dev 14 points 6 hours ago

The thing is, we had regulatory bodies that did that. Then citizens united happened and now companies can sue the government for infringing on their rights as “people” since clearly our constitution meant corporations are people. As a result every single regulatory body has to fight every single change in court.

[–] paraphrand@lemmy.world 8 points 6 hours ago

Parents, right? That’s always the solution to platforms.

[–] Korkki@lemmy.ml 7 points 6 hours ago (1 children)

They are going to play the same old "freedom of choice" defense... aren't they.

It's not our fault we made it purposefully addictive, you could just not watch it. Hasn't this been the case with every tobacco-, soda-, fast food-, etc company. For example: the whole mainstream idea that weight gain is about caloric imbalance and not consuming what you eat. That is the mainstream because is helps the food companies sway public opinion for their cause. It's not our food that is horrible slop, disruptive to metabolism and engineered to make people eat more and more and still crave more, it's the people who could just not eat it and if they do eat it they could like run 10km to sweat off the effects of like one sandwich.

They always shift the responsibility to the individuals when they are pressed on their wrongdoings. "The freedom of choice" at large is the great lie that at large keeps society running and is the main defense against any complain why something is systematically shit and fundamentally inhuman, from food to labor markets.

[–] ZeDoTelhado@lemmy.world 3 points 4 hours ago (1 children)

I think the food analogy is a good one here. I have debated personally a lot about this false sense of choice, when in reality you are bombarded with every psychological tactic to keep you hooked. Instagram in this sense is no different. If it lawsuit leads to somewhere, I do not know, however at some point the whole manipulative algorithms should be addressed (but by who and when are the biggest questions)

[–] Korkki@lemmy.ml 2 points 3 hours ago* (last edited 3 hours ago)

at some point the whole manipulative algorithms should be addressed (but by who and when are the biggest questions)

I have long since been of the opinion that all the big social multinational media should be seen as global technical, communication and media infrastructure. All the companies should be seized and put under some global foundation or the UN, everything open sourced, costs paid by member states and the platforms forced to remain impartial and to be organized for improving human condition, development, communication and understanding. If there is no need for profit then there is no need for entrapping users in toxic swamps of algorithm hell for more platform engagement.

[–] WalrusDragonOnABike@reddthat.com 4 points 6 hours ago (2 children)

Is anyone at threat of prison time or at least fines measured in the 100's of billions, if not trillions?

[–] PhoenixDog@lemmy.world 2 points 4 hours ago

A couple million bucks to charity and an assurance they'll do better next time.

[–] john_t@piefed.ee 5 points 5 hours ago (1 children)

Don't worry, nothing will happen as long as they "gift" Trump his 10%.

[–] jubilationtcornpone@sh.itjust.works 1 points 59 minutes ago

Don't even need to spend that much. Trump accepts fake peace prizes.

[–] MrSulu@lemmy.ml 3 points 6 hours ago

And the $multi billion companies will use every bent strategy available to delay, prevent, obfuscate evidence, attack & destroy witnesses etc. They will water down the impact to harm minimilisation outcome and so set out the precedent for how bad companies can be and get away with it. We really need that precedent to be seriously strong.