HN.zip

Meta and YouTube Found Negligent in Landmark Social Media Addiction Case

241 points by mrjaeger - 103 comments
krunck [3 hidden]5 mins ago
strongpigeon [3 hidden]5 mins ago
There is a fairly low amount of details about the case in the article. This NPR article [0] has a bit more, but it's still fairly sparse. Though it's interesting how Zuckerberg thought it was a good idea to say: "If people feel like they're not having a good experience, why would they keep using the product?".

Given that this is a case about addiction, that feels like a shockingly bad thing to say in defense of your product. Can you imagine saying the same thing about oxycodone or cigarettes?

[0] https://www.npr.org/2026/03/25/nx-s1-5746125/meta-youtube-so...

Vegenoid [3 hidden]5 mins ago
> Can you imagine saying the same thing about oxycodone or cigarettes?

No, but unfortunately I can very easily imagine people saying it, just like the people who made loads of money from pushing those products did. Also just like the people who are profiting from the spread of gambling are saying now.

Why would someone choose to do a thing if it harms them? There are good arguments against laws that restrict personal freedoms, but this isn't one of them.

strongpigeon [3 hidden]5 mins ago
But what if we're talking about a product that you're giving away to children? I agree that for adults, cigarettes are fine. But in this case, you're actively designing to maximize tweens and teens engagement and the end result is them saying that they wan't to stop but can't.

Though to be fair, I was mostly pointing out the fact that this was a pretty dumb thing to say for a case like this, especially in a jury trial.

twoodfin [3 hidden]5 mins ago
As someone who values a liberal society, I hope we’d be exceedingly careful in what we label “addictive” in the same bucket as oxy or nicotine.

I also hope the reasons are obvious.

nkrisc [3 hidden]5 mins ago
Keep in mind that this case is about about a minor, not an adult. I don't think it's fair to ask children to resist social media through sheer willpower when there are legions of highly educated adults on the other side trying to increase engagement.

It should be no surprise that children can be manipulated by highly intelligent adults.

Barbing [3 hidden]5 mins ago
>[There are] legions of highly educated adults [at Meta] trying to increase [child] engagement

Why is this not only OK but the best way for Mark to spend every waking moment of his life?

Money thing? But often would he think about his bank account versus his products, maybe it’s pure drive?

hnlmorg [3 hidden]5 mins ago
We already have a distinction because it’s been known for decades already that some things are addictive purely through reinforcement psychology and some things lock people into a chemical dependence.

For example see the glossary in https://en.wikipedia.org/wiki/Substance_dependence

JKCalhoun [3 hidden]5 mins ago
"I hope we’d be exceedingly careful in what we label “addictive”…"

To be sure. But still an obviously dumb thing for a CEO to say though.

yabutlivnWoods [3 hidden]5 mins ago
Mmhmm those are words. Words that are hand wavy pretexts for conservatism rather than liberalism; as a lover of liberal society you hope it acts conservatively!

This just comes off as poorly obfuscated self selection. You own a bunch of Meta, Alphabet and other media stocks?

anon84873628 [3 hidden]5 mins ago
I don't think the reasons are obvious. Where do you put gambling on the spectrum?
twoodfin [3 hidden]5 mins ago
If something compels behavior vs. behavior remaining a free choice, a liberal society can and should treat it like any other source of compulsion.

Personally, I am leery of any technical definition of “addictive” that operates outside the traditional chemical influences on physiology. So I would not describe gambling in that sense.

One might have a malady that causes gambling to take on the same physiological vibe for you, but that’s not what it means for gambling itself to be addictive.

jwardbond [3 hidden]5 mins ago
I am not a neuroscientist, but I thought the actual physiological cause of addiction was similar in both nicotine and gambling: you crave the predictable release of dopamine.

If that is the (heavily simplified) case, is there a distinction for you between a chemically-induced dopamine release from smoking and, say, and a button you can press that magically releases dopamine in your brain?

SauntSolaire [3 hidden]5 mins ago
You're missing the negative affect node of the Koob addiction cycle, which exists for gambling but to a lesser degree than for nicotine.
twoodfin [3 hidden]5 mins ago
I don’t gamble, but if I did, I am fairly certain it would release little to no dopamine for me, win or lose.

I don’t smoke, but if I did, I’m also fairly certain I would find it hard to stop.

ses1984 [3 hidden]5 mins ago
And yet, some people find themselves compelled to continue gambling long after they’re drowning in debt.

If you don’t want to call that addiction, fine, but you can’t deny that it happens.

SoftTalker [3 hidden]5 mins ago
You seem to be differentiating between physical and psychological addiction, and saying that only physical addiction meets the technical definition of addiction?
twoodfin [3 hidden]5 mins ago
I’m saying society should tread extremely carefully in attempting to regulate citizens’ potentially psychologically addictive behaviors.
dylan604 [3 hidden]5 mins ago
We already have a category called addictive personality disorder where someone is much more prone to being addicted to pretty much anything.

In the US, regardless of what type of addiction you have, it is considered mental health. Open market insurance like ACA does not cover mental health, so there is no addiction treatment available. Sure, you can be addicted to a substance where your body needs a fix, but it is still treated as mental care. This seems to go directly against what your thoughts are on addiction, but that doesn't say much as you're just some rando on the interweb expressing their untrained opinions. So am I, but I'm not the spouting differing opinions with nothing more to back them up than how you feel.

tqi [3 hidden]5 mins ago
Where would you put 24x7 political content?
toyg [3 hidden]5 mins ago
That's more like perversion...
Sir_Twist [3 hidden]5 mins ago
I feel like people use the word “addiction” to refer to both chemical addiction and behavioral addiction, and that people understand that the latter is (usually) far less serious than the former.
Zigurd [3 hidden]5 mins ago
Dark patterns are real. Deceptive advertising is real. So-called prediction markets amount to unregulated gambling on any proposition. Many online businesses are whale hunts and the whales are often addicts.
joecool1029 [3 hidden]5 mins ago
> I hope we’d be exceedingly careful in what we label “addictive” in the same bucket as oxy or nicotine.

Not careful enough apparently: Nicotine isn't that addictive on its own, tobacco is.

vjulian [3 hidden]5 mins ago
Be aware, the vast majority of people who have ever smoked cigarettes occasionally never became addicted. They were not labeled as “smokers”. A non-trivial number of people today continue to smoke cigarettes on occasion. I like to have one on my birthday. Then again, I’m able to eat a chip and not consume the entire bag. I’m not convinced of these social science studies, and when digging into individual studies I’m sure the replication crisis comes into play.
thewebguyd [3 hidden]5 mins ago
Tobacco may be the most* addictive delivery method, but nicotine alone is also addictive. To say its not is misinformation. Consistent use of nicotine still leads to upregulation, which does cause irritability, brain fog, cravings when you stop.

* I'd even change this to say modern nicotine salts in vapes are likely to lead to dependency faster than tobacco. A 5% nicotine salt pod will contain as much nicotine as a full pack of cigarettes, and so vapers tend to consume far more nicotine in a single sitting than they ever could with a cigarette. That combined withe constant availability means users of nicotine vapes & pouches (aka, no tobacco) are likey to have a more difficult time quitting than cigarette smokers.

Bottom line, its still dangerous to dismiss nicotine's addictive potential with or without tobacco as a delivery method.

SauntSolaire [3 hidden]5 mins ago
> Not careful enough apparently: Nicotine isn't that addictive on its own, tobacco is.

That is a very strong claim to make when the current scientific consensus strongly disagrees.

mrintegrity [3 hidden]5 mins ago
How does that work when nicotine products that are every bit as addictive as tobacco exist, maybe you're just not aware of them? Sitting here with non tobacco snus (Swedish nicotine pouch) under my top lip, something I have been utterly unable to quit. I believe "nicotine free" tobacco would be completely non addictive.
btmiller [3 hidden]5 mins ago
There’s a big distance between libertarian and liberal societies. The libertarian tendencies of corporations are what tend to cause more harm.
cyanydeez [3 hidden]5 mins ago
Social media is addictive the same way anorexia is. If you think Anorexia isn't a form of addiction, then sure, you got your 'safety'.
easytiger [3 hidden]5 mins ago
Why not make personal responsibility illegal whilst we are at it. It is egregious that an individual can be held accountable for their own behaviours.
cyanydeez [3 hidden]5 mins ago
"If people didn't like destroying the environment, why would they let lobbiests run their government"

-- Billionaires

hash872 [3 hidden]5 mins ago
At least even money that an appellate court throws this verdict out entirely. Reminder that the US is the only developed country that uses juries for civil trials- everywhere else, complex issues of business litigation are generally left to a panel of judges. It's not that hard to rile up a bunch of randomly impaneled jurors against Big Bad Corporation. The US is kind of infamous for its very large, very unpredictable civil verdicts. There's an incredibly long history of juries racking up shockingly large verdicts against companies, only for an appellate court to throw the whole case out as unreasonable. Not even close to the final word in the American judicial system.

Edit to include: I mean this is coming the same day as the Supreme Court throwing out the piracy case against Cox Communications 9-0. Remember that this case originated with $1 billion dollar jury verdict against them! Was reversed by an appeals court 5 years later and completely invalidated today. Juries should not handle complex civil litigation, I'm sorry

aprilthird2021 [3 hidden]5 mins ago
Thanks for this take. Also explains why this did not result in much stock price movement today
zahlman [3 hidden]5 mins ago
Also at least partially explained by being priced in. The trial was known about and given the conditions described in GP it's not surprising that the verdict went this way.
pow_ext [3 hidden]5 mins ago
Apps like instagram and YouTube should be required at least to give an option to disable reels and shorts
ryandvm [3 hidden]5 mins ago
There should be a law to require the ability to disable algorithmic customization of content. If these apps are so compelling it shouldn't take a Spark cluster riffing on my private viewing habits to come up with content for me.

I don't recall a lot of complaints about Facebook or Instagram when it was actually your friends' content. But now it's force-feeding everybody their own "guilty pleasure" viewing material 24 hours a day. It's fucking sick.

polskibus [3 hidden]5 mins ago
Don’t forget WhatsApp. Kids are allowed to have WhatsApp as messaging but they get fed videos there too. There is no way to really disable them . Also this be allowed as parental supervision, not something that kids can override.
fraywing [3 hidden]5 mins ago
I'd hope the next iteration of social media tools humanity builds are less about reinforcing the individual ego and more about collective improvement, learning, and supporting the health of our species.

Anecdote, but it does seem like a lot of younger folks I speak with are exhausted by the dark patterns and dopamine extraction that top-k social media platforms create.

If agents/AI/bots inadvertently destroy the current incarnation of social media through noise, I think we'll be better for it.

iamnothere [3 hidden]5 mins ago
> I'd hope the next iteration of social media tools humanity builds are less about reinforcing the individual ego and more about collective improvement, learning, and supporting the health of our species.

To me this statement reads as both inaccurate and ignorant of human nature. Social media was actually better when it was about individual ego (Myspace/LiveJournal); as obnoxious as that can be, today everything is worse because of petty tribalism. Most conflicts on social media are inter-tribal, whether it’s racial, political, national, or feuding “stan” culture groups. The worst problems come from groups who organize on platforms like Discord or Kiwi Farms to direct harassment campaigns against perceived enemies (or random “lolcow” victims).

Simple observation of the present world and history will tell you that a platform focused on “collective improvement” will only appeal to a small subset of potential users. Of course such a platform would not be a bad thing. Places like this (such as The WELL) used to be common when the internet was dominated by academics, futurists, and tech enthusiasts. But average people are not interested in this kind of platform, and will not participate in good faith in such an environment.

fraywing [3 hidden]5 mins ago
> To me this statement reads as both inaccurate and ignorant of human nature

> But average people are not interested in this kind of platform, and will not participate in good faith in such an environment.

I'm not ignorant of human nature and tribalistic tendencies. The undercurrent of my comment is of an optimistic hope (or cope) that we can move past competitive individual validation programming. I'm aware that it's due to our nature, but also aware that it's exploited by dark patterns and extraction at scale through software.

amelius [3 hidden]5 mins ago
> I'd hope the next iteration of social media tools humanity builds are less about reinforcing the individual ego and more about collective improvement, learning, and supporting the health of our species.

This sounds like the original internet.

Before adtech took over.

0x3f [3 hidden]5 mins ago
The original internet wasn't about that at all, it was just in limbo while people were figuring out what it was going to be. It wasn't developed or optimized enough to be _anything_.
asim [3 hidden]5 mins ago
It will come. The problem is. So will the addictive stuff. The key is going to be real meaningful connection. Social media wasn't about community. Web 2.0 was. In 2005 we were connecting with real people we knew and probably up until 2011-2012 maybe we still were, but I guess friends of friends, colleagues, people in our network. Then it got really bad.

Getting back to community is key.

andai [3 hidden]5 mins ago
I hear word that in some countries, the government makes it so that screen time is limited, and algorithms promote educational content. Fortunately we civilized peoples are free of such a brutal oppression ;)
idle_zealot [3 hidden]5 mins ago
> I'd hope the next iteration of social media tools humanity builds are less about reinforcing the individual ego and more about collective improvement, learning, and supporting the health of our species

Do you have a mechanism for this in mind, incentives-wise? I can't see this making money.

Zigurd [3 hidden]5 mins ago
A $4.99/mo subscription would yield more revenue than Facebook makes in ARPU from all that fancy, creepy, and intrusive ad tech. Paying YouTube to not advertise to you makes it a 10X better experience.
myroon5 [3 hidden]5 mins ago
> $4.99/mo subscription would yield more revenue than Facebook makes in ARPU

Even ignoring the adverse selection of who'd subscribe, their ARPU is higher than that in North America: https://www.statista.com/statistics/251328/facebooks-average...

andai [3 hidden]5 mins ago
Well, another example comes to mind. Coordinated efforts to preserve the biosphere for all mankind are probably not going to be great for GDP.

We've tied our incentives to a structure which is not in alignment with continued survival. The real question is how can we incentivize ourselves to continue to exist?

The "the incentive structure says we should all destroy our brains" thing is just a small aspect of that.

0x3f [3 hidden]5 mins ago
Actually that's probably really good for GDP, just not over the kind of time periods an individual human deals with or cares about.

> We've tied our incentives to a structure which is not in alignment with continued survival. The real question is how can we incentivize ourselves to continue to exist?

The continued survival of individuals or humanity as a whole? The individuals seem to survive OK, and arguably there's nothing that could convince them to prefer the survival of the amorphous group, save for some kind of brainwashing.

benoau [3 hidden]5 mins ago
I guess the real question is whether a website where you communicate with friends and close ones needs to be a multi-trillion dollar company in the first place... historically most of them have not been worth very much at all.
pixl97 [3 hidden]5 mins ago
The question then becomes how can you make a website with all your friend (and by association all their friends) make enough profit to run itself?
andai [3 hidden]5 mins ago
You mean, how can my friends and I fundraise my $3 VPS? It's going to be rough, but I think we'll find a way ;)

(If we hit the stretch goal, we can upgrade to a raspberry pi!)

pixl97 [3 hidden]5 mins ago
This is a bit of a silly response on your part. You're not answering the question of WHY people are on FB and not on the little sites like existed 20 years ago before FB. It's called the network effect. You have friends, your friends have friend, those friends have friends. Rather than there being 30 bajillion separate sites representing these friends connections, people go "hey, why not one site with everyone there".

Said little sites may run for a bit and die, and the massive monolith remains, at least until another monolith replaces them.

sosborn [3 hidden]5 mins ago
Early Facebook was kind of a great mix. It had enough people on it, it was making money, and the advertising was much more reasonable. At the time it really was a place to connect with IRL friends.
aprilthird2021 [3 hidden]5 mins ago
It needs enough revenue to fund its operations. And most people won't pay for such a website, so if you want one place where most people you know are, then...
bogwog [3 hidden]5 mins ago
Come on, don't hand wave over the obvious. Think about how much it would actually cost to run a social media website that competes with the big social media on the core product of sharing and communicating with friends. It would be extremely realistic to build something that's both free and sustainable with just regular ads, as was done decades before.

(EDIT: to clarify, I don't mean to build an alternative monopoly, I mean to build alternatives that are big enough to survive as a business, and big enough to be useful; A few million users as opposed to the few billions Facebook and Youtube (allegedly) have)

The reason it's hard to imagine such a thing today is because the tech giants have illegally suppressed competition for so long. If Google or Meta were ordered to break up, and Facebook/Youtube forced to try and survive as standalone businesses, all the weaknesses in their products would manifest as actual market consequences, creating opportunity for competitors to win market share. Anybody with basic coding skills or money to invest would be tripping over themselves to build competing products which actually focus on the things people want or need, because consumers will be able to choose the ones they like.

hatsunearu [3 hidden]5 mins ago
I feel like discord is kind of like this used correctly, but with the recent drama and such it feels terrible
2OEH8eoCRo0 [3 hidden]5 mins ago
Ads were profitable before the outrage optimized flamebait internet era.
slopinthebag [3 hidden]5 mins ago
It doesn't need to make money directly (and probably shouldn't).

The incentives would be those which have motivated people throughout history: to create something which benefits humanity.

pixl97 [3 hidden]5 mins ago
Ah yes, I too love free servers and bandwidth.
slopinthebag [3 hidden]5 mins ago
Lol, it doesn't have to run for free and servers are really powerful these days (especially if you don't use a slow language). There are other monetisation strategies besides exploiting users for profit.
pixl97 [3 hidden]5 mins ago
It doesn't have to run for free, but if you're competing against anyone else running for free you've already lost the game as they suck the air out of the room with the network effect.

Next, text only platforms are nice, but niche on the modern internet. People seem to love multimedia which takes tons of bandwidth/cpu.

Paid for services don't mean spam free either. If it's worth people to pay for, it's worth spammers paying to get in and spam.

Then you have all the questions on what happens if you grow, how do you deal with working with all the laws around the world, how do you deal with other legal issues.

Having a site/service of any size can quickly become an expensive mess.

aprilthird2021 [3 hidden]5 mins ago
> If agents/AI/bots inadvertently destroy the current incarnation of social media through noise, I think we'll be better for it.

They are going to be (and AI slop already is) so much worse. Once they get ads to work well / seem natural the dark patterns will pop right back up and the money spigot will keep flowing upwards

xvxvx [3 hidden]5 mins ago
In before someone says ‘blame the parents’ and not the multi-billion dollar companies who’ve spent decades targeting children for lifelong addiction, ignoring the negative effects on their mental health.
dmitrygr [3 hidden]5 mins ago
It need not be either-or.

The guy who made the drugs is guilty. The guy who sold the drugs to kids is guilty. But parents who failed to warn kids about drugs and to oversee them properly are also guilty...

gusgus01 [3 hidden]5 mins ago
Generally in an article about arresting or sentencing a drug dealer, people don't bring up that the drug users are actually to blame.

Now if we're in a discussion around the cartels, plenty of people do bring up (and there's also those that get annoyed by it) that the drug users are actually the ones funding the cartels via their drug use.

Along these lines, I think another fun comparison might be opioid use and Purdue.

dmitrygr [3 hidden]5 mins ago
I think that that is actually an oversight. One needs to consider the entire chain. For example, with proper parenting, there would be a lot less youth demand for drugs. It doesn’t make what a drug dealer does any less bad, nor does it make the efforts of the police to arrest the drug dealer any less important. But it’s suboptimal to consider a small piece of a system, without thinking of the whole.
psychoslave [3 hidden]5 mins ago
So is the judicial system that is not making this illegal or don't enforce laws to prevent people targeting kids to create early dependence on drugs.
dmitrygr [3 hidden]5 mins ago
That is a fair point, I did attempt to make a complete list, of course, but you are right, there are more layers that could be named. All valid. The point I was making is that parents are also responsible.

eg: I grew up in a very nasty place. My neighborhood had a few pregnant 13 year old girls and a lot of drunks and smokers, including kids in their early teens. My parents kept me away from it all, while also both having full-time jobs. They put a lot of work into filtering whom I could be friends with and where I was allowed to be. THAT is the job of a parent.

macintux [3 hidden]5 mins ago
I agree it's the job of a parent, but two parents (and with only a single job each) is sadly not the norm in many challenging environments.
kakacik [3 hidden]5 mins ago
The thing is, it should be both. Parents often give too little fucks for long term welfare of their children, often also guilty of same vices. Issue is, these addictions are way more destructive to young forming mind than to adults. Nobody having small kids now had fb or instagram access when they were 5, did they.

Maybe you don't do this. Certainly I don't. But when looking around, its much less rosy and... lets say in blue collar families its too common to drug kids with screens so parents have off time. Heck, some are even proud how modern parents they are. Any good advice is successfully ignored, and ideas of passing some proper time with kids instead are skillfully avoided. People got lazy and generally expect miracles from life without putting in any miracle-worth efforts.

Companies just maximize their profits till laws allows them (and then some more), and expecting nice moral behavior by default is dangerously naive and never true.

strongpigeon [3 hidden]5 mins ago
dzink [3 hidden]5 mins ago
Read the book “Careless People” if you have a chance - according to the book, social media companies figured out they have real leverage with politicians since they can influence elections. As a result they are actively pushing for far right candidates to reduce their own taxation and regulation.
Zigurd [3 hidden]5 mins ago
I don't think this accelerationism/fascism hobby of many tech bros is going to age well.
ApolloFortyNine [3 hidden]5 mins ago
This just seems ripe for selective enforcement if not codified in law. I agree the algorithm they use can be addicting, but it's because it's simply good at providing content the user wants to consume.

Besides a general 'don't be too good' I'm really not sure what companies should do about it. It just seems like it'll lead to some judges allowing rulings against companies they don't like.

Television's goal was always viewer retention as well, they were just never able to target as well as you can on the internet.

kelseyfrog [3 hidden]5 mins ago
I see it as similar to the public health crisis created when protonated nicotine salts made their way into vapes along with flavors allowing 2-10x more nicotine to be delivered and the innovation that made Juul so popular with children.

The subsequent effects - namely being easier to consume and more addictive - eventually resulted in legislation catching up, and restrictions on what Juul could do. It being "too good" of a product parallels what we're seeing in social media seven years later.

Like most[all] all public health problems we see individualization of responsibility touted as a solution. If individualization worked, it would have already succeeded. Nothing prevents individualization except its failure of efficacy.

What does work is systems-level thinking and considering it an epidemiological problem rather than a problem of responsibility. Responsibility didn't work with the AIDS crisis, it didn't work on Juul, and it's not going to work on social media.

It is ripe for public health strategies. The biggest impediment to this is people who mistakingly believe that negative effects represent a personal moral failure.

jdasdf [3 hidden]5 mins ago
thats the point
tmpz22 [3 hidden]5 mins ago
Lets just be honest, if you make enough money its legal in America.

Unless you hurt children, then its mostly legal and a slap on the wrist.

carabiner [3 hidden]5 mins ago
Nukes are the same as knives, just different in magnitude. Should one have special rules?
woah [3 hidden]5 mins ago
Are there any takeaways here for builders of social media applications who are not Facebook or Google? Is this a warning to not make your newsfeed algorithm "too engaging" or is it only really relevant for big companies?
vaylian [3 hidden]5 mins ago
I'm not an authority on this matter. But if you say "I can stop any time", and it is not true, then you have a problem.
mikece [3 hidden]5 mins ago
A good time to (re-)recommend the movie "The Social Dilemma".
dlcarrier [3 hidden]5 mins ago
This is the kind of stuff that is causing them to push for mandatory identity verification laws. If they are being held liable for the the desires of their users, they're being forced micromanage the affairs of their customers, which preclude anonymous usage.
intended [3 hidden]5 mins ago
Meta is not pushing for mandatory age verification laws.They are pushing for age verification burdens to be pushed to the OS / App Store layer.
jmyeet [3 hidden]5 mins ago
I believe social media is on a collision course with an iceberg called Section 230.

Broadly speaking, Section 230 differentiates between publishers and platforms. A platform is like Geocities (back in the day) where the platform provider isn't liable for the content as long as they staisfy certain requirements about havaing processes for taking down content when required. A bit like the Cox decision today, you're broadly not responsible for the actions of people using your service unless your service is explicitly designed for such things.

A publisher (in the Section 230 sense) is like any media outlet. The publisher is liable for their content but they can say what they want, basically. It's why publishers tend to have strict processes around not making defamatory or false statements, etc.

I believe that any site that uses an algorithmic news feed is, legally speaking, a publisher acting like a platform.

Example: let's just say that you, as Twitter, FB, IG or Youtube were suddenly pro-Russian in the Ukraine conflict. You change your algorithm to surface and distribute pro-Russian content and suppress pro-Ukraine content. Or you're pro-Ukrainian and you do the reverse.

How is this different from being a publisher? IMHO it isn't. You've designed your algorithm knowingly to produce a certain result.

I believe that all these platforms will end up being treated like publishers for this reason.

So, with today's ruling about platforms creating addiction, (IMHO) it's no different to surfacing content. You are choosing content to produce a certain outcome. Intentionally getting someone addicted is funtionally no different to changing their views on something.

I actually blame Google for all this because they very successfully sold the idea that "the algorithm" ranks search results like it's some neutral black box but every behavior by an algorithm represents a choice made by humans who created that algorithm.

timdev2 [3 hidden]5 mins ago
Why do you believe that "Section 230 differentiates between publishers and platforms"?
ChrisArchitect [3 hidden]5 mins ago
Notably a different case from the other one in New Mexico:

Jury finds Meta liable in case over child sexual exploitation on its platforms

https://news.ycombinator.com/item?id=47509984

SpicyLemonZest [3 hidden]5 mins ago
And one with much deeper implications on how they operate. It's easy for Meta to just hire more moderators or treat reports of exploitation with higher priority; if this verdict stands, I think they have no realistic choice but to abandon usage targets.
aprilthird2021 [3 hidden]5 mins ago
Realistically they will hire expensive lawyers, pay out hundreds of millions to billions in settlements, fire lots of people (workforce is predominantly American), etc.

Even if they do what you're saying, lots of people who've used any Meta property in the last 15 years has a potentially viable case, and no future work can swat those away

Handy-Man [3 hidden]5 mins ago
IMO, parents share just as much blame here, if not more. Giving your kids independence doesn't mean being oblivious to what they're doing online. Too many parents confuse hands-off parenting with not parenting at all.
bluedevil2k [3 hidden]5 mins ago
Have you met kids? They’re devious, tech knowledgeable, and scheming and can find ways around any rule. Plus, no matter how good of a parent you are, you’re somewhat at the mercy of their friends’ parents as well. I can block TikTok from my daughter’s phone, but can’t block her from watching her friend’s phone while she’s out of the house.
intended [3 hidden]5 mins ago
I dont think parents going up against psychologists, data scientists, product managers and software engineers with the best pay in the world is any kind of fair fight.
2OEH8eoCRo0 [3 hidden]5 mins ago
Huge if upheld. This was the bellwether case for thousands of other similar cases.
aprilthird2021 [3 hidden]5 mins ago
I can't help but feel these are "revenge" verdicts. Public perception of these companies is dirt low, and there are so few levers the average person has to change what they feel is an increase in atomization, loneliness, breakdown of civic discourse, Cambridge Analytica level political targeting, misinformation, etc.

Maybe the social media companies could do more to combat all these. They certainly have a level of profit compared to what they provide to the average person that makes people squirm.

But does anyone believe for a second that YouTube is responsible for a person's internet / video watching addiction? It's like saying cable television is responsible for people who binge watch TV.

It's hard to square this circle while sports gambling apps and Polymarket / Kalshi are tearing through the landscape right now with no real pushback

bitwank [3 hidden]5 mins ago
>But does anyone believe for a second that YouTube is responsible for a person's internet / video watching addiction?

Yes? Is there an algorithm or not?

apopapo [3 hidden]5 mins ago
Will they also find liable all the companies that produce addictive food by injecting sugar into everything?

What about the "infinite" broadcasts found on all television channels?

This is ridiculous and pathetic.

btmiller [3 hidden]5 mins ago
A full sentence answer for you: yes.
BoredPositron [3 hidden]5 mins ago
In other countries that's the case so I don't know why it shouldn't be applicable in the US?
richwater [3 hidden]5 mins ago
People provide proof that other companies apply punitive damages to food companies knowinly adding sugar to food
BoredPositron [3 hidden]5 mins ago
8 countries in Europe, 4 in South America and 3 in Asia have an added sugar tax. So yes they did.
pixl97 [3 hidden]5 mins ago
"Libertarian demands companies have unlimited freedom until a corporation with unlimited freedom repeatedly eats their face with no consequences, wonders why the face eating leopards they voted for are actually allowed"
dmix [3 hidden]5 mins ago
> During his first-ever appearance before a jury in February, Meta's chairman and chief executive, Mark Zuckerberg, relied on his company's longstanding policy of not allowing users under the age of 13 on any of its platforms.

> When presented with internal research and documents showing that Meta knew young children were in fact using its platforms, Zuckerberg said he "always wished" for faster progress to identify users under 13. He insisted the company had reached the "right place over time".

Soon there will be government IDs required to use social media sites because parent's can't take phones away from their kids.