
A jury in a Los Angeles court has delivered a surprising verdict against Meta and YouTube, finding the two tech companies responsible for creating platforms that contributed to a young woman’s social media addiction and harm to her health.
The decision, which follows a pivotal pre-trial ruling allowing the case to proceed, could reshape how courts treat social media platforms in liability cases.
· A Los Angeles jury awarded $6 million to a plaintiff who developed a social media addiction as a child
According to a BBC report, a Los Angeles jury has found that Meta (the company that owns Instagram and Facebook) and Google (which owns YouTube) have played a direct role in harming a young woman’s mental health by exposing her to addictive platform designs from an early age.
She started using YouTube at the age of six and Instagram at nine, eventually developing anxiety, depression and body dysmorphia.
The jury concluded that the companies acted with enough wrongdoing to justify both compensatory and punitive damages, awarding a total of $6 million. Meta was assigned 70% of the liability, with Google responsible for the remaining 30%.
Traditionally, social media companies defend themselves by saying that they just host user-generated content. In this case, that particular argument failed.
The court explicitly stated that harm can stem from a platform's design, not just from the content users see.
Instead of asking whether harmful posts caused damage, the court examined whether platform mechanics encouraged compulsive use that led to harm.
The judge pointed to features like infinite scrolling as potentially harmful on their own, even when the content itself is not illegal or inappropriate.
Section 230 of the Communications Decency Act has shielded tech companies from liability for user-generated content in the past. Meta and Google relied heavily on this defense.
The court ruled that Section 230 does not apply when claims target product design decisions rather than third-party content.
In practical terms, this means platforms can no longer rely on blanket immunity if their systems are engineered to harm users.
The court allowed expert testimony arguing that social media platforms use psychological mechanisms similar to those found in gambling systems. These include:
Experts argued these features are engineered to maximize engagement, especially among younger users, and can increase the risk of anxiety, depression and compulsive behavior.
Yes. This case shows that courts may allow claims if they focus on addictive design features, not just content.
Section 230 protects platforms from liability for user-generated content. However, it does not protect product design decisions.
Features like infinite scroll, autoplay, and algorithmic recommendations were central to the argument.
Very likely. Governments are already exploring restrictions for minors, and this ruling strengthens the legal basis for action.
tags
Silviu is a seasoned writer who followed the technology world for almost two decades, covering topics ranging from software to hardware and everything in between.
View all posts