A jury in New Mexico has delivered a pivotal verdict against Meta, the parent company of platforms such as Instagram, Facebook, and WhatsApp, determining that the tech giant knowingly endangered children’s mental health and concealed information regarding child sexual exploitation on its platforms. This landmark decision follows a nearly seven-week trial and illustrates a growing inclination towards accountability for technology companies in the context of child safety.
During the proceedings, state prosecutors asserted that Meta prioritized profit over the wellbeing and safety of its younger users. The jury found that Meta had violated aspects of New Mexico’s Unfair Practices Act by allegedly hiding crucial information about the dangers associated with child sexual exploitation and the negative effects on minors’ mental health. Additionally, the jury agreed with the claim that Meta engaged in “unconscionable” trade practices that preyed on the vulnerabilities and inexperience of children.
Despite the gravity of the verdict, the jury has calculated that Meta will be liable for $375 million, a figure significantly lower than the amount sought by prosecutors, which highlighted the complexities and challenges in adjudicating such cases. Meta, valued at approximately $1.5 trillion, saw a 5% increase in its stock price during after-hours trading following the verdict, indicating a positive reaction from shareholders despite the legal consequences.
A critical next step involves the determination of whether Meta’s practices have created a public nuisance, which will be decided by a judge in May. This phase may also address potential financial responsibilities for public programs aimed at mitigating the harm caused by the company’s operations. In response to the verdict, a spokesperson for Meta expressed disagreement and stated the company would be appealing the decision. The spokesperson emphasized ongoing efforts to maintain safety on its platforms, underscoring the difficulty in identifying harmful content.
New Mexico’s lawsuit is part of a broader wave of litigation targeting social media companies over their impacts on children. More than 40 state attorneys general have joined in various lawsuits against Meta, alleging that its platforms contribute to a mental health crisis among young individuals due to their addictive features. Advocates for oversight, such as Sacha Haworth, executive director of The Tech Oversight Project, are voicing hope that this verdict marks a shift towards greater accountability for tech companies.
The New Mexico case drew from detailed undercover investigations where agents created childlike social media accounts to document solicitations and examined Meta’s responses. The lawsuit, initiated by Attorney General Raúl Torrez, also contends that Meta has not sufficiently acknowledged or addressed the dangers of social media addiction. Although executives at Meta acknowledged “problematic use” of their platforms, they declined to accept that social media addiction exists as a definitive issue.
As the jury deliberated, they reviewed extensive internal communications from Meta, expert testimonies, and observed the struggles faced by public school educators confronting social media-related disruptions, including potentially harmful operations targeting children.
In their discussions, jurors utilized a list of allegations regarding Meta’s safety disclosures and the detrimental implications of its algorithms. One juror stated that they reached a consensus on the maximum penalty per child affected, reflecting a strong sentiment about the importance of holding the company accountable.
Families affected by social media-related harms expressed gratitude for the verdict, calling it a “watershed moment” in the ongoing battle for accountability from tech giants regarding their impact on children’s lives. As this case highlights the emerging legal challenges facing tech companies, further developments and ongoing scrutiny in the realm of child safety and mental health in relation to social media are anticipated.


