Yazılar

U.S. Appeals Court Upholds Most of California’s Child Social Media Law

A U.S. federal appeals court has largely upheld California’s law restricting social media platforms from offering “addictive feeds” to children without parental consent, in a ruling that could reshape how tech giants design online experiences for minors.

What the Law Does

  • Applies to: Social media companies like Google, Meta (Facebook, Instagram), Netflix, and X (formerly Twitter).

  • Core restriction: Makes it illegal for platforms to serve algorithmically personalized feeds to children unless parents explicitly approve.

  • Rationale: California lawmakers argue such feeds can harm children’s mental health by encouraging compulsive scrolling and social comparison.

Court’s Ruling

  • The 9th U.S. Circuit Court of Appeals rejected most of NetChoice’s claims that the law violates the First Amendment by limiting how companies “speak” to children.

  • Judge Ryan Nelson, writing for the panel, said NetChoice failed to show that unconstitutional applications of the law outweighed constitutional ones.

  • Age verification rules (taking effect in 2027) were deemed too early to challenge.

  • However, the court did block a requirement that default settings hide likes and comments from children, saying it was not the least restrictive way to protect mental health.

Industry Pushback

  • NetChoice, a trade group representing 41 major tech companies, said it was “largely disappointed” by the decision.

    • Paul Taske, its litigation co-director, argued the law “usurps the role of parents” and expands government control over lawful online speech.

  • NetChoice has also filed lawsuits against similar state-level internet restrictions across the U.S.

What’s Next

  • The case now returns to U.S. District Judge Edward Davila in San Jose, who had previously blocked parts of the law.

  • For now, California retains one of the strongest legal frameworks in the U.S. aimed at curbing social media’s impact on children.

Broader Context

  • The ruling adds momentum to state-led efforts to regulate youth access to social media amid rising concerns over depression, anxiety, and addiction linked to digital platforms.

  • Tech firms argue these laws could fragment the internet and undermine innovation, while advocates say they are essential to protect minors in an era of algorithm-driven engagement.

U.S. Supreme Court Lets Mississippi Social Media Age-Check Law Stand for Now

The U.S. Supreme Court on Thursday declined to temporarily block a Mississippi law requiring social media users to verify their age and obtain parental consent for minors, in a challenge filed by NetChoice, a trade group representing companies including Meta (META.O), Alphabet’s YouTube (GOOGL.O), and Snapchat (SNAP.N). The law remains in effect while NetChoice’s broader legal challenge, which argues it violates the First Amendment, continues in lower courts.

Justice Brett Kavanaugh noted in a statement that the Mississippi law is likely unconstitutional but said NetChoice had not met the high standard needed to halt enforcement at this stage. Paul Taske, co-director of the NetChoice Litigation Center, described the Supreme Court’s decision as “an unfortunate procedural delay” but expressed confidence that the challenge would ultimately succeed.

Mississippi’s attorney general welcomed the order, saying it allows “thoughtful consideration” of the law. The legislation, passed unanimously by the state legislature, requires platforms to obtain “express consent” from a parent or guardian before a minor can open an account and mandates “commercially reasonable” age verification. Violations can carry civil penalties of up to $10,000 per incident and potential criminal penalties under state deceptive trade practices laws.

The case comes after U.S. District Judge Halil Suleyman Ozerden initially blocked enforcement for some NetChoice members, but the 5th U.S. Circuit Court of Appeals allowed the law to take effect. Similar measures have been blocked in courts in seven other states. Technology companies maintain that their platforms already include extensive content moderation and parental controls to protect minors.

Mississippi defended the law as a “common” method to safeguard children online, emphasizing parental consent and age verification as key protective measures.