California lawmakers on Tuesday sent Gov. Gavin Newsom two groundbreaking bills intended to limit the downside of social media, as they faulted Congress for failing to act on the problem.
A first-of-its kind measure would require social media companies to make public their policies for removing disturbing content and provide details on how and when they remove it.
The second bill would require companies that provide online services attractive to children to follow age-appropriate design code principles aimed at keeping children safe. That includes not profiling a child or using the child’s personal information in a way that could harm the child’s physical or mental health or well-being.
“The online world has created tremendous opportunities, but also real and proximate threats to kids, to vulnerable communities and to American democracy as we know it,” said Democratic Assemblyman Jesse Gabriel, author of the first bill.
“We believe that California has a special obligation and a special opportunity to lead on these issues,” Gabriel added during a news conference Tuesday. “We’re proud of our technology economy, and we know that many of the companies that these bills would regulate are homegrown California companies. But with dysfunction in Washington, D.C., we believe that California must step up and lead.”
His measure would require companies to say how they regulate their own content under their social media terms of service. It stalled last year over free speech issues before clearing the Senate on a 33-3 vote and the Assembly, 48-0. It says it is “the intent of the Legislature” that the state attorney general or a city attorney take civil action against violators.
While the measure had bipartisan support, Republican Sen. Melissa Melendez in opposition worried that it could be used to punish legitimate but unpopular content, particularly as Attorney General Rob Bonta is a progressive Democrat.
“I can’t help but wonder if this is not in fact an attempt for the attorney general to perhaps harass the citizens of California, particularly those who have an opposing viewpoint, and I don’t think it is appropriate that the state attorney general get involved in any attempt to censor speech,” she said during debate Monday night.
But Democratic Sen. Thomas Umberg, who carried the bill, said the measure “does not basically censor content … If they have no policy, they have nothing to report. If they do have a policy then they need to report how they’re implementing that policy.”
Democratic Sen. Scott Wiener said the bill sought by the Anti-Defamation League is particularly important to the Legislature’s Jewish Caucus, “given the rampant anti-Semitism on social media.”
Opponents include the California Chamber of Commerce, Computer and Communications Industry Association, Consumer Technology Association, Internet Coalition, Netchoice and TechNet.
A coalition of the opponents said companies already must make public their content moderation policies, and the bill goes too far by requiring them to disclose to the attorney general “sensitive information about how we implement policies, detect activity, train employees, and use technology to detect content in need of moderation.”
The second bill intended specifically to protect children from inappropriate online content cleared the Senate 33-0 also with bipartisan support, though seven Republicans did not vote. It cleared the Assembly, 60-0.
The measure “will represent a major positive step toward creating a global standard for the protection of youth online. That’s an aspiration about which I think we can all agree,” said Democratic Sen. Josh Newman, who carried the bill in the Senate.
It is modeled after a similar measure in the United Kingdom. It also is opposed by the Chamber of Commerce and some of the tech industry associations. A coalition including the Entertainment Software Association said the bill includes “an over-inclusive standard and would capture far more websites and platforms than necessary.”
A third measure working its way through the Legislature would require large social media platforms to disclose to the public starting in mid-2023 statistics on content that violated its policies that were recommended or otherwise amplified by the platform’s algorithms.
Another more controversial measure failed in the gatekeeper Senate Appropriations Committee earlier this month after it was heavily opposed by the influential tech industry. It would have subjected some popular social media platforms like Instagram and TikTok to fines for using features they know can harm children.
Suggest a Correction