ARTICLE
2 October 2024

Setting Some Ground Rules: Gavin Newsom Signs Bills Regulating AI In Political Campaigns

B
BakerHostetler

Contributor

Recognized as one of the top firms for client service, BakerHostetler is a leading national law firm that helps clients around the world address their most complex and critical business and regulatory issues. With five core national practice groups — Business, Labor and Employment, Intellectual Property, Litigation, and Tax — the firm has more than 970 lawyers located in 14 offices coast to coast. BakerHostetler is widely regarded as having one of the country’s top 10 tax practices, a nationally recognized litigation practice, an award-winning data privacy practice and an industry-leading business practice. The firm is also recognized internationally for its groundbreaking work recovering more than $13 billion in the Madoff Recovery Initiative, representing the SIPA Trustee for the liquidation of Bernard L. Madoff Investment Securities LLC. Visit bakerlaw.com
What would you do if you received a call from the sitting president of the United States telling you not to vote in your state's upcoming party primary election?
United States Technology

What would you do if you received a call from the sitting president of the United States telling you not to vote in your state's upcoming party primary election?

In January 2024, many New Hampshire voters faced this exact question when they received a call from a former New Hampshire Democratic Party chair with a message from President Joe Biden telling them that they should not vote in their primary. The problem? The call came from inside the house. All kidding aside, it did come from an unexpected source – artificial intelligence (AI). The former chair never made the call, and Biden never made the statement.

This story set the news ablaze early in the 2024 election cycle. The public began to witness the potential of AI to manipulate voters in the coming election. Since then, further instances have continued to illustrate potential dangers:

  • On August 13, 2024, Elon Musk's AI chatbot, Grok, began allowing users to create AI-generated images from text prompts. Soon thereafter, X (formerly Twitter) users inundated the social media platform with fake photorealistic political images, including those of both presidential candidates, Vice President Kamala Harris and former President Donald Trump. The flood of potentially dangerous misinformation prompted X to implement restrictions on Grok's text-to-image ability only three days later.
  • In July 2024, a video of a voice seeming to be that of Harris went viral. The video placed the voice over a series of short videos, some containing Harris and some not. At some points, the video was broken up by actual videos of Harris speaking. Throughout the video, the voice made statements such as "I was selected because I am the ultimate diversity hire," and "I had four years under the tutelage of the ultimate deep-state puppet." Although the video was originally titled "Kamala Harris Campaign Ad PARODY," this title got lost as X users reposted the video on and off X.
  • In September 2024, when Taylor Swift formally endorsed Harris, she cited her concerns about Trump's use of several fake AI-created images of her and her fans to claim he had her support.

Sparked by these concerns, the California Legislature sprang into action, and California Governor Gavin Newsom subsequently signed a suite of three bills into law on September 17. The laws aim to protect voters from "deepfakes" – fake photorealistic images or videos of real people.

  • AB 2355 requires political campaigns to provide a disclosure in the advertisement of the use of AI in making the advertisement. This disclosure must be written or oral, depending on the advertising medium.
  • AB 2839 prohibits knowingly distributing materially deceptive content involving candidates for office or elected officials made with AI within 120 days of a California election and, in some cases, for 60 days after. The law allows a judge to issue an injunction requiring the content to be taken down and potentially for the judge to grant civil damages.
  • AB 2655 requires "large online platforms" to remove or label deepfakes within 72 hours of a user report. The law defines "large online platform" as "a public-facing internet website, web application, or digital application, including a social media platform ... , video sharing platform, advertising network, or search engine that had at least 1,000,000 California users during the preceding 12 months."

In signing the bills, Newsom stated, "Safeguarding the integrity of elections is essential to democracy, and it's critical that we ensure AI is not deployed to undermine the public's trust through disinformation – especially in today's fraught political climate."

Although the laws explicitly exempt parody and/or comedy, critics worry the new laws could encroach on individuals' freedom of speech. Those in support note that placing guardrails on potentially harmful disinformation is long overdue, especially as the 2024 election is only 36 days away.

Newsom vetoed the more comprehensive AI regulation, SB 1047 (the "Safe and Secure Innovation for Frontier Artificial Intelligence Models Act"), which passed the California Legislature last month. SB 1047 would have imposed significant requirements on AI developers. Lawmakers and AI and tech companies were similarly divided over whether such legislation would establish much-needed safeguards or put a halt to AI development.

Co-author: Noah J. Morris

The content of this article is intended to provide a general guide to the subject matter. Specialist advice should be sought about your specific circumstances.

Mondaq uses cookies on this website. By using our website you agree to our use of cookies as set out in our Privacy Policy.

Learn More