#HelloWorld. It’s 2024 and we… are…back. Lots to catch up on. AI legal developments worldwide show no signs of letting up, so here’s our New Year’s resolution: We’re redoubling efforts to serve concise, focused guidance of immediate use to the legal and business community. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)
The AI Update | October 5, 2023
#HelloWorld. Fall begins, and the Writers Guild strike ends. In this issue, we look at what that means for AI in Hollywood. We also run through a dizzying series of self-regulating steps AI tech players have undertaken. As the smell of pumpkin spice fills the air, let’s stay smart together. (Subscribe to the mailing list to receive future issues.)
AI and the Writers Guild. Screenwriters ended a nearly five-month strike with a tentative new agreement good through mid-2026. The Minimum Basic Agreement (MBA) includes multiple AI-centric provisions. The AI-related highlights: Continue reading “The AI Update | October 5, 2023”
The AI Update | August 10, 2023
#HelloWorld. In this issue, the state of state AI laws (disclaimer: not our original phrase, although we wish it were). Deals for training data are in the works. And striking actors have made public their AI-related proposals—careful about those “Digital Replicas.” It’s August, but we’re not stopping. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)
States continue to pass and propose AI bills. Sometimes you benefit from the keen, comprehensive efforts of others. In the second issue of The AI Update, we summarized state efforts to legislate in the AI space. Now, a dedicated team at EPIC, the Electronic Privacy Information Center, spent all summer assembling an update, “The State of State AI Laws: 2023,” a master(ful) list of all state laws enacted and bills proposed touching on AI. We highly recommend reading their easy-to-navigate online site, highlights below:
The AI Update | July 27, 2023
#HelloWorld. Copyright suits are as unrelenting as the summer heat, with no relief in the forecast. AI creators are working on voluntary commitments to watermark synthetic content. And meanwhile, is ChatGPT getting “stupider”? Lots to explore. Let’s stay smart together. (Subscribe to the mailing list to receive future issues).
Big names portend big lawsuits. Since ChatGPT’s public launch in November 2022, plaintiffs have filed eight major cases in federal court—mostly in tech-centric Northern California—accusing large language models and image generators of copyright infringement, Digital Millennium Copyright Act violations, unfair competition, statutory and common law privacy violations, and other assorted civil torts. (Fancy a summary spreadsheet? Drop us a line.)
Here comes another steak for the grill: This month, on CBS’ “Face the Nation,” IAC’s chairman Barry Diller previewed that “leading publishers” were constructing copyright cases against generative AI tech companies, viewing it as a lynchpin for arriving at a viable business model: “yes, we have to do it. It’s not antagonistic. It’s to stake a firm place in the ground to say that you cannot ingest our material without figuring out a business model for the future.” Semafor later reported that The New York Times, News Corp., and Axel Springer were all among this group of likely publishing company plaintiffs, worried about the loss of website traffic that would come from generative AI answers replacing search engine results and looking for “billions, not millions, from AI.”
The AI Update | July 13, 2023
#HelloWorld. Pushback and disruption are the themes of this edition as we look at objections to proposed regulation in Europe, an FTC investigation, the growing movement in support of uncensored chatbots, and how AI is disrupting online advertising. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)
Pushback against AI regulation. The AI Update has followed closely the progress of the European Union’s proposed AI Act. Today we report on pushback in the form of an open letter from representatives of companies that operate in Europe expressing “serious concerns” that the AI Act would “jeopardise Europe’s competitiveness and technological sovereignty without effectively tackling the challenges we are and will be facing.” The letter takes aim in particular at the proposed “high risk” treatment of generative AI models, worrying that “disproportionate compliance costs and disproportionate liability risks” will push companies out of Europe and harm the ability of the EU to be at the forefront of AI development. The ask from the signatories is that European legislation “confine itself to stating broad principles in a risk-based approach.” As we have explained, there is a long road and many negotiations ahead before any version of the AI Act becomes the law in Europe. So it remains to be seen whether any further revisions reflect these concerns. Continue reading “The AI Update | July 13, 2023”
The AI Update | June 29, 2023
#HelloWorld. In the midst of summer, the pace of significant AI legal and regulatory news has mercifully slackened. With room to breathe, this issue points the lens in a different direction, at some of our persistent AI-related obsessions and recurrent themes. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)
Stanford is on top of the foundation model evaluation game. Dedicated readers may have picked up on our love of the Stanford Center for Research on Foundation Models. The Center’s 2021 paper, “On the Opportunities and Risks of Foundation Models,” is long, but it coined the term “foundation models” to cover the new transformer LLM and diffusion image generator architectures dominating the headlines. The paper exhaustively examines these models’ capabilities; underlying technologies; applications in medicine, law, and education; and potential social impacts. In a downpour of hype and speculation, the Center’s empirical, fact-forward thinking provides welcome shelter.
Now, like techno-Britney Spears, the Center has done it again. (The AI Update’s human writers can, like LLMs, generate dad jokes.) With the European Parliament’s mid-June adoption of the EU AI Act (setting the stage for further negotiation), researchers at the Center asked this question: To what extent would the current LLM and image-generation models be compliant with the EU AI Act’s proposed regulatory rules for foundation models, mainly set out in Article 28? The answer: None right now. But open-source start-up Hugging Face’s BLOOM model ranked highest under the Center’s scoring system, getting 36 out of 48 total possible points. The scores of Google’s PaLM 2, OpenAI’s GPT-4, Stability.ai’s Stable Diffusion, and Meta’s LLaMA models, in contrast, all hovered in the 20s.