The AI Update | September 18, 2023

#HelloWorld. In this issue, the Copyright Office asks all the right questions—but will it do something interesting with the answers? Microsoft and Adobe offer clever ideas of their own. And, surprise (not really): Two new lawsuits against AI developers. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)

The Copyright Office has questions. Since the spring, the U.S. Copyright Office has devoted considerable effort to its AI Initiative, launching an AI webpage, holding four public listening sessions, and hosting educational webinars. In what it calls a “critical next step,” the Office on August 30 published a notice of inquiry asking for written comments (due October 18) on around 66 wide-ranging AI-related questions. Continue reading “The AI Update | September 18, 2023”

The Future of Digital Art as Training Material For Generative Artificial Intelligence Models

Before ChatGPT and other artificial intelligence (AI) large language models exploded on the scene last fall, there were AI art generators, based on many of the same technologies. Simplifying, in the context of art generation, these technologies involve a company first setting up a software-based network loosely modeled on the brain with millions of artificial “neurons.” […] This article has two goals: to provide a reader-friendly introduction to the copyright and right-of-publicity issues raised by such AI model training, and to offer practical tips about what art owners can do, currently, if they want to keep their works away from such training uses. […]

Read the full Art Business News article. 

Safeguarding Companies’ Online Data in the AI Era

The rapidly evolving landscape of advanced technology renders data one of the most valuable commodities today. This is especially true for artificial intelligence (AI), which can advance significantly in capability and complexity by learning from massive data sets used as training data. …

[This article identifies] considerations companies should account for when undertaking efforts to protect their online data based on an analysis of legal protections applicable to companies’ online data against unauthorized use.

Read the full article by Agatha H. Liu and Ariel Seidner.

The AI Update | August 29, 2023

#HelloWorld. In this issue, ChatGPT cannot be the next John Grisham, the secret is out on The New York Times’ frustrations with generative AI, and YouTube looks to a technological fix for voice replicas. Summer may soon be over, but AI issues are not going anywhere. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)

AI cannot be a copyright author—for now. In one of the most-awaited copyright events of the summer (not Barbie-related), the federal district court in D.C. held that an AI system could not be deemed the author of a synthetically-generated artwork. This was a test case brought by Stephen Thaler, a computer scientist and passionate advocate for treating AI as both copyright author and patent inventor, notwithstanding its silicon- and software-based essence. The D.C. district court, however, held firm to the policy position taken by the U.S. Copyright Office—copyright protects humans alone. In the words of the court: “human authorship is an essential part of a valid copyright claim.” Those who have followed Thaler’s efforts will remember that, about a year ago, the Federal Circuit similarly rejected Thaler’s attempt to list an AI model as an “inventor” on a patent application, holding instead that an inventor must be a “natural person.” Continue reading “The AI Update | August 29, 2023”

AI Software Settlement Highlights Risk in Hiring Decisions

In Equal Employment Opportunity Commission v. ITutorGroup, Inc., et al., No. 1:22-CV-2565 (E.D.N.Y. Aug. 9, 2023), the EEOC and a tutoring company filed a Joint Settlement Agreement and Consent Decree in the U.S. District Court for the Eastern District of New York, memorializing a $365,000 settlement for claims involving hiring software that automatically rejected applicants based on their age. This is first EEOC settlement involving artificial intelligence (“AI”) software bias.

Read more on the Class Action Defense Blog.

The AI Update | August 10, 2023

#HelloWorld. In this issue, the state of state AI laws (disclaimer: not our original phrase, although we wish it were). Deals for training data are in the works. And striking actors have made public their AI-related proposals—careful about those “Digital Replicas.” It’s August, but we’re not stopping. Let’s stay smart together. (Subscribe to the mailing list to receive future issues.)

States continue to pass and propose AI bills. Sometimes you benefit from the keen, comprehensive efforts of others. In the second issue of The AI Update, we summarized state efforts to legislate in the AI space. Now, a dedicated team at EPIC, the Electronic Privacy Information Center, spent all summer assembling an update, “The State of State AI Laws: 2023,” a master(ful) list of all state laws enacted and bills proposed touching on AI. We highly recommend reading their easy-to-navigate online site, highlights below:

Continue reading “The AI Update | August 10, 2023”

AI Tools in the Workplace and the Americans with Disabilities Act

On July 26, 2023, the EEOC issued a new Guidance entitled “Visual Disabilities in the Workplace and the Americans with Disabilities Act” (the “Guidance”).  This document is an excellent resource for employers, and provides insight into how to handle situations that may arise with job applicants and employees that have visual disabilities. Notably, for employers that use algorithms or artificial intelligence (“AI”) as a decision-making tool, the Guidance makes clear that employers have an obligation to make reasonable accommodations for applicants or employees with visual disabilities who request them in connection with these technologies.

Read more on the Class Action Defense Blog.

 

The AI Update | July 27, 2023

#HelloWorld. Copyright suits are as unrelenting as the summer heat, with no relief in the forecast. AI creators are working on voluntary commitments to watermark synthetic content. And meanwhile, is ChatGPT getting “stupider”? Lots to explore. Let’s stay smart together. (Subscribe to the mailing list to receive future issues).

Big names portend big lawsuits. Since ChatGPT’s public launch in November 2022, plaintiffs have filed eight major cases in federal court—mostly in tech-centric Northern California—accusing large language models and image generators of copyright infringement, Digital Millennium Copyright Act violations, unfair competition, statutory and common law privacy violations, and other assorted civil torts. (Fancy a summary spreadsheet? Drop us a line.)

Here comes another steak for the grill: This month, on CBS’ “Face the Nation,” IAC’s chairman Barry Diller previewed that “leading publishers” were constructing copyright cases against generative AI tech companies, viewing it as a lynchpin for arriving at a viable business model: “yes, we have to do it. It’s not antagonistic. It’s to stake a firm place in the ground to say that you cannot ingest our material without figuring out a business model for the future.” Semafor later reported that The New York Times, News Corp., and Axel Springer were all among this group of likely publishing company plaintiffs, worried about the loss of website traffic that would come from generative AI answers replacing search engine results and looking for “billions, not millions, from AI.”

Continue reading “The AI Update | July 27, 2023”

Taking Heed of AI Contracts

Duane Morris partner Neville M. Bilimoria authored the McKnight’s Long-Term Care article, “AI is everywhere! Addressing the legal risks through contracting.”

Mr. Bilimoria writes:

You can’t look in the news or see social media posts each day without hearing about artificial intelligence in healthcare. In fact, the advancements in AI in healthcare are making leaps and bounds, seemingly with each day that goes by.

But nursing homes and assisted living providers need to understand not just the benefits of how AI can improve quality of resident care and improved operations, but also the legal issues surrounding AI in your facility.

Read the full article on the McKnight’s Long-Term Care website.

© 2009- Duane Morris LLP. Duane Morris is a registered service mark of Duane Morris LLP.

The opinions expressed on this blog are those of the author and are not to be construed as legal advice.

Proudly powered by WordPress