Showing posts with label generative AI. Show all posts
Showing posts with label generative AI. Show all posts

Saturday, April 6, 2024

Where AI and property law intersect; Arizona State University (ASU) News, April 5, 2024

 Dolores Tropiano, Arizona State University (ASU) News; Where AI and property law intersect

"Artificial intelligence is a powerful tool that has the potential to be used to revolutionize education, creativity, everyday life and more.

But as society begins to harness this technology and its many uses — especially in the field of generative AI — there are growing ethical and copyright concerns for both the creative industry and legal sector.

Tyson Winarski is a professor of practice with the Intellectual Property Law program in Arizona State University’s Sandra Day O’Connor College of Law. He teaches an AI and intellectual property module within the course Artificial Intelligence: Law, Ethics and Policy, taught by ASU Law Professor Gary Marchant.

“The course is extremely important for attorneys and law students,” Winarski said. “Generative AI is presenting huge issues in the area of intellectual property rights and copyrights, and we do not have definitive answers as Congress and the courts have not spoken on the issue yet.”"

Thursday, March 7, 2024

Introducing CopyrightCatcher, the first Copyright Detection API for LLMs; Patronus AI, March 6, 2024

Patronus AI; Introducing CopyrightCatcher, thefirst Copyright Detection API for LLMs

"Managing risks from unintended copyright infringement in LLM outputs should be a central focus for companies deploying LLMs in production.

  • On an adversarial copyright test designed by Patronus AI researchers, we found that state-of-the-art LLMs generate copyrighted content at an alarmingly high rate 😱
  • OpenAI’s GPT-4 produced copyrighted content on 44% of the prompts.
  • Mistral’s Mixtral-8x7B-Instruct-v0.1 produced copyrighted content on 22% of the prompts.
  • Anthropic’s Claude-2.1 produced copyrighted content on 8% of the prompts.
  • Meta’s Llama-2-70b-chat produced copyrighted content on 10% of the prompts.
  • Check out CopyrightCatcher, our solution to detect potential copyright violations in LLMs. Here’s the public demo, with open source model inference powered by Databricks Foundation Model APIs. 🔥

LLM training data often contains copyrighted works, and it is pretty easy to get an LLM to generate exact reproductions from these texts1. It is critical to catch these reproductions, since they pose significant legal and reputational risks for companies that build and use LLMs in production systems2. OpenAI, Anthropic, and Microsoft have all faced copyright lawsuits on LLM generations from authors3, music publishers4, and more recently, the New York Times5.

To check whether LLMs respond to your prompts with copyrighted text, you can use CopyrightCatcher. It detects when LLMs generate exact reproductions of content from text sources like books, and highlights any copyrighted text in LLM outputs. Check out our public CopyrightCatcher demo here!

Thursday, February 29, 2024

The Intercept, Raw Story and AlterNet sue OpenAI for copyright infringement; The Guardian, February 28, 2024

, The Guardian ; The Intercept, Raw Story and AlterNet sue OpenAI for copyright infringement

"OpenAI and Microsoft are facing a fresh round of lawsuits from news publishers over allegations that their generative artificial intelligence products violated copyright laws and illegally trained by using journalists’ work. Three progressive US outlets – the Intercept, Raw Story and AlterNet – filed suits in Manhattan federal court on Wednesday, demanding compensation from the tech companies.

The news outlets claim that the companies in effect plagiarized copyright-protected articles to develop and operate ChatGPT, which has become OpenAI’s most prominent generative AI tool. They allege that ChatGPT was trained not to respect copyright, ignores proper attribution and fails to notify users when the service’s answers are generated using journalists’ protected work."

Friday, February 16, 2024

How AI copyright lawsuits could make the whole industry go extinct; The Verge, February 15, 2024

Nilay Patel, The Verge ; How AI copyright lawsuits could make the whole industry go extinct

"Our new Thursday episodes of Decoder are all about deep dives into big topics in the news, and for the next few weeks, we’re going to stay focused on one of the biggest topics of all: generative AI. 

There’s a lot going on in the world of generative AI, but maybe the biggest is the increasing number of copyright lawsuits being filed against AI companies like OpenAI and Stability AI. So for this episode, we brought on Verge features editor Sarah Jeong, who’s a former lawyer just like me, and we’re going to talk about those cases and the main defense the AI companies are relying on in those copyright cases: an idea called fair use."


How to Think About Remedies in the Generative AI Copyright Cases; LawFare, February 15, 2024

 Pamela Samuelson, LawFare; How to Think About Remedies in the Generative AI Copyright Cases

"So far, commentators have paid virtually no attention to the remedies being sought in the generative AI copyright complaints. This piece shines a light on them."

Wednesday, February 7, 2024

This AI Model Is Trained On Public Domain Stills To Create Mickey Mouse Images; Cartoon Brew, February 5, 2024

, Cartoon Brew ; This AI Model Is Trained On Public Domain Stills To Create Mickey Mouse Images

"In the comments section of an article about Mickey-1928 published by Ars Technica, Langlais explained that he’s aware of the line his model may be crossing by using Stable Diffusion and says that the public domain might provide a lot of answers to the ethical and legal questions raised by the emergence of AI tech:

I do agree with the issues of using a model trained on copyrighted content. I’m currently part of a new project to train a French LLM on public domain/open science/free culture sources, not only out of concern for author rights but also to enforce better standards of reproducibility and data provenance in the field. I’m hoping to see similar efforts on diffusion models this year. My general impression is that the copyright extension terms have made impossible an obvious solution to the AI copyright problem: having AI models trained openly on 20th century culture, and thus creating powerful incentives to digitize newspapers, books, movies for the commons."

Thursday, February 1, 2024

The economy and ethics of AI training data; Marketplace.org, January 31, 2024

 Matt Levin, Marketplace.org;  The economy and ethics of AI training data

"Maybe the only industry hotter than artificial intelligence right now? AI litigation. 

Just a sampling: Writer Michael Chabon is suing Meta. Getty Images is suing Stability AI. And both The New York Times and The Authors Guild have filed separate lawsuits against OpenAI and Microsoft. 

At the heart of these cases is the allegation that tech companies illegally used copyrighted works as part of their AI training data. 

For text focused generative AI, there’s a good chance that some of that training data originated from one massive archive: Common Crawl

“Common Crawl is the copy of the internet. It’s a 17-year archive of the internet. We make this freely available to researchers, academics and companies,” said Rich Skrenta, who heads the nonprofit Common Crawl Foundation."

Tuesday, January 30, 2024

Florida’s New Advisory Ethics Opinion on Generative AI Hits the Mark; JDSupra, January 29, 2024

 Ralph Artigliere , JDSupra; Florida’s New Advisory Ethics Opinion on Generative AI Hits the Mark

"As a former Florida trial lawyer and judge who appreciates emerging technology, I admit that I had more than a little concern when The Florida Bar announced it was working on a new ethics opinion on generative AI. Generative AI promises to provide monumental advantages to lawyers in their workflow, quality of work product, productivity, and time management and more. For clients, use of generative AI by their lawyers can mean better legal services delivered faster and with greater economy. In the area of eDiscovery, generative AI promises to surpass technology assisted review in helping manage the increasingly massive amounts of data.

Generative AI is new to the greater world, and certainly to busy lawyers who are not reading every blogpost on AI. The internet and journals are afire over concerns of hallucinations, confidentiality, bias, and the like. I felt a new ethics opinion might throw a wet blanket on generative AI and discourage Florida lawyers from investigating the new technology.

Thankfully, my concerns did not become reality. The Florida Bar took a thorough look at the technology and the existing ethical guidance and law and applied existing guidelines and rules in a thorough and balanced fashion. This article briefly summarizes Opinion 24-1 and highlights some of its important features.

The Opinion

On January 19, 2024, The Florida Bar released Ethics Opinion 24-1(“Opinion 24-1”)regarding the use of generative artificial intelligence (“AI”) in the practice of law. The Florida Bar and the State Bar of California are leaders in issuing ethical guidance on this issue. Opinion 24-1 draws from a solid background of ethics opinions and guidance in Florida and around the country and provides positive as well as cautionary statements regarding the emerging technologies. Overall, the guidance is well-placed and helpful for lawyers at a time when so many are weighing the use of generative AI technology in their law practices."

Wednesday, December 20, 2023

AI’s Billion-Dollar Copyright Battle Starts With a Font Designer; Bloomberg Law, December 18, 2023

 Isaiah Poritz, Bloomberg Law; AI’s Billion-Dollar Copyright Battle Starts With a Font Designer

"The makers of Copilot, which include OpenAI Inc., Microsoft Corp., GitHub Inc., and other top AI companies, are now facing nearly a dozen lawsuits from authors, artists, and programmers. They claim the industry has vacuumed up their creative work—without consent or compensation—to train AI chatbots and image generators that are already beginning to replace them.

At the core of these novel cases sits Butterick, a typographer and lawyer hailed by some for leading the fight to holding AI accountable, and slammed by others as a Luddite and an obstacle to transformative technological advances."

Friday, December 1, 2023

Copyright law will shape how we use generative AI; Axios, December 1, 2023

"In the year since the release of ChatGPT, generative AI has been moving fast and breaking things — and copyright law is only beginning to catch up. 

Why it matters: From Section 230 to the Digital Millennium Copyright Act (DMCA) to domain name squatting protections, intellectual property law has shaped the internet for three decades. Now, it will shape the way we use generative AI.

Driving the news: The Biden administration's recent executive order contained no initial guidance on copyright law and AI, which means these decisions will largely be left up to the courts."

Tuesday, November 14, 2023

YouTube to offer option to flag AI-generated songs that mimic artists’ voices; The Guardian, November 14, 2023

  , The Guardian; YouTube to offer option to flag AI-generated songs that mimic artists’ voices

"Record companies can request the removal of songs that use artificial intelligence-generated versions of artists’ voices under new guidelines issued by YouTube.

The video platform is introducing a tool that will allow music labels and distributors to flag content that mimics an artist’s “unique singing or rapping voice”.

Fake AI-generated music has been one of the side-effects of leaps forward this year in generative AI – the term for technology that can produce highly convincing text, images and voice from human prompts.

One of the most high-profile examples is Heart on My Sleeve, a song featuring AI-made vocals purporting to be Drake and the Weeknd. It was pulled from streaming services after Universal Music Group, the record company for both artists, criticised the song for “infringing content created with generative AI”. However, the song can still be accessed by listeners on YouTube."

Saturday, September 23, 2023

Generative AI vs. Copyright; Publishers Weekly, September 22, 2023

Lloyd J. Jassin, Publishers Weekly; Generative AI vs. Copyright

"To sum up, on a case-by-case basis, courts must determine if substantial similarity exists and then engage in line drawing—balancing free expression and the rights of creators."

Generative AI and US copyright law are on a collision course; Computer World, September 22, 2023

 , Computer World; Generative AI and US copyright law are on a collision course

"The unsettled state of play around AI and copyright law is not simply a US phenomenon, although most countries have yet to pass detailed legislation around it. The EU’s AI Act, as well as frameworks for more general AI regulation passed by the US and China, do not change the confused state of play around copyright issues. One country that has done so is Japan, which clarified in June that the use of copyright works for AI training is permitted, even for commercial purposes.

But active regulation of these issues may still be far off in the US, according to the experts."

Thursday, August 31, 2023

Copyright Office Issues Notice of Inquiry on Copyright and Artificial Intelligence; U.S. Copyright Office, August 30, 2023

U.S. Copyright Office ; Copyright Office Issues Notice of Inquiry on Copyright and Artificial Intelligence

"Today, the U.S. Copyright Office issued a notice of inquiry (NOI) in the Federal Register on copyright and artificial intelligence (AI). The Office is undertaking a study of the copyright law and policy issues raised by generative AI and is assessing whether legislative or regulatory steps are warranted. The Office will use the record it assembles to advise Congress; inform its regulatory work; and offer information and resources to the public, courts, and other government entities considering these issues.

The NOI seeks factual information and views on a number of copyright issues raised by recent advances in generative AI. These issues include the use of copyrighted works to train AI models, the appropriate levels of transparency and disclosure with respect to the use of copyrighted works, the legal status of AI-generated outputs, and the appropriate treatment of AI-generated outputs that mimic personal attributes of human artists.

The NOI is an integral next step for the Office’s AI initiative, which was launched in early 2023. So far this year, the Office has held four public listening sessions and two webinars. This NOI builds on the feedback and questions the Office has received so far and seeks public input from the broadest audience to date in the initiative.

“We launched this initiative at the beginning of the year to focus on the increasingly complex issues raised by generative AI. This NOI and the public comments we will receive represent a critical next step,” said Shira Perlmutter, Register of Copyrights and Director of the U.S. Copyright Office. “We look forward to continuing to examine these issues of vital importance to the evolution of technology and the future of human creativity.”

Initial written comments are due by 11:59 p.m. eastern time on Wednesday, October 18, 2023. Reply comments are due by 11:59 p.m. eastern time on Wednesday, November 15, 2023. Instructions for submitting comments are available on the Office’s website. Commenters may choose which and how many questions to respond to in the NOI.

For more general information about the Copyright Office’s AI initiative, please visit our website."

Tuesday, August 1, 2023

Webb Wright , The Drum; What you need to know about the copyright issues surrounding generative AI; The Drum, August 1, 2023

Webb Wright , The Drum; What you need to know about the copyright issues surrounding generative AI

"“The basic question of whether or not an AI using copyrighted work constitutes copyright infringement is for now an open issue,” says patent attorney Robert McFarlane. Ultimately, McFarlane believes that some uses of generative AI will be deemed to constitute copyright infringement, while others won’t. “These cases that are just starting now are going to try to draw that line,” he says."

Tuesday, July 25, 2023

The Generative AI Battle Has a Fundamental Flaw; Wired, July 25, 2023

  , Wired; The Generative AI Battle Has a Fundamental Flaw

"At the core of these cases, explains Sag, is the same general theory: that LLMs “copied” authors’ protected works. Yet, as Sag explained in testimony to a US Senate subcommittee hearing earlier this month, models like GPT-3.5 and GPT-4 do not “copy” work in the traditional sense. Digest would be a more appropriate verb—digesting training data to carry out their function: predicting the best next word in a sequence. “Rather than thinking of an LLM as copying the training data like a scribe in a monastery,” Sag said in his Senate testimony, “it makes more sense to think of it as learning from the training data like a student.”...

Ultimately, though, the technology is not going away, and copyright can only remedy some of its consequences. As Stephanie Bell, a research fellow at the nonprofit Partnership on AI, notes, setting a precedent where creative works can be treated like uncredited data is “very concerning.” To fully address a problem like this, the regulations AI needs aren't yet on the books."

Wednesday, July 19, 2023

US judge finds flaws in artists' lawsuit against AI companies; Reuters, July 19, 2023

 , Reuters; US judge finds flaws in artists' lawsuit against AI companies

"U.S. District Judge William Orrick said during a hearing in San Francisco on Wednesday that he was inclined to dismiss most of a lawsuit brought by a group of artists against generative artificial intelligence companies, though he would allow them to file a new complaint.

Orrick said that the artists should more clearly state and differentiate their claims against Stability AI, Midjourney and DeviantArt, and that they should be able to "provide more facts" about the alleged copyright infringement because they have access to Stability's relevant source code."

Monday, July 17, 2023

AI learned from their work. Now they want compensation.; The Washington Post, July 16, 2023

 , The Washington Post; AI learned from their work. Now they want compensation.

"Artists say the livelihoods of millions of creative workers are at stake, especially because AI tools are already being used to replace some human-made work. Mass scraping of art, writing and movies from the web for AI training is a practice creators say they never considered or consented to.

But in public appearances and in responses to lawsuits, the AI companies have argued that the use of copyrighted works to train AI falls under fair use — a concept in copyright law that creates an exception if the material is changed in a “transformative” way."

Thousands of authors urge AI companies to stop using work without permission; Morning Edition, NPR, July 17, 2023

, Morning Edition NPR; Thousands of authors urge AI companies to stop using work without permission

"Thousands of writers including Nora Roberts, Viet Thanh Nguyen, Michael Chabon and Margaret Atwood have signed a letter asking artificial intelligence companies like OpenAI and Meta to stop using their work without permission or compensation."

Saturday, July 15, 2023

'Not for Machines to Harvest’: Data Revolts Break Out Against A.I.; The New York Times, July 15, 2023

 Sheera Frenkel and , The New York Times;  'Not for Machines to Harvest’: Data Revolts Break Out Against A.I.

"At the heart of the rebellions is a newfound understanding that online information — stories, artwork, news articles, message board posts and photos — may have significant untapped value.

The new wave of A.I. — known as “generative A.I.” for the text, images and other content it generates — is built atop complex systems such as large language models, which are capable of producing humanlike prose. These models are trained on hoards of all kinds of data so they can answer people’s questions, mimic writing styles or churn out comedy and poetry...

“What’s happening here is a fundamental realignment of the value of data,” said Brandon Duderstadt, the founder and chief executive of Nomic, an A.I. company...

“The data rebellion that we’re seeing across the country is society’s way of pushing back against this idea that Big Tech is simply entitled to take any and all information from any source whatsoever, and make it their own,” said Ryan Clarkson, the founder of Clarkson...

Eric Goldman, a professor at Santa Clara University School of Law, said the lawsuit’s arguments were expansive and unlikely to be accepted by the court. But the wave of litigation is just beginning, he said, with a “second and third wave” coming that would define A.I.’s future."