Showing posts with label AI training data. Show all posts
Showing posts with label AI training data. Show all posts

Wednesday, October 1, 2025

Disney Sends Cease And Desist Letter To Character.ai For Copyright Infringement As Studios Move To Protect IP; Deadline, September 30, 2025

Jill Goldsmith, Deadline; Disney Sends Cease And Desist Letter To Character.ai For Copyright Infringement As Studios Move To Protect IP

"Walt Disney sent a cease-and-desist letter to Character.AI, a “personalized superintelligence platform” that the media giant says is ripping off copyrighted characters without authorization.

The AI startup offers users the ability to create customizable, personalized AI companions that can be totally original but in some cases are inspired by existing characters, including, it seems, Disney icons from Spider-Man and Darth Vader to Moana and Elsa.

The letter is the latest legal salvo by Hollywood as studios begin to step up against AI. Disney has also sued AI company Midjourney for allegedly improper use and distribution of AI-generated characters from Disney films. Disney, Warner Bros. and Universal Pictures this month sued Chinese AI firm MiniMax for copyright infringement."

Monday, September 29, 2025

I Sued Anthropic, and the Unthinkable Happened; The New York Times, September 29, 2025

 , The New York Times; I Sued Anthropic, and the Unthinkable Happened

"In August 2024, I became one of three named plaintiffs leading a class-action lawsuit against the A.I. company Anthropic for pirating my books and hundreds of thousands of other books to train its A.I. The fight felt daunting, almost preposterous: me — a queer, female thriller writer — versus a company now worth $183 billion?

Thanks to the relentless work of everyone on my legal team, the unthinkable happened: Anthropic agreed to pay authors and publishers $1.5 billion in the largest copyright settlement in history. A federal judge preliminarily approved the agreement last week.

This settlement sends a clear message to the Big Tech companies splashing generative A.I. over every app and page and program: You are not above the law. And it should signal to consumers everywhere that A.I. isn’t an unstoppable tsunami about to overwhelm us. Now is the time for ordinary Americans to recognize our agency and act to put in place the guardrails we want.

The settlement isn’t perfect. It’s absurd that it took an army of lawyers to demonstrate what any 10-year-old knows is true: Thou shalt not steal. At around $3,000 per work, shared by the author and publisher, the damages are far from life-changing (and, some argue, a slap on the wrist for a company flush with cash). I also disagree with the judge’s ruling that, had Anthropic acquired the books legally, training its chatbot on them would have been “fair use.” I write my novels to engage human minds — not to empower an algorithm to mimic my voice and spit out commodity knockoffs to compete directly against my originals in the marketplace, nor to make that algorithm’s creators unfathomably wealthy and powerful.

But as my fellow plaintiff Kirk Wallace Johnson put it, this is “the beginning of a fight on behalf of humans that don’t believe we have to sacrifice everything on the altar of A.I.” Anthropic will destroy its trove of illegally downloaded books; its competitors should take heed to get out of the business of piracy as well. Dozens of A.I. copyright lawsuits have been filed against OpenAI, Microsoft and other companies, led in part by Sylvia Day, Jonathan Franzen, David Baldacci, John Grisham, Stacy Schiff and George R. R. Martin. (The New York Times has also brought a suit against OpenAI and Microsoft.)

Though a settlement isn’t legal precedent, Bartz v. Anthropic may serve as a test case for other A.I. lawsuits, the first domino to fall in an industry whose “move fast, break things” modus operandi led to large-scale theft. Among the plaintiffs of other cases are voice actors, visual artists, record labels, YouTubers, media companies and stock-photo libraries, diverse stakeholders who’ve watched Big Tech encroach on their territory with little regard for copyright law...

Now the book publishing industry has sent a message to all A.I. companies: Our intellectual property isn’t yours for the taking, and you cannot act with impunity. This settlement is an opening gambit in a critical battle that will be waged for years to come."

Saturday, September 27, 2025

Judge approves $1.5 billion copyright settlement between AI company Anthropic and authors; AP, September 25, 2025

BARBARA ORTUTAY , AP; Judge approves $1.5 billion copyright settlement between AI company Anthropic and authors

" A federal judge on Thursday approved a $1.5 billion settlement between artificial intelligence company Anthropic and authors who allege nearly half a million books had been illegally pirated to train chatbots.

U.S. District Judge William Alsup issued the preliminary approval in San Francisco federal court Thursday after the two sides worked to address his concerns about the settlement, which will pay authors and publishers about $3,000 for each of the books covered by the agreement. It does not apply to future works."

Tuesday, September 23, 2025

Screw the money — Anthropic’s $1.5B copyright settlement sucks for writers; TechCrunch, September 5, 2025

Amanda Silberling , TechCrunch; Screw the money — Anthropic’s $1.5B copyright settlement sucks for writers

"But writers aren’t getting this settlement because their work was fed to an AI — this is just a costly slap on the wrist for Anthropic, a company that just raised another $13 billion, because it illegally downloaded books instead of buying them.

In June, federal judge William Alsup sided with Anthropic and ruled that it is, indeed, legal to train AI on copyrighted material. The judge argues that this use case is “transformative” enough to be protected by the fair use doctrine, a carve-out of copyright law that hasn’t been updated since 1976.

“Like any reader aspiring to be a writer, Anthropic’s LLMs trained upon works not to race ahead and replicate or supplant them — but to turn a hard corner and create something different,” the judge said.

It was the piracy — not the AI training — that moved Judge Alsup to bring the case to trial, but with Anthropic’s settlement, a trial is no longer necessary."

Tuesday, September 9, 2025

Judge Delays Preliminary Approval in Anthropic Copyright Settlement; Publishers Weekly, September 9, 2025

 Jim Milliot, Publishers Weekly; Judge Delays Preliminary Approval in Anthropic Copyright Settlement

"Alsup signaled his discomfort with the proposal in a filing released the evening before the September 8 hearing, writing that he was “disappointed” that attorneys representing the author plaintiffs had left “important questions to be answered in the future, including respecting the Works List, Class List, Claim Form." He was especially concerned for works with multiple claimants with regards to the notification process, voicing worry over what would happen if one party wanted to opt-out of the settlement and the other did not...

In a statement, Authors Guild CEO Mary Rasenberger said the Guild was “confused” by the court’s suggestion that the Guild and AAP were working behind the scenes in ways that could pressure authors to accept the settlement “when that is precisely the opposite of our proposed role as informational advisors to the working group.”

The goal of the working group, which had been proposed by lawyers for the class, “is to ensure that authors’ interests are fully represented and to bring our expertise... to the discussions with complete transparency,” Rasenberger continued. “There are industry norms that we want to make sure are accounted for.”...

AAP CEO Maria Pallante offered an even more vigorous explanation of AAP’s role, as well as the role of the Guild, in the proceedings. “The Association of American Publishers and the Authors’ Guild are not-for-profits that have worked hard to support counsel in the case and to make sure that authors and publishers have the information they need,” Pallante said in a statement. “Unfortunately, the Court today demonstrated a lack of understanding of how the publishing industry works.”"

Monday, September 8, 2025

Class-Wide Relief:The Sleeping Bear of AI Litigation Is Starting to Wake Up; Intellectual Property & Technology Law Journal, October 2025

Anna B. Naydonov, Mark Davies and Jules Lee, Intellectual Property &Technology Law Journal; Class-Wide Relief:The Sleeping Bear of AI Litigation Is Starting to Wake Up

"Probably no intellectual property (IP) topic in the last several years has gotten more attention than the litigation over the use of the claimed copyrighted content in training artificial intelligence (AI) models.The issue of whether fair use applies to save the day for AI developers is rightfully deemed critical, if not existential, for AI innovation. But whether class relief – and the astronomical damages that may come with it – is available in these cases is a question of no less significance."

Saturday, September 6, 2025

Big Questions About AI and the Church Video; August 25, 2025

Big Questions About AI and the Church Video

Kip Currier: This Big Questions About AI and the Church video (1:12:14) was created by the members of my cohort and me (Cohort 7). Our cohort emanated from the groundbreaking August 2024 ecumenical AI & The Church Summit in Seattle that we all attended.

Perhaps raising more questions than providing answers, the video's aim is to encourage reflection and discussion of the many-faceted issues and concerns at the nexus of AI, faith communities, and our broader societies.

Many thanks to our cohort member Rev. Dr. Andy P. Morgan for spearheading, synthesizing, and uploading this video to YouTube. 

Anthropic settles with authors in first-of-its-kind AI copyright infringement lawsuit; NPR, September 5, 2025

  , NPR; Anthropic settles with authors in first-of-its-kind AI copyright infringement lawsuit

"In one of the largest copyright settlements involving generative artificial intelligence, Anthropic AI, a leading company in the generative AI space, has agreed to pay $1.5 billion to settle a copyright infringement lawsuit brought by a group of authors.

If the court approves the settlement, Anthropic will compensate authors around $3,000 for each of the estimated 500,000 books covered by the settlement.

The settlement, which U.S. Senior District Judge William Alsup in San Francisco will consider approving next week, is in a case that involved the first substantive decision on how fair use applies to generative AI systems. It also suggests an inflection point in the ongoing legal fights between the creative industries and the AI companies accused of illegally using artistic works to train the large language models that underpin their widely-used AI systems.

The fair use doctrine enables copyrighted works to be used by third parties without the copyright holder's consent in some circumstances, such as when illustrating a point in a news article. AI companies trying to make the case for the use of copyrighted works to train their generative AI models commonly invoke fair use. But authors and other creative industry plaintiffs have been pushing back.

"This landmark settlement will be the largest publicly reported copyright recovery in history," the settlement motion states, arguing that it will "provide meaningful compensation" to authors and "set a precedent of AI companies paying for their use of pirated websites."

"This settlement marks the beginning of a necessary evolution toward a legitimate, market-based licensing scheme for training data," said Cecilia Ziniti, a tech industry lawyer and former Ninth Circuit clerk who is not involved in this specific case but has been following it closely. "It's not the end of AI, but the start of a more mature, sustainable ecosystem where creators are compensated, much like how the music industry adapted to digital distribution.""

Saturday, August 30, 2025

Anthropic’s settlement with authors may be the ‘first domino to fall’ in AI copyright battles; Fortune, August 27, 2025

BEATRICE NOLAN, Fortune; Anthropic’s settlement with authors may be the ‘first domino to fall’ in AI copyright battles

"The amount of the settlement was not immediately disclosed, but legal experts not involved in the case said the figure could easily reach into the hundreds of millions. It’s also still unclear how the settlement will be distributed among various copyright holders, which could include large publishing houses as well as individual authors.

The case was the first certified class action against an AI company over the use of copyrighted materials, and the quick settlement, which came just one month after the judge ruled the case could proceed to trial as a class action, is a win for the authors, according to legal experts."

Friday, August 29, 2025

Anthropic Settles High-Profile AI Copyright Lawsuit Brought by Book Authors; Wired, August 26, 2025

Kate Knobs, Wired ; Anthropic Settles High-Profile AI Copyright Lawsuit Brought by Book Authors

"ANTHROPIC HAS REACHED a preliminary settlement in a class action lawsuit brought by a group of prominent authors, marking a major turn in one of the most significant ongoing AI copyright lawsuits in history. The move will allow Anthropic to avoid what could have been a financially devastating outcome in court."

Thursday, August 28, 2025

Anthropic’s surprise settlement adds new wrinkle in AI copyright war; Reuters, August 27, 2025

 , Reuters; Anthropic’s surprise settlement adds new wrinkle in AI copyright war

"Anthropic's class action settlement with a group of U.S. authors this week was a first, but legal experts said the case's distinct qualities complicate the deal's potential influence on a wave of ongoing copyright lawsuits against other artificial-intelligence focused companies like OpenAI, Microsoft and Meta Platforms.

Amazon-backed Anthropic was under particular pressure, with a trial looming in December after a judge found it liable for pirating millions of copyrighted books. The terms of the settlement, which require a judge's approval, are not yet public. And U.S. courts have just begun to wrestle with novel copyright questions related to generative AI, which could prompt other defendants to hold out for favorable rulings."

Monday, August 25, 2025

Who owns the copyright for AI work?; Financial Times, August 24, 2025

 , Financial Times; Who owns the copyright for AI work?

"Generative artificial intelligence poses two copyright puzzles. The first is the widely discussed question of compensation for work used to train AI models. The second, which has yet to receive as much attention, concerns the work that AI produces. Copyright is granted to authors. So what happens to work that has no human author?"

Sunday, August 24, 2025

Using AI for Work Could Land You on the Receiving End of a Nasty Lawsuit; Futurism, August 23, 2025

 JOE WILKINS , Futurism; Using AI for Work Could Land You on the Receiving End of a Nasty Lawsuit

"For all its hype, artificial intelligence isn't without its psychologicalenvironmental, and even spiritual hazards.

Perhaps the most pressing concern on an individual level, though, is that it puts users on the hook for a nearly infinite number of legal hazards — even at work, as it turns out.


A recent breakdown by The Register highlights the legal dangers of AI use, especially in corporate settings. If you use generative AI software to spit out graphics, press releases, logos, or videos, you and your employer could end up facing six-figure damages, the publication warns.


This is thanks to the vast archive of copyrighted data that virtually all commercial generative AI models are trained on.


The Register uses Nintendo's Mario as a prime example of how one might stumble, intentionally or not, into a massive copyright lawsuit, regardless of intent to infringe: if you use AI to generate a cutesy mascot for your plumbing company that looks too much like the iconic videogame character, you could easily find yourself in the legal crosshairs of the notoriously litigious corporation.


"The real harm comes from the attorney's fees that you can get saddled with," intellectual property lawyer Benjamin Bedrava told the publication. "Because you could have a hundred and fifty thousand dollars in attorney's fees over something where the license would have been fifteen hundred dollars.""

Saturday, August 23, 2025

Watering down Australia’s AI copyright laws would sacrifice writers’ livelihoods to ‘brogrammers’; The Guardian, August 11, 2025

 Tracey Spicer, The Guardian; Watering down Australia’s AI copyright laws would sacrifice writers’ livelihoods to ‘brogrammers’

"My latest book, which is about artificial intelligence discriminating against people from marginalised communities, was composed on an Apple Mac.

Whatever the form of recording the first rough draft of history, one thing remains the same: they are very human stories – stories that change the way we think about the world.

A society is the sum of the stories it tells. When stories, poems or books are “scraped”, what does this really mean?

The definition of scraping is to “drag or pull a hard or sharp implement across (a surface or object) so as to remove dirt or other matter”.

A long way from Brisbane or Bangladesh, in the rarefied climes of Silicon Valley, scrapers are removing our stories as if they are dirt.

These stories are fed into the machines of the great god: generative AI. But the outputs – their creations – are flatter, less human, more homogenised. ChatGPT tells tales set in metropolitan areas in the global north; of young, cishet men and people living without disability.

We lose the stories of lesser-known characters in remote parts of the world, eroding our understanding of the messy experience of being human.

Where will we find the stories of 64-year-old John from Traralgon, who died from asbestosis? Or seven-year-old Raha from Jaipur, whose future is a “choice” between marriage at the age of 12 and sexual exploitation?

OpenAI’s creations are not the “machines of loving grace” envisioned in the 1967 poem by Richard Brautigan, where he dreams of a “cybernetic meadow”.

Scraping is a venal money grab by oligarchs who are – incidentally – scrambling to protect their own intellectual property during an AI arms race.

The code behind ChatGPT is protected by copyright, which is considered to be a literary work. (I don’t know whether to laugh or cry.)

Meta has already stolen the work of thousands of Australian writers.

Now, our own Productivity Commission is considering weakening our Copyright Act to include an exemption for text and data mining, which may well put us out of business.

In its response, The Australia Institute uses the analogy of a car: “Imagine grabbing the keys for a rental car and just driving around for a while without paying to hire it or filling in any paperwork. Then imagine that instead of being prosecuted for breaking the law, the government changed the law to make driving around in a rental car legal.”

It’s more like taking a piece out of someone’s soul, chucking it into a machine and making it into something entirely different. Ugly. Inhuman.

The commission’s report seems to be an absurdist text. The argument for watering down copyright is that it will lead to more innovation. But the explicit purpose of the Copyright Act is to protect innovation, in the form of creative endeavour.

Our work is being devalued, dismissed and destroyed; our livelihoods demolished.

In this age of techno-capitalism, it appears the only worthwhile innovation is being built by the “brogrammers”.

US companies are pinching Australian content, using it to train their models, then selling it back to us. It’s an extractive industry: neocolonialism, writ large."

Thursday, August 14, 2025

Japan’s largest newspaper, Yomiuri Shimbun, sues AI startup Perplexity for copyright violations; NiemanLab, August 11, 2025

ANDREW DECK  , NiemanLab; Japan’s largest newspaper, Yomiuri Shimbun, sues AI startup Perplexity for copyright violations

"The Yomiuri Shimbun, Japan’s largest newspaper by circulation, has sued the generative AI startup Perplexity for copyright infringement. The lawsuit, filed in Tokyo District Court on August 7, marks the first copyright challenge by a major Japanese news publisher against an AI company.

The filing claims that Perplexity accessed 119,467 articles on Yomiuri’s site between February and June of this year, based on an analysis of its company server logs. Yomiuri alleges the scraping has been used by Perplexity to reproduce the newspaper’s copyrighted articles in responses to user queries without authorization.

In particular, the suit claims Perplexity has violated its “right of reproduction” and its “right to transmit to the public,” two tenets of Japanese law that give copyright holders control over the copying and distribution of their work. The suit seeks nearly $15 million in damages and demands that Perplexity stop reproducing its articles...

Japan’s copyright law allows AI developers to train models on copyrighted material without permission. This leeway is a direct result of a 2018 amendment to Japan’s Copyright Act, meant to encourage AI developmentin the country’s tech sector. The law does not, however, allow for wholesale reproduction of those works, or for AI developers to distribute copies in a way that will “unreasonably prejudice the interests of the copyright owner."

Wednesday, August 13, 2025

Judge rejects Anthropic bid to appeal copyright ruling, postpone trial; Reuters, August 12, 2025

 , Reuters; Judge rejects Anthropic bid to appeal copyright ruling, postpone trial

"A federal judge in California has denied a request from Anthropic to immediately appeal a ruling that could place the artificial intelligence company on the hook for billions of dollars in damages for allegedly pirating authors' copyrighted books.

U.S. District Judge William Alsup said on Monday that Anthropic must wait until after a scheduled December jury trial to appeal his decision that the company is not shielded from liability for pirating millions of books to train its AI-powered chatbot Claude."

Monday, August 11, 2025

Boston Public Library aims to increase access to a vast historic archive using AI; NPR, August 11, 2025

 , NPR ; Boston Public Library aims to increase access to a vast historic archive using AI

"Boston Public Library, one of the oldest and largest public library systems in the country, is launching a project this summer with OpenAI and Harvard Law School to make its trove of historically significant government documents more accessible to the public.

The documents date back to the early 1800s and include oral histories, congressional reports and surveys of different industries and communities...

Currently, members of the public who want to access these documents must show up in person. The project will enhance the metadata of each document and will enable users to search and cross-reference entire texts from anywhere in the world. 

Chapel said Boston Public Library plans to digitize 5,000 documents by the end of the year, and if all goes well, grow the project from there...

Harvard University said it could help. Researchers at the Harvard Law School Library's Institutional Data Initiative are working with libraries, museums and archives on a number of fronts, including training new AI models to help libraries enhance the searchability of their collections. 

AI companies help fund these efforts, and in return get to train their large language models on high-quality materials that are out of copyright and therefore less likely to lead to lawsuits. (Microsoft and OpenAI are among the many AI players targeted by recent copyright infringement lawsuits, in which plaintiffs such as authors claim the companies stole their works without permission.)"

Saturday, August 9, 2025

News Corp CEO Robert Thomson slams AI firms for stealing copyrighted material like Trump’s ‘Art of the Deal’; New York Post, August 6, 2025

Ariel Zilber, New York Post ; News Corp CEO Robert Thomson slams AI firms for stealing copyrighted material like Trump’s ‘Art of the Deal’

"The media executive said the voracious appetite of the AI firms to train their bots on proprietary content without paying for it risks eroding America’s edge over rival nations.

“Much is made of the competition with China, but America’s advantage is ingenuity and creativity, not bits and bytes, not watts but wit,” he said.

“To undermine that comparative advantage by stripping away IP rights is to vandalize our virtuosity.”"

AI industry horrified to face largest copyright class action ever certified; Ars Technica, August 8, 2025

ASHLEY BELANGER, Ars Technica ; AI industry horrified to face largest copyright class action ever certified

"AI industry groups are urging an appeals court to block what they say is the largest copyright class action ever certified. They've warned that a single lawsuit raised by three authors over Anthropic's AI training now threatens to "financially ruin" the entire AI industry if up to 7 million claimants end up joining the litigation and forcing a settlement.

Last week, Anthropic petitioned to appeal the class certification, urging the court to weigh questions that the district court judge, William Alsup, seemingly did not. Alsup allegedly failed to conduct a "rigorous analysis" of the potential class and instead based his judgment on his "50 years" of experience, Anthropic said.

If the appeals court denies the petition, Anthropic argued, the emerging company may be doomed. As Anthropic argued, it now "faces hundreds of billions of dollars in potential damages liability at trial in four months" based on a class certification rushed at "warp speed" that involves "up to seven million potential claimants, whose works span a century of publishing history," each possibly triggering a $150,000 fine.

Confronted with such extreme potential damages, Anthropic may lose its rights to raise valid defenses of its AI training, deciding it would be more prudent to settle, the company argued. And that could set an alarming precedent, considering all the other lawsuits generative AI (GenAI) companies face over training on copyrighted materials, Anthropic argued."

Tuesday, August 5, 2025

Robot Art Riles Artists; ABA Litigation Section, June 25, 2025

 James Michael Miller, ABA Litigation Section; Robot Art Riles Artists

"Visual artists have survived a motion to dismiss their class claims brought against generative artificial intelligence (AI) companies related to the companies’ use of the artists’ visual works without consent. The plaintiffs claimed that the defendants’ text-to-image AI products were trained in part on their copyrighted works. ABA Litigation Section leaders agree that the case sets up a showdown between copyright interests and the “democratization” of art through AI."