Showing posts with label AI chatbots. Show all posts
Showing posts with label AI chatbots. Show all posts

Monday, July 28, 2025

Your employees may be leaking trade secrets into ChatGPT; Fast Company, July 24, 2025

KRIS NAGEL , Fast Company; Your employees may be leaking trade secrets into ChatGPT

"Every CEO I know wants their team to use AI more, and for good reason: it can supercharge almost every area of their business and make employees vastly more efficient. Employee use of AI is a business imperative, but as it becomes more common, how can companies avoid major security headaches? 

Sift’s latest data found that 31% of consumers admit to entering personal or sensitive information into GenAI tools like ChatGPT, and 14% of those individuals explicitly reported entering company trade secrets. Other types of information that people admit to sharing with AI chatbots include financial details, nonpublic facts, email addresses, phone numbers, and information about employers. At its core, it reveals that people are increasingly willing to trust AI with sensitive information."

Wednesday, July 23, 2025

AI chatbots remain overconfident -- even when they’re wrong; EurekAlert!, July 22, 2025

  CARNEGIE MELLON UNIVERSITY, EurekAlert!; AI chatbots remain overconfident -- even when they’re wrong

"Artificial intelligence chatbots are everywhere these days, from smartphone apps and customer service portals to online search engines. But what happens when these handy tools overestimate their own abilities? 

Researchers asked both human participants and four large language models (LLMs) how confident they felt in their ability to answer trivia questions, predict the outcomes of NFL games or Academy Award ceremonies, or play a Pictionary-like image identification game. Both the people and the LLMs tended to be overconfident about how they would hypothetically perform. Interestingly, they also answered questions or identified images with relatively similar success rates.

However, when the participants and LLMs were asked retroactively how well they thought they did, only the humans appeared able to adjust expectations, according to a study published today in the journal Memory & Cognition.

“Say the people told us they were going to get 18 questions right, and they ended up getting 15 questions right. Typically, their estimate afterwards would be something like 16 correct answers,” said Trent Cash, who recently completed a joint Ph.D. at Carnegie Mellon University in the departments of Social Decision Science and Psychology. “So, they’d still be a little bit overconfident, but not as overconfident.”

“The LLMs did not do that,” said Cash, who was lead author of the study. “They tended, if anything, to get more overconfident, even when they didn’t do so well on the task.”

The world of AI is changing rapidly each day, which makes drawing general conclusions about its applications challenging, Cash acknowledged. However, one strength of the study was that the data was collected over the course of two years, which meant using continuously updated versions of the LLMs known as ChatGPT, Bard/Gemini, Sonnet and Haiku. This means that AI overconfidence was detectable across different models over time.

“When an AI says something that seems a bit fishy, users may not be as skeptical as they should be because the AI asserts the answer with confidence, even when that confidence is unwarranted,” said Danny Oppenheimer, a professor in CMU’s Department of Social and Decision Sciences and coauthor of the study."

Sunday, July 20, 2025

AI guzzled millions of books without permission. Authors are fighting back.; The Washington Post, July 19, 2025

  , The Washington Post; AI guzzled millions of books without permission. Authors are fighting back.


[Kip Currier: I've written this before on this blog and I'll say it again: technology companies would never allow anyone to freely vacuum up their content and use it without permission or compensation. Period. Full Stop.]


[Excerpt]

"Baldacci is among a group of authors suing OpenAI and Microsoft over the companies’ use of their work to train the AI software behind tools such as ChatGPT and Copilot without permission or payment — one of more than 40 lawsuits against AI companies advancing through the nation’s courts. He and other authors this week appealed to Congress for help standing up to what they see as an assault by Big Tech on their profession and the soul of literature.

They found sympathetic ears at a Senate subcommittee hearing Wednesday, where lawmakers expressed outrage at the technology industry’s practices. Their cause gained further momentum Thursday when a federal judge granted class-action status to another group of authors who allege that the AI firm Anthropic pirated their books.

“I see it as one of the moral issues of our time with respect to technology,” Ralph Eubanks, an author and University of Mississippi professor who is president of the Authors Guild, said in a phone interview. “Sometimes it keeps me up at night.”

Lawsuits have revealed that some AI companies had used legally dubious “torrent” sites to download millions of digitized books without having to pay for them."

Tuesday, June 24, 2025

Anthropic wins key US ruling on AI training in authors' copyright lawsuit; Reuters, June 24, 2025

, Reuters; Anthropic wins key US ruling on AI training in authors' copyright lawsuit

 "A federal judge in San Francisco ruled late on Monday that Anthropic's use of books without permission to train its artificial intelligence system was legal under U.S. copyright law.

Siding with tech companies on a pivotal question for the AI industry, U.S. District Judge William Alsup said Anthropic made "fair use" of books by writers Andrea Bartz, Charles Graeber and Kirk Wallace Johnson to train its Claude large language model.

Alsup also said, however, that Anthropic's copying and storage of more than 7 million pirated books in a "central library" infringed the authors' copyrights and was not fair use. The judge has ordered a trial in December to determine how much Anthropic owes for the infringement."

Study: Meta AI model can reproduce almost half of Harry Potter book; Ars Technica, June 20, 2025

TIMOTHY B. LEE  , Ars Techcnica; Study: Meta AI model can reproduce almost half of Harry Potter book

"In recent years, numerous plaintiffs—including publishers of books, newspapers, computer code, and photographs—have sued AI companies for training models using copyrighted material. A key question in all of these lawsuits has been how easily AI models produce verbatim excerpts from the plaintiffs’ copyrighted content.

For example, in its December 2023 lawsuit against OpenAI, The New York Times Company produced dozens of examples where GPT-4 exactly reproduced significant passages from Times stories. In its response, OpenAI described this as a “fringe behavior” and a “problem that researchers at OpenAI and elsewhere work hard to address.”

But is it actually a fringe behavior? And have leading AI companies addressed it? New research—focusing on books rather than newspaper articles and on different companies—provides surprising insights into this question. Some of the findings should bolster plaintiffs’ arguments, while others may be more helpful to defendants.

The paper was published last month by a team of computer scientists and legal scholars from Stanford, Cornell, and West Virginia University. They studied whether five popular open-weight models—three from Meta and one each from Microsoft and EleutherAI—were able to reproduce text from Books3, a collection of books that is widely used to train LLMs. Many of the books are still under copyright."

Thursday, May 22, 2025

A.I.-Generated Reading List in Chicago Sun-Times Recommends Nonexistent Books; The New York Times, May 21, 2025

  , The New York Times; A.I.-Generated Reading List in Chicago Sun-Times Recommends Nonexistent Books

"The summer reading list tucked into a special section of The Chicago Sun-Times and The Philadelphia Inquirer seemed innocuous enough.

There were books by beloved authors such as Isabel Allende and Min Jin Lee; novels by best sellers including Delia Owens, Taylor Jenkins Reid and Brit Bennett; and a novel by Percival Everett, a recent Pulitzer Prize winner.

There was just one issue: None of the book titles attributed to the above authors were real. They had been created by generative artificial intelligence.

It’s the latest case of bad A.I. making its way into the news. While generative A.I. has improved, there is still no way to ensure the systems produce accurate information. A.I. chatbots cannot distinguish between what is true and what is false, and they often make things up. The chatbots can spit out information and expert names with an air of authority."

Wednesday, May 21, 2025

Most AI chatbots easily tricked into giving dangerous responses, study finds; The Guardian, May 21, 2025

   , The Guardian; Most AI chatbots easily tricked into giving dangerous responses, study finds

"Hacked AI-powered chatbots threaten to make dangerous knowledge readily available by churning out illicit information the programs absorb during training, researchers say.

The warning comes amid a disturbing trend for chatbots that have been “jailbroken” to circumvent their built-in safety controls. The restrictions are supposed to prevent the programs from providing harmful, biased or inappropriate responses to users’ questions.

The engines that power chatbots such as ChatGPT, Gemini and Claude – large language models (LLMs) – are fed vast amounts of material from the internet.

Despite efforts to strip harmful text from the training data, LLMs can still absorb information about illegal activities such as hacking, money laundering, insider trading and bomb-making. The security controls are designed to stop them using that information in their responses.

In a report on the threat, the researchers conclude that it is easy to trick most AI-driven chatbots into generating harmful and illegal information, showing that the risk is “immediate, tangible and deeply concerning”...

The research, led by Prof Lior Rokach and Dr Michael Fire at Ben Gurion University of the Negev in Israel, identified a growing threat from “dark LLMs”, AI models that are either deliberately designed without safety controls or modified through jailbreaks. Some are openly advertised online as having “no ethical guardrails” and being willing to assist with illegal activities such as cybercrime and fraud."

Monday, April 21, 2025

This Is How Meta AI Staffers Deemed More Than 7 Million Books to Have No “Economic Value”; Vanity Fair, April 15, 2025

, Vanity Fair; This Is How Meta AI Staffers Deemed More Than 7 Million Books to Have No “Economic Value”

"Last month, I prompted Meta AI with a question: “Is it okay to use pirated books to train AI?” The machine’s response was unequivocal. “Using pirated books to train AI is not okay,” it admonished. “In fact, it’s a serious violation of copyright laws and can have significant consequences for authors and publishers. Recently, it’s been reported that Meta used millions of pirated books to develop its AI programs, which has sparked widespread outrage and condemnation from the writing community.”"

Thursday, April 17, 2025

The philosopher’s machine: my conversation with Peter Singer’s AI chatbot; The Guardian, April 17, 2025

  , The Guardian; The philosopher’s machine: my conversation with Peter Singer’s AI chatbot

"Singer has had a long and high-profile career in philosophy. He has only recently retired from Princeton University, where he was professor of bioethics. His CV is extensive (it runs to a full 102 pages). He has a Substack, a podcast, a Bluesky account, a journal he co-edits, and a polished online presence. As I’m scrolling through the storied septuagenarian philosopher’s blog, the chatbot pops up in the corner of the page. I can’t help but feeling like I am being sold something.

In launching his chatbot, Singer described the tool as “employing the Socratic method” – that is, dialogue – to “[guide] users through a process of critical thinking and self-examination”. But I have questions about that.

Me: Do you think it is philosophically justifiable to outsource the discussion of ethics to a chatbot?

Peter Singer AI: Outsourcing ethical discussions to a chatbot can be philosophically justifiable if viewed as a tool to enhance and facilitate ethical reflection. Chatbots can provide insights, pose questions, and offer perspectives that might help individuals think more deeply about ethical issues.

However, it’s crucial to remember that chatbots lack human experience, empathy, and the ability to understand context fully. They should be used as a complement to human judgment and conversation, not a replacement.

Engaging with a chatbot can stimulate thinking and provide a starting point for deeper exploration, but ultimately, ethical decision-making involves personal reflection and discussion with others.

My most memorable and exciting philosophy lessons were the ones in which, as a class, we hotly debated ideas in which we were all materially and emotionally invested. The AI floats loftily above the emotional fray, coolly dispensing pre-programmed wisdom from the cloud, while we grapple with the contradictions and complications of actually living. Between the chatbot and me, there is only one of us for whom the dialogue elicits an emotional response, because for only one of us does the conversation actually matter." 

Thursday, March 27, 2025

Judge allows 'New York Times' copyright case against OpenAI to go forward; NPR, March 27, 2025

, NPR ; Judge allows 'New York Times' copyright case against OpenAI to go forward

"A federal judge on Wednesday rejected OpenAI's request to toss out a copyright lawsuit from The New York Times that alleges that the tech company exploited the newspaper's content without permission or payment.

In an order allowing the lawsuit to go forward, Judge Sidney Stein, of the Southern District of New York, narrowed the scope of the lawsuit but allowed the case's main copyright infringement claims to go forward.

Stein did not immediately release an opinion but promised one would come "expeditiously."

The decision is a victory for the newspaper, which has joined forces with other publishers, including The New York Daily News and the Center for Investigative Reporting, to challenge the way that OpenAI collected vast amounts of data from the web to train its popular artificial intelligence service, ChatGPT."

Wednesday, March 26, 2025

Anthropic wins early round in music publishers' AI copyright case; Reuters, March 26, 2025

 , Reuters; Anthropic wins early round in music publishers' AI copyright case

"Artificial intelligence company Anthropic convinced a California federal judge on Tuesday to reject a preliminary bid to block it from using lyrics owned by Universal Music Group and other music publishers to train its AI-powered chatbot Claude.

U.S. District Judge Eumi Lee said that the publishers' request was too broad and that they failed to show Anthropic's conduct caused them "irreparable harm."

Friday, December 27, 2024

Character.AI Confirms Mass Deletion of Fandom Characters, Says They're Not Coming Back; Futurism, November 27, 2024

 MAGGIE HARRISON DUPRÉ , Futurism; Character.AI Confirms Mass Deletion of Fandom Characters, Says They're Not Coming Back

"The embattled AI companion company Character.AI confirmed to Futurism that it removed a large number of characters from its platform, citing its adherence to the Digital Millennium Copyright Act (DCMA) and copyright law, but failing to say whether the deletions were proactive or in response to requests from the holders of the characters' intellectual property rights...

That's not surprising: Character.AI is currently facing a lawsuit brought by the family of a 14-year-old teenager in Florida who died by suicide after forming an intense relationship with a Daenerys Targaryen chatbot on its platform...

It's been a bad few months for Character.AI. In October, shortly before the recent lawsuit was filed, it was revealed that someone had created a chatbot based on a murdered teenager without consent from the slain teen's family. (The character was removed and Character.AI apologized, as AdWeek first reported.) And in recent weeks, we've reported on disturbing hordes of suicidepedophilia, and eating disorder-themed chatbots hosted by the platform, all of which were freely accessible to Character.AI users of all ages."

Sunday, December 8, 2024

Google CEO: AI development is finally slowing down—‘the low-hanging fruit is gone’; CNBC, December 8, 2024

Megan Sauer , CNBC; Google CEO: AI development is finally slowing down—‘the low-hanging fruit is gone’;

"Now, with the industry’s competitive landscape somewhat established — multiple big tech companies, including Google, have competing models — it’ll take time for another technological breakthrough to shock the AI industry into hyper-speed development again, Pichai said at the New York Times’ DealBook Summit last week.

“I think the progress is going to get harder. When I look at [2025], the low-hanging fruit is gone,” said Pichai, adding: “The hill is steeper ... You’re definitely going to need deeper breakthroughs as we get to the next stage.”...

Some tech CEOs, like Microsoft’s Satya Nadella, agree with Pichai. “Seventy years of the Industrial Revolution, there wasn’t much industry growth, and then it took off ... it’s never going to be linear,” Nadella saidat the Fast Company Innovation Festival 2024 in October.

Others disagree, at least publicly. OpenAI CEO Sam Altman, for example, posted “there is no wall” on social media platform X in November — a response to reports that the recently released ChatGPT-4 was only moderately better than previous models."

In Wisconsin, Professors Worry AI Could Replace Them; Inside Higher Ed, December 6, 2024

 Kathryn Palmer, Inside Higher Ed; In Wisconsin, Professors Worry AI Could Replace Them

"Faculty at the cash-strapped Universities of Wisconsin System are pushing back against a proposed copyright policy they believe would cheapen the relationship between students and their professors and potentially allow artificial intelligence bots to replace faculty members...

The policy proposal is not yet final and is open for public comment through Dec. 13. ..

Natalia Taft, an associate professor of biological sciences at the University of Wisconsin–Parkside who signed the open letter, told Inside Higher Ed that she believes the policy proposal “is part of the trend of the corporatization of academia.”...

Jane Ginsburg, a professor of literary and artistic property law at Columbia University School of Law, said the university has the law on its side. 

Under the 1976 Copyright Act, “course material prepared by employees, including professors, as part of their jobs comes within the definition of a ‘work made for hire,’ whose copyright vests initially in the employer (the University), not the employee (the professor).”"

The Copyrighted Material Being Used to Train AI; The Bulwark, December 7, 2024

 SONNY BUNCH, The Bulwark; The Copyrighted Material Being Used to Train AI

"On this week’s episode, I talked to Alex Reisner about his pieces in the Atlantic highlighting the copyrighted material being hoovered into large language models to help AI chatbots simulate human speech. If you’re a screenwriter and would like to see which of your work has been appropriated to aid in the effort, click here; he has assembled a searchable database of nearly 140,000 movie and TV scripts that have been used without permission. (And you should read his other stories about copyright law reaching its breaking point and “the memorization problem.”) In this episode, we also got into the metaphysics of art and asked what sort of questions need to be asked as we hurtle toward the future. If you enjoyed this episode, please share it with a friend!"

Monday, November 4, 2024

What AI knows about you; Axios, November 4, 2024

 Ina Friend, Axios; What AI knows about you

"Most AI builders don't say where they are getting the data they use to train their bots and models — but legally they're required to say what they are doing with their customers' data.

The big picture: These data-use disclosures open a window onto the otherwise opaque world of Big Tech's AI brain-food fight.

  • In this new Axios series, we'll tell you, company by company, what all the key players are saying and doing with your personal information and content.

Why it matters: You might be just fine knowing that picture you just posted on Instagram is helping train the next generative AI art engine. But you might not — or you might just want to be choosier about what you share.

Zoom out: AI makers need an incomprehensibly gigantic amount of raw data to train their large language and image models. 

  • The industry's hunger has led to a data land grab: Companies are vying to teach their baby AIs using information sucked in from many different sources — sometimes with the owner's permission, often without it — before new laws and court rulings make that harder. 

Zoom in: Each Big Tech giant is building generative AI models, and many of them are using their customer data, in part, to train them.

  • In some cases it's opt-in, meaning your data won't be used unless you agree to it. In other cases it is opt-out, meaning your information will automatically get used unless you explicitly say no. 
  • These rules can vary by region, thanks to legal differences. For instance, Meta's Facebook and Instagram are "opt-out" — but you can only opt out if you live in Europe or Brazil.
  • In the U.S., California's data privacy law is among the laws responsible for requiring firms to say what they do with user data. In the EU, it's the GDPR."

Friday, October 18, 2024

Penguin Random House underscores copyright protection in AI rebuff; The Bookseller, October 18, 2024

 MATILDA BATTERSBY, The Bookseller; Penguin Random House underscores copyright protection in AI rebuff

"The world’s biggest trade publisher has changed the wording on its copyright pages to help protect authors’ intellectual property from being used to train large language models (LLMs) and other artificial intelligence (AI) tools, The Bookseller can exclusively reveal.

Penguin Random House (PRH) has amended its copyright wording across all imprints globally, confirming it will appear “in imprint pages across our markets”. The new wording states: “No part of this book may be used or reproduced in any manner for the purpose of training artificial intelligence technologies or systems”, and will be included in all new titles and any backlist titles that are reprinted.

The statement also “expressly reserves [the titles] from the text and data mining exception”, in accordance with a European Parliament directive.

The move specifically to ban the use of its titles by AI firms for the development of chatbots and other digital tools comes amid a slew of copyright infringement cases in the US and reports that large tranches of pirated books have already been used by tech companies to train AI tools. In 2024, several academic publishers including Taylor & Francis, Wiley and Sage have announced partnerships to license content to AI firms.

PRH is believed to be the first of the Big Five anglophone trade publishers to amend its copyright information to reflect the acceleration of AI systems and the alleged reliance by tech companies on using published work to train language models."

Wednesday, October 16, 2024

His daughter was murdered. Then she reappeared as an AI chatbot.; The Washington Post, October 15, 2024

  , The Washington Post; His daughter was murdered. Then she reappeared as an AI chatbot.

"Jennifer’s name and image had been used to create a chatbot on Character.AI, a website that allows users to converse with digital personalities made using generative artificial intelligence. Several people had interacted with the digital Jennifer, which was created by a user on Character’s website, according to a screenshot of her chatbot’s now-deleted profile.

Crecente, who has spent the years since his daughter’s death running a nonprofit organization in her name to prevent teen dating violence, said he was appalled that Character had allowed a user to create a facsimile of a murdered high-schooler without her family’s permission. Experts said the incident raises concerns about the AI industry’s ability — or willingness — to shield users from the potential harms of a service that can deal in troves of sensitive personal information...

The company’s terms of service prevent users from impersonating any person or entity...

AI chatbots can engage in conversation and be programmed to adopt the personalities and biographical details of specific characters, real or imagined. They have found a growing audience online as AI companies market the digital companions as friends, mentors and romantic partners...

Rick Claypool, who researched AI chatbots for the nonprofit consumer advocacy organization Public Citizen, said while laws governing online content at large could apply to AI companies, they have largely been left to regulate themselves. Crecente isn’t the first grieving parent to have their child’s information manipulated by AI: Content creators on TikTok have used AI to imitate the voices and likenesses of missing children and produce videos of them narrating their deaths, to outrage from the children’s families, The Post reported last year.

“We desperately need for lawmakers and regulators to be paying attention to the real impacts these technologies are having on their constituents,” Claypool said. “They can’t just be listening to tech CEOs about what the policies should be … they have to pay attention to the families and individuals who have been harmed.”

Tuesday, October 1, 2024

Fake Cases, Real Consequences [No digital link as of 10/1/24]; ABA Journal, Oct./Nov. 2024 Issue

 John Roemer, ABA Journal; Fake Cases, Real Consequences [No digital link as of 10/1/24]

"Legal commentator Eugene Volokh, a professor at UCLA School of Law who tracks AI in litigation, in February reported on the 14th court case he's found in which AI-hallucinated false citations appeared. It was a Missouri Court of Appeals opinion that assessed the offending appellant $10,000 in damages for a frivolous filing.

Hallucinations aren't the only snag, Volokh says. "It's also with the output mischaracterizing the precedents or omitting key context. So one still has to check that output to make sure it's sound, rather than just including it in one's papers.

Echoing Volokh and other experts, ChatGPT itself seems clear-eyed about its limits. When asked about hallucinations in legal research, it replied in part: "Hallucinations in chatbot answers could potentially pose a problem for lawyers if they relied solely on the information provided by the chatbot without verifying its accuracy."