AIs can generate near-verbatim copies of novels from training data

The world’s top AI models can be prompted to generate near-verbatim copies of bestselling novels, raising fresh questions about the industry’s claim that its systems do not store copyrighted works. A series of recent studies has shown that large language models from OpenAI, Google, Meta, Anthropic, and xAI memorize far more of their training data than previously thought.

Source: AIs can generate near-verbatim copies of novels from training data

Get the latest RightsTech news and analysis delivered directly in your inbox every week
We respect your privacy.