Just as long as you don't in any way absorb the information into your brain it's okay.
I understand what authors are trying to say, but legally I don't quite get how they can say that an AI is doing anything other than reading the books, which obviously they're allowed to do.
It is not as if they are making the books publicly available for free, and it's not as if writing in the style of another author is illegal, so I'm not quite sure what law has been broken here.
It's baffling to me seeing comments like this as if the 'AI' is some natural intelligence just hanging out going around reading books it's interested in for the hell of it.. No. These are software companies illegally using artists works (which we require licensing for commercial use) to develop a commercial, profit generating product. Whatever the potential outputs of the AI are is irrelevant when the sources used to train it were obtained illegally.
These are software companies illegally using artists works
There is nothing illegal about what they're doing. You may want it to be illegal, but it's not illegal until laws are actually passed to make it illegal. Things are not illegal by default.
Copyright only prevents copying works. Not analyzing them. The results of the analysis are not the same as the original work.
It is illegal. As an artist, if another individual or company wants to use my work for their own commercial purposes in any way, even if just to 'analyze' (since the analysis is part of their private commercial product), they still need to pay for a license to do so. Otherwise it's an unauthorized use and theft. Copyright doesn't even play into it at that point, and would be a separate issue.
As an artist, if another individual or company wants to use my work for their own commercial purposes in any way, even if just to 'analyze', they still need to pay for a license to do so.
I think you need to review the relevant laws, that's not true.
For example, your comment that I'm responding to is copyrighted and you own the copyright. I just quoted part of it in my response without your permission, and that's an entirely legal fair use. I also pasted your comment into Notepad++ and did a word count, there are 64 words in it. That didn't break any laws either.
A lot of people have very expansive and incorrect ideas about how intellectual property works.
First of all, a random online comment is not protected by copyright law afaik.
Secondly, if you did take something protected by copyright and then used it for commercial purposes (to make money off it), like these LLMs do, then you would be breaking the law.
In short, I'd say you are using a flawed analogy from the start.
Also copyright is not about just copying but also distributing as well. Playing.(radio) songs in your coffee shop for clients is treated differently than you listening to it at home. You generally can't just profit off someone else's work without them allowing it.
First of all, a random online comment is not protected by copyright law afaik.
You got a fundamental aspect of copyright law wrong right in the first line.
Your comments are indeed protected by copyright.
Secondly, if you did take something protected by copyright and then used it for commercial purposes
That's wrong too. Whether or not someone's making money off of a copyright violation will affect the damages you can sue them for, but it's copyright violation either way.
Also copyright is not about just copying but also distributing as well. Playing.(radio) songs in your coffee shop for clients is treated differently than you listening to it at home.
Technically true, but what does it have to do with these circumstances?
You generally can't just profit off someone else's work without them allowing it.
Generally speaking, sure you can. Why couldn't you? People do work that other people profit off of all the time. If a carpenter builds a desk and then I go sit at it while doing my job and earning millions of dollars, I don't need to ask the carpenter's permission.
Copyright has a few extra limitations, but those limitations are on copying stuff without permission.
I keep rereading this comment and as someone in R&D... I'm so astonished that people think that companies just spontaneously come up with everything they produce without looking around. Companies start off almost every venture by analyzing any work in the field that's been done and reverse engineering it. It's how basically anyone you've heard of works. It goes double for art. Inspiration is key for art. Composers will break down the sheet music of great compositions, graphic designers will have walls full of competitors designs, cinematographers will study movies frame by frame.
With great respect I believe that to be a gross simplification of what an LLMs does. There is no training set stored in the LLM, only statistics about what word set is likely to follow what word set. There is not regurgitation of the date - if that was the case, they temperature parameter wouldn’t matter when it very much does.
A slightly compressed JPG of an oil painting is still, at least for purposes of intellectual property rights, not distinct from the original work on canvas. Sufficiently complex and advanced statistics on a work are not substantially different from the work itself. It's just a different way of storing a meaningful representation.
These LLMs are all more or less black boxes. We really cannot conclusively say one way or another whether they are storing and using the full original work in some form or another. We do know that they can be coaxed into spitting out the original work, though, which sure implies it is in there.
And if the work of a human that needs to be fed is being used by one of these bots -- which is pretty much by definition a commercial purpose given that all the relevant bots are operated as such -- then that human should be getting paid.
We do know that they can be coaxed into spitting out the original work, though, which sure implies it is in there.
Only very rarely, under extreme cases of overfitting. Overfitting is a failure state that LLM trainers want to avoid anyway, for reasons unrelated to copyright.
There simply isn't enough space in a LLM's neural network to be storing actual copies of the training data. It's impossible, from a data compression perspective, to fit it in there.
It's not just style. From what I understand (I've never used any of the generative AI tools), they supposedly can and do output chunks of text verbatim from copyrighted works.
Yeah, and even if it WERE truly intelligent -- which these SALAMIs are almost certainly not -- it doesn't even matter.
A human and a robot are not the same. They have different needs and must be afforded different moral protections. Someone can buy a book, read it, learn from it, and incorporate things it learned from that experience into their own future work. They may transform it creatively or it may plagiarize or it may rest in some grey area in-between where it isn't 100% clear if it was novel or plagiarized. All this is also true for a LLM "AI". -- But whether or not this process is fundamentally the same or not isn't even a relevant question.
Copyright law isn't something that exists because it is a pure moral good to protect the creative output of a person from theft. It would be far more ethical to say that all the outputs of human intellect should be shared freely and widely for all people to use, unencumbered by such things. But if creativity is rewarded with only starvation, creativity will go away, so copyright exists as a compromise to try and ensure there is food in the bellies of artists. And with it, we have an understanding that there is a LOT of unclear border space where one artist may feed on the output of another to hopefully grow the pot for everyone.
The only way to fit generative bots into the philosophical framework of copyright is to demand that the generative bots keep food in the bellies of the artists. Currently, they threaten it. It's just that simple. People act like it's somehow an important question whether they "learn" the same way people do, but the question doesn't matter at all. Robots don't get the same leeway and protection afforded to humans because robots do not need to eat.
But if creativity is rewarded with only starvation, creativity will go away, so copyright exists as a compromise to try and ensure there is food in the bellies of artists
no. it exists to stop 18th century london print shops from breaking each others' knees, and subsequently has been expanded to continue to serve the interests of industry.
Maybe so, but the big industry leaders who misuse copyright are the ones who are going to benefit from LLMs the most. Why would profit driven conpanies pay creatives and other employees when a tiny access fee to LLMs will cover almost everything? I think that was admiralteal's point.
I think it's a pretty important question whether we're reaching the end of the distinction between human and machine. People will begin to use machine minds more and more as part of their work. Tying strings now to the works of machines is screwing the creators of tomorrow. The line between what a person creates and what a machine creates WILL evaporate. It's not a matter of if, but when.
Imagine we put a ton of regulations on people who use power tools to do carpentry. I'm sure the carpenters around the time power tools were created figured "That's not true craftsmanship. They shouldn't be able to make a living off that!" But the carpenters of today would be screwed by these regulations because of course they have to use the latest technology to stay competitive.
As for the argument that we're taking the food out of creative's mouths: I don't think anyone is not buying Stephen King novels now because they can just ask for a Stephen King style novel from ChatGPT. You can pirate Stephen King already. People aren't fascinated by LLMs because of how well they plagiarize. They're fascinated by them because they're capable of transformative works, not unlike humans. Nobody is typing "Write a Stephen King Novel" they're typing, "Harold and Kumar go to White Castle but it's Snoop Dogg and Betty White in the style of Stephen King." As much as I'm sure King would love to suck up all royalties for these stories, there's no universe where it makes sense that he should. You don't own what you inspire.
LLMs have been caught plagiarising works, by the simple nature of how they function. They predict the next word based on an assumed context of the previous words, they're very good at constructing sentences but often the issue is "where is it getting its information from?" Authors never consented to their works being fed into an optimisation algorithm and neither did artists when DALL E was created.
For authors, you buy the book and thus the author is paid but that's not what happened with ChatGPT.
Authors never consented to their works being fed into an optimisation algorithm
Yeah I know they didn't but at worst the company owes them 30 bucks for the licence. I don't think copyright law gives authors the right to say who can buy their works, so at the absolute worst, the AI company's stole a book.
To be clear I'm not saying that this should be allowed, I'm just saying that under the current legal system I'm not sure they actually committed that much of a crime. Obviously it needs to be updated, but you do that through political reform (and good luck with that because AI is big bucks), not through the courts.
Copyright Law doesn't talk about who can consume the work. ChatGPT's theft is no different to piracy and companies have gotten very pissy about their shit being pirated but when ChatGPT does it (because the piracy is hidden behind its training), it's fine. The individual authors and artists get shafted in the end because their work has been weaponised against them.
I'm highly skeptical about GPT4 having been directly trained on copyrighted material by Stephen King. Simply by all the sheer information about his works, including summaries, themes, characters, and critical analyses that are publicly available, a good LLM can appear to be able to plagiarize these works, while it doesn't.
If I'm right, there is no leverage for creators to complain. Just accept that that's the world we're living in now. I don't see why this world will stop the sales of books or movie rights on books, etc.
Especially since copyright only protects human authored works. Meaning anything created by an LLM is in the public domain, and the publisher using it loses control of the work.
Of course, this has the potential to be a significant issue, as I can take a copyrighted work, train an LLM using it, and then get it to generate a similar but unique work that is in the public domain. This new work will likely impact the original author’s ability to profit off their original work, thus decreasing supply of human created works in the long run.
Sure, it can plagiarize works it has been trained on. They didn't show in the study, however, that this has occurred for copyright protected material like fiction books.
I saw a comment, probably on Mastodon, from an author saying that (I believe) ChatGPT had plagiarized some of his work verbatim. I don't recall if it was a work of fiction or not, although for the purpose of copyright it doesn't matter.
I wouldn't be surprised if it's trained on works of fiction just as much as non-fiction though. I think that from what I've heard, you can ask ChatGPT to write something in the style of particular writers? If it's possible to give a very specific prompt for it to write something with the same plot points as a Stephen King story in the style of Stephen King, I wonder just how close it would look like the original?