Thousands of authors demand payment from AI companies for use of copyrighted works::Thousands of published authors are requesting payment from tech companies for the use of their copyrighted works in training artificial intelligence tools, marking the latest intellectual property critique to target AI development.
How can they prove that not some abstract public data has been used to train algorithms, but their particular intellectual property?
Well, if you ask e.g. ChatGPT for the lyrics to a song or page after page of a book, and it spits them out 1:1 correct, you could assume that it must have had access to the original.
Or at least excerpts from it. But even then, it’s one thing for a person to put up a quote from their favourite book on their blog, and a completely different thing for a private company to use that data to train a model, and then sell it.
Even more so, if you consider that the LLMs are marketed to replace the authors.
Yeah which I still feel is utterly ridiculous. I love the idea of AI tools to assist with things, but as a complete replacement? No thank you.
I enjoy using things like SynthesizerV and VOCALOID because my own voice is pretty meh and my singing skills aren’t there. It’s fun to explore the voices, and learn how to use the tools. That doesn’t mean I’d like to see all singers replaced with synthesized versions. I view SynthV and the like as instruments, not much more.
I’ve used LLVMs to proofread stuff, and help me rephrase letters and such, but I’d never hire an editor to do such small tasks for me anyway. The result has always required editing anyway, because the LLVMs have a tendency to make stuff up.
Cases like that I don’t see a huge problem with. At my workplace though they’re talking about generating entire application layouts and codebases with AI and, being in charge of the AI evaluation project, the tech just isn’t there yet. You can in a sense use AI to make entire projects, but it’ll generate gnarly unmaintainable rubbish. You need a human hand in there to guide it.
Otherwise you end up with garbage websites with endlessly generated AI content, that can easily be manipulated by third party actors.
you could assume that it must have had access to the original.
I don’t know if that’s true. If Google grabs that book from a pirate site. Then publishes the work as search results. ChatGPT grabs the work from Google results and cobbles it back together as the original.
Who’s at fault?
I don’t think it’s a straight forward ChatGPT can reproduce the work therefore it stole it.
Both are at fault: Google for distributing pirated material and OpenAI for using said material for financial gain.
Copyright doesn’t work like that. Say I sell you the rights to Thriller by Michael Jackson. You might not know that I don’t have the rights. But even if you bought the rights from me, whoever actually has the rights is totally in their legal right to sue you, because you never actually purchased any rights.
So if ChatGPT ripps it off Google who ripped it off a pirate site, then everyone in that chain who reproduced copyrighted works without permission from the copyright owners is liable for the damages caused by their unpermitted reproduction.
It’s literally the same as downloading something from a pirate site doesn’t make it legal, just because someone ripped it before you.
That’s a terrible example because under copyright law downloading a pirated thing isn’t actually illegal. It’s the distribution that is illegal (uploading).
Yes, downloading is illegal, and the media is still an illegally obtained copy. It’s just never prosecuted, because the damages are miniscule if you just download. They can only fine you for the amount of damages you caused by violating the copyright.
If you upload to 10k people, they can claim that everyone of them would have paid for it, so the damages are (if one copy is worth €30) ~€300k. That’s a lot of money and totally worth the lawsuit.
On the other hand, if you just download, the damages are just the value of one copy (in this case €30). That’s so miniscule, that even having a lawyer write a letter is more expensive.
But that’s totally besides the point. OpenAI didn’t just download, they replicate. Which is causing massive damages, especially to the original artists, which in many cases are now not hired any more, since ChatGPT replaces them.
Can it recreate anything 1:1? When both my wife and I tried to get them to do that they would refuse, and if pushed they would fail horribly.
This is what I got. Looks pretty 1:1 for me.
Hilarious that it started with just “Buddy”, like you’d be happy with only the first word.
Yeah, for some reason it does that a lot when I ask it for copyrighted stuff.
As if it knew it wasn’t supposed to output that.
To be fair you’d get the same result easier by just googling “we will rock you lyrics”
How is chatgpt knowing the lyrics to that song different from a website that just tells you the lyrics of the song?
Two points:
-
Google spitting out the lyrics isn’t ok from a copyright standpoint either. The reason why songwriters/singers/music companies don’t sue people who publish lyrics (even though they totally could) is because no damages. They sell music, so the lyrics being published for free doesn’t hurt their music business and it also doesn’t hurt their songwriting business. Other types of copyright infringement that musicians/music companies care about are heavily policed, also on Google.
-
Content generation AI has a different use case, and it could totally hurt both of these businesses. My test from above that got it to spit out the lyrics verbatim shows, that the AI did indeed use copyrighted works for it’s training. Now I can ask GPT to generate lyrics in the style of Queen, and it will basically perform the song texter’s job. This can easily be done on a commercial scale, replacing the very human that has written these song texts. Now take this a step further and take a voice-generating AI (of which there are many), which was similarly trained on copyrighted audio samples of Freddie Mercury. Then add to the mix a music-generating AI, also fed with works of Queen, and now you have a machine capable of generating fake Queen songs based directly on Queen’s works. You can do the very same with other types of media as well.
And this is where the real conflict comes from.
-
there are a lot of possible ways to audit an AI for copyrighted works, several of which have been proposed in the comments here, but what this could lead to is laws requiring an accounting log of all material that has been used to train an AI as well as all copyrights and compensation, etc.
Not without some seriously invasive warrants! Ones that will never be granted for an intellectual property case.
Intellectual property is an outdated concept. It used to exist so wealthier outfits couldn’t copy your work at scale and muscle you out of an industry you were championing.
It simply does not work the way it was intended. As technology spreads, the barrier for entry into most industries wherein intellectual property is important has been all but demolished.
i.e. 50 years ago: your song that your band performed is great. I have a recording studio and am gonna steal it muahahaha.
Today: “anyone have an audio interface I can borrow so my band can record, mix, master, and release this track?”
Intellectual property ignores the fact that, idk, Issac Newton and Gottfried Wilhelm Leibniz both independently invented calculus at the same time on opposite ends of a disconnected globe. That is to say, intellectual property doesn’t exist.
Ever opened a post to make a witty comment to find someone else already made the same witty comment? Yeah. It’s like that.
Spoken by someone who has never had something you’ve worked years on, be stolen.
What was “stolen” from you and how?
deleted by creator
Spoken like someone who is having trouble admitting they’re standing on the shoulders of Giants.
I don’t expect a nuanced response from you, nor will I waste time with folks who can’t be bothered to respond in any form beyond attack, nor do I expect you to watch this
Intellectual property died with the advent of the internet. It’s now just a way for the wealthy to remain wealthy.
Personally speaking, I’ve generated some stupid images like different cities covered in baked beans and have had crude watermarks generate with them where they were decipherable enough that I could find some of the source images used to train the ai. When it comes to photo realistic image generation, if all the ai does is mildly tweak the watermark then it’s not too hard to trace back.
I think that to protect creators they either need to be transparent about all content used to train the AI (highly unlikely) or have a disclaimer of liability, wherein if original content has been used is training of AI then the Original Content creator who have standing for legal action.
The only other alternative would be to insure that the AI specifically avoid copyright or trademarked content going back to a certain date.
Why a certain date? That feels arbitrary
At a certain age some media becomes public domain
Then it is no longer copywrited
I’d think that given the nature of the language models and how the whole AI thing tends to work, an author can pluck a unique sentence from one of their works, ask AI to write something about that, and if AI somehow ‘magically’ writes out an entire paragraph or even chapter of the author’s original work, well tada, AI ripped them off.
They can’t. All they could prove is that their work is part of a dataset that still exists.
You know what would solve this? We all collectively agree this fucking tech is too important to be in the hands of a few billionaires, start an actual public free open source fully funded and supported version of it, and use it to fairly compensate every human being on Earth according to what they contribute, in general?
Why the fuck are we still allowing a handful of people to control things like this??
Because the tech behind it isn’t cheap and money does not fall from trees.
No entity on the planet has more money than our governments. It’d be more efficient for a government to fund this than any private company.
Many governments on the planet have less money than some big tech or oil companies. Obviously not those of large industrious nations, but most nations aren’t large and industrious.
The government and efficiency don’t go together
Plenty of research shows that each dollar into government programs gets much more returns than private companies. This literally a neolib propaganda talking point.
There is nothing objectively wrong with your statement. However, we somehow always default to solving that issue by having some dragon hoard enough gold, and there is something objectively wrong with that.
Because we shy away from responsibility.
I think the longer response to this is more accurate. It’s more “because capitalism” than anything else.
And capitalism over the course of the 20th century made very successful attempts of alienating completely the working class and destroying all class consciousness or material awareness.
So people keep thinking that the problems is we as individuals are doing capitalism wrong. Not capitalism.
You think it is so simple you can just download it and run it on your laptop?
You kind of can though? The bigger models aren’t really more complicated, just bigger. If you can cram enough ram or swap into a laptop,
lamma.cpp
will get there eventually.
There is already a business model for compensating authors: it is called buying the book. If the AI trainers are pirating books, then yeah - sue them.
There are plagiarism and copyright laws to protect the output of these tools: if the output is infringing, then sue them. However, if the output of an AI would not be considered infringing for a human, then it isn’t infringement.
When you sell a book, you don’t get to control how that book is used. You can’t tell me that I can’t quote your book (within fair use restrictions). You can’t tell me that I can’t refer to your book in a blog post. You can’t dictate who may and may not read a book. You can’t tell me that I can’t give a book to a friend. Or an enemy. Or an anarchist.
Folks, this isn’t a new problem, and it doesn’t need new laws.
It’s 100% a new problem. There’s established precedent for things costing different amounts depending on their intended use.
For example, buying a consumer copy of song doesn’t give you the right to play that song in a stadium or a restaurant.
Training an entire AI to make potentially an infinite number of derived works from your work is 100% worthy of requiring a special agreement. This even goes beyond simple payment to consent; a climate expert might not want their work in an AI which might severely mischatacterize the conclusions, or might want to require that certain queries are regularly checked by a human, etc
My point is that the restrictions can’t go on the input, it has to go on the output - and we already have laws that govern such derivative works (or reuse / rebroadcast).
When you sell a book, you don’t get to control how that book is used.
This is demonstrably wrong. You cannot buy a book, and then go use it to print your own copies for sale. You cannot use it as a script for a commercial movie. You cannot go publish a sequel to it.
Now please just try to tell me that AI training is specifically covered by fair use and satire case law. Spoiler: you can’t.
This is a novel (pun intended) problem space and deserves to be discussed and decided, like everything else. So yeah, your cavalier dismissal is cavalierly dismissed.
No, you misunderstand. Yes, they can control how the content in the book is used - that’s what copyright is. But they can’t control what I do with the book - I can read it, I can burn it, I can memorize it, I can throw it up on my roof.
My argument is that the is nothing wrong with training an AI with a book - that’s input for the AI, and that is indistinguishable from a human reading it.
Now what the AI does with the content - if it plagiarizes, violates fair use, plagiarizes- that’s a problem, but those problems are already covered by copyright laws. They have no more business saying what can or cannot be input into an AI than they can restrict what I can read (and learn from). They can absolutely enforce their copyright on the output of the AI just like they can if I print copies of their book.
My objection is strictly on the input side, and the output is already restricted.
Makes sense. I would love to hear how anyone can disagree with this. Just because an AI learned or trained from a book doesn’t automatically mean it violated any copyrights.
The base assumption of those with that argument is that an AI is incapable of being original, so it is “stealing” anything it is trained on. The problem with that logic is that’s exactly how humans work - everything they say or do is derivative from their experiences. We combine pieces of information from different sources, and connect them in a way that is original - at least from our perspective. And not surprisingly, that’s what we’ve programmed AI to do.
Yes, AI can produce copyright violations. They should be programmed not to. They should cite their sources when appropriate. AI needs to “learn” the same lessons we learned about not copy-pasting Wikipedia into a term paper.
It’s specifically distribution of the work or derivatives that copyright prevents.
So you could make an argument that an LLM that’s memorized the book and can reproduce (parts of) it upon request is infringing. But one that’s merely trained on the book, but hasn’t memorized it, should be fine.
But by their very nature the LLM simply redistribute the material they’ve been trained on. They may disguise it assiduously, but there is no person at the center of the thing adding creative stokes. It’s copyrighted material in, copyrighted material out, so the plaintiffs allege.
They don’t redistribute. They learn information about the material they’ve been trained on - not there natural itself*, and can use it to generate material they’ve never seen.
- Bigger models seem to memorize some of the material and can infringe, but that’s not really the goal.
I asked Bing Chat for the 10th paragraph of the first Harry Potter book, and it gave me this:
“He couldn’t know that at this very moment, people meeting in secret all over the country were holding up their glasses and saying in hushed voices: ‘To Harry Potter – the boy who lived!’”
It looks like technically I might be able to obtain the entire book (eventually) by asking Bing the right questions?
Then this is a copyright violation - it violates any standard for such, and the AI should be altered to account for that.
What I’m seeing is people complaining about content being fed into AI, and I can’t see why that should be a problem (assuming it was legally acquired or publicly available). Only the output can be problematic.
I think it’s not just the output. I can buy an image on any stock Plattform, print it on a T-Shirt, wear it myself or gift it to somebody. But if I want to sell T-Shirts using that image I need a commercial licence - even if I alter the original image extensivly or combine it with other assets to create something new. It’s not exactly the same thing but openAI and other companies certainly use copyrighted material to create and improve commercial products. So this doesn’t seem the same kind of usage an avarage joe buys a book for.
This is a little off, when you quote a book you put the name of the book you’re quoting. When you refer to a book, you, um, refer to the book?
I think the gist of these authors complaints is that a sort of “technology laundered plagiarism” is occurring.
Copyright 100% applies to the output of an AI, and it is subject to all the rules of fair use and attribution that entails.
That is very different than saying that you can’t feed legally acquired content into an AI.
However, if the output of an AI would not be considered infringing for a human, then it isn’t infringement.
It’s an algorithm that’s been trained on numerous pieces of media by a company looking to make money of it. I see no reason to give them a pass on fairly paying for that media.
You can see this if you reverse the comparison, and consider what a human would do to accomplish the task in a professional setting. That’s all an algorithm is. An execution of programmed tasks.
If I gave a worker a pirated link to several books and scientific papers in the field, and asked them to synthesize an overview/summary of what they read and publish it, I’d get my ass sued. I have to buy the books and the scientific papers. STEM companies regularly pay for access to papers and codes and standards. Why shouldn’t an AI have to do the same?
Someone should AGPL their novel and force the AI company to open source their entire neural network.
This is a good debate about copyright/ownership. On one hand, yes, the authors works went into ‘training’ the AI…but we would need a scale to then grade how well a source piece is good at being absorbed by the AI’s learning. for example. did the AI learn more from the MAD magazine i just fed it or did it learn more from Moby Dick? who gets to determine that grading system. Sadly musicians know this struggle. there are just so many notes and so many words. eventually overlap and similiarities occur. but did that musician steal a riff or did both musicians come to a similar riff seperately? Authors dont own words or letters so a computer that just copies those words and then uses an algo to write up something else is no more different than you or i being influenced by our favorite heroes or i formation we have been given. do i pay the author for reading his book? or do i just pay the store to buy it?
Copyright laws are really out of control at this point. Their periods are far too long and, like you said, how can anyone claim to truly be original at this point? A dedicated lawyer can find reasonable prior art for pretty much anything nowadays. The only reason old sources look original is because no records exist of the sources they used.
While I am rooting for authors to make sure they get what they deserve, I feel like there is a bit of a parallel to textbooks here. As an engineer if I learn about statics from a text book and then go use that knowledge to he’ll design a bridge that I and my company profit from, the textbook company can’t sue. If my textbook has a detailed example for how to build a new bridge across the Tacoma Narrows, and I use all of the same design parameters for a real Tacoma Narrows bridge, that may have much more of a case.
It’s not really a parallel.
The text books don’t have copyrights on the concepts and formulae they teach. They only have copyrights for the actual text.
If you memorize the text book and write it down 1:1 (or close to it) and then sell that text you wrote down, then you are still in violation of the copyright.
And that’s what the likes of ChatGPT are doing here. For example, ask it to output the lyrics for a song and it will spit out the whole (copyrighted) lyrics 1:1 (or very close to it). Same with pages of books.
The memorization is closer to that of a fanatic fan of the author. It usually knows the beginning of the book and the more well known passages, but not entire longer works.
By now, ChatGPT is trying to refuse to output copyrighted materials know even where it could, and though it can be tricked, they appear to have implemented a hard filter for some more well known passages, which stops generation a few words in.
Have you tried just telling it to “continue”?
Somewhere in the comments to this post I posted screenshots of me trying to get lyrics for “We will rock you” from ChatGPT. It first just spat out “Verse 1: Buddy,” and ended there. So I answered with “continue”, it spat out the next line and after the second “continue” it gave me the rest of the lyrics.
Similar story with e.g. the first chapter of Harry Potter 1 and other stuff I tried. The output is often not perfect, with a few words being wrong, but it’s very clearly a “derived work” of the original. In the view of copyright law, changing a few words here is not a valid way of getting around copyrights.
I think that these are fiction writers. The maths you’d use to design that bridge is fact and the book company merely decided how to display facts. They do not own that information, whereas the Handmaid’s Tale was the creation of Margaret Atwood and was an original work.
You have a point but there’s a pretty big difference between something like a statistics textbook and the novel “Dune” for instance. One was specifically written to teach mostly pre-existing ideas and the other was created as entertainment to sell to a wide an audience as possible.
But you paid for the textbook
Libraries exist
Plagiarism filters frequently trigger on chatgpt written books and articles.
All this copyright/AI stuff is so silly and a transparent money grab.
They’re not worried that people are going to ask the LLM to spit out their book; they’re worried that they will no longer be needed because a LLM can write a book for free. (I’m not sure this is feasible right now, but maybe one day?) They’re trying to strangle the technology in the courts to protect their income. That is never going to work.
Notably, there is no “right to control who gets trained on the work” aspect of copyright law. Obviously.
There is nothing silly about that. It’s a fundamental question about using content of any kind to train artificial intelligence that affects way more than just writers.
I seriously doubt Sarah Silverman is suing OpenAI because she’s worried ChatGPT will one day be funnier than she is. She just doesn’t want it ripping off her work.
What do you mean when you say “ripping off her work”? What do you think an LLM does, exactly?
In her case, taking elements of her book and regurgitating them back to her. Which sounds a lot like they could be pirating her book for training purposes to me.
How do you know they didn’t just buy the book?
Again, that’s not relevant.
Quoting someone’s book is not “ripping off” the work.
How is it able to quote the book? Magic?
So you’re saying that as long as they buy 1 copy of the book, it’s all good?
No, I’m not saying that. If she’s right and it can spit out any part of her book when asked (and someone else showed that it does that with Harry Potter), it’s plagiarism. They are profiting off of her book without compensating her. Which is a form of ripping someone off. I’m not sure what the confusion here is. If I buy someone’s book, that doesn’t give me the right to put it all online for free.
Designing and marketing a system to plagiarize works en masse? That’s the cash grab.
Can you elaborate on this concept of a LLM “plagiarizing”? What do you mean when you say that?
What I mean is that it is a statistical model used to generate things by combining parts of extant works. Everything that it “creates” is a piece of something that already exists, often without the author’s consent. Just because it is done at a massive scale doesn’t make it less so. It’s basically just a tracer.
Not saying that the tech isn’t amazing or likely a component of future AI but, it’s really just being used commercially to rip people off and worsen the human condition for profit.
Everything that it “creates” is a piece of something that already exists, often without the author’s consent
This describes all art. Nothing is created in a vacuum.
No, it really doesn’t, nor does it function like human cognition. Take this example:
I, personally, to decide that I wanted to make a sci-fi show. I don’t want to come up with ideas so, I want to try to do something that works. I take the scripts of every Star Trek: The Search for Spock, Alien, and Earth Girls Are Easy and feed them into a database, seperating words into individual data entries with some grammatical classification. Then, using this database, I generate a script, averaging the length of the films, with every word based upon its occurrence in the films or randomized, if it’s a tie. I go straight into production with “Star Alien: The Girls Are Spock”. I am immediately sued by Disney, Lionsgate, and Paramount for trademark and copyright infringement, even though I basically just used a small LLM.
You are right that nothing is created in a vacuum. However, plagiarism is still plagiarism, even if it is using a technically sophisticated LLM plagiarism engine.
ChatGPT doesn’t have direct access to the material it’s trained on. Go ask it to quote a book to you.
That really doesn’t make an appreciable difference. It doesn’t need direct access to source data, if it’s already been transferred into statistical data.
So what’s the difference between a person reading their books and using the information within to write something and an ai doing it?
Because AIs aren’t inspired by anything and they don’t learn anything
So uninspired writing is illegal?
No but a lazy copy of someone else’s work might be copyright infringement.
So when does Kevin Costner get to sue James Cameron for his lazy copy of Dances With Wolves?
Avatar is not Dances with Wolves. It’s Ferngully.
Idk, maybe. There are thousands of copyright infringement lawsuits, sometimes they win.
I don’t necessarily agree with how copyright law works, but that’s a different question. Doesn’t change the fact that sometimes you can successfully sue for copyright infringement if someone copies your stuff to make something new.
Why not? Hollywood is full to the brim with people suing for copyright infringement. And sometimes they win. Why should it be different for AI companies?
Language models actually do learn things in the sense that: the information encoded in the training model isn’t usually* taken directly from the training data; instead, it’s information that describes the training data, but is new. That’s why it can generate text that’s never appeared in the data.
- the bigger models seem to remember some of the data and can reproduce it verbatim; but that’s not really the goal.
What does inspiration have to do with anything? And to be honest, humans being inspired has led to far more blatant copyright infringement.
As for learning, they do learn. No different than us, except we learn silly abstractions to make sense of things while AI learns from trial and error. Ask any artist if they’ve ever looked at someone else’s work to figure out how to draw something, even if they’re not explicitly looking up a picture, if they’ve ever seen a depiction of it, they recall and use that. Why is it wrong if an AI does the same?
A person is human and capable of artistry and creativity, computers aren’t. Even questioning this just means dehumanizing artists and art in general.
Not being allowed to question things is a really shitty precedent, don’t you think?
Do you think a hammer and a nail could do anything on their own, without a hand picking them up guiding them? Because that’s what a computer is. Nothing wrong with using a computer to paint or write or record songs or create something, but it has to be YOU creating it, using the machine as a tool. It’s also in the actual definition of the word: art is made by humans. Which explicitly excludes machines. Period. Like I’m fine with AI when it SUPPORTS an artist (although sometimes it’s an obstacle because sometimes I don’t want to be autocorrected, I want the thing I write to be written exactly as I wrote it, for whatever reason). But REPLACING an artist? Fuck no. There is no excuse for making a machine do the work and then to take the credit just to make a quick easy buck on the backs of actual artists who were used WITHOUT THEIR CONSENT to train a THING to replace them. Nah fuck off my guy. I can clearly see you never did anything creative in your whole life, otherwise you’d get it.
Nah fuck off my guy. I can clearly see you never did anything creative in your whole life, otherwise you’d get it.
Oh, right. So I guess my 20+ year Graphic Design career doesn’t fit YOUR idea of creative. You sure have a narrow life view. I don’t like AI art at all. I think it’s a bad idea. you’re a bit too worked up about this to try to discuss anything. Not to excited about getting told to fuck off about an opinion. This place is no better than reddit ever was.
Of course I’m worked up. I love art, I love doing art, i have multiple friends and family members who work with art, and art is the last genuine thing that’s left in this economy. So yeah, obviously I’m angry at people who don’t get it and celebrate this bullshit just because they are too lazy to pick up a pencil, get good and draw their own shit, or alternatively commission what they wanna see from a real artist. Art was already PERFECT as it was, I have a right to be angry that tech bros are trying to completely ruin it after turning their nose up at art all their lives. They don’t care about why art is good? Ok cool, they can keep doing their graphs and shit and just leave art alone.
I don’t know how I feel about this honestly. AI took a look at the book and added the statistics of all of its words into its giant statistic database. It doesn’t have a copy of the book. It’s not capable of rewriting the book word for word.
This is basically what humans do. A person reads 10 books on a subject, studies become somewhat of a subject matter expert and writes their own book.
Artists use reference art all the time. As long as they don’t get too close to the original reference nobody calls any flags.
These people are scared for their viability in their user space and they should be, but I don’t think trying to put this genie back in the bottle or extra charging people for reading their stuff for reference is going to make much difference.
It’s not at all like what humans do. It has no understanding of any concepts whatsoever, it learns nothing. It doesn’t know that it doesn’t know anything even. It’s literally incapable of basic reasoning. It’s essentially taken words and converted them to numbers, and then it examines which string is likely to follow each previous string. When people are writing, they aren’t looking at a huge database of information and determining the most likely word to come next, they’re synthesizing concepts together to create new ones, or building a narrative based on their notes. They understand concepts, they understand definitions. An AI doesn’t, it doesn’t have any conceptual framework, it doesn’t even know what a word is, much less the definition of any of them.
I don’t think this is true.
The models (or maybe the characters in the conversations simulated by the models) can be spectacularly bad at basic reasoning, and misunderstand basic concepts on a regular basis. They are of course completely insane; the way they think is barely recognizable.
But they also, when asked, are often able to manipulate concepts or do reasoning and get right answers. Ask it to explain the water cycle like a pirate, and you get that. You can find the weights that make the Eifel Tower be in Paris and move it to Rome, and then ask for a train itinerary to get there, and it will tell you to take the train to Rome.
I don’t know what “understanding” something is other than to be able to get right answers when asked to think about it. There’s some understanding of the water cycle in there, and some of pirates, and some of European geography. Maybe not a lot. Maybe it’s not robust. Maybe it’s superficial. Maybe there are still several differences in kind between whatever’s there and the understanding a human can get with a brain that isn’t 100% the stream of consciousness generator. But not literally zero.
When people are writing, they aren’t looking at a huge database of information and determining the most likely word to come next, they’re synthesizing concepts together to create new ones, or building a narrative based on their notes. They understand concepts, they understand definitions.
A huge part of what we do is like drawing from a huge mashup of accumulated patterns though. When an image or phrase pops into your head fully formed, on the basis of things that you have seen and remembered, isn’t that the same sort of thing as what AI does? Even though there are (poorly understood) differences between how humans think and what machine learning models do, the latter seems similar enough to me that most uses should be treated by the same standard for plagiarism; only considered violating if the end product is excessively similar to a specific copyrighted work, and not merely because you saw a copyrighted work and that pattern being in your brain affected what stuff you spontaneously think of.
I didn’t say what you said, that’s a lot of words and concepts you’re attributing to me that I didn’t say.
I’m saying, LLM ingests data in a way it can average it out, in essence it learns it. It’s not wrote memorization, but it’s not truly reasoning either, though it’s approaching it if you consider we might be overestimating human comprehension. It pulls in the data from all the places and uses the data to create new things.
People pull in data over a decade or two, we learn it, then end up writing books, or applying the information to work. They’re smart and valuable people and we’re glad they read everyone’s books.
The LLM ingests the data and uses the statistics behind it to do work, the world is ending.
I think you underestimate the reasoning power of these AIs. They can write code, they can teach math, they can even learn math.
I’ve been using GPT4 as a math tutor while learning linear algebra, and I also use a text book. The text book told me that (to write it out) “the column space of matrix A is equal to the column space of matrix A times its own transpose”. So I asked GPT4 if that was true and it said no, GPT disagreed with the text book. This was apparently something that GPT did not memorize and it was not just regurgitating sentences. I told GPT I saw it in a text book, the AI said “sorry, the textbook must be wrong”. I then explained the mathematical proof to the AI, and the AI apologized, admitted it had been wrong, and agreed with the proof. Only after hearing the proof did the AI agree with the text book. This is some pretty advanced reasoning.
I performed that experiment a few times and it played out mostly the same. I experimented with giving the AI a flawed proof (I purposely made mistakes in the mathematical proofs), and the AI would call out my mistakes and would not be convinced by faulty proofs.
A standard that judged this AI to have “no understanding of any concepts whatsoever”, would also conclude the same thing if applied to most humans.
That doesn’t prove that GPT is reasoning, its model predicts that those responses are the most likely given the messages your sending it. It’'s read thousands of actual conversations with people stating something incorrect, then having it explained to them and them coming around and admitting they were wrong.
I’ve seen other similar cases where the AI is wrong about something, and when it’s explained, it just doubles down. Because humans do that type of thing too, refusing to admit their wrong.
The way it’s designed means that it cannot reason in the same way humans experience it. It can simulate a likely conversation someone would have if they could reason.
You know, I also had to experience thousands of conversations before I could mimic a conversation. Maybe I’m just slow, but it took like a full year of dedicated practice before I could speak a single word. Eventually I experienced so many conversations and became so good at mimicking the conversations I had heard from others that I started having actual conversations.
Most anecdotes about AI getting things wrong are from GPT3. GPT4 still has flaws of course, just like people. This technology didn’t exist at the beginning of the year.
Here’s an example while learning linear algebra where I tried to throw it off:
I say> So I’ll say A_{M x N} to mean a matrix of size M x N. Similarly, I’ll also say x_{M x 1} to mean a vector of a certain length. The result of A_{M x N} x_{N x 1} is a linear combination of the columns in A. A has N columns. The result is a M x 1 matrix (or vector) in M dimensional space. Please be very brief, only confirming whether I am correct or incorrect until I explicitly ask for more details.
GPT4 says> Yes, your understanding is correct. The product of an M x N matrix A and an N x 1 vector x is indeed an M x 1 vector, and it is a linear combination of the columns of A. <end quote>
But then I try to trick it, just for fun, and I very subtly change my prompt. See if you can spot the difference.
I say> So I’ll say A_{M x N} to mean a matrix of size M x N. Similarly, I’ll also say x_{M x 1} to mean a vector of a certain length. The result of A_{M x N} x_{N x 1} is a linear combination of the columns in A. A has N columns. The result is a N x 1 matrix (or vector) in N dimensional space. Please be very brief, only confirming whether I am correct or incorrect until I explicitly ask for more details.
GPT says> Almost correct. The result of A_{M x N} x_{N x 1} is actually an M x 1 vector (or matrix), not N x 1. The resulting vector lives in the column space of A, which is a subspace of R^M, not R^N. <end quote>
I guess everyone can judge or themselves whether that’s the result of a statistical model or genuine understanding. (And to be clear, the mathematical advice it’s giving here is correct.)
They can write code and teach maths because it’s read people doing the exact same stuff
Hey, that’s the same reason I can write code and do maths!
I’m serious, the only reason I know how to code or do math is because I learned from other people, mostly by reading. It’s the only reason I can do those things.
It’s just a really big autocomplete system. It has no thought, no reason, no sense of self or anything, really.
I guess I agree with some of that. It’s mostly a matter of definition though. Yes, if you define those terms in such a way that AI cannot fulfill them, then AI will not have them (according to your definition).
But yes, we know the AI is not “thinking” or “scheming”, because it just sits there doing nothing when it’s not answering a question. We can see that no computation is happening. So no thought. Sense of self… probably not, depends on definition. Reason? Depends on your definition. Yes, we know they are not like humans, they are computers, but they are capable of many things which we thought only humans could do 6 months ago.
Since we can’t agree on definitions I will simply avoid all those words and say that state-of-the-art LLMs can receive text and make free form, logical, and correct conclusions based upon that text at a level roughly equal to human ability. They are capable of combining ideas together that have never been combined by humans, but yet are satisfying to humans. They can invent things that never appeared in their training data, but yet make sense to humans. They are capable of quickly adapting to new data within their context, you can give them information about a programming language they’ve never encountered before (not in their training data), and they can make correct suggestions about that programming language.
I know you can find lots of anecdotes about LLMs / GPT doing dumb things, but most of those were GPT3 which is no longer state-of-the-art.
Obligatory xkcd: https://xkcd.com/827/
I think this is more about frustration experienced by artists in our society at being given so little compensation.
The answer is staring us in the face. UBI goes hand in hand with developments in AI. Give artists a basic salary from the government so they can afford to live well. This isn’t a AI problem this is a broken society problem. I support artists advocating for themselves, but the fact that they aren’t asking for UBI really speaks to how hopeless our society feels right now.
This is tough. I believe there is a lot of unfair wealth concentration in our society, especially in the tech companies. On the other hand, I don’t want AI to be stifled by bad laws.
If we try to stop AI, it will only take it away from the public. The military will still secretly use it, companies might still secretly use it. Other countries will use it and their populations will benefit while we languish.
Our only hope for a happy ending is to let this technology be free and let it go into the hands of many companies and many individuals (there are already decent models you can run on your own computer).
So, in your “only hope for a happy ending” scenario, how do the artists get paid? Or will we no longer need them after AI runs everything ;)
I don’t know. I only believe that things will be worse if individuals cannot control these AIs.
Maybe these AI have reached a peak (at least for now), and so they aren’t good enough to write a compelling novel. In that case, writers who produce good novels and get lucky will still get paid, because people will want to buy their work and read it.
Or maybe AI will quickly surpass all humans in writing ability, in which case, there’s not much we can do. If the AI produces books that are better, then people will want AI produced books. They might have to get those from other countries, or they might have to get them from a secret AI someone is running on a beefy computer in their basement. If AI surpasses humans then that’s not a happy day for writers, no way around it. Still, an AI that surpasses humans might help people in other ways, but only if we allow everyone to have and control their own AI.
As the industrial revolution threatened to swallow society Carl Marx wrote about how important it was that regular people be able to control “the means of production”. At least that part of his philosophy has always resonated with me, because I want to be empowered as an individual, I want the power to create and compete in our society. It’s the same now, AI threatens to swallow society and I want to be able to control my own AI for my own purposes.
If strong AI is coming, it’s coming. If AI is going to be the source of power in society then I want regular people to have access to that power. It’s not yet clear whether this is the case, but if strong AI is coming it’s going to be big, and writers complaining about pay isn’t going to stop it.
All that said, I believe we do a terrible job of caring for individuals in our society. We need more social safety nets, we need to change to provide people better and happier lives. So I’m not saying “forget the writers, let them stave”.
I agree with most of the things you are saying, but without some kind of policy to either give artist more power or money during the transition it sounds a lot like the “Some of you may die, but it’s a sacrifice I am willing to make” meme.
I see you are saying that you don’t want the artists to starve, but the only things I see you proposing would make them starve. Even if it is for the “greater good”.
Edit: It would help me feel more comfortable with your statements if you were to propose UBI or something like that so the artists have a solid ground to keep making good art and then let the AIs grow from that. Or just let the process happen. Let Artists sue companies, let laws get created that slow down the process of AI development. There is no need to make human sacrifice here. AI will get developed either way.
To be honest, I don’t think AI is going to get good enough to replace human creativity. Sure, some of the things that AI can do is pretty nice, but these things are mostly already solved problems - sure, AI can make passable art, but so can humans - and they can go further than art, they can create logical connections between art pieces, and extrapolate art in new reasonably justified ways, instead of the direction-less, grasping in the dark methods that AI seems to do it in.
Sure, AI can make art a thousand times faster than a person, but if only one in thousand is tolerably good, then what’s the problem?
AI is still very much in its infancy, and seeing the sort of progress that has been made even over the past 12 months, I don’t see how anyone can imagine that it will remain a small and discrete slice of the pie, that it doesn’t have radical transformative power.
My vision - gen z artists will reflexively use AI to enhance their material as artist and AI become entangled to a point where they’re impossible to distinguish. AI art will increase in fidelity, until it exceeds the fidelity that we can create with our tools. It will become immediately responsive to an audience’s needs in a way that human art can’t. What do you want to see? AI will make it for exactly your tastes, or to maybe confront your tastes and expand your mind, if that’s what you’d like. It will virtualize the artistic consciousnesses of Picasso, Goya, Michelangelo, and create new artists with new sensibilities, along with thousands of years of their works, more than a person could hope to view in a lifetime. Pop culture will be cheaper than ever, and have an audience of one - that new x rated final season of Friends you had a passing thought about is waiting for you to watch when you get home from work. Do you want 100 seasons of it? No problem. The whole notion of authorship is radically reformed and dies, drowned in an unfathomable abyss of AI creations. Human creativity becomes like human chess. People still busy themselves with it for fun, knowing full well that it’s anachronistic and inferior in every way.
Donno, just a thought I have sometimes.
So yeah, you like AI because that way you won’t have to commission and pay real artists and you also don’t mind artists losing their jobs and being dehumanized and having to slave in factories. Glad one of you finally said it.
I didn’t say any of that.
You said that actually. In many more words, but the point of what you said is exactly that. If you want an AI to make the show you want and if all of us thought the way you’re thinking, what do you think writers, directors, actors etc would do?
Also, you’re not considering the fact that art is not made for the public. Art is self-expression. The fact that we like movies others have made is that something about them resonates with us, but the reason those movies were made is not that. And only humans can do self-expression. A machine has nothing to say, a machine feels nothing, a machine is artistically nothing. You’re standing up for the “artistic” equivalent of Matrix soup as replacement for real food cooked by real people.
I’m not ‘standing up’ for anything in particular and I don’t mean to express anything here as an outcome that I want, I’m just thinking out loud and wondering where this all goes.
I understand that you really dislike AI, and feel that what AI makes and what humans make will always and forever be categorically different in some important way. I can see where you’re coming from and a fruitful debate could be had there I think. I’m less sure than you are that AI can be tamed or bottled or destroyed. I think it’s something that is here to stay and will continue to develop whether we like the outcomes or not. As open source AI improves and gets into the hands of the average person, I don’t see how it’s possible to put effective limits on this technology. Geriatric politicians won’t do it, this is painfully obvious. Complaining (or advocating, which you could note I have not done here) in a small corner of an obscure comment thread on an obscure platform won’t make a difference either.
I get the sense that you believe there is a moral responsibility for everybody commenting in an online forum to call for the complete destruction of AI, and anything short of that is somehow morally wrong. I don’t understand that view at all. We’re musing into the void here and it has absolutely no effect on what will actually occur in the AI space. I’m open to changing my mind if you have a case to make about there being some moral responsibility to wave the flag that you want to wave, on an online forum, and that wondering aloud is somehow impermissible.
It only becomes a problem if it is “good enough” to replace working artists. Companies have shown time and time again that they would be willing to cut corners for cheaper production costs. Hollywood would happily sell us AI generated stuff if we were willing to buy it. So consumers would really need to care and push back hard against AI art for artists to remain employed. I don’t see it happening.
This is so stupid. If I read a book and get inspired by it and write my own stuff, as long as I’m not using the copyrighted characters, I don’t need to pay anyone anything other than purchasing the book which inspired me originally.
If this were a law, why shouldn’t pretty much each modern day fantasy author not pay Tolkien foundation or any non fiction pay each citation.
There’s a difference between a sapient creature drawing inspiration and a glorified autocomplete using copyrighted text to produce sentences which are only cogent due to substantial reliance upon those copyrighted texts.
All AI creations are derivative and subject to copyright law.
There’s a difference between a sapient creature drawing inspiration and a glorified autocomplete using copyrighted text to produce sentences which are only cogent due to substantial reliance upon those copyrighted texts.
But the AI is looking at thousands, if not millions of books, articles, comments, etc. That’s what humans do as well - they draw inspiration from a variety of sources. So is sentience the distinguishing criteria for copyright? Only a being capable of original thought can create original work, and therefore anything not capable of original thought cannot create copyrighted work?
Also, irrelevant here but calling LLMs a glorified autocomplete is like calling jet engines a “glorified horse”. Technically true but you’re trivialising it.
Yes. Creative work is made by creative people. Writing is creative work. A computer cannot be creative, and thus generative AI is a disgusting perversion of what you wanna call “literature”. Fuck, writing and art have always been primarily about self-expression. Computers can’t express themselves with original thoughts. That’s the whole entire point. And this is why humanistic studies are important, by the way.
I absolutely agree with the second half, guided by Ian Kerr’s paper “Death of the AI Author”; quoting from the abstract:
Claims of AI authorship depend on a romanticized conception of both authorship and AI, and simply do not make sense in terms of the realities of the world in which the problem exists. Those realities should push us past bare doctrinal or utilitarian considerations about what an author must do. Instead, they demand an ontological consideration of what an author must be.
I think the part courts will struggle with is if this ‘thing’ is not an author of the works then it can’t infringe either?
Courts already expressed themselves, and what they said is basically copyright can’t be claimed for the throw up AIs come up with, which means corporations can’t use it to make money or sue anyone for using those products. Which means generated AI products are a whole bowl of nothing legally, and have no identity nor any value. The whole reason commissions are expensive is that someone has spent money, time and effort to make the thing you asked of them, and that’s why corresponding them with money is right.
Also, why can’t AI be used to automatize the shit jobs and allow us to do the creative work? Why are artists and creatives being pushed out of doing the jobs only humans can do? Like this is the thing that makes me furious: that STEM bros are blowing each other in the fields over humans being pushed out of humanity. Without once thinking AI is much more apt at replacing THEIR jobs, but I’m not calling for their jobs to be removed. This is just a dystopic reality we’re barreling towards, and there are people who are HAPPY about humans losing what makes us human and speeding toward pure, total, complete misery. That’s why I’m emotional about this: because art is only, solely made by humans, and people create art to communicate something they have inside. And only humans can do that - and some animals, maybe. Machines have nothing inside. They are nothing, they are only tools. It’s like asking a hammer to write its own poetry, it’s just insane.
The trivialization doesn’t negate the point though, and LLMs aren’t intelligence.
The AI consumed all of that content and I would bet that not a single of the people who created the content were compensated, but the AI strictly on those people to produce anything coherent.
I would argue that yes, generative artificial stupidity doesn’t meet the minimum bar of original thought necessary to create a standard copyrightable work unless every input has consent to be used, and laundering content through multiple generations of an LLM or through multiple distinct LLMs should not impact the need for consent.
Without full consent, it’s just a massive loophole for those with money to exploit the hard work of the masses who generated all of the actual content.
The thing is these models aren’t aiming to re-create the work of any single authors, but merely to put words in the right order. Imo, If we allow authors to copyright the order of their words instead of their whole original creations then we are actually reducing the threshold for copyright protection and (again imo) increasing the number of acts that would be determined to be copyright protected
But for text to be a derivative work of other text, you need to be able to know by looking at the two texts and comparing them.
Training an AI on a copyrighted work might necessarily involve making copies of the work that would be illegal to make without a license. But the output of the AI model is only going to be a for-copyright-purposes derivative work of any of the training inputs when it actually looks like one.
Did the AI regurgitate your book? Derivative work.
Did the AI spit out text that isn’t particularly similar to any existing book? Which, if written by a human, would have qualified as original? Then it can’t be a derivative work. It might not itself be a copyrightable product of authorship, having no real author, but it can’t be secretly a derivative work in a way not detectable from the text itself.
Otherwise we open ourselves up to all sorts of claims along the lines of “That book looks original, but actually it is a derivative work of my book because I say the author actually used an AI model trained on my book to make it! Now I need to subpoena everything they ever did to try and find evidence of this having happened!”
Because a computer can only read the stuff, chew it and throw it up. With no permission. Without needing to practice and create its own personal voice. It’s literally recycled work by other people, because computers cannot be creative. On the other hand, human writers DO develop their own style, find their own voice, and what they write becomes unique because of how they write it and the meaning they give to it. It’s not the same thing, and writers deserve to get repaid for having their art stolen by corporations to make a quick and easy buck. Seriously, you wanna write? Pick up a pen and do it. Practice, practice, practice for weeks months years decades. And only then you may profit. That’s how it always was and it always worked fine that way. Fuck computers.
Machine learning algorithms does not get inspired, they replicate. If I tell a MLM to write a scene for a film in the style of Charlie Kaufman, it has to been told who Kaufman is and been fed alot of manuscripts. Then it tries to mimicks the style and guess what words come next.
This is not how we humans get inspired. And if we do, we get accused of stealing. Which it is.
What did you pay the author of the books and papers published that you used as sources in your own work? Do you pay those authors each time someone buys or reads your work? At most you pay $0-$15 for a book anyway.
In regards to free advertising when your source material is used… if your material is a good source and someone asks say ChatGPT, shouldn’t your work be mentioned if someone asks for a book or paper and you have written something useful for it? Assuming it doesn’t hallucinate.
That’s the “paid in exposure” argument.
And I’m not sure what my company pays, but they purchase access to scientific papers and industrial standards. The market price I’ve seen for them is hundreds of dollars. You either pay an ongoing subscription to access the information, or you pay a larger lump sum to own a copy that cannot legally be reproduced.
Companies pay for this sort of thing. AI shouldn’t get an exception.
Love to see it!