A lot of people would be very pleased if this leads to Zuckerberg getting even the statutory minimum damages ($750?) on each infringement.
The previous infringement case with Anthropic said that while training an AI was transformative and not itself an infringement, pirating works for that purpose still was definitely infringement all by itself. The settlement was $1.5bn, so close to $3k for each of the 500k they pirated, so if Zuckerberg pirated "millions" (plural) it is quite plausible his settlement could be $6bn.
I had to block meta's ASN on my personal cgit server a few weeks ago because they were ignoring robots.txt and torching it. Like hundreds of megabytes of access logs just from them, spread around different network blocks to clearly try and defeat IP based limiting. I couldn't believe it.
I take issue with the use of tense used in this framing. Its not 'infringed' its 'infringing' and to say that it happened is wrong, its happening and happening continuously in these models that are in use. To say a one time payment settles it is missing the whole scope of this theft.
Royalties are owed and continuously owed as these models are deployed and doing inference. How is it any different to paying a small pittance to someone every time a song is played?
Royalties for inference are unrealistic in a way that even royalties for training aren't.
The LLaMA models were released openly. Copies exist everywhere in the world. You aren't going to be able to charge someone for running `llama.cpp`; a court order ceases to have practical relevance at that point.
> a Meta spokesperson said, “AI is powering transformative innovations, productivity and creativity for individuals and companies, and courts have rightly found that training AI on copyrighted material can qualify as fair use. We will fight this lawsuit aggressively.”
> Authors have sued AI companies for copyright infringement before - and lost.
they'll litigate how meta acquired those materials to train. you can do whatever you want with a book after it's in your house. but how did it get there?
Just gonna say... Aaron Swartz faced years of prison time and ultimately decided to take his own life... for downloading scientific journal articles... to share freely with the world (aka not even profiting from it).
But a multi-billion dollar corporation downloading millions of copyrighted creative works so that they can reshape the entire labor market by training a new type of artificial intelligence model on that data set? Meh, sounds like Silicon Valley disruption, give the man a medal!
I'm gonna have to go dig up the link, but isn't there a guy that Nintendo basically has on indentured servitude for the rest of his life?
Ah, found it:
>In April 2023, a 54-year-old programmer named Gary Bowser was released from prison having served 14 months of a 40-month sentence. Good behaviour reduced time behind bars, but now his options are limited. For a while he was crashing on a friend’s couch in Toronto. The weekly physical therapy sessions, which he needs to ease chronic pain, were costing hundreds of dollars every week, and he didn’t have a job. And soon, he would need to start sending cheques to Nintendo. Bowser owes the makers of Super Mario $14.5m (£11.5m), and he’s probably going to spend the rest of his life paying it back.
I'm not even a tiny bit supportive, but there is precedent.
American executives have been pushing to criminalise copyright infringement for decades, and America has worked hard to pressure countries all round the world to do this as part of trade deals. There is, for example, a Brit serving an eleven year sentence right now *.
The non-strawman way to interpret the parent comment is that they want them to be treated the same as normal copyright violators. Jail is a common result of (criminal) copyright prosecution, with 44% of convicted offenders being imprisoned, averaging 25 months [0].
Now, I personally find the idea of imprisoning people for copyright offenses horrific, but I don't think it's remotely insane that someone else might come to that conclusion, given that we broadly accept it as a society.
From [0]: "In fiscal year 2017, there were 80 copyright/trademark infringement offenders who accounted for 0.1% of all offenders sentenced under the guidelines." This is such a low number that I assume most prosecuted cases are settled without ever making it to sentencing, or alternatively copyright infringement is just hardly ever prosecuted criminally at all.
> I'm all for strong justice, but you want to imprison an executive for decades for copyright violations?
They stole the life's work of millions of people.
In less civilized times, they likely would have been drawn and quartered by strong horses, and had their limbs drug to the 4 corners of the continent as a warning to anyone else that would consider doing it again.
There aren't enough things an executive can go to jail for.
Fines don't do anything to deter bad behavior. Either:
* The company pays
* They pay and the company mysteriously increases next year's comp / grants a "loan" / etc
* D&O insurer pays
In all three cases the money comes out of the shareholders' hides. It provides zero personal deterrence. The payoff matrix, as seen by a sociopath, makes it rational to always defect against the common good.
The only punishment that can really focus attention is physical imprisonment in a facility they can't choose.
SOX did this for financial reporting and gee shucks it turned out executives can follow the law after all!
The human savant will remember where they read it and give you credit. It might lead more people to read your work, and ultimately you make money.
The AI won't even know where the page of text it's seeing came from, and people will avoid your book as they can just ask the AI. So you make less money. (Talking about specialized technical books here.)
There's a huge difference in scale. The human mind can only process a limited portion of all works available over a lifetime. Human learning is therefore naturally limited to small-scale reuse, which serves to keep it proportional.
A machine training on all copyrighted materials in the world for commercial purposes at an industrial scale makes it disproportionate.
It would hardly make a dent. And if you hired hundreds of savants, the knowledge would still be spread over hundreds of separate minds.
And even if we grant that those savants are also very skilled at creating "market substitutes" based on their training that are capable of competing with the original works, their maximum creative output would only be a relatively small number of new works, because they can only work at human speed.
No one is asking human savants about what they read 1 million times per day.
Suppose they did, and some guy was filling stadiums regularly to hear him recite an entire audio book. That would probably get the attention of someone's lawyers.
I don't think anyone is arguing that the consumption is illegal. It's the reproduction that is illegal.
Read a book, that's fine. Write a book, that's fine. Read a book and then write a book that is 99.9% the same as the book that you read and sell it for profit without a license from the original author, that's infringement.
No, if you read the article, the point is in the training, not the reproduction.
That's what all these lawsuits are about - it's the training not the reproduction. I already agreed in my first comment that the reproduction is off limits.
In this case, it appears that Meta torrented illegal copies of the work to do the training. Obviously that's bad. But conflating that with training itself doesn't follow.
If copyright law doesn't extend to the works being used for training, why should it extend to the model that is produced as a result? AI model creators have set up an ethical scenario where the right thing to do is ignore copyright laws when it comes to AI, which includes model use. It might never be legal, but it has become ethical to pirate models, distill them against ToS, etc.
We're not talking about rights, we're talking about illegal acts. If it's illegal for a machine to do it, how can it be ok for a human?
Just from a rational argumentation point of view. Clearly if a law is written saying as much, then sure. But there is no such copyright law like that yet.
The issue is certainly not so simple. But it seems to me, purely theoretically, that the rules don't necessarily have to be the same for living people and non-living machines.
Well - actually - it is pretty simple. For something to be illegal, there must be a law saying it's illegal. There are no laws distinguishing humans from machines in copyright law.
So is it a problem when humans produce and monetize competing works? My understanding is that there quite an industry in humans reading books and synthesizing their points. Cliff's Notes, for example.
I did some quick googling and most of cliffs notes guides are on public domain works so no problem there, they've also paid to license content, and also have been protected by fair use as parody
The problem is people at large companies creating these AI models, wanting the freedom to copy artists’ works when using it, but these large companies also want to keep copyright protection intact, for their regular business activities. They want to eat the cake and have it too. And they are arguing for essentially eliminating copyright for their specific purpose and convenience, when copyright has virtually never been loosened for the public’s convenience, even when the exceptions the public asks for are often minor and laudable. If these companies were to argue that copyright should be eliminated because of this new technology, I might not object. But now that they come and ask… no, they pretend to already have, a copyright exception for their specific use, I will happily turn around and use their own copyright maximalist arguments against them.
A lot of people would be very pleased if this leads to Zuckerberg getting even the statutory minimum damages ($750?) on each infringement.
The previous infringement case with Anthropic said that while training an AI was transformative and not itself an infringement, pirating works for that purpose still was definitely infringement all by itself. The settlement was $1.5bn, so close to $3k for each of the 500k they pirated, so if Zuckerberg pirated "millions" (plural) it is quite plausible his settlement could be $6bn.
Nothing will happen to him/Meta while DJT is president.
He bought the best protection around for breaking the law.
When you're a big Trump donor they let you do it.
For context, his net worth is ~$220 billion.
And meta's worth is much more than that. He's not personally paying.
I had to block meta's ASN on my personal cgit server a few weeks ago because they were ignoring robots.txt and torching it. Like hundreds of megabytes of access logs just from them, spread around different network blocks to clearly try and defeat IP based limiting. I couldn't believe it.
I take issue with the use of tense used in this framing. Its not 'infringed' its 'infringing' and to say that it happened is wrong, its happening and happening continuously in these models that are in use. To say a one time payment settles it is missing the whole scope of this theft.
Royalties are owed and continuously owed as these models are deployed and doing inference. How is it any different to paying a small pittance to someone every time a song is played?
Royalties for inference are unrealistic in a way that even royalties for training aren't.
The LLaMA models were released openly. Copies exist everywhere in the world. You aren't going to be able to charge someone for running `llama.cpp`; a court order ceases to have practical relevance at that point.
These models can provide citations so I don't see why they can't tick a royalty owed. I'm sure many here could help build this pipeline.
If you steal a book and read it, should you have to pay every time you use the knowledge gained or recall parts of it from memory?
If I perform a song in public then yes, I should pay the creator every time I play it. I fail to see the difference here.
No. People are not LLMs. And even if some argue that they are mechanically similar, they are legally distinct.
What if you steal a CD and then play it on your radio station each morning?
If I charged people for the privilege of listening to me recite relevant parts of the book to them for profit? Yes. Depending on the copyright.
So... "move fast and steal things"?
Always Has Been
> a Meta spokesperson said, “AI is powering transformative innovations, productivity and creativity for individuals and companies, and courts have rightly found that training AI on copyrighted material can qualify as fair use. We will fight this lawsuit aggressively.”
> Authors have sued AI companies for copyright infringement before - and lost.
So, basically nothing will come out of this
they'll litigate how meta acquired those materials to train. you can do whatever you want with a book after it's in your house. but how did it get there?
Rules for thee but not for me.
The behavior will continue until a consequence is imposed.
Just gonna say... Aaron Swartz faced years of prison time and ultimately decided to take his own life... for downloading scientific journal articles... to share freely with the world (aka not even profiting from it).
But a multi-billion dollar corporation downloading millions of copyrighted creative works so that they can reshape the entire labor market by training a new type of artificial intelligence model on that data set? Meh, sounds like Silicon Valley disruption, give the man a medal!
Had Aaron copied Snapchat 5 times the DOJ would've been fine with it all. His fault for not having the foresight
I would rather Zuckerberg do 6 months in jail and probation than fine Meta.
You aren't going to be able to make me anti-piracy just because some corpo benefits from it too.
I agree, time to start handing out real punishments, I think 6 months is way to small.
If this was you or me, we would be in prison for decades and have a fine in the millions. Time for these people to feel consequences.
As someone said, they will probably settle for around 6 billion, that is the same as say a $100 fine for us.
This comment could get its own DSM classification for how insane it is.
I'm all for strong justice, but you want to imprison an executive for decades for copyright violations?
I'm gonna have to go dig up the link, but isn't there a guy that Nintendo basically has on indentured servitude for the rest of his life?
Ah, found it:
>In April 2023, a 54-year-old programmer named Gary Bowser was released from prison having served 14 months of a 40-month sentence. Good behaviour reduced time behind bars, but now his options are limited. For a while he was crashing on a friend’s couch in Toronto. The weekly physical therapy sessions, which he needs to ease chronic pain, were costing hundreds of dollars every week, and he didn’t have a job. And soon, he would need to start sending cheques to Nintendo. Bowser owes the makers of Super Mario $14.5m (£11.5m), and he’s probably going to spend the rest of his life paying it back.
I'm not even a tiny bit supportive, but there is precedent.
https://www.theguardian.com/games/2024/feb/01/the-man-who-ow...
American executives have been pushing to criminalise copyright infringement for decades, and America has worked hard to pressure countries all round the world to do this as part of trade deals. There is, for example, a Brit serving an eleven year sentence right now *.
Why should Zuckerberg be exempt?
* https://www.bbc.co.uk/news/uk-65697595
Facebook isn't one of the companies that's been pushing for that.
The non-strawman way to interpret the parent comment is that they want them to be treated the same as normal copyright violators. Jail is a common result of (criminal) copyright prosecution, with 44% of convicted offenders being imprisoned, averaging 25 months [0].
Now, I personally find the idea of imprisoning people for copyright offenses horrific, but I don't think it's remotely insane that someone else might come to that conclusion, given that we broadly accept it as a society.
[0] https://www.ussc.gov/sites/default/files/pdf/research-and-pu...
From [0]: "In fiscal year 2017, there were 80 copyright/trademark infringement offenders who accounted for 0.1% of all offenders sentenced under the guidelines." This is such a low number that I assume most prosecuted cases are settled without ever making it to sentencing, or alternatively copyright infringement is just hardly ever prosecuted criminally at all.
> I'm all for strong justice, but you want to imprison an executive for decades for copyright violations?
They stole the life's work of millions of people.
In less civilized times, they likely would have been drawn and quartered by strong horses, and had their limbs drug to the 4 corners of the continent as a warning to anyone else that would consider doing it again.
I would prefer a harsher punishment, but I would begrudgingly accept throwing him in jail for decades.
I always heard that criminals should be thrown in jail, it's time we started doing it to the real criminals.
Is this controversial? Executives should be held liable, certainly moreso than just regular people sharing files.
There aren't enough things an executive can go to jail for.
Fines don't do anything to deter bad behavior. Either:
* The company pays
* They pay and the company mysteriously increases next year's comp / grants a "loan" / etc
* D&O insurer pays
In all three cases the money comes out of the shareholders' hides. It provides zero personal deterrence. The payoff matrix, as seen by a sociopath, makes it rational to always defect against the common good.
The only punishment that can really focus attention is physical imprisonment in a facility they can't choose.
SOX did this for financial reporting and gee shucks it turned out executives can follow the law after all!
I know people really hate AI training on their work - but is it really any different than a human reading it?
I know there's a complaint that AI can verbatim repeat that work. But so can human savants. No one is suing human savants for reading their books.
Producing copyrighted material, of course. Training on copyrighted material... I just don't see it.
EDIT: Making a perfectly valid point, but it's unpopular, so down I go.
I had to buy the copyrighted material before reading it... Meta apparently operates in a different legal system than me. That's my issue with it.
Yes, I have no objection to that part. It's the arguments that training itself is the problem.
Sarah Silverman as the most prominent example.
The human savant will remember where they read it and give you credit. It might lead more people to read your work, and ultimately you make money.
The AI won't even know where the page of text it's seeing came from, and people will avoid your book as they can just ask the AI. So you make less money. (Talking about specialized technical books here.)
Not necessarily.
There's a huge difference in scale. The human mind can only process a limited portion of all works available over a lifetime. Human learning is therefore naturally limited to small-scale reuse, which serves to keep it proportional.
A machine training on all copyrighted materials in the world for commercial purposes at an industrial scale makes it disproportionate.
I see that as a distinction - but does it make a difference?
If a company hired hundreds of savants, then it would be illegal for them to read books?
I don't follow.
It would hardly make a dent. And if you hired hundreds of savants, the knowledge would still be spread over hundreds of separate minds.
And even if we grant that those savants are also very skilled at creating "market substitutes" based on their training that are capable of competing with the original works, their maximum creative output would only be a relatively small number of new works, because they can only work at human speed.
No one is asking human savants about what they read 1 million times per day.
Suppose they did, and some guy was filling stadiums regularly to hear him recite an entire audio book. That would probably get the attention of someone's lawyers.
I don't see your point. The problem is producing the copyrighted work, not processing it beforehand.
If it's illegal for AIs it should be illegal for humans, too. Is that really what you're arguing? It should be illegal for savants to read books?
I don't think anyone is arguing that the consumption is illegal. It's the reproduction that is illegal.
Read a book, that's fine. Write a book, that's fine. Read a book and then write a book that is 99.9% the same as the book that you read and sell it for profit without a license from the original author, that's infringement.
No, if you read the article, the point is in the training, not the reproduction.
That's what all these lawsuits are about - it's the training not the reproduction. I already agreed in my first comment that the reproduction is off limits.
In this case, it appears that Meta torrented illegal copies of the work to do the training. Obviously that's bad. But conflating that with training itself doesn't follow.
If copyright law doesn't extend to the works being used for training, why should it extend to the model that is produced as a result? AI model creators have set up an ethical scenario where the right thing to do is ignore copyright laws when it comes to AI, which includes model use. It might never be legal, but it has become ethical to pirate models, distill them against ToS, etc.
I'm not sure I follow. Can you say it a different way?
It’s different.
Hm. I'm not sure I follow your logic.
Why should an AI have the same rights as a human?
How about then to grant AI all other rights, for example, to allow voting?(sarcasm)
We're not talking about rights, we're talking about illegal acts. If it's illegal for a machine to do it, how can it be ok for a human?
Just from a rational argumentation point of view. Clearly if a law is written saying as much, then sure. But there is no such copyright law like that yet.
The issue is certainly not so simple. But it seems to me, purely theoretically, that the rules don't necessarily have to be the same for living people and non-living machines.
Well - actually - it is pretty simple. For something to be illegal, there must be a law saying it's illegal. There are no laws distinguishing humans from machines in copyright law.
reading it after stealing it: gray area. producing & monetizing competing works devaluing the original is a problem
So is it a problem when humans produce and monetize competing works? My understanding is that there quite an industry in humans reading books and synthesizing their points. Cliff's Notes, for example.
I did some quick googling and most of cliffs notes guides are on public domain works so no problem there, they've also paid to license content, and also have been protected by fair use as parody
To Kill a Mockingbird, The Catcher in the Rye, Beloved, The Kite Runner, The Handmaid's Tale are all copyrighted works with a Cliff's Notes guide.
HN really loves the copyright lobby when it's against someone they hate, huh
The problem is people at large companies creating these AI models, wanting the freedom to copy artists’ works when using it, but these large companies also want to keep copyright protection intact, for their regular business activities. They want to eat the cake and have it too. And they are arguing for essentially eliminating copyright for their specific purpose and convenience, when copyright has virtually never been loosened for the public’s convenience, even when the exceptions the public asks for are often minor and laudable. If these companies were to argue that copyright should be eliminated because of this new technology, I might not object. But now that they come and ask… no, they pretend to already have, a copyright exception for their specific use, I will happily turn around and use their own copyright maximalist arguments against them.
(Copied from a comment of mine written more than three years ago: <https://news.ycombinator.com/item?id=33582047>)