Hacker Newsnew | past | comments | ask | show | jobs | submit | luke5441's commentslogin

GPL works via copyright. Since AI companies claim fair use no copyright applies. There is no fixing this. The only option is not to publish.

There are non-US jurisdictions where you have some options, but since most of them are trained in the US that won't help much.


> Since AI companies claim fair use no copyright applies. There is no fixing this.

They can claim whatever they want. You can still try to stop it via lawsuits and make them claim it in court. Granted, I believe there's already been some jurisdictions that have sided with fair use in those particular cases.


Laws can be changed. This is right now a trillion dollar industry, perhaps later it could even become a billion dollar industry. Either way, it's very important.

Strict copyright enforcement is a competitive disadvantage. Western countries lobbied for copyright enforcement in the 20th century because it was beneficial. Now the tables have turned, don't hold your breath for copyright enforcement against the wishes of the markets. We are all China now.


Yes, I think Japan added an AI friendly copyright law. If there were problems in the US, they'd just move training there.

Moving training won't help them if their paying customers are in jurisdictions which do respect copyright as written and intended.

OPs idea is about having a new GPL like license with a "may not be used for LLM training" clause.

That the LLM itself is not allowed to produce copyrighted work (e.g. just copies of works or too structurally similar) without using a license for that work is something that is probably currently law. They are working around this via content filters. They probably also have checks during/after training that it does not reproduce work that is too similar. There are law suits about this pending if I remember correctly e.g. with the New York Times.


The issue is that everyone is focusing on verbatim (or "too similar") reproduction.

LLMs themselves are compressed models of the training data. The trick is the compression is highly lossy by being able to detect higher-order patterns instead of fucusing on the first-order input tokens (or bytes). If you look at how, for example, any of the Lempel-Ziv algorithms work, they also contain patterns from the input and they also predict the next token (usually byte in their case), except they do it with 100% probability because they are lossless.

So copyright should absolutely apply to the models themselves and if trained on AGPL code, the models have to follow the AGPL license and I have the right to see their "source" by just being their user.

And if you decompress a file from a copyrighted archive, the file is obviously copyrighted. Even if you decompress only a part. What LLMs do is another trick - by being lossy, they decompress probabilistically based on all the training inputs - without seeing the internals, nobody can prove how much their particular work contributed to the particular output.

But it is all mechanical transformation of input data, just like synonym replacement, just more sophisticated, and the same rules regarding plagiarism and copyright infringement should apply.

---

Back to what you said - the LLM companies use fancy language like "artificial intelligence" to distract from this so they can they use more fancy language to claim copyright does not apply. And in that case, no license would help because any such license fundamentally depends on copyright law, which as they claim does not apply.

That's the issue with LLMs - if they get their way, there's no way to opt out. If there was, AGPL would already be sufficient.


I agree with your view. One just has to go into courts and somehow get the judges to agree as well.

An open question would be if there is some degree of "loss" where copyright no longer applies. There is probably case law about this in different jurisdictions w.r.t. image previews or something.


I don't think copyright should be binary or should work the way it does not. It's just the only tool we have now.

There should be a system which protects all work (intellectual and physical) and makes sure the people doing it get rewarded according to the amount of work and skill level. This is a radical idea and not fully compatible with capitalism as implemented today. I have a lot on my to-read list and I don't think I am the first to come up with this but I haven't found anyone else describing it, yet.

And maybe it's broken by some degenerate case and goes tits up like communism always did. But AFAICT, it's a third option somewhere in between, taking the good parts of each.

For now, I just wanna find ways to stop people already much richer than me from profiting from my work without any kind of compensation for me. I want inequality to stop worsening but OTOH, in the past, large social change usually happened when things got so bad people rejected the status quo and went to the streets, whether with empty hands or not. And that feels like where we're headed and I don't know whether I should be exited or worried.


I recall a basics of law class saying that in some countries (e.g. Czech Republic), open source contributors have the right to small compensation if their work is used to a large financial benefit.

At some point, I'll have to look it up because if that's right, the billionaires and wannabe-trillionaires owe me a shitton of money.


With which version did they enable/implement the CPU mitigations (Spectre etc.)?


Is it 1 million USD plus statehood? Or 1 million per head to become a US colony?

At the very least don't complain about it publicly. Using diplomatic channels I think things like this can quietly go away. But if you make it an issue in the public court of opinion people usually support their own democratically elected government compared to the increasingly hostile foreign regime.

They can use an European alternative like bunny.net. It's cheaper anyway.

I don't get how such idiotic people get into those kinds of positions.

I'd guess you can also easily track down star link terminals via drones (fly high up until you find a signal, then go lower while keeping the signal active)

Possibly a difficult ask for a regime under fairly extreme pressure (but not impossible, for sure, and they'd only need to do it once per system).

It's not about a world view. It is about watching a video and seeing what is happening.

Unless you mean "world view" is my knowledge about how cars and guns work.


I watched the video and I saw Renee attempt to hit the ICE agent with her car. After doing that, lethal force is unfortunately on the table. The officer doesn’t have time after that to wait and see if she is going to try to kill him again or not.

How is she going to try to kill him again? Drive away then come back to ram?

I'm getting really depressed here that people watch those videos and actually think this was not murder.

It's like they live in a different reality or something.


My baseline (as an outside observer) is still that they are too stupid and uncoordinated to do a facist take-over. But that a significant possibility even exists that there will be a facist take-over is awful.

(The main ingredient for a facist take-over is control over the military and media and he is not creating many loyalists in either)


Mussolini was not the brightest bulb in the tanning bed, but it worked for him.

He served in the military, though. Idk if he obviously was not loyal to anyone like Trump is, I'd guess not.

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: