No disclosure for LLM-generated patch?
No disclosure for LLM-generated patch?
Posted Jun 27, 2025 10:51 UTC (Fri) by mb (subscriber, #50428)In reply to: No disclosure for LLM-generated patch? by Funcan
Parent article: Supporting kernel development with large language models
If you as a human learn from proprietary code and then write Open Source with that knowledge, it's not copying unless you actually copy code sections. Same goes for LLMs. If it produces a copy, then it copied. Otherwise it didn't.
Posted Jun 27, 2025 11:47 UTC (Fri)
by laarmen (subscriber, #63948)
[Link] (1 responses)
Posted Jun 27, 2025 12:57 UTC (Fri)
by mb (subscriber, #50428)
[Link]
It's in no way required to avoid copyright problems.
And you can also use that concept with LLMs, if you want.
Posted Jul 1, 2025 9:51 UTC (Tue)
by cyphar (subscriber, #110703)
[Link]
You could just as easily argue that LLMs produce something equivalent to a generative collage of all of their training data, which (given the current case law on programs and copyright) would mean that the copyright status of the training data would be transferred to the collage. You would thus need to make an argument for a fair use exemption for the output, which your example would not pass muster.
However, this is not the only issue at play here -- to submit code to Linux you need to sign the DCO, which the commit author did with their Signed-off-by line. However, none of the sections of the DCO can be applied to LLM-produced code, and so the Signed-off-by is invalid regardless of the legal questions about copyright and LLM code.
No disclosure for LLM-generated patch?
No disclosure for LLM-generated patch?
Just don't copy and then you are safe.
Learning is not copying.
Just feed the output from one LLM into the input of another LLM and you basically get the same thing as with two human clean-room teams.
No disclosure for LLM-generated patch?