| ▲ | kshri24 3 hours ago | ||||||||||||||||||||||||||||
> The ownership void: If the code is truly a “new” work created by a machine, it might technically be in the public domain the moment it’s generated, rendering the MIT license moot. How would that work? We still have no legal conclusion on whether AI model generated code, that is trained on all publicly available source (irrespective of type of license), is legal or not. IANAL but IMHO it is totally illegal as no permission was sought from authors of source code the models were trained on. So there is no way to just release the code created by a machine into public domain without knowing how the model was inspired to come up with the generated code in the first place. Pretty sure it would be considered in the scope of "reverse engineering" and that is not specific only to humans. You can extend it to machines as well. EDIT: I would go so far as to say the most restrictive license that the model is trained on should be applied to all model generated code. And a licensing model with original authors (all Github users who contributed code in some form) should be setup to be reimbursed by AI companies. In other words, a % of profits must flow back to community as a whole every time code-related tokens are generated. Even if everyone receives pennies it doesn't matter. That is fair. Also should extend to artists whose art was used for training. | |||||||||||||||||||||||||||||
| ▲ | kouteiheika 3 hours ago | parent | next [-] | ||||||||||||||||||||||||||||
> I would go so far as to say the most restrictive license that the model is trained on should be applied to all model generated code. That license is called "All Rights Reserved", in which case you wouldn't be able to legally use the output for anything. There are research models out there which are trained on only permissively licensed data (i.e. no "All Rights Reserved" data), but they're, colloquially speaking, dumb as bricks when compared to state-of-art. But I guess the funniest consequence of the "model outputs are a derivative work of their training data" would be that it'd essentially wipe out (or at very least force a revert to a pre-AI era commit) every open source project which may have included any AI-generated or AI-assisted code, which currently pretty much includes every major open source project out there. And it would also make it impossible to legally train any new models whose training data isn't strictly pre-AI, since otherwise you wouldn't know whether your training data is contaminated or not. | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||
| ▲ | adrianN 3 hours ago | parent | prev | next [-] | ||||||||||||||||||||||||||||
We‘ll have to wait until the technology progresses sufficiently that AI cuts into Disney’s profit. | |||||||||||||||||||||||||||||
| ▲ | thedevilslawyer 3 hours ago | parent | prev [-] | ||||||||||||||||||||||||||||
That's unpractical enough that you might as well wish for UBI and world peace rather than this. | |||||||||||||||||||||||||||||
| |||||||||||||||||||||||||||||