▲ | chii 7 months ago | ||||||||||||||||
> If an LLM does it how should we deal with that? why not deal with it the same way as humans have been dealt with in the past? If you copied an art piece using photoshop, you would've violated copyright. Photoshop (and adobe) itself never committed copyright violations. Somehow, if you swap photoshop with openAI and chatGPT, then people claim that the actual application itself is a copyright violation. | |||||||||||||||||
▲ | dijksterhuis 7 months ago | parent [-] | ||||||||||||||||
this isn’t the same. > If you copied an art piece using photoshop, you would've violated copyright. Photoshop (and adobe) itself never committed copyright violations. the COPYing is happening on your local machine with non-cloud versions of Photoshop. you are making a copy, using a tool, and then distributing that copy. in music royalty terms, the making a copy is the Mechanical right, while distributing the copy is the Performing right. and you are liable in this case. > Somehow, if you swap photoshop with openAI and chatGPT, then people claim that the actual application itself is a copyright violation OpenAI make a copy of the original works to create training data. when the original works are reproduced verbatim (memorisation in LLMs is a thing), then that is the copyrighted work being distributed. mechanical and performing rights, again. but the twist is that ChatGPT does the copying on their servers and delivers it to your device. they are creating a new copy and distributing that copy. which makes them liable. — you are right that “ChatGPT” is just a tool. however, the interesting legal grey area with this is — are ChatGPT model weights an encoded copy of the copyrighted works? that’s where the conversation about the tool itself being a copyright violation comes in. photoshop provides no mechanism to recite The Art Of War out of the box. an LLM could be trained to do so (like, it’s a hypothetical example but hopefully you get the point). | |||||||||||||||||
|