logoalt Hacker News

golem14yesterday at 10:57 PM1 replyview on HN

Care to explain? I see that statement in this thread, but I am not sure where this is grounded in fact.

This is very interesting, because there must be a line here that AI is crossing, and the line is not clearly determined yet.

Is linting code crossing the line?

Is re-factoring code with automated tools like bicycle repair man crossing the line ?

Is AI doing a code review and suggesting the code crossing the line ?

Is writing code with a specific prompt and sample code crossing the line?

Is producing a high level spec and let the AI design details and code the whole thing crossing the line ?

So, where exactly is this line ?

The next interesting question is how this could even be enforced. It's going to be hard to prove AI use when using strictly local models. Maybe they could embed some watermark like thing, but I am not sure this can't be circumvented.

Would really like to see some legal opinions on this ( unlikely to happen :)

The best I found is here: https://copyrightlately.com/thaler-is-dead-ai-copyright-ques...


Replies

heavyset_gotoday at 12:36 AM

Here's what a Red Hat/IBM IP lawyer said about the chardet situation: https://github.com/chardet/chardet/issues/334#issuecomment-4...

Here's what the US Copyright Office says: https://newsroom.loc.gov/news/copyright-office-releases-part...