Skip to content

Commit ab50154

Browse files
authored
Apply suggestion from @lwasser
1 parent 6dd3a9a commit ab50154

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

_posts/2025-11-18-generative-ai-peer-review.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -88,7 +88,7 @@ The problem? LLMs sometimes spit out near-exact copies of that training data, bu
8888
Why this matters:
8989

9090
* Using LLM output verbatim could violate the original code's license
91-
* You might accidentally commit plagiarism or copyright infringement by using that output verbatim in your code
91+
* License conflicts can occur if your package's license (e.g., MIT) is incompatible with code patterns that the LLM learned on such as code licensed as GPL or Apache-2.0.
9292
* Due diligence is nearly impossible since you can't trace what the LLM "learned from" (most LLM's are black boxes)
9393

9494
When licenses clash, it gets messy. Say your package uses an MIT license (common in scientific Python), but an LLM outputs Apache-2.0 or GPL code—those licenses aren't compatible. You can't just add attribution to fix it. Technically, you'd have to delete everything and rewrite it from scratch to comply with the licensing requirements.

0 commit comments

Comments
 (0)