this post was submitted on 10 May 2026
50 points (94.6% liked)

Ask Lemmy

39476 readers
2355 users here now

A Fediverse community for open-ended, thought provoking questions


Rules: (interactive)


1) Be nice and; have funDoxxing, trolling, sealioning, racism, toxicity and dog-whistling are not welcomed in AskLemmy. Remember what your mother said: if you can't say something nice, don't say anything at all. In addition, the site-wide Lemmy.world terms of service also apply here. Please familiarize yourself with them


2) All posts must end with a '?'This is sort of like Jeopardy. Please phrase all post titles in the form of a proper question ending with ?


3) No spamPlease do not flood the community with nonsense. Actual suspected spammers will be banned on site. No astroturfing.


4) NSFW is okay, within reasonJust remember to tag posts with either a content warning or a [NSFW] tag. Overtly sexual posts are not allowed, please direct them to either !asklemmyafterdark@lemmy.world or !asklemmynsfw@lemmynsfw.com. NSFW comments should be restricted to posts tagged [NSFW].


5) This is not a support community.
It is not a place for 'how do I?', type questions. If you have any questions regarding the site itself or would like to report a community, please direct them to Lemmy.world Support or email info@lemmy.world. For other questions check our partnered communities list, or use the search function.


6) No US Politics.
Please don't post about current US Politics. If you need to do this, try !politicaldiscussion@lemmy.world or !askusa@discuss.online


Reminder: The terms of service apply here too.

Partnered Communities:

Tech Support

No Stupid Questions

You Should Know

Reddit

Jokes

Ask Ouija


Logo design credit goes to: tubbadu


founded 2 years ago
MODERATORS
top 17 comments
sorted by: hot top controversial new old

What you're thinking of is this: Malus.

It's an example of Clean-room design, basically you have a LLM read the code and write a specification, which a second LLM uses to rewrite the code without access to the original.

Although since the original code was most likely included in the LLMs training data, this might not really be really true.

[–] slazer2au@lemmy.world 42 points 17 hours ago (1 children)

Why are you using the future tense when AI companies have already scraped the ever living crap out of every git forge it can access?

[–] ComradePenguin@lemmy.ml 5 points 17 hours ago (1 children)

Updated question to clarify. Was thinking more of specific repos being laundered, not entirety

[–] slazer2au@lemmy.world 15 points 17 hours ago (2 children)

When you use LLMs to regurgitate code, you do not get ownership of the code as you did not produce it. so using llms to "launder" code doesn't accomplish anything.

[–] ComradePenguin@lemmy.ml 4 points 14 hours ago (1 children)

From my understanding it does allow me to use the code for any purpose regardless of the license, does it not? Even if I dont own the LLM written code?

[–] Ephera@lemmy.ml 6 points 13 hours ago

Yeah, but you also have to be aware that companies rarely care to (fully) comply with licenses to begin with, if their own code isn't publicly accessible.

Basically:

  • If they actually open-source their own code, they have to fully comply (though the worst consequence is often just having to open-source your own code, which it already is, so it might not always be the highest priority either).
  • If they build a frontend, they generally do want to comply, because someone might be able to decompile the software and prove that licensed code is used inappropriately.
  • If they build a backend or build tooling or the like, GPL and AGPL is often still prohibited due to the high impact, but other than that, complying with licenses is seen as reducing risk for something that's pretty unlikely to affect them. The chance of them being sued for code that no one sees is just practically 0, so it's usually treated as an acceptable legal risk to not give a fuck.
[–] lIlIlIlIlIlIl@lemmy.world 1 points 16 hours ago (2 children)

Can you explain? I’ve seen vibe coded projects with licenses. Are they not valid?

[–] PlzGibHugs@piefed.ca 4 points 15 hours ago

Even before getting into the copyrightability of code, at the very least, any LLM-produced parts are not copyrightable. They are public domain.

That said, if its a mix of LLM code and human code, things get pretty messy. From my understanding, if the human expanded on or modified AI code, its public domain. If they wrote a section fully independently, they absolutely own the copyright. If its an unclear mix, it would have to be proven on a case-by-case basis with the onus being on the AI user to provide solid evidence that the code copied isn't AI generated.

[–] naught101@lemmy.world 2 points 15 hours ago

Not if the code is not copyrightable, and it may not be

[–] MajorHavoc@programming.dev 9 points 14 hours ago

Sure!

But I don't expect it to change much.

I could already do that, by hand as well.

It's a bit like how there's so many different superheroes who are obviously just off-brand SuperMan or off-brand Captain America.

Minor changes to avoid intellectual property law and branding has always been an option.

And I suppose all of these are easier now with the remixing slop-o-trons.

But they weren't terribly difficult, or particularly uncommon, even before.

[–] lime@feddit.nu 10 points 17 hours ago (1 children)

pretty much, with the caveat that code that has gone through an llm can't ever be licensed or copyrighted. it's basically a public domainifyer.

[–] Korhaka@sopuli.xyz 5 points 16 hours ago (1 children)

That doesn't seem to stop corporations assuming their software is still theirs even when an LLM wrote a lot of it.

[–] lime@feddit.nu 6 points 16 hours ago (1 children)

until it goes to court. i am giddy for the day.

[–] Korhaka@sopuli.xyz 1 points 16 hours ago

If that happens I will be so excited that I will stand as a tripod waiting for the results.

[–] Paranoidfactoid@lemmy.world 5 points 15 hours ago

Now you know why Microsoft bought github.

[–] tal@lemmy.today 1 points 12 hours ago* (last edited 12 hours ago)

If you mean "can I just recreate an existing copyrighted work without it being copyright-infringing", no. You're still liable if you recreate text that would be considered a derivative work. You don't have a fantastic mechanism to avoid that with existing LLMs, though I would guess that you most-likely aren't going to generate infringing code randomly.

Same thing for images or other media.

[–] PlzGibHugs@piefed.ca 1 points 15 hours ago

From my understanding, code is still covered by copyright. This means that copied code, even if run through an intermediary like an AI, is still copyright infringement. In the same way, even if an image generator recreates a character or movie frame, it isn't made public domain (the default state of AI Output), its just that the AI ingringed on someone else's copyright. If the code or image is then used, you can still be sued.