362
submitted 5 months ago by lautan@lemmy.ca to c/technology@lemmy.world

New development policy: code generated by a large language model or similar technology (e.g. ChatGPT, GitHub Copilot) is presumed to be tainted (i.e. of unclear copyright, not fitting NetBSD's licensing goals) and cannot be committed to NetBSD.

https://www.NetBSD.org/developers/commit-guidelines.html

all 25 comments
sorted by: hot top controversial new old
[-] omgitsaheadcrab@sh.itjust.works 89 points 5 months ago

Ok but how is anyone meant to know if you generated your docstrings using copilot?

[-] Terces@lemmy.world 110 points 5 months ago

How do they know that you wrote it yourself and didn't just steal it?

This is a rule to protect themselves. If there is ever a case around this, they can push the blame to the person that committed the code for breaking that rule.

[-] sirico@feddit.uk 35 points 5 months ago

This is the only reason rules exist, not to stop people doing a thing but to be able to enforce or defect responsibility when they do.

[-] ripcord@lemmy.world 7 points 5 months ago

I mean, generally rules at least are to strongly discourage people from doing a thing, or to lead to things that WOULD prevent people from doing a thing.

A purely conceptual rule by itself would not magically stop someone from doing a thing, but that's kind of a weird way to think about it.

[-] mannycalavera@feddit.uk 40 points 5 months ago

They'll use AI to detect it.... obviously. ☺️

[-] ceasarlegsvin@kbin.social 11 points 5 months ago

Because they'll be shit?

Docstrings based on the method signature and literal contents of a method or class are completely pointless, and that's all copilot can do. It can't Intuit anything that docstrings are actually there for.

[-] SMillerNL@lemmy.world 11 points 5 months ago

Are they long, super verbose and often incorrect?

[-] Xantar@lemmy.dbzer0.com 7 points 5 months ago

Magic, I guess ?

[-] todd_bonzalez@lemm.ee 62 points 5 months ago

Lots of stupid people asking "how would they know?"

That's not the fucking point. The point is that if they catch you they can block future commits and review your past commits for poor quality code. They're setting a quality standard, and establishing consequences for violating it.

If your AI generated code isn't setting off red flags, you're probably fine, but if something stupid slips through and the maintainers believe it to be the result of Generative AI, they will remove your code from the codebase and you from the project.

It's like laws against weapons. If you have a concealed gun on your person and enter a public school, chances are that nobody will know and you'll get away with it over and over again. But if anyone ever notices, you're going to jail, you're getting permanently trespassed from school grounds, and you're probably not going to be allowed to own guns for a while.

And, it's a message to everyone else quietly breaking the rules that they have something to lose if they don't stop.

[-] homesweethomeMrL@lemmy.world 13 points 5 months ago

Lots of stupid people asking "how would they know?"

That's not the fucking point.

Okay, easy there, Chief. We were just trying to figure out how it worked. Sorry.

[-] NotMyOldRedditName@lemmy.world 12 points 5 months ago* (last edited 5 months ago)

It was a fair question, but this is just going to turn out like universities failing or expelling people for alleged AI content in papers.

They can't prove it. They try to use AI tools to prove it, but those same tools will say a thesis paper from a decade ago is also AI generated. Pretty sure I saw a story of a professor accusing someone based off a tool having his own past paper fail the same tool

Short of an admission of guilt, it's a witch hunt.

[-] WalnutLum@lemmy.ml 34 points 5 months ago

This is a good move for international open source projects, with multiple lawsuits in multiple countries around the globe currently ongoing, the intellectual property nature of code made using AI isn't really secure enough to open yourself up to the liability.

I've done the same internally at our company. You're free to use whatever tool you want but if the tool you use spits out copyrighted code, and the law eventually has decided that model users instead of model trainers are liable for model output, then that's on you buddy.

[-] sugar_in_your_tea@sh.itjust.works 11 points 5 months ago

Yup. We don't allow AI tools on our codebase, but I allow it for interviews. I honestly haven't been impressed by it at all, it just encourages not understanding the code.

[-] brbposting@sh.itjust.works 2 points 5 months ago

Does this mean you have indicated to your employees and/or contractors that you intend to hold them legally liable in the case someone launches litigation against you?

[-] elias_griffin@lemmy.world 22 points 5 months ago* (last edited 5 months ago)

So proud of you NetBSD, this is why I sponsor you, slam dunk for the future. I'm working on a NetBSD hardening script and Rice as we speak, great OS with some fantastically valuable niche applications and I think, a new broad approach I'm cooking up, a University Edition. I did hardening for all the other BSD, I saved the best for last!

[EDIT 5/16/2024 15:04 GMT -7] NetBSD got Odin lang support yesterday. That totally seals the NetBSD deal for me if I can come up with something cool for my workstation with Odin.

If you would like to vote on whether, or by what year, AI will be in the Linux Kernel on Infosec.space:

https://infosec.space/@wravoc/112441828127082611

[-] ripcord@lemmy.world 2 points 5 months ago

Thanks for your efforts Elias!

[-] jaybone@lemmy.world 18 points 5 months ago

I was hoping they ban it because it’s shit, but banning it for copyright reasons is fine too.

[-] bamboo@lemm.ee 11 points 5 months ago

I can understand why a project might want to do this until the law is fully implemented and testing in court, but I can tell most of the people in this thread haven’t actually figured out how to effectively use LLMs productively. They’re not about to replace software engineers, but as a software engineer, tools like GitHub copilot and ChatGPT are excellent at speeding up a workflow. ChatGPT for example is an excellent search engine that can give you a quick understanding of a topic. It’ll generate small amounts of code more quickly than I could write it by hand. Of course I’m still going to review that code to ensure it is to the same quality that hand written code would be, but overall this is still a much faster problem.

The luddites who hate on LLMs would have complained about the first compilers too, because they could write marginally faster assembly by hand.

[-] Tja@programming.dev 5 points 5 months ago

Same for intellisense, IDEs, Debuggers, linters, static analyzers, dynamic languages, garbage collection, NoSQL databases...

[-] Anti_Iridium@lemmy.world 3 points 5 months ago* (last edited 5 months ago)

Why use a computer to do the work when I could do myself? /s

[-] diffusive@lemmy.world 8 points 5 months ago

I never felt so close to try NetBSD as after reading this 😃

[-] possiblylinux127@lemmy.zip 5 points 5 months ago

We need to see more of this

[-] hperrin@lemmy.world 3 points 5 months ago* (last edited 5 months ago)

Hell yeah! Get that shit… OUTTA HERE!!!

Ok but seriously, that is a very good reason to ban it. Who knows what would happen if the AI just fully ripped someone else’s code off that’s supposed to be like GPL licensed or something. If humans can plagiarize, than AIs can plagiarize.

But also, how are they still using CVS? CVS is so slow and so bad. Even Subversion would be an upgrade.

this post was submitted on 16 May 2024
362 points (100.0% liked)

Technology

58998 readers
3221 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS