Sadly, it seems like Lemmy is going to integrate LLM code going forward: https://github.com/LemmyNet/lemmy/issues/6385 If you comment on the issue, please try to make sure it’s a productive and thoughtful comment and not pure hate brigading.

Consider upvoting the issue to show community interest.

Edit: perhaps I should also mention this one here as a similar discussion: https://github.com/sashiko-dev/sashiko/issues/31 This one concerns the Linux kernel. I hope you’ll forgive me this slight tangent, but more eyes could benefit this one too.

  • Rentlar@lemmy.ca
    link
    fedilink
    arrow-up
    1
    ·
    2 months ago

    Code written with the help of LLM and being reviewed is different than like what was happening with Lutris where the developer decided to obfuscate their use of AI-generated code.

    The approach you suggest to totally ban it, while in principle can agree and I think that’s noble, it could lead to people accusing each other of using AI code where it may or may not have happened, or others just hiding it and trying to submit anyway without the reviewers knowing, which is just counter-productive.

    I’ve followed Lemmy development now for 3 years, the devs approach is slow and steady, to a fault in some people’s views. I think it’s a better use of open source resources if we encourage candor and honesty. If the repo gets spammed with AI-generated PRs, then it will probably be blanket banned, but contributors accurately documenting and reporting their usage of AI will help direct reviewers attention to ensure the code is not slop quality or full of hallucinations.

    • wheezy@lemmy.ml
      link
      fedilink
      English
      arrow-up
      1
      ·
      2 months ago

      Great perspective and response. Far too many “fuck AI” people are literally advocating for the equivalent of “fuck computers” and “more tedious labor please!”

      The reason you should hate AI should be related to it’s exploitation of labor and it’s over use leading to energy and environmental impacts. Trying to ban AI for all applications is just counter productive and impossible. If the anti AI crowd is just filled with people that want it banned outright for everything, well, then the pro AI crowd that wants to slam it into anything and everything will win out.

      We need to be pointing to good applications of AI that can benefit open source projects in a responsible way as examples of how it should be used. Not spamming them with hate comments because “AI bad”

    • ell1e@leminal.spaceOP
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      2 months ago

      In my opinion, this argument is exactly the same as saying “we can’t enforce people not stealing GPL-licensed code and copy&pasting it into our project, so we might as well allow it and ask them to disclose it.”

      You can try to argue AI may actually be useful, which seems like what they did, and that would more fairly inform a policy in my opinion. I think your argument doesn’t.

      • Rentlar@lemmy.ca
        link
        fedilink
        arrow-up
        1
        ·
        2 months ago

        My argument is that a total ban on AI use is more comparable to saying “Code from any other coding project is not allowed”. It will start unproductive arguments over boilerplate, struct definitions and other commonly used code.

        The broadness and vaagueness of “no AI whatsoever” or “no code from any other projects whatsoever” will be more confusing than saying, “if you do copy any code from another project, let us know where from”. Then the PR can be evaluated, rejected if it’s nonfree or just poor quality, rather than incentivizing people to pretend other people’s code is their own, risking bigger consequences for the whole project. People can be honest if they got inspiration from stackoverflow, a reference book, or another project, if they are allowed to be.

        I’m not saying AI should be blanket allowed, the submitter needs to understand the code, enough to be able to revise it for errors themselves if the devs point out something. They can’t just say “I asked AI and it’s confident that the code does this and is bug free”.

      • hitmyspot@aussie.zone
        link
        fedilink
        arrow-up
        1
        ·
        2 months ago

        Not all ai, or rather, llm output is slop. Some is useful. The reason for review is to differentiate. I’m not just talking about coding. I’m talking about their actual useful functionality.

        It would be great if they didn’t hallucinate, or produce slop. It would also be great if the fact that companies use them instead of workers meant we worked less hours and had more leisure time rather than less paying jobs and more stress. The llm is not at fault for the structure of society.

        Llm and ai is a tool. If used appropriately, there should be no issue. Of used inappropriately, it should be called out. Certainly where there is a risk of it appearing on the surface to be good, but not actually good,.like AI generated codez then marking it as such seems reasonable. Banning it doesn’t get rid of it. It hides it. It exists and is now in the world. We need to have policies that support appropriate use.