The GNU Compiler Collection (GCC) developers now have a need to set a policy whether AI / Large Language Model (LLM) generated patches will be accepted for this open-source compiler stack.

The GCC compiler doesn’t currently have a policy in place whether to permit AI/LLM-generated patches. But within a bug report today there is a patch posted by a user in trying to fix a GCC 16 compiler regression.

  • mech@feddit.org
    link
    fedilink
    arrow-up
    32
    arrow-down
    1
    ·
    1 day ago

    A lot of open source development now is done by employees at big corps like IBM, Red Hat, Google and Microsoft.
    Those companies heavily push towards using AI for development, some even enforcing it.
    Will be interesting to see what happens when more and more OSS projects outright ban AI contributions, while more and more companies mandate using AI.

  • frongt@lemmy.zip
    link
    fedilink
    arrow-up
    26
    arrow-down
    1
    ·
    1 day ago

    If it’s maintainable, I don’t think there’s an issue (aside from the usual license violations in LLM training). The big problem is when people send giant unmaintainable patches that they can’t even explain themselves.

    • mech@feddit.org
      link
      fedilink
      arrow-up
      21
      ·
      1 day ago

      No, the license violations are the main problem.
      AI will steal all publicly visible code regardless of its license as “training data” and then put out the same code almost verbatim when asked to solve the same problem. And projects need to prevent that to avoid costly legal battles they can’t afford.

  • DaGeek247@fedia.io
    link
    fedilink
    arrow-up
    15
    ·
    1 day ago

    Enforce the usual quality rules and most all llm results will be banned by default. Technically, you could try banning llm work by default, but that just results in people lying, and having to check over llm work anyways.

  • mrmaplebar@fedia.io
    link
    fedilink
    arrow-up
    7
    ·
    1 day ago

    Well… I’d be curious to know how anyone would argue that AI generated code is GPL safe, especially considering they are often using black box, binary blob models trained on mystery code from all over the internet, with zero attribution or license information, and with some of it almost certainly being proprietary.

    Are we going to start ignoring the very licenses that we created.

    • grue@lemmy.world
      link
      fedilink
      arrow-up
      8
      ·
      23 hours ago

      Alternatively, I would argue that all LLM code output must be GPL, since it was trained partially on GPL code.

      Either that, or LLM code output cannot be used for any purpose at all, by anyone.

  • magic_lobster_party@fedia.io
    link
    fedilink
    arrow-up
    8
    ·
    1 day ago

    IMO I think it’s fine as long the submitter understand all their changes. This is easiest done if the submitter writes all the code themselves though.

    In this case I would reject this. The submitter doesn’t seem to be qualified to understand the code change, and just let ChatGPT explain the code change for them. ChatGPT is a master in being confidently incorrect.

    • A_A@lemmy.world
      link
      fedilink
      arrow-up
      3
      ·
      1 day ago

      out of context joke
      640KB ought to be enough for everybody - - Bill Gates