• Alex@lemmy.ml
    link
    fedilink
    English
    arrow-up
    3
    ·
    6 days ago

    Where are you seeing the 2-10% figure?

    In my experience code generation is most affected by the local context (i.e. the codebase you are working on). On top of that a lot of code is purely mechanical - code generally has to have a degree of novelty to be protected by copyright.

      • Alex@lemmy.ml
        link
        fedilink
        English
        arrow-up
        1
        ·
        2 days ago

        If the 2-10% is just boilerplate syscall number defines or trivial MIN/MAX macros then it’s just the common way to do things.

        • ell1e@leminal.space
          link
          fedilink
          English
          arrow-up
          1
          ·
          2 days ago

          So do you want to legally review every line by an LLM to see if it meets the fair use criterion, since you have to assume it was probably stolen? And would you do this for a known plagiarizing human contributor too…?

          • Alex@lemmy.ml
            link
            fedilink
            English
            arrow-up
            1
            ·
            2 days ago

            No, that’s why the author asserts that with their signed-of-by. It’s what I do if I use any LLM content as the basis of my patches.

            • ell1e@leminal.space
              link
              fedilink
              English
              arrow-up
              1
              ·
              edit-2
              1 day ago

              So what does the signed-off-by magically solve here, that doesn’t require either you or the contributor to legally review every line by an LLM? If you’re not a lawyer, is your contributor going to be one?

              • Alex@lemmy.ml
                link
                fedilink
                English
                arrow-up
                1
                ·
                1 day ago

                They don’t have to be. They know what they asked the LLM to do. They know how much they adapted the output. You usually have to work to get the models to spit out significant chunks of memorised text.

                • ell1e@leminal.space
                  link
                  fedilink
                  English
                  arrow-up
                  1
                  ·
                  1 day ago

                  I don’t have much more to say other than I doubt the data backs up what you’re saying at all.

    • Danquebec@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      1
      arrow-down
      1
      ·
      6 days ago

      Imagine how broken it would be otherwise. The first person to write a while loop in any given language would be the owner of it. Anyone else using the same concept would have to write an increasingly convoluted while loop with extra steps.

      • sloppy_diffuser@sh.itjust.works
        link
        fedilink
        English
        arrow-up
        1
        ·
        6 days ago

        Anyone else using the same concept would have to write an increasingly convoluted while loop with extra steps.

        Sounds like an origin story for recursion.