• Lemming6969@lemmy.world
    link
    fedilink
    English
    arrow-up
    2
    arrow-down
    1
    ·
    1 day ago

    Except learning in this context is building a probability map reinforcing the exact text of the book. Given the right prompt, no new generative concepts come out, just the verbatim book text trained on.

    So it depends on the model I suppose and if the model enforces generative answers and blocks verbatim recitation.

    • FaceDeer@fedia.io
      link
      fedilink
      arrow-up
      1
      ·
      24 hours ago

      Again, you should read the ruling. The judge explicitly addresses this. The Authors claim that this is how LLMs work, and the judge says “okay, let’s assume that their claim is true.”

      Fourth, each fully trained LLM itself retained “compressed” copies of the works it had trained upon, or so Authors contend and this order takes for granted.

      Even on that basis he still finds that it’s not violating copyright to train an LLM.

      And I don’t think the Authors’ claim would hold up if challenged, for that matter. Anthropic chose not to challenge it because it didn’t make a difference to their case, but in actuality an LLM doesn’t store the training data verbatim within itself. It’s physically impossible to compress text that much.