• FaceDeer@fedia.io
    link
    fedilink
    arrow-up
    1
    arrow-down
    4
    ·
    5 hours ago

    A lot of the griping about AI training involves data that’s been freely published. Stable Diffusion, for example, trained on public images available on the internet for anyone to view, but led to all manner of ill-informed public outrage. LLMs train on public forums and news sites. But people have this notion that copyright gives them some kind of absolute control over the stuff they “own” and they suddenly see a way to demand a pound of flesh for what they previously posted in public. It’s just not so.

    I have the right to analyze what I see. I strongly oppose any move to restrict that right.

    • kittenzrulz123@lemmy.blahaj.zone
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 hour ago

      Publically available =/= freely published

      Many images are made and published with anti AI licenses or are otherwise licensed in a way that requires attribution for derivative works.

      • FaceDeer@fedia.io
        link
        fedilink
        arrow-up
        1
        arrow-down
        1
        ·
        46 minutes ago

        The problem with those things is that the viewer doesn’t need that license in order to analyze them. They can just refuse the license. Licenses don’t automatically apply, you have to accept them. And since they’re contracts they need to offer consideration, not just place restrictions.

        An AI model is not a derivative work, it doesn’t include any identifiable pieces of the training data.

    • AwesomeLowlander@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      5
      ·
      2 hours ago

      It’s also pretty clear they used a lot of books and other material they didn’t pay for, and obtained via illegal downloads. The practice of which I’m fine with, I just want it legalised for everyone.