“Suno’s training data includes essentially all music files of reasonable quality that are accessible on the open internet.”

“Rather than trying to argue that Suno was not trained on copyrighted songs, the company is instead making a Fair Use argument to say that the law should allow for AI training on copyrighted works without permission or compensation.”

Archived (also bypass paywall): https://archive.ph/ivTGs

    • Dr. Moose@lemmy.world
      link
      fedilink
      English
      arrow-up
      0
      ·
      edit-2
      5 months ago

      You’re free to learn from any piece of music too. Whether AI is actually learning is still debatable but you have the same rights right now.

      I’m still on the edge tbh I feel like it is learning and it is transformative but it’s just too powerful for our current copyright framework.

      Either way, that’ll be such a headache for the transformative work clause of copyright for years to come. Also policing training would be completely unenforcable so any decision here would be rather moot in real world practice either way.

      • just another dev@lemmy.my-box.dev
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        Also policing training would be completely unenforcable

        That’s where laws would come in. Obviously it would have civil law, not criminal law, but making sure it would be enforceable would have to be part of such laws. For example, forcing model makers to disclose their training dataset in one way or another.

        • Dr. Moose@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 months ago

          But you can already train models at home also you can just extend existing models with new training data. Will that be regulated too? How?

          • CosmoNova@lemmy.world
            link
            fedilink
            English
            arrow-up
            0
            ·
            edit-2
            5 months ago

            They‘re literally already about to heavily regulate hobby AI to ensure giant corporations that hoard all our information get to make even more mountains of money with it. The idea that anyone gets to use any media for machine learning is already a relict of the past and in fact not remotely comparable to learning things for yourself. Especially not in the legal sense. Did you really naively believe AI will democratize anything for even a second?

      • jadelord@discuss.tchncs.de
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        We are free to learn, but learning is not free.

        Freedom vs cost. One cannot pickup a skill without time, effort and more importantly access to guidance and a vast library of content. Same applies to man or machine. The difference is how corporations have essentially reinvented piracy to facilitate their selfish ends after decades of dictating what’s right with DMCA, DRM and what not.

        • credo@lemmy.world
          link
          fedilink
          English
          arrow-up
          0
          ·
          5 months ago

          If you get an idea from a song, you are 1000% free to turn that into new art. This is the fair use argument.

          • Telorand@reddthat.com
            link
            fedilink
            English
            arrow-up
            0
            ·
            5 months ago

            I agree with the logic, but I don’t think it should apply to LLMs—a humans-only law, if you will.

            • ClamDrinker@lemmy.world
              link
              fedilink
              English
              arrow-up
              0
              ·
              edit-2
              5 months ago

              That’s sort of currently the law with copyright in the US. You can’t get a copyright on material made completely by an AI. Only if a human interfered can you get a copyright, and most likely only on the parts that the human interfered with.

              Source: https://www.copyright.gov/ai/ai_policy_guidance.pdf (See header II. The Human Authorship Requirement)

              TL;DR

              the Office states that “to qualify as a work of ‘authorship’ a work must be created by a human being” and that it “will not register works produced by a machine or mere mechanical process that operates randomly or automatically without any creative input or intervention from a human author.”

            • The Quuuuuill@slrpnk.net
              link
              fedilink
              English
              arrow-up
              0
              ·
              5 months ago

              I think the key is that an LLM can’t have ideas. Its creative endeavors aren’t creative. Art is about the craft and the message and an LLM lacks that context. Like. The best an LLM can do is produce the kinds of music Drake does that is meant to pacify people into continuous consumption

              • Telorand@reddthat.com
                link
                fedilink
                English
                arrow-up
                0
                ·
                5 months ago

                I had a similar thought after I wrote this. LLMs aren’t creating anything so much as style-copying. They’re unique productions, insomuch as rearranging notes or pixels makes something unique, but I think creativity requires conscious agency, which LLMs definitely do not have.

                Also, I don’t need to copy the entirety of Drake’s discography to produce music like his, which is an aspect of human creativity that LLMs currently lack.

      • PerogiBoi@lemmy.ca
        link
        fedilink
        English
        arrow-up
        0
        ·
        5 months ago

        One has to pay a very high cost to do this. These AI companies did not pay. Why do AI companies get a pass on copyrighted material that the rest of us are getting sued, imprisoned, and fined for accessing?