In its submission to the Australian government’s review of the regulatory framework around AI, Google said that copyright law should be altered to allow for generative AI systems to scrape the internet.

    • Phanatik@kbin.social
      link
      fedilink
      arrow-up
      15
      ·
      edit-2
      1 year ago

      This is stupid and I’ll tell you why.
      As humans, we have a perception filter. This filter is unique to every individual because it’s fed by our experiences and emotions. Artists make great use of this by producing art which leverages their view of the world, it’s why Van Gogh or Picasso is interesting because they had a unique view of the world that is shown through their work.
      These bots do not have perception filters. They’re designed to break down whatever they’re trained on into numbers and decipher how the style is constructed so it can replicate it. It has no intention or purpose behind any of its decisions beyond straight replication.
      You would be correct if a human’s only goal was to replicate Van Gogh’s style but that’s not every artist. With these art bots, that’s the only goal that they will ever have.

      I have to repeat this every time there’s a discussion on LLM or art bots:
      The imitation of intelligence does not equate to actual intelligence.

      • frog 🐸@beehaw.org
        link
        fedilink
        arrow-up
        13
        ·
        1 year ago

        Absolutely agreed! I think if the proponents of AI artwork actually had any knowledge of art history, they’d understand that humans don’t just iterate the same ideas over and over again. Van Gogh, Picasso, and many others, did work that was genuinely unique and not just a derivative of what had come before, because they brought more to the process than just looking at other artworks.

          • frog 🐸@beehaw.org
            link
            fedilink
            arrow-up
            7
            ·
            1 year ago

            My feeling is that the vast majority of pro-AI techbros come from a computer science, finance, or business background; undoubtedly intelligent people, but completely and utterly lacking in any appreciation or understanding of what actually goes into creative work. I’m sure they genuinely believe that there’s no difference between what a human does and what an AI does, because they think art (or writing, music, etc) are just the product of an algorithm.

            • Phanatik@kbin.social
              link
              fedilink
              arrow-up
              2
              ·
              1 year ago

              Ironically, my background is in mathematics but I also happen to be a writer so I see both sides of the argument. I just see the utter lack of compassion people have for those who produce creative work and the same people believe that if it can be automated, it should be automated.

      • davehtaylor@beehaw.org
        link
        fedilink
        arrow-up
        9
        ·
        1 year ago

        I really, really, really wish people would understand this.

        AI can only create a synthesis of exactly what it’s fed. It has no life experience, no emotional experience, no nurture-related experiences, no cultural experiences that color it’s thinking, because it isn’t thinking.

        The “AI are only doing what humans do” is such a brain-dead line of thinking, to the point that it almost feels like it’s 100% in bad faith whenever it’s brought up.

        • Phanatik@kbin.social
          link
          fedilink
          arrow-up
          1
          ·
          edit-2
          1 year ago

          I wasn’t talking about copyright law in regards to the model itself.

          I was talking about what is/isn’t grounds for plagiarism. I strongly disagree with the idea that artists and art bots go through the same process. They don’t and it’s reductive to claim otherwise. It negatively impacts the perception of artists’ work to assert that these models can automate a creative process which might not even involve looking at other artists’ work because humans are able to create on their own.

          A person who has never looked upon a single painting in their life can still produce a piece but the same cannot be said for an art bot. A model must be trained on work that you want the model to be able to imitate.

          This is why ChatGPT required the internet to do what it does (the privacy violation is another big concern there). The model needed vast quantities of information to be sufficiently trained because language is difficult to decipher. Languages evolved by getting in contact with other languages and organically making new words. ChatGPT will never invent a new word because it’s not intelligent, it is merely imitating intelligence.

            • Phanatik@kbin.social
              link
              fedilink
              arrow-up
              1
              ·
              1 year ago

              Once again, being reductive about artists’ work. Jackson Pollock’s entire career was smashing colours on a canvas. If you want to argue that Pollock had to look at thousands of paintings before making his, I honestly can’t take you seriously at that point.

              A computer could easily generate such “work” as well with no training data at all.

              Yes and in the eyes of its creators, that was deemed a failure which is why Midjourney and Dall-E are the way they are. These bots don’t want to create art, they want to imitate it.

              Children have barely any experiences and can still create something. You might not deem it worthy of calling it art but they created something despite their limited knowledge and life experience.

              Of course, you’d need books to read and write. The words have to be written and you need to see the words in written form if you also want to write them. But one thing you don’t take into account is handwriting. Another thing that is unique to every individual. Some have worse handwriting than others and with practice (like any muscle) it can be improved but you haven’t had to have seen handwritten text before writing it yourself. You only need to be taught how to hold a pen and you can write.

              Novels are complex structures of language just like poetry. In order to write novels, you have to consume novels because it’s well understood that to find your own narrative voice you must see how others express theirs. Stories are told in unique ways and it’s crucial as a writer to understand and break these concepts down. Intention and purpose form a core part of storytelling and an LLM cannot and will not be able to express those things.

              They’re written in certain ways because the author intended them to be that way, such as Cormac McCarthy deciding to be very minimalist with his punctuation.
              I would love to see you make a point that an LLM without being specifically prompted to do so would make that stylistic decision. An LLM can’t make that decision because unless you specify a style it is aware of, it won’t organically do it.

              I am also a writer. I’ve written a short story. One of my stylistic choices is that I don’t use dialogue tags like “said”. An LLM won’t make that choice because it isn’t designed to do so, it won’t decide to minimise its use of dialogue tags to improve the flow of the narrative unless you told it to.

              It’s also completely ignoring the fact that you had to previously learn the spoken language as well (which is a vast quantity of information that takes a human decades to acquire proficiency in even with daily practice).

              Yes, in order to learn a spoken language you have to have heard it. However, languages evolve over time. You develop regional accents and dialects. All of the UK speaks English but no two towns speak the same way.

                • Phanatik@kbin.social
                  link
                  fedilink
                  arrow-up
                  1
                  ·
                  1 year ago

                  An LLM or art creation tool is barely equatable to one person. The difference between a child and an art creation tool is that you could show a child a single picture of a bunny, a bike and a carrot then ask them to draw an orange bunny riding a bike and they could draw something resembling that. An art bot would require hundreds to thousands of images of each object to understand what it is before it can even make a reasonable attempt. It’s not even comparable the level of training required.

                  At least the child’s drawing will have some personality in it, every output from an art bot ends up looking soulless. The reason for that is the simple concept that an art bot only imitates what it’s been trained on and an artist draws on inspiration before applying the two things an art bot will never have; intent or purpose.

      • acastcandream@beehaw.org
        link
        fedilink
        arrow-up
        2
        ·
        1 year ago

        this is stupid I’ll tell you why

        Not sure why you think anyone would read anything if that’s how you start it.

    • 50gp@kbin.social
      link
      fedilink
      arrow-up
      10
      ·
      1 year ago

      a human does not copy previous work exactly like these algorithms, whats this shit take?

        • nickwitha_k (he/him)@lemmy.sdf.org
          link
          fedilink
          arrow-up
          7
          ·
          1 year ago

          And LLMs and related technologies, by themselves, are artificial but not intelligent. So, the facts are not in favor of your argument to allow commercial parasitism on creative works.

            • nickwitha_k (he/him)@lemmy.sdf.org
              link
              fedilink
              arrow-up
              2
              ·
              edit-2
              1 year ago

              Nah. You’re missing the forest for the trees. Let’s get abstract:

              Person A makes a living by making product X and selling it.

              Person B makes a living by making product Y and selling it.

              Both A and B are in the same industry.

              Person C uses a machine to extract the essence of product X and Y and blend them. Person C then claims authorship and sells it as product Z, which they sell in competition to X and Y.

              Person C has not created anything. Their machine does not have value in the absence of products X and Y, yet received no permission, offers no credit nor compensation. In addition, they are competing for the same customers and harming the livelihoods of A and B. Person C is acting in a purely parasitic manner that cannot be seen as ethical in any widely accepted definition of the word.

                • nickwitha_k (he/him)@lemmy.sdf.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  1 year ago

                  First, feeding something into a machine is not the same as looking at it. Person C literally creates nothing. They are a parasite. There’s far more to creating than using statistical modeling algorithms. One cannot claim that that’s what people studying a style and then creating someone are doing because it is empirically false.

                  Second, the scope of the discussion is not just “can someone legally get in trouble”.

                • nickwitha_k (he/him)@lemmy.sdf.org
                  link
                  fedilink
                  arrow-up
                  2
                  ·
                  edit-2
                  1 year ago

                  The scope here is not limited to “can someone legally get in trouble under current law” (which, seems likely but is still working its way through courts). The discussion is specifically discussing ethics. Person C has created nothing. They should have no product to sell, if not for persons A and B. Their competition with those that their product is derived from is a parasitic relationship, plain and simple. They are performing an act of exploitation with measurable harm both to persons A and B but also to further development of their craft by destroying any incentive to continue it.

                  Now, in some sort of alternate economic system, where one’s livelihood is not tied to their vocation, sure, it’s possibly not problematic because the economic harm is removed. However, in current capitalist systems that are in place where LLMs are heavily hyped, it’s an ethically bankrupt action to take.

                  ETA: No amount of mental gymnastics can change the fact that use of others’ works without their consent to train a model, then claiming authorship and competing IS plainly theft of the labor that went into creating the original works.

                  That’s not too say that LLMs and they like don’t have value or often require effort to produce something worthwhile. Just that they need to be used in an ethical manner that improves the human condition, not as another tool to rob others of the fruit of their labors.

        • Zapp@beehaw.org
          link
          fedilink
          English
          arrow-up
          3
          ·
          1 year ago

          The goal of AI is fictional, and there’s no solid evidence today that it will ever stop being fiction.

          What at have today are stupid learning algorithms that are surprisingly good at mimicing intelligent people.

          The most apt comparison today is a particularly clever parrot.

          I’m all for having the discussion about how to handle AI when we have it, but it’s bad faith to apply it to what we have today.

          Critically, what we have today will never ever go on strike, or really make any kind of correct moral decision on it’s own. We must treat it like dumb automation, because it is dumb automation.

        • acastcandream@beehaw.org
          link
          fedilink
          English
          arrow-up
          1
          ·
          1 year ago

          the fact of the matter is that the goal of AI is literally to replicate the function of a human brain

          …says who? That’s absolutely your feeling and not facts.