Greg Rutkowski, a digital artist known for his surreal style, opposes AI art but his name and style have been frequently used by AI art generators without his consent. In response, Stable Diffusion removed his work from their dataset in version 2.0. However, the community has now created a tool to emulate Rutkowski’s style against his wishes using a LoRA model. While some argue this is unethical, others justify it since Rutkowski’s art has already been widely used in Stable Diffusion 1.5. The debate highlights the blurry line between innovation and infringement in the emerging field of AI art.

  • Rygel the Dom
    link
    fedilink
    English
    482 years ago

    What blurry line? An artist doesn’t what his art stolen from him. Seems pretty cut and dry to me.

    • fades
      link
      fedilink
      English
      36
      edit-2
      2 years ago

      I don’t disagree but stolen is a bit of a stretch

    • @teichflamme@lemm.ee
      link
      fedilink
      222 years ago

      Nothing was stolen.

      Drawing inspiration from someone else by looking at their work has been around for centuries.

      Imagine if the Renaissance couldn’t happen because artists didn’t want their style stolen.

    • @KoboldCoterie@pawb.social
      link
      fedilink
      English
      222 years ago

      I don’t fully understand how this works, but if they’ve created a way to replicate his style that doesn’t involve using his art in the model, how is it problematic? I understand not wanting models to be trained using his art, but he doesn’t have exclusive rights to the art style, and if someone else can replicate it, what’s the problem?

      This is an honest question, I don’t know enough about this topic to make a case for either side.

      • jamesravey
        link
        fedilink
        English
        32
        edit-2
        2 years ago

        TL;DR The new method still requires his art.

        LoRA is a way to add additional layers to a neural network that effectively allow you to fine tune it’s behaviour. Think of it like a “plugin” or a “mod”

        LoRas require examples of the thing you are targeting. Lots of people in the SD community build them for particular celebrities or art styles by collecting examples of the that celebrity or whatever from online.

        So in this case Greg has asked Stable to remove his artwork which they have done but some third party has created an unofficial LoRA that does use his artwork to mod the functionality back in.

        In the traditional world the rights holder would presumably DMCA the plugin but the lines are much blurrier with LoRA models.

      • @delollipop@beehaw.org
        link
        fedilink
        11
        edit-2
        2 years ago

        Do you know how they recreated his style? I couldn’t find such information or frankly have enough understanding to know how.

        But if they either use his works directly or works created by another GAI with his name/style in the prompt, my personal feeling is that would still be unethical, especially if they charge money to generate his style of art without compensating him.

        Plus, I find that the opt-out mentality really creepy and disrespectful

        “If he contacts me asking for removal, I’ll remove this.” Lykon said. “At the moment I believe that having an accurate immortal depiction of his style is in everyone’s best interest.”

        • fsniper
          link
          fedilink
          202 years ago

          I still have trouble understanding the distinction between “a human consuming different artists, and replicating the style” vs “software consuming different artists, and replicating the style”.

        • SweetAIBelle
          link
          fedilink
          82 years ago

          Generally speaking, the way training works is this:
          You put together a folder of pictures, all the same size. It would’ve been 1024x1024 in this case. Other models have used 768z768 or 512x512. For every picture, you also have a text file with a description.

          The training software takes a picture, slices it into squares, generates a square the same size of random noise, then trains on how to change that noise into that square. It associates that training with tokens from the description that went with that picture. And it keeps doing this.

          Then later, when someone types a prompt into the software, it tokenizes it, generates more random noise, and uses the denoising methods associated with the tokens you typed in. The pictures in the folder aren’t actually kept by it anywhere.

          From the side of the person doing the training, it’s just put together the pictures and descriptions, set some settings, and let the training software do its work, though.

          (No money involved in this one. One person trained it and plopped it on a website where people can download loras for free…)

        • @KoboldCoterie@pawb.social
          link
          fedilink
          62 years ago

          Do you know how they recreated his style? I couldn’t find such information or frankly have enough understanding to know how.

          I don’t, but another poster noted that it involves using his art to create the LoRA.

          Plus, I find that the opt-out mentality really creepy and disrespectful

          I don’t know about creepy and disrespectful, but it does feel like they’re saying “I know the artist doesn’t want me to do this, but if he doesn’t specifically ask me personally to stop, I’m going to do it anyway.”

        • Rhaedas
          link
          fedilink
          62 years ago

          they charge money to generate his style of art without compensating him.

          That’s really the big thing, not just here but any material that’s been used to train on without permission or compensation. The difference is that most of it is so subtle it can’t be picked out, but an artist style is obviously a huge parameter since his name was being used to call out those particular training aspects during generations. It’s a bit hypocritical to say you aren’t stealing someone’s work when you stick his actual name in the prompt. It doesn’t really matter how many levels the art style has been laundered, it still originated from him.

          • conciselyverbose
            link
            fedilink
            102 years ago

            It is unconditionally impossible to own an artistic style. “Stealing a style” cannot be done.

            • Peanut
              link
              fedilink
              6
              edit-2
              2 years ago

              Just wait until you can copywrite a style. Guess who will end up owning all the styles.

              Spoiler, it’s wealthy companies like Disney and Warner. Oh you used cross hatching? Disney owns the style now you theif.

              Copyright is fucked. Has been since before the Mickey mouse protection act. Our economic system is fucked. People would rather fight each other and new tools instead of rallying against the actual problem, and it’s getting to me.

              • @Pseu@beehaw.org
                link
                fedilink
                52 years ago

                You’re right, copyright won’t fix it, copyright will just enable large companies to activate more of their work extract more from the creative space.

                But who will benefit the most from AI? The artists seem to be getting screwed right now, and I’m pretty sure that Hasbro and Disney will love to cut costs and lay off artists as soon as this blows over.

                Technology is capital, and in a capitalist system, that goes to benefit the holders of that capital. No matter how you cut it, laborers including artists are the ones who will get screwed.

                • @TheBurlapBandit@beehaw.org
                  link
                  fedilink
                  42 years ago

                  Me, I’ll benefit the most. I’ve been using a locally running instance of the free and open source AI software Stable Diffusion to generate artwork for my D&D campaigns and they’ve never looked more beautiful!

                  • FaceDeer
                    link
                    fedilink
                    32 years ago

                    Same here. It’s awesome being able to effectively “commission” art for any random little thing the party might encounter. And sometimes while generating images there’ll be surprising details that give me new ideas, too. It’s like brainstorming with ChatGPT but in visual form.

            • Rhaedas
              link
              fedilink
              62 years ago

              And yet the artist’s name is used to push the weights towards pictures in their style. I don’t know what the correct semantics are for it, nor the legalities. That’s part of the problem, the tech is ahead of our laws, as is usually the case.

              • conciselyverbose
                link
                fedilink
                82 years ago

                And yet the artist’s name is used to push the weights towards pictures in their style.

                That’s not even vaguely new in the world of art.

                Imitating style is the core of what art is. It’s absolutely unconditionally protected by copyright law. It’s not even a .01 out of 10 on the scale of unethical. It’s what’s supposed to happen.

                The law might not cover this yet, but any law that restricts the fundamental right to build off of the ideas of others that are the core of the entirety of human civilization is unadulterated evil. There is no part of that that could possibly be acceptable to own.

                • Rhaedas
                  link
                  fedilink
                  12 years ago

                  I totally agree with you on protecting the basics of creativity and growth. I think the core issue is using “imitate” here. Is that what the LLM is doing, or is that an anthropomorphism of some sense that there’s intelligence guiding the process? I know it seems like I’m nitpicking things to further my point, but the fact that this is an issue to many even outside artwork says there is a question here of what is and isn’t okay.

                  • conciselyverbose
                    link
                    fedilink
                    5
                    edit-2
                    2 years ago

                    The AI is not intelligent. That doesn’t matter.
                    Nothing anyone owns is being copied or redistributed. The creator isn’t the tool; it’s the person using the tool.

                    AI needs two things to work, an algorithm and data. If training is allowed to anyone, anyone can create their own algorithms and use the AI as a tool to create innovative new messages with some ideas borrowed from other work.

                    If data is proprietary, they cannot. But Disney still can. They’ll just as successfully flood out all the artists who can’t use AI because they don’t have a data set, but now they and the two other companies in the world who own IP are basically a monopoly (or tri- or whatever) and everyone else is screwed.

              • Altima NEO
                link
                fedilink
                72 years ago

                It’s only using his name because the person who created the LORA trained it with his name. They could have chosen any other word.

                • Rhaedas
                  link
                  fedilink
                  12 years ago

                  True, and then because it’s a black box there wouldn’t be a known issue at all. Or maybe it would be much less of an issue because the words might have blended others into the mix, and his style wouldn’t be as obvious in the outputs, and/or it would be easier to dismiss. Did the training involve actual input of his name, or was that pulled from the source trained on? How much control was in the training?

            • snooggums
              link
              fedilink
              42 years ago

              Is drawing Mickey Mouse in a new pose copying the style or copying Mickey Mouse?

              • @ricecake@beehaw.org
                link
                fedilink
                52 years ago

                You said it yourself. You’re drawing Micky mouse in a new pose, so you’re copying Mickey mouse.

                Drawing a cartoon in the style of Mickey mouse isn’t the same thing.

                You can’t have a copyright on “big oversized smile, exaggerated posture, large facial features, oversized feet and hands, rounded contours and a smooth style of motion”.

              • conciselyverbose
                link
                fedilink
                12 years ago

                The second.

                I’m not sure how that’s relevant here, though. There is nothing at all being copied but an aesthetic.

                • tqgibtngo
                  link
                  fedilink
                  1
                  edit-2
                  2 years ago

                  There is nothing at all being copied but an aesthetic.

                  Although to me it is interesting that, even without literal copying, a generator might be capable of potentially emulating some key features of a specified source. Can this sometimes arguably extend beyond just “an aesthetic”? We’ve all seen examples similar to this one (from the SD online demo, default setting, with a familiar public-domain source) — https://i.imgur.com/PUJs3RL.png

      • @Hubi@feddit.de
        link
        fedilink
        7
        edit-2
        2 years ago

        You’re pretty spot on. It’s not much different from a human artist trying to copy his style by hand but without reproducing the actual drawings.

    • falsem
      link
      fedilink
      192 years ago

      If I look at someone’s paintings, then paint something in a similar style did I steal their work? Or did I take inspiration from it?

      • @Pulse@dormi.zone
        link
        fedilink
        152 years ago

        No, you used it to inform your style.

        You didn’t drop his art on to a screenprinter, smash someone else’s art on top, then try to sell t-shirts.

        Trying to compare any of this to how one, individual, human learns is such a wildly inaccurate way to justify stealing a someone’s else’s work product.

        • falsem
          link
          fedilink
          142 years ago

          If it works correctly it’s not a screenprinter, it’s something unique as the output.

          • @Pulse@dormi.zone
            link
            fedilink
            182 years ago

            The fact that folks can identify the source of various parts of the output, and that intact watermarks have shown up, shows that it doesn’t work like you think it does.

            • FaceDeer
              link
              fedilink
              112 years ago

              They can’t, and “intact” watermarks don’t show up. You’re the one who is misunderstanding how this works.

              When a pattern is present very frequently the AI can learn to imitate it, resulting in things that closely resemble known watermarks. This is called “overfitting” and is avoided as much as possible. But even in those cases, if you examine the watermark-like pattern closely you’ll see that it’s usually quite badly distorted and only vaguely watermark-like.

              • @Pulse@dormi.zone
                link
                fedilink
                112 years ago

                Yes, because “imitate” and “copy” are different things when stealing from someone.

                I do understand how it works, the “overfitting” was just laying clear what it does. It copies but tries to sample things in a way that won’t look like clear copies. It had no creativity, it is trying to find new ways of making copies.

                If any of this was ethical, the companies doing it would have just asked for permission. That they didn’t says a everything you need to know.

                I don’t usually have these kinds discussions anymore, I got tired of conversations like this back in 2016, when it became clear that people will go to the ends of the earth to justify unethical behavior as long as the people being hurt by it are people they don’t care about.

                • FaceDeer
                  link
                  fedilink
                  52 years ago

                  And we’re back to you calling it “stealing”, which it certainly is not. Even if it was copyright violation, copyright violation is not stealing.

                  You should try to get the basic terminology right, at the very least.

                  • @Pulse@dormi.zone
                    link
                    fedilink
                    62 years ago

                    Just because you’ve redefined theft in a way that makes you feel okay about it doesn’t change what they did.

                    They took someone else’s work product, fed it into their machine then used that to make money.

                    They stole someone’s labor.

            • @jarfil@beehaw.org
              link
              fedilink
              42 years ago

              Does that mean the AI is not smart enough to remove watermarks, or that it’s so smart it can reproduce them?

              • falsem
                link
                fedilink
                72 years ago

                It means that it’s stupid enough that it reproduces them - poorly.

              • nickwitha_k (he/him)
                link
                fedilink
                42 years ago

                LLMs and directly related technologies are not AI and possess no intelligence or capability to comprehend, despite the hype. So, they are absolutely the former, though it’s rather like a bandwagon sort of thing (x number of reference images had a watermark, so that’s what the generated image should have).

                • @jarfil@beehaw.org
                  link
                  fedilink
                  32 years ago

                  LLMs […] no intelligence or capability to comprehend

                  That’s debatable. LLMs have shown emergent behaviors aside from what was trained, and they seem to be capable of comprehending relationships between all sorts of tokens, including multi-modal ones.

                  Anyway, Stable diffusion is not an LLM, it’s more of a “neural network hallucination machine” with some cool hallucinations, that sometimes happen to be really close to some or parts of the input data. It still needs to be “smart” enough to decompose the original data into enough and the right patterns, that it can reconstruct part of the original from the patterns alone.

                  • nickwitha_k (he/him)
                    link
                    fedilink
                    32 years ago

                    Thanks for the clarification!

                    LLMs have indeed shown interesting behaviors but, from my experience with the technology and how it works, I would say that any claims of intelligence being possessed by a system that is only an LLM would be suspect and require extraordinary evidence to prove that it is not mistaken anthropomorphizing.

              • @Swedneck@discuss.tchncs.de
                link
                fedilink
                12 years ago

                It’s like staring yourself blind at artworks with watermarks until you start seeing artworks with blurry watermarks in your dreams