Greg Rutkowski, a digital artist known for his surreal style, opposes AI art but his name and style have been frequently used by AI art generators without his consent. In response, Stable Diffusion removed his work from their dataset in version 2.0. However, the community has now created a tool to emulate Rutkowski’s style against his wishes using a LoRA model. While some argue this is unethical, others justify it since Rutkowski’s art has already been widely used in Stable Diffusion 1.5. The debate highlights the blurry line between innovation and infringement in the emerging field of AI art.

  • delollipop@beehaw.org
    link
    fedilink
    arrow-up
    4
    ·
    edit-2
    1 year ago

    Do you know how they recreated his style? I couldn’t find such information or frankly have enough understanding to know how.

    But if they either use his works directly or works created by another GAI with his name/style in the prompt, my personal feeling is that would still be unethical, especially if they charge money to generate his style of art without compensating him.

    Plus, I find that the opt-out mentality really creepy and disrespectful

    “If he contacts me asking for removal, I’ll remove this.” Lykon said. “At the moment I believe that having an accurate immortal depiction of his style is in everyone’s best interest.”

    • fsniper@kbin.social
      link
      fedilink
      arrow-up
      5
      ·
      1 year ago

      I still have trouble understanding the distinction between “a human consuming different artists, and replicating the style” vs “software consuming different artists, and replicating the style”.

    • Rhaedas@kbin.social
      link
      fedilink
      arrow-up
      3
      ·
      1 year ago

      they charge money to generate his style of art without compensating him.

      That’s really the big thing, not just here but any material that’s been used to train on without permission or compensation. The difference is that most of it is so subtle it can’t be picked out, but an artist style is obviously a huge parameter since his name was being used to call out those particular training aspects during generations. It’s a bit hypocritical to say you aren’t stealing someone’s work when you stick his actual name in the prompt. It doesn’t really matter how many levels the art style has been laundered, it still originated from him.

        • Peanut@sopuli.xyz
          link
          fedilink
          arrow-up
          1
          ·
          1 year ago

          Just wait until you can copywrite a style. Guess who will end up opening all the styles.

          Spoiler, it’s wealthy companies like Disney and Warner. Oh you used cross hatching? Disney owns the style now you theif.

          Copyright is fucked. Has been since before the Mickey mouse protection act. Our economic system is fucked. People would rather fight each other and new tools instead of rallying against the actual problem, and it’s getting to me.

    • KoboldCoterie@pawb.social
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Do you know how they recreated his style? I couldn’t find such information or frankly have enough understanding to know how.

      I don’t, but another poster noted that it involves using his art to create the LoRA.

      Plus, I find that the opt-out mentality really creepy and disrespectful

      I don’t know about creepy and disrespectful, but it does feel like they’re saying “I know the artist doesn’t want me to do this, but if he doesn’t specifically ask me personally to stop, I’m going to do it anyway.”

    • SweetAIBelle@kbin.social
      link
      fedilink
      arrow-up
      1
      ·
      1 year ago

      Generally speaking, the way training works is this:
      You put together a folder of pictures, all the same size. It would’ve been 1024x1024 in this case. Other models have used 768z768 or 512x512. For every picture, you also have a text file with a description.

      The training software takes a picture, slices it into squares, generates a square the same size of random noise, then trains on how to change that noise into that square. It associates that training with tokens from the description that went with that picture. And it keeps doing this.

      Then later, when someone types a prompt into the software, it tokenizes it, generates more random noise, and uses the denoising methods associated with the tokens you typed in. The pictures in the folder aren’t actually kept by it anywhere.

      From the side of the person doing the training, it’s just put together the pictures and descriptions, set some settings, and let the training software do its work, though.

      (No money involved in this one. One person trained it and plopped it on a website where people can download loras for free…)