• NaibofTabr@infosec.pub
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    1 day ago

    Oh this is just nonsense. This isn’t “gatekeeping being an artist”. You want to be an artist? Great! learn some skills and make some art (you know, your own art, which you make yourself). And yes I know “all art is derivative”. That is entirely beside the point.

    Machine learning is a vacuum connected to a blender. It ingests information which it combines with statistical analyses and then predicts an output based on an algorithm generated from the statistical model. There is nothing “avant-garde” here because all it can do is regurgitate existing material which it has ingested. There’s no inspiration, it can’t make anything new, and it can only make any product by ripping off someone else’s work.

    • VintageGenious@sh.itjust.works
      link
      fedilink
      English
      arrow-up
      3
      ·
      1 day ago

      Sure the style isn’t new, but you can make it work in new pieces that didn’t exist before, you can also merge art styles and combine concepts not blended before. There have been many innovating art kinds from generative ai, like infinitely zooming pieces or beat-synced deformation of faces or working qr code art pieces, mix use of 3d modeling then controlnet to make custom scenes, many things too detailed to be done by a human in a reasonable time.

      • NaibofTabr@infosec.pub
        link
        fedilink
        English
        arrow-up
        3
        arrow-down
        1
        ·
        1 day ago

        We’re not talking about a “style”, we’re talking about producing finished work. The image generation models aren’t style guides, they output final images which are produced from the ingestion of other images as training data. The source material might be actual art (or not) but it is generally the product of a real person (because ML ingesting its own products is very much a garbage-in garbage-out system) who is typically not compensated for their work. So again, these generative ML models are ripoff systems, and nothing more. And no, typing in a prompt doesn’t count as innovation or creativity.

        • VintageGenious@sh.itjust.works
          link
          fedilink
          English
          arrow-up
          3
          arrow-down
          2
          ·
          17 hours ago

          Generative ai is not only prompting, which shows you don’t know. Who are you to decide what is creativity and innovation? Are you Mr Art?

          Anyway, it is not ingesting images and photobashing them into a final picture, that’s not how it works. It has no memory of training data images, instead it learned to generate images by trying and when similar to a training data image going more in that direction. So it has the ability to create in the same style, but the original images it doesn’t have them

          • NaibofTabr@infosec.pub
            link
            fedilink
            English
            arrow-up
            2
            arrow-down
            1
            ·
            5 hours ago

            I see, so your argument is that because the training data is not stored in the model in its original form, it doesn’t count as a copy, and therefore it doesn’t constitute intellectual property theft. I had never really understood what the justification for this point of view was, so thanks for that, it’s a bit clearer now. It’s still wrong, but at least it makes some kind of sense.

            If the model “has no memory of training data images”, then what effect is it that the images have on the model? Why is the training data necessary, what is its function?

            • barsoap@lemm.ee
              link
              fedilink
              English
              arrow-up
              2
              ·
              edit-2
              4 hours ago

              the training data is not stored in the model in its original form,

              It is not stored in the model, period. Same as you do not store the shape of the letters you’re reading right now, not even the words, but their overall meaning. Remembering the meaning of what I write here, you can then produce words and letters again and you might be close but even with this short paragraph you’ll find it very hard to make an exact replica. That’s because you did not store it in its original form, not even compressed, you re-encoded it using your own understanding of language, of the world, of everything.