• nicetriangle@kbin.social
    link
    fedilink
    arrow-up
    10
    ·
    1 year ago

    If it works anything like the other supposed AI image protector tool I’m aware of (Glaze) then it’s not gonna look great and I would not call it a practical way to go. Everything I’ve seen run through glaze looks objectively worse than the original.

    Also in the long run this is just an arms race and it’s just a matter of time before models learn to subvert these kinds of tools. And if that’s the case that means every time someone figures out how to get over these hurdles, anyone looking to protect their images will have to go back and replace every online instance of those images when the protection tool comes out with a fix. Back and forth forever.

    And that’s just ridiculous and basically impossible when you realize that stuff gets reposted all over the net all the time and can’t be controlled.

    • FaceDeer@kbin.social
      link
      fedilink
      arrow-up
      14
      ·
      edit-2
      1 year ago

      every time someone figures out how to get over these hurdles, anyone looking to protect their images will have to go back and replace every online instance of those images when the protection tool comes out with a fix.

      And if those older versions got downloaded and saved by a trainer there’s nothing at all they can do to replace those.

      This all feels a lot like the DRM treadmill, which has never done much to actually prevent piracy. Just made things annoying for everyone else.

    • hypnicjerk@lemmy.world
      link
      fedilink
      English
      arrow-up
      6
      ·
      1 year ago

      Zhao’s team also developed Glaze,

      from the article, so it’s likely they run on similar principles.