• 0 Posts
  • 28 Comments
Joined 9 months ago
cake
Cake day: September 27th, 2023

help-circle




  • I hope you’re right because this article says they used a spray can.

    Which brings me back to the last point in my comment. I also hope I’m right but the two times I looked into it (right after the attack and before writing my comment) both came up with that result and also it seems that English Heritage came out today saying there was “No visible damage”.

    As I said, I’m not writing to defend the action, just pointing out that the OP article is, willfully or not, omitting certain aspects that could make JSO look a little bit better.


  • but we did damage a 5000-year-old monument

    As far as I could find out, they used orange cornflour that will just wash off the next time it rains. The most amount of damage anyone could seriously bring up was that it could harm/displace the lichen on the henge.

    That’s not to say that I specifically condone the action, but it’s a lot less bad than this article makes it sound. It’s the same with the soup attack on one of van Gogh’s painting, which had protective glass on it. So far all the JSO actions targeting cultural/historical things (at least the ones that made it to the big news) have been done in a way that makes them sound awful at first hearing, but intentionally did not actually damage the targeted cultural/historical thing.

    I think the biases of the journalist/news outlet/etc. are somewhat exposed by which parts they focus on and which they downplay or omit entirely.



  • so the names of the ai characters HAVE to be stored in game…

    Some games also generate names oh the fly based on rules. For example, KSP stitches names together based on a pre- and suffix and then rejects a few unfortunate possible combinations such as Dildo, prompting a reroll.

    I suspect with your game, they just fed it a dictionary of common words though without properly vetting it.




  • I think the humor is meant to be in the juxtaposition between “reference” in media contexts (e.g. “I am your father”) and “reference” in programming contexts and applying the latter context to the former one.

    What does “I’m your father” mean if the movie is jaws?

    I think the absurdity of that question is part of said humor. That being said, I didn’t find it funny either.


  • That was a response I got from ChatGPT with the following prompt:

    Please write a one sentence answer someone would write on a forum in a response to the following two posts:
    post 1: “You sure? If it’s another bot at the other end, yeah, but a real person, you recognize ChatGPT in 2 sentences.”
    post 2: “I was going to disagree with you by using AI to generate my response, but the generated response was easily recognizable as non-human. You may be onto something lol”

    It’s does indeed have an AI vibe, but I’ve seen scammers fall for more obvious pranks than this one, so I think it’d be good enough. I hope it fooled at least a minority of people for a second or made them do a double take.



  • It’s not as accurate as you’d like it to be. Some issues are:

    • It’s quite lossy.
    • It’ll do better on images containing common objects vs rare or even novel objects.
    • You won’t know how much the result deviates from the original if all you’re given is the prompt/conditioning vector and what model to use it on.
    • You cannot easily “compress” new images, instead you would have to either finetune the model (at which point you’d also mess with everyone else’s decompression) or do an adversarial attack onto the model with another model to find the prompt/conditioning vector most likely to create something as close as possible to the original image you have.
    • It’s rather slow.

    Also it’s not all that novel. People have been doing this with (variational) autoencoders (another class of generative model). This also doesn’t have the flaw that you have no easy way to compress new images since an autoencoder is a trained encoder/decoder pair. It’s also quite a bit faster than diffusion models when it comes to decoding, but often with a greater decrease in quality.

    Most widespread diffusion models even use an autoencoder adjacent architecture to “compress” the input. The actual diffusion model then works in that “compressed data space” called latent space. The generated images are then decompressed before shown to users. Last time I checked, iirc, that compression rate was at around 1/4 to 1/8, but it’s been a while, so don’t quote me on this number.

    edit: fixed some ambiguous wordings.




  • I think it’s much more likely whatever scraping they used to get the training data snatched a screenshot of the movie some random internet user posted somewhere. (To confirm, I typed “joaquin phoenix joker” into Google and this very image was very high up in the image results) And of course not only this one but many many more too.

    Now I’m not saying scraping copyrighted material is morally right either, but I’d doubt they’d just feed an entire movie frame by frame (or randomly spaced screenshots from throughout a movie), especially because it would make generating good labels for each frame very difficult.