There are a bunch of reasons why this could happen. First, it’s possible to “attack” some simpler image classification models; if you get a large enough sample of their outputs, you can mathematically derive a way to process any image such that it won’t be correctly identified. There have also been reports that even simpler processing, such as blending a real photo of a wall with a synthetic image at very low percent, can trip up detectors that haven’t been trained to be more discerning. But it’s all in how you construct the training dataset, and I don’t think any of this is a good enough reason to give up on using machine learning for synthetic media detection in general; in fact this example gives me the idea of using autogenerated captions as an additional input to the classification model. The challenge there, as in general, is trying to keep such a model from assuming that all anime is synthetic, since “AI artists” seem to be overly focused on anime and related styles…
My wife once hit me in front of my kids because she didn’t like my pointing out a double standard in how she was treating them. The one she was favoring recently started hitting the other one in a similar manner–basically just to silence her when she said something he didn’t like–and when I pointed out the similarity to my wife’s actions and suggested he had learned it from her she got mad and claimed that rather than hitting me she had “hit my hand away” which is a lie and she knows it. It is 100% classic spousal abuse and gaslighting, and yet due to the sheer size difference between us–I’m a foot taller–I feel ridiculous calling it that, and don’t want to find out what else my son learns is OK from his mom if I’m not around, so here I am still married to her, mostly trying to forget the abuse when it’s not actively happening. She’s been abusive, but I’m not really in any physical danger, so staying seems like the rational option in my situation… I imagine that’s relatively common among men.