Abstract:
Modern generative models produce images that are virtually indistinguishable from human-created ones, posing serious challenges for content verification. As machine-generated content is increasingly integrated into professional workflows, the task of reliably detecting such content becomes critically important. Existing detectors of machine-generated images do not generalize well to new generative models and visual domains. This work investigates the ability of current detectors of machine-generated images to recognize new generative models and images from different domains not represented in the training data. The objects of study include popular architectures, such as a combination of pre-trained CLIP with an MLP classifier and a model based on a mixture of experts. Particular attention is paid to analyzing current limitations and the reliability of both closed and open solutions, especially in the context of emerging new generative methods and specific types of images. Experimental results demonstrate significant limitations of existing approaches: models exhibit low generalization ability not only to new generators but also when working with images from new domains.