With the rapid advancement of generative models, powerful image editing methods now enable diverse and highly realistic image manipulations that far surpass traditional deepfake techniques, posing new challenges for manipulation detection. Existing image manipulation detection and localization (IMDL) benchmarks suffer from limited content diversity, narrow generative-model coverage, and insufficient interpretability, which hinders the generalization and explanation capabilities of current manipulation detection methods. To address these limitations, we introduce \textbf{ManipBench}, a large-scale benchmark for image manipulation detection and localization focusing on AI-edited images. ManipBench contains over 450K manipulated images produced by 25 state-of-the-art image editing models across 12 manipulation categories, among which 100K images are further annotated with bounding boxes, judgment cues, and textual explanations to support interpretable detection. Building upon ManipBench, we propose \textbf{ManipShield}, an all-in-one model based on a Multimodal Large Language Model (MLLM) that leverages contrastive LoRA fine-tuning and task-specific decoders to achieve unified image manipulation detection, localization, and explanation. Extensive experiments on ManipBench and several public datasets demonstrate that ManipShield achieves state-of-the-art performance and exhibits strong generality to unseen manipulation models. Both ManipBench and ManipShield will be released upon publication.
翻译:随着生成模型的快速发展,强大的图像编辑方法现已能够实现多样且高度逼真的图像篡改,其效果远超传统的深度伪造技术,为篡改检测带来了新的挑战。现有的图像篡改检测与定位(IMDL)基准存在内容多样性有限、生成模型覆盖范围狭窄以及可解释性不足等问题,这制约了当前篡改检测方法的泛化与解释能力。为应对这些局限,我们引入了\textbf{ManipBench},一个专注于AI编辑图像的大规模图像篡改检测与定位基准。ManipBench包含超过45万张由25个最先进的图像编辑模型生成的篡改图像,涵盖12种篡改类别,其中10万张图像进一步标注了边界框、判断线索和文本解释,以支持可解释的检测。基于ManipBench,我们提出了\textbf{ManipShield},一个基于多模态大语言模型(MLLM)的一体化模型,它利用对比LoRA微调和任务特定解码器,实现了统一的图像篡改检测、定位与解释。在ManipBench及多个公开数据集上的大量实验表明,ManipShield达到了最先进的性能,并对未见过的篡改模型展现出强大的泛化能力。ManipBench和ManipShield均将在发表后开源。