Skip to content

Commit

Permalink
add new papers
Browse files Browse the repository at this point in the history
  • Loading branch information
greatzh committed Apr 4, 2024
1 parent 1fa6f01 commit f3b764e
Show file tree
Hide file tree
Showing 2 changed files with 47 additions and 5 deletions.
18 changes: 14 additions & 4 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -43,11 +43,16 @@ description: >-
<details open>
<summary>2024</summary>

* [ ] PROMPT-IML: Image Manipulation Localization with Pre-trained Foundation Models Through Prompt Tuning _(arXiv '24)_ [**[Paper]**](https://arxiv.org/abs/2401.00653)
* [ ] Diffusion models meet image counter-forensics _(WACV '24)_ **[[Paper](https://arxiv.org/abs/2311.13629)]** **[[Code](https://github.com/mtailanian/diff-cf)]**
* [ ] Research about the Ability of LLM in the Tamper-Detection Area _(arXiv '24)_ [**[Paper]**](https://arxiv.org/abs/2401.13504)
* [ ] Deep Image Composition Meets Image Forgery _(arXiv '24)_ [**[Paper]**](https://arxiv.org/abs/2404.02897) **[[Code](https://github.com/99eren99/DIS25k)]**
* [ ] Fusion Transformer with Object Mask Guidance for Image Forgery Analysis **[[Paper](https://arxiv.org/abs/2403.12229)]**
* [ ] Exploring Multi-Modal Fusion for Image Manipulation Detection and Localization _(MMM '24)_ [![arXiv](https://img.shields.io/badge/arXiv-b31b1b.svg)](https://arxiv.org/abs/2312.01790) [![GitHub](https://img.shields.io/github/stars/idt-iti/mmfusion-iml?style=flat)](https://github.com/idt-iti/mmfusion-iml)
* [ ] A New Benchmark and Model for Challenging Image Manipulation Detection _(AAAI '24)_ [![arXiv](https://img.shields.io/badge/arXiv-b31b1b.svg)](https://arxiv.org/abs/2311.14218)
* [ ] MGQFormer: Mask-Guided Query-Based Transformer for Image Manipulation Localization _(AAAI '24)_ [![arXiv](https://img.shields.io/badge/News-4096ff.svg)](https://dml.fudan.edu.cn/d1/65/c35285a643429/page.htm)
* [ ] Learning Discriminative Noise Guidance for Image Forgery Detection and Localization _(AAAI '24)_
* [ ] CatmullRom Splines-Based Regression for Image Forgery Localization _(AAAI '24)_
* [ ] MGQFormer: Mask-Guided Query-Based Transformer for Image Manipulation Localization _(AAAI '24)_ [![arXiv](https://img.shields.io/badge/News-4096ff.svg)](https://dml.fudan.edu.cn/d1/65/c35285a643429/page.htm) **[[Paper](https://ojs.aaai.org/index.php/AAAI/article/view/28520)]**
* [ ] Learning Discriminative Noise Guidance for Image Forgery Detection and Localization _(AAAI '24)_ **[[Paper](https://ojs.aaai.org/index.php/AAAI/article/view/28608)]**
* [ ] CatmullRom Splines-Based Regression for Image Forgery Localization _(AAAI '24)_ **[[Paper](https://ojs.aaai.org/index.php/AAAI/article/view/28548)]**
* [ ] UnionFormer: Unified-Learning Transformer with Multi-View Representation for Image Manipulation Detection and Localization _(CVPR '24)_
* [ ] Towards Modern Image Manipulation Localization: A Large-Scale Dataset and Novel Methods _(CVPR '24)_
* [ ] EditGuard: Versatile Image Watermarking for Tamper Localization and Copyright Protection _(CVPR '24)_ [![arXiv](https://img.shields.io/badge/arXiv-b31b1b.svg)](https://arxiv.org/abs/2312.08883) [![GitHub](https://img.shields.io/github/stars/xuanyuzhang21/EditGuard?style=flat)](https://github.com/xuanyuzhang21/EditGuard)
Expand All @@ -60,7 +65,8 @@ description: >-
<details open>
<summary>2023</summary>

* [ ] PL-GNet: Pixel Level Global Network for detection and localization of image forgeries _(IMAGE '23)_ **[[Paper](https://www.sciencedirect.com/science/article/pii/S092359652300111X)]** **[[Code](https://github.com/znshi/PL-GNet)]**
* [ ] A survey on deep learning-based image forgery detection _(PR '23)_ **[[Paper](https://www.sciencedirect.com/science/article/pii/S0031320323004764)]**
* [ ] PL-GNet: Pixel Level Global Network for detection and localization of image forgeries _(IMAGE '23)_ [![paper](https://img.shields.io/badge/ScienceDirect-ff6a5a.svg)](https://www.sciencedirect.com/science/article/pii/S092359652300111X) [![GitHub](https://img.shields.io/github/stars/znshi/PL-GNet?style=flat)](https://github.com/znshi/PL-GNet)
* [ ] Pixel-Inconsistency Modeling for Image Manipulation Localization _(arXiv '23)_ **[[Paper](https://arxiv.org/abs/2310.00234)]**
* [ ] Progressive Feedback-Enhanced Transformer for Image Forgery Localization _(arXiv '23)_ **[[Paper](https://arxiv.org/abs/2311.08910)]** **[[Code(N/A)](https://github.com/multimediaFor/ProFact)]**
* [ ] Secondary Labeling A Novel Labeling Strategy for Image Manipulation Detection _(MM '23)_ **[[Paper](https://doi.org/10.1145/3581783.3613839)]**
Expand Down Expand Up @@ -168,6 +174,7 @@ _Some of the above papers also contain methods to detect tampered images generat
<details open>
<summary>2024</summary>

* [ ] Feature Aggregation and Region-Aware Learning for Detection of Splicing Forgery _(SPL '24)_ **[[Paper](https://ieeexplore.ieee.org/abstract/document/10378732/)]**
* [ ] Towards Effective Image Forensics via A Novel Computationally Efficient Framework and A New Image Splice Dataset _(arXiv '24)_ **[[Paper](https://arxiv.org/abs/2401.06998)]**

</details>
Expand Down Expand Up @@ -247,6 +254,7 @@ _Some of the above papers also contain methods to detect tampered images generat

**人脸篡改**,篡改方法以及检测问题

* [ ] MCS-GAN: A Different Understanding for Generalization of Deep Forgery Detection (_TMM '23_) **[[Paper](https://ieeexplore.ieee.org/document/10141892)]**
* [ ] Exposing the Deception: Uncovering More Forgery Clues for Deepfake Detection _(AAAI '24)_ **[[Paper](https://arxiv.org/abs/2403.01786)]** **[[Code](https://github.com/QingyuLiu/Exposing-the-Deception)]**
* [ ] Contrastive Learning for DeepFake Classification and Localization via Multi-Label Ranking _(CVPR '24)_
* [ ] Transcending Forgery Specificity with Latent Space Augmentation for Generalizable Deepfake Detection _(CVPR '24)_ **[[Paper](https://arxiv.org/abs/2311.11278)]**
Expand Down Expand Up @@ -275,6 +283,8 @@ _Some of the above papers also contain methods to detect tampered images generat

**复制移动篡改定位**问题

* [ ] Object-level Copy-Move Forgery Image Detection based on Inconsistency Mining (_arXiv '24_) **[[Paper](https://arxiv.org/abs/2404.00611)]**
* [ ] UCM-Net: A U-Net-Like Tampered-Region-Related Framework for Copy-Move Forgery Detection (_TMM '24_) **[[Paper](https://ieeexplore.ieee.org/document/10109179)]**
* [ ] An effective image copy-move forgery detection using entropy image _(arXiv '23)_ **[[Paper](https://arxiv.org/abs/2312.11793)]**
* [ ] CMFDFormer: Transformer-based Copy-Move Forgery Detection with Continual Learning _(arXiv '23)_ **[[Paper](https://arxiv.org/abs/2311.13263)]**
* [ ] An approach for copy-move image multiple forgery detection based on an optimized pre-trained deep learning model _(KBS '23)_ **[[Paper](https://www.sciencedirect.com/science/article/pii/S0950705123002587)]**
Expand Down
34 changes: 33 additions & 1 deletion image-forgery/2023/trufor.md
Original file line number Diff line number Diff line change
Expand Up @@ -4,14 +4,46 @@
![title](https://s2.loli.net/2023/06/10/XtwLjhaKW8YIGyc.png)

现在(2023)提出的篡改检测都必须能够有应对基于深度学习的篡改检测的能力,作者提出的方法是之前被广泛采用的,将RGB特征和noise-sensitive fingerprint,两个特征经Transformer的结构融合起来,能够实现对传统的图像编辑篡改(cheapfakes)和最近几年的基于深度学习(deepfakes)的篡改。不过作者这里使用的noise fingerprint是经过学习之后的noise fingerprint,它通过自我监督的方式,使用真实数据进行训练,学习嵌入和相机内部及外部处理相关的伪影artifact。
现在(2023)提出的篡改检测都必须能够有应对基于深度学习的篡改检测的能力,作者提出的方法是之前被广泛采用的,将RGB特征和noise-sensitive fingerprint,两个特征经Transformer的结构融合起来,能够实现对传统的图像编辑篡改(cheapfakes)和最近几年的基于深度学习(deepfakes)的篡改。不过作者这里使用的noise fingerprint是经过学习之后的noise fingerprint,它通过自我监督的方式,使用真实数据进行训练,学习嵌入和相机内部及外部处理相关的伪影artifact,所以与传统的 RGB + handcrafted feature 不一样。同时为了提供更可靠的检测结果,作者在给出localization mask之外还引入了 confidence map 来展示可能的误报。

![image-20240330101228161](https://s2.loli.net/2024/03/30/cXJMzf2qH4UYARk.png)

## 方法与实现

![framework](https://s2.loli.net/2023/06/10/AyzgIiMlUBXothR.png)

### 出发点

1. 有限的泛化能力,只针对某一种篡改类型,只针对某一种图片格式,只采用某一个low feature,在真实场景下的表现不佳。
2. 有限的鲁棒性
3. 检测能力不足,目前的方法都把定位当作重点,而没有重视检测,检测的结果往往就是根据定位的heatmap来计算,但是在现实世界里面,这样的检测准确性较差,误报高,应该考虑直接地解决检测任务。

### 概述

一样采用了fusion,但是使用了跨模态的特征校准来结合noise和RGB特征,而为了直接考虑检测这一问题,作者设计了一个**篡改检测模块**,通过输入异常分析图和置信度图来获得最后的检测结果。

展开来讲,输入 RGB 图片,通过一个**Noiseprint++ Extractor** 得到图片的noise fingerprint,然后将RGB图片和noise fingerprint一起输入到encoder网络种提取密集特征,再将提取到的特征分别输入到两个decoder网络种,一个用来提取异常map(localization mask),一个则是提取置信map(区分篡改区域的有效预测和随机异常);提取到的异常分析map将结合置信度map确定池化的权重,得到一个紧凑的描述符,通过篡改检测器后得到一个完整性的评分。

### Noiseprint++

![image-20240330142947467](https://s2.loli.net/2024/03/30/wFBj6shLvdSM7ca.png)

作者曾在其2018年的文章种提出过[基于深度学习的方法](https://github.com/grip-unina/noiseprint)来提取图片的noiseprint,当时采用的方法是使用原图来进行自监督训练,而没有考虑到相机之外的后处理所带来的鲁棒性问题;而这次提出的 Noiseprint++则是不仅考虑了相机内的流程,硬件处理信息(不同型号)等,也考虑到了相机外可能的篡改痕迹历史等,来提高可靠性。具体的实现如上图所示,在extractor中,作者采用了自监督的对比学习,与之前的方法一样,使用[DnCNN](https://github.com/cszn/DnCNN)网络架构,[InfoNCE](https://github.com/HobbitLong/SupContrast)对比损失,将来自相同相机模型,相同位置,相同编辑历史的patch输出之间的距离最小化。

作者为了确保Noiseprint++是在未经修改的图片上进行训练,对相机模型,手机图片的解析度,JPEG的压缩量化矩阵,元数据的统一等进行控制变量。训练中,在每个batch中包含160个64x64像素的patch,这160个patch是由4张不同的图片经5个相机模型拍摄得到,共20张,接着这20张图片再经过4种不同的编辑方式,随机的大小裁剪,压缩,对比度和亮度的调整,总共有512种可能的编辑结合方式。

![image-20240330172718888](https://s2.loli.net/2024/03/30/Y3xwVhPQcJeWz8L.png)

### 异常定位map



## 结果

## 总结

1. 作者表示该方法还无法检测完全由 ai 生成的图片;其次,训练的异常map和最后的检测得分是分开的,并不是端到端的训练;最后还提到需要在更多的由生成模型局部编辑的图片数据集上进行泛化性的评估。

## GitHub

Project page: https://grip-unina.github.io/TruFor/
Expand Down

0 comments on commit f3b764e

Please sign in to comment.