CAMOUFLAGE_Light / README.md
MalumaDev's picture
Update README.md
038336a verified
# CAMOUFLaGE: Controllable AnoniMizatiOn throUgh diFfusion-based image coLlection GEneration
Code [Here](https://gitlab.com/grains2/camouflage)
Official implementations of ["Latent Diffusion Models for Attribute-Preserving Image Anonymization"](#latent-diffusion-models-for-attribute-preserving-image-anonymization)
and ["Harnessing Foundation Models for Image Anonymization"](#harnessing-foundation-models-for-image-anonymization).
## Latent Diffusion Models for Attribute-Preserving Image Anonymization
[[Paper]](https://arxiv.org/abs/2403.14790)
This paper presents, to the best of our knowledge, the first approach to image anonymization based on
Latent Diffusion Models (LDMs). Every element of a scene is maintained to convey the same meaning, yet
manipulated in a way that makes re-identification difficult. We propose two LDMs for this purpose:
- *CAMOUFLaGE-Base*
- *CAMOFULaGE-Light*
The former solution achieves superior performance on most metrics and benchmarks, while the latter cuts
the inference time in half at the cost of fine-tuning a lightweight module.
Compared to state-of-the-art, we anonymize complex scenes by introducing variations in the faces, bodies,
and background elements.
#### CAMOUFLaGE-Base
CAMOUFLaGE-Base exploits a combination of pre-trained ControlNets and introduces an anonymizazion guidance based on
the original image.
![Architecture_Base](images/camouflage-base.jpg)
More details on its usage can be found [here](CAMOUFLaGE-Base-v1-0).
#### CAMOUFLaGE-Light
CAMOUFLaGE-Light trains a lightweight IP-Adapter to encode key elements of the scene and facial attributes of each
person.
![Architecture_Light](images/camouflage-light.jpg)
More details on its usage can be found [here](CAMOUFLaGE_light).
## Harnessing Foundation Models for Image Anonymization
[[Paper]]()
We explore how foundation models can be leveraged to solve tasks, specifically focusing on anonymization,
without the requirement for training or fine-tuning. By bypassing traditional pipelines, we demonstrate the
efficiency and effectiveness of this approach in achieving anonymization objectives directly from the
foundation model’s inherent knowledge.
#### CAMOUFLaGE-Face
We examine how foundation models can generate anonymized images directly from textual descriptions. Two models
were employed for information extraction: FACER, used to identify the 40 CelebA-HQ attributes, and DeepFace,
used to determine ethnicity and age. Using this rich information, we craft captions to guide the generation process.
Classifier-free guidance was employed to push the image content in the direction of the positive prompt P and far
from the negative prompt ¬P.
![Architecture-Face](images/camouflage-face.jpg)
More details on its usage can be found [here](GEM2024).
## Citation
If you find CAMOUFLaGE-Base and/or CAMOUFLaGE-Light useful, please cite:
```
@misc{camouflage,
title={Latent Diffusion Models for Attribute-Preserving Image Anonymization},
author={Luca Piano and Pietro Basci and Fabrizio Lamberti and Lia Morra},
year={2024},
eprint={2403.14790},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
```
If you find CAMOUFLaGE-Face useful, please cite:
```
@inproceedings{pianoGEM24,
title={Harnessing Foundation Models for Image Anonymization},
author={Piano, Luca and Basci, Pietro and Lamberti, Fabrizio and Morra, Lia},
booktitle={2024 IEEE CTSoc Gaming, Entertainment and Media},
year={2024},
organization={IEEE}
}
```