metadata
license: apache-2.0
A Task is Worth One Word: Learning with Task Prompts for High-Quality Versatile Image Inpainting
Project Page | Paper | Online Demo(OpenXlab)
This README provides a step-by-step guide to download the repository, set up the required virtual environment named "PowerPaint" using conda, and run PowerPaint with or without ControlNet.
Feel free to try it and give it a star!:star:
π News
May 22, 2024:fire:
April 7, 2024:fire:
April 6, 2024:
- We have retrained a new PowerPaint, taking inspiration from Brushnet. The Online Demo has been updated accordingly. We plan to release the model weights and code as open source in the next few days.
- Tips: We preserve the cross-attention layer that was deleted by BrushNet for the task prompts input.
December 22, 2023:wrench:
- The logical error in loading ControlNet has been rectified. The
gradio_PowerPaint.py
file and Online Demo have also been updated.
December 18, 2023
Enhanced PowerPaint Model
- We are delighted to announce the release of more stable model weights. These refined weights can now be accessed on Hugging Face. The
gradio_PowerPaint.py
file and Online Demo have also been updated as part of this release.
Getting Started
# Clone the Repository
git clone https://github.com/zhuang2002/PowerPaint.git
# Navigate to the Repository
cd projects/powerpaint
# Create Virtual Environment with Conda
conda create --name PowerPaint python=3.9
conda activate PowerPaint
# Install Dependencies
pip install -r requirements.txt
PowerPaint v2
python gradio_PowerPaint_BrushNet.py
PowerPaint v1
# Create Models Folder
mkdir models
# Set up Git LFS
git lfs install
# Clone PowerPaint Model
git lfs clone https://huggingface.co/JunhaoZhuang/PowerPaint-v1/ ./models
python gradio_PowerPaint.py
This command will launch the Gradio interface for PowerPaint.
Feel free to explore and edit images with PowerPaint!
BibTeX
@misc{zhuang2023task,
title={A Task is Worth One Word: Learning with Task Prompts for High-Quality Versatile Image Inpainting},
author={Junhao Zhuang and Yanhong Zeng and Wenran Liu and Chun Yuan and Kai Chen},
year={2023},
eprint={2312.03594},
archivePrefix={arXiv},
primaryClass={cs.CV}
}