Spaces:
Running
on
A10G
Running
on
A10G
## Get Started | |
1. Install ProPainter Dependencies | |
You can follow the [Dependencies and Installation](https://github.com/Luo-Yihang/ProPainter-pr/tree/dev_yihang#dependencies-and-installation) | |
2. Install Demo Dependencies | |
```shell | |
cd web-demos/hugging_face | |
# install python dependencies | |
pip3 install -r requirements.txt | |
# Run the demo | |
python app.py | |
``` | |
## Usage Guidance | |
* Step 1: Upload your video and click the `Get video info` button. | |
![Step 1](./assets/step1.png) | |
* Step 2: | |
1. *[Optional]* Specify the tracking period for the currently added mask by dragging the `Track start frame` or `Track end frame`. | |
2. Click the image on the left to select the mask area. | |
3. - Click `Add mask` if you are satisfied with the mask, or | |
- *[Optional]* Click `Clear clicks` if you want to reselect the mask area, or | |
- *[Optional]* Click `Remove mask` to remove all masks. | |
4. *[Optional]* Go back to step 2.1 to add another mask. | |
![Step 2](./assets/step2.png) | |
* Step 3: | |
1. Click the `Tracking` button to track the masks for the whole video. | |
2. *[Optional]* Select the ProPainter parameters if the `ProPainter Parameters` dropdown. | |
2. Then click `Inpainting` to get the inpainting results. | |
![Step 3](./assets/step3.png) | |
*You can always refer to the `Highlighted Text` box on the page for guidance on the next step!* | |
## Citation | |
If you find our repo useful for your research, please consider citing our paper: | |
```bibtex | |
@inproceedings{zhou2023propainter, | |
title={{ProPainter}: Improving Propagation and Transformer for Video Inpainting}, | |
author={Zhou, Shangchen and Li, Chongyi and Chan, Kelvin C.K and Loy, Chen Change}, | |
booktitle={Proceedings of IEEE International Conference on Computer Vision (ICCV)}, | |
year={2023} | |
} | |
``` | |
## License | |
This project is licensed under <a rel="license" href="./LICENSE">NTU S-Lab License 1.0</a>. Redistribution and use should follow this license. | |
## Acknowledgements | |
The project harnesses the capabilities from [Track Anything](https://github.com/gaomingqi/Track-Anything), [Segment Anything](https://github.com/facebookresearch/segment-anything), [Cutie](https://github.com/hkchengrex/Cutie), and [E2FGVI](https://github.com/MCG-NKU/E2FGVI). Thanks for their awesome works. | |