ShineOn: Illuminating Design Choices for Practical Video-based Virtual Clothing Try-on
This repository contains the code for our paper accepted at the
Generation of Human Behavior Workshop at WACV 2021.
- Scientific experiments built from the ground-up to isolate effects of each method
- Empirically show DensePose results in better quality than CocoPose
- Add self-attention layers
- Find that GeLU show best results
How To Use This Repository
The point of entry of this repository is
test.py. We have organized
our code into these main folders:
datasets folder contains several custom defined datasets. To create your own custom tryon dataset,
please refer to the Documentation IV below.
models folder contains several models, such as the warp model and U-Net model
that we used during virtual try-on work. Inside the
networks sub-folder, we include several
utility networks that we make use of.
options folder contains several of the options we use at train and test time.
These options allows our code to flexible, and run experiments easily.
Qualitative Comparison with FW-GAN and CP-VTON
Qualitative Comparison of Pose and Self-Attention
Qualitative Comparison of Activation Functions
Qualitative Comparison of Optical Flow
Acknowledgements and Related Code
- This code is based in part on Sergey Wong's stellar CP-VTON repository. Thank you very much,
Sergey, for your hard work.
- Thank you Haoye Dong and his team for hosting the VUHCS competition at CVPR 2020,
providing the VVT Dataset, and giving access to the FW-GAN reference code.
- Thank you NVIDIA's team for their work on Vid2Vid and FlowNet2.
- Credits to David Park's Self-Attention GAN implementation for attention layers reference.
- Credits to Self-Corrective Human-Parsing for easy parsing of LIP clothing labels.
- Credits to the detectron2 repository for Densepose annotations.