Skip to content

Latest commit

 

History

History
35 lines (30 loc) · 2.07 KB

README.md

File metadata and controls

35 lines (30 loc) · 2.07 KB

Worlds first API for Deep Virtual Try On cloth powered by Pytorch

made-with-python license

Note

Currently it is on research stage and I am improving the code. Soon I will release the notebook for training and inference.

API

alt text

Android Application(Adobe XD wireframe)

alt text

Results

alt text alt text alt text alt text alt text alt text

Training

Download the dataset

Download the MPV dataset from Image-based Multi-pose Virtual Try On and put the dataset under "./dataset/images/".

Select postive perspective images, create dataset split file 'data_pair.txt', and put it under "./dataset/".

Dataset preprocessing

Pose keypoints. Use the Openpose, and put the keypoints file in "./dataset/pose_coco".

Semantic parsing. Use the CIHP_PGN, and put the parsing results in "./dataset/parse_cihp".

Cloth mask. Use the removebg api or holistically-nested-edge-detection for the cloth mask, and put the mask in "./dataset/cloth_mask"..

Coarse-to-fine training

Download the VGG19 pretrained checkpoint

cd vgg_model/

Set different configuration based on the "config.py". Then run

sh train.sh