Paper #1 - M2E-Try On Net: Fashion from Model to Everyone
Key Notes
Texture refinement network (TRN) to enrich the textures and logo patterns to the desired clothes
Texture details, region of texture, binary mask with the same size. Merged images while still preserving the textual details on the garments
Fitting network (FTN) to merge the transferred garments to the target person images. Generative network to generate fashion images from textual inputs
Fitting Network is an encoder-decoder network, including three convolution layers as the encoder, six residual blocks for feature learning, followed by two deconvolution layers and one convolution layer as the decoder
Code - https://github.com/shionhonda/viton-gan
Paper #2 - VITON-GAN: Virtual Try-on Image Generator Trained with Adversarial Loss
Key Notes
Try-on module (TOM)
Key Notes
Keep Thinking!!!
Key Notes
- Pose alignment network (PAN)
- Texture refinement network (TRN)
- Fitting network (FTN)
Texture refinement network (TRN) to enrich the textures and logo patterns to the desired clothes
Texture details, region of texture, binary mask with the same size. Merged images while still preserving the textual details on the garments
Fitting network (FTN) to merge the transferred garments to the target person images. Generative network to generate fashion images from textual inputs
Fitting Network is an encoder-decoder network, including three convolution layers as the encoder, six residual blocks for feature learning, followed by two deconvolution layers and one convolution layer as the decoder
Code - https://github.com/shionhonda/viton-gan
Paper #2 - VITON-GAN: Virtual Try-on Image Generator Trained with Adversarial Loss
Key Notes
- U-net generator and thin plate spline (TPS)
- Human parser
- Pose estimator
Try-on module (TOM)
- Trained adversarially against the discriminator that uses the TOM result image
- Person representation as inputs and judges whether the result is real or fake VITON-GAN generated hands and arms
Key Notes
- GAN - a model G that learns a data distribution from example data, and a discriminator D that attempts to distinguish generated from training data
- These models learn loss functions that adapt to the data, and this makes them perfectly suitable for image-to-image translation tasks
- (cGAN) learns to generate images as function of conditioning information from a dataset, instead of random noise from a prior, as in standard GAN
- does an image x look reasonable, i.e. indistinguishable from the training distribution of human images {xi}?
- does the article y look well-painted on the human model image x
Keep Thinking!!!
No comments:
Post a Comment