Recent Advances in Unsupervised Image-to-image Tranformation

 Today's talk is by Xun Huang from Cornell University (and Nvidia) on 'Recent Advances in Unsupervised Image to Image Translation' and was given at Microsoft Research in Sept 2019. 

Note that for the title of the post we're using HTC preferred terminology of Transformation for this subject area, since Translation is typically thought of as an affine transformation.

The project page for the talk is available here.  Here's a link to some of Xun's publications.  He also has a personal page with a lot of info on it.

Abstract for this talk:
Unsupervised image-to-image translation aims to map an image drawn from one distribution to an analogous image in a different distribution, without seeing any example pairs of analogous images. For example, given an image of a landscape taken in the summer, one may want to know what it would look like in the winter. There is not just a single answer. One could imagine many possibilities due to differences in weather, timing, lighting, etc.

1:  What he is really covering in this talk is why they are interested in the Unit GAN architecture in stead of the CycleGAN architecture.  He is not the world's greatest speaker, and the small Microsoft audience keep interrupting him with particularly stupid questions, but what he has to say when he gets around to it is interesting.

2: Here are some associated references for the talk.

Here's his paper on MultiModal Unsupervised Image-to-Image Translation (MUNIT).

Original Unsupervised Image-to-Image Translation Networks (UNIT) Nvidia paper is here.

Proposes coupled GANs architecture.

Here's a comparison of CycleGan and Unit github project.

From the following paper.

3:  On a totally different topic, his PointFlow paper on 3D Point Cloud generation is pretty interesting.


Popular posts from this blog

Pix2Pix: a GAN architecture for image to image transformation

CycleGAN: a GAN architecture for learning unpaired image to image transformations

Smart Fabrics