# UNIT **Repository Path**: BolinLi-s/UNIT ## Basic Information - **Project Name**: UNIT - **Description**: No description available - **Primary Language**: Unknown - **License**: Not specified - **Default Branch**: master - **Homepage**: None - **GVP Project**: No ## Statistics - **Stars**: 0 - **Forks**: 1 - **Created**: 2020-02-14 - **Last Updated**: 2024-10-17 ## Categories & Tags **Categories**: Uncategorized **Tags**: None ## README [![License CC BY-NC-SA 4.0](https://img.shields.io/badge/license-CC4.0-blue.svg)](https://raw.githubusercontent.com/NVIDIA/FastPhotoStyle/master/LICENSE.md) ![Python 2.7](https://img.shields.io/badge/python-2.7-green.svg) ## UNIT: UNsupervised Image-to-image Translation Networks ### License Copyright (C) 2018 NVIDIA Corporation. All rights reserved. Licensed under the CC BY-NC-SA 4.0 license (https://creativecommons.org/licenses/by-nc-sa/4.0/legalcode). ### Code usage - Please check out our [tutorial](TUTORIAL.md). - For multimodal (or many-to-many) image translation, please check out our new work on [MUNIT](https://github.com/NVlabs/MUNIT). ### What's new. - 05-02-2018: We now adapt [MUNIT](https://github.com/NVlabs/MUNIT) code structure. For reproducing experiment results in the NIPS paper, please check out [version_02 branch](https://github.com/mingyuliutw/UNIT/tree/version_02). - 12-21-2017: Release pre-trained synthia-to-cityscape image translation model. See [USAGE.md](TUTORIAL.md) for usage examples. - 12-14-2017: Added multi-scale discriminators described in the [pix2pixHD](https://arxiv.org/pdf/1711.11585.pdf) paper. To use it simply make the name of the discriminator COCOMsDis. ### Paper [Ming-Yu Liu, Thomas Breuel, Jan Kautz, "Unsupervised Image-to-Image Translation Networks" NIPS 2017 Spotlight, arXiv:1703.00848 2017](https://arxiv.org/abs/1703.00848) #### Two Minute Paper Summary [![](./docs/two-minute-paper.png)](https://youtu.be/dqxqbvyOnMY) (We thank the Two Minute Papers channel for summarizing our work.) #### The Shared Latent Space Assumption [![](./docs/shared-latent-space.png)](https://www.youtube.com/watch?v=nlyXoX2aIek) #### Result Videos More image results are available in the [Google Photo Album](https://photos.app.goo.gl/5x7oIifLh2BVJemb2). *Left: input.* **Right: neural network generated.** Resolution: 640x480 ![](./docs/snowy2summery.gif) *Left: input.* **Right: neural network generated.** Resolution: 640x480 ![](./docs/day2night.gif) ![](./docs/dog_breed.gif) ![](./docs/cat_species.gif) - [Snowy2Summery-01](https://youtu.be/9VC0c3pndbI) - [Snowy2Summery-02](https://youtu.be/eUBiiBS1mj0) - [Day2Night-01](https://youtu.be/Z_Rxf0TfBJE) - [Day2Night-02](https://youtu.be/mmj3iRIQw1k) - [Translation Between 5 dog breeds](https://youtu.be/3a6Jc7PabB4) - [Translation Between 6 cat species](https://youtu.be/Bwq7BmQ1Vbc) #### Street Scene Image Translation From the first row to the fourth row, we show example results on day to night, sunny to rainy, summery to snowy, and real to synthetic image translation (two directions). For each image pair, *left is the input image*; **right is the machine generated image.** ![](./docs/street_scene.png) #### Dog Breed Image Translation ![](./docs/dog_trans.png) #### Cat Species Image Translation ![](./docs/cat_trans.png) #### Attribute-based Face Image Translation ![](./docs/faces.png)