Project Summary

Our project’s goal is to implement an image-to-image translation, in which a screenshot of the blocks-based world in Minecraft could be converted into an actual photo composed by items and scenes similar to those in the real world. We use CycleGAN, an approach for learning to map an image from the source domain of Minecraft to one from the target domain of actual environments. We would consider the translation to be “successful” if generated pictures are composed of recognizable items and scenes as same as those in input image but with smooth edges and authentic textures. By implementing the translation, Minecraft users are able to enjoy the creating or survival experience in a more realistic scene. It could be applied to other industries as well, for example, blueprint design.

Approach

In this project, we use CycleGAN, which was introduced in the paper from UC Berekey, “Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks”.

Model Description

CycleGAN is a Generative Adversarial Network (GAN) that uses two generators and two discriminators. Let’s denote that is the domain of Minecraft images and is the domain of real-world images. The generator converts images from domain to domain , and the generator converts images from domain to domain :

And each generator has a corresponding discriminator which distinguishes real images from fake ones:

Objective Function

Implementation

We use the network architectures described in the original paper. We also follow some parts of the authors’ training procedures to train our model.

Evaluation

Remaining Goals and Challenges

  1. We are not very satistied with the quality of the generated image. And we are looking for further improvement. Since currently we build the model from scratch, we plan to use the implementation provided by the authors to compare with the result. Another reason of the shortage might be the limitation of the model mentioned in the original paper. We will look for another state-of-the-art model to meet our goals.

  2. We are still trying to generate the real-time transformation if time permits.

References

  1. Jun-Yan Zhu, Taesung Park, Phillip Isola, Alexei A. Efros. Unpaired Image-to-Image Translation using Cycle-Consistent Adversarial Networks

  2. PyTorch 1.1: https://pytorch.org

  3. Urban and Natural Scene Datasets: http://cvcl.mit.edu/database.htm