GANcraft: Unsupervised 3D Neural Rendering of Minecraft Worlds
Zekun Hao, Arun Mallya, Serge Belongie, Ming-Yu Liu
4/15/2021
Keywords: Generalization, Generative Models, Data-Driven Method, Local Conditioning, Voxel Grid, Large-Scale Scenes
Venue: ICCV 2021
Bibtex:
@inproceedings{hao2021gancraft,
url = {http://arxiv.org/abs/2104.07659v1},
year = {2021},
title = {GANcraft: Unsupervised 3D Neural Rendering of Minecraft Worlds},
author = {Zekun Hao and Arun Mallya and Serge Belongie and Ming-Yu Liu},
booktitle = {Proceedings of the IEEE International Conference on Computer Vision (ICCV)},
entrytype = {inproceedings},
id = {hao2021gancraft}
}
Abstract
We present GANcraft, an unsupervised neural rendering framework for generating photorealistic images of large 3D block worlds such as those created in Minecraft. Our method takes a semantic block world as input, where each block is assigned a semantic label such as dirt, grass, or water. We represent the world as a continuous volumetric function and train our model to render view-consistent photorealistic images for a user-controlled camera. In the absence of paired ground truth real images for the block world, we devise a training technique based on pseudo-ground truth and adversarial training. This stands in contrast to prior work on neural rendering for view synthesis, which requires ground truth images to estimate scene geometry and view-dependent appearance. In addition to camera trajectory, GANcraft allows user control over both scene semantics and output style. Experimental results with comparison to strong baselines show the effectiveness of GANcraft on this novel task of photorealistic 3D block world synthesis. The project website is available at https://nvlabs.github.io/GANcraft/ .
Citation Graph
(Double click on nodes to open corresponding papers' pages)
(Double click on nodes to open corresponding papers' pages)
* Showing citation graph for papers within our database. Data retrieved from Semantic Scholar. For full citation graphs, visit ConnectedPapers.