X-Fields: Implicit Neural View-, Light- and Time-Image Interpolation

ACM Transactions on Graphics (SIGGRAPH Asia 2020)

Mojtaba Bemana1     Karol Myszkowski1     Hans-Peter Seidel1     Tobias Ritschel2

1 MPI Informatik    2 University College London

The processing pipeline.

Abstract

We suggest to represent an X-Field —a set of 2D images taken across different view, time or illumination conditions, i.e., video, light field, reflectance fields or combinations thereof—by learning a neural network (NN) to map their view, time or light coordinates to 2D images. Executing this NN at new coordinates results in joint view, time or light interpolation. The key idea to make this workable is a NN that already knows the “basic tricks” of graphics (lighting, 3D projection, occlusion) in a hard-coded and differentiable form. The NN represents the input to that rendering as an implicit map, that for any view, time, or light coordinate and for any pixel can quantify how it will move if view, time or light coordinates change (Jacobian of pixel position with respect to view, time, illumination, etc.). Our X-Field representation is trained for one scene within minutes, leading to a compact set of trainable parameters and hence real-time navigation in view, time and illumination.

Materials

Gallery

Video


Citation

Mojtaba Bemana, Karol Myszkowski, Hans-Peter Seidel, Tobias Ritschel
X-Fields: Implicit Neural View-, Light- and Time-Image Interpolation
To appear in: ACM Transactions on Graphics (Proc. SIGGRAPH Asia 2020)

@article{Bemana2020xfields,
	author	= {Bemana, Mojtaba and Myszkowski, Karol and Seidel, Hans-Peter and Ritschel, Tobias},
	title	= {X-Fields: Implicit Neural View-, Light- and Time-Image Interpolation},
	journal	= {ACM Transactions on Graphics (Proc. SIGGRAPH Asia 2020)},
	year	= {2020},
	volume = {39},
	number = {6},
	doi = {10.1145/3414685.3417827}
	}