Neural scene representation and rendering

S. M. Ali Eslami, Danilo Jimenez Rezende, Frederic Besse, Fabio Viola, Ari S. Morcos, Marta Garnelo, Avraham Ruderman, Andrei A. Rusu, Ivo Danihelka, Karol Gregor, David P. Reichert, Lars Buesing, Theophane Weber, Oriol Vinyals, Dan Rosenbaum, Neil Rabinowitz, Helen King, Chloe Hillier, Matt Botvinick, Daan WierstraKoray Kavukcuoglu, Demis Hassabis

Research output: Contribution to journalArticlepeer-review

Abstract

Scene representation—the process of converting visual sensory data into concise descriptions—is a requirement for intelligent behavior. Recent work has shown that neural networks excel at this task when provided with large, labeled datasets. However, removing the reliance on human labeling remains an important open problem. To this end, we introduce the Generative Query Network (GQN), a framework within which machines learn to represent scenes using only their own sensors. The GQN takes as input images of a scene taken from different viewpoints, constructs an internal representation, and uses this representation to predict the appearance of that scene from previously unobserved viewpoints. The GQN demonstrates representation learning without human labels or domain knowledge, paving the way toward machines that autonomously learn to understand the world around them.

Original languageEnglish
Pages (from-to)1204-1210
Number of pages7
JournalScience
Volume360
Issue number6394
DOIs
StatePublished - 15 Jun 2018
Externally publishedYes

Bibliographical note

Publisher Copyright:
2017 © The Authors

ASJC Scopus subject areas

  • General

Fingerprint

Dive into the research topics of 'Neural scene representation and rendering'. Together they form a unique fingerprint.

Cite this