The highly structured nature of the visual world has inspired many computer vision researchers to pursue an analysis-by-synthesis approach: to understand an image, one should be able to reproduce it with a model. A good model should also be able to extrapolate into unseen space or time: given a 2D or 2.5D image of a partially occluded object, what is the full 3D extent? Given a fragment of a video, how does the remainder play out? Generative models of images, video, and 3D data have made great strides in recent years, but their utility as causal or interpretable models has not always advanced in step. For example, while GANs can currently generate beautiful images, they do not necessarily learn a latent space of graphics-like or semantically-interpretable elements. In this workshop, we aim to explore how generative models can facilitate perception, and in particular, how to design and use structured generative models (of images, video, and 3D data) for computer vision inference applications.
We are soliciting original contributions in computer vision, robotics, and machine learning relating to the following topics:
Submission deadline: August 20. We encourage authors to submit earlier, anytime between August 1 and August 20, to help us spread out the reviewing work.
Submit your paper to our OpenReview site, using the ECCV 2020 "final copy" latex kit.
Submissions should be 4 pages long, including references. The 4-page limit helps eliminate dual-submission conflicts with ECCV and other conferences. (E.g., even papers accepted to ECCV may be dual-submitted here, provided that they are shortened to 4 pages.)
The workshop organizers will review the papers in a single-blind fashion. All accepted papers will be included in a poster presentation session. Accepted papers will be published in the proceedings.
Contact: Adam Harley (aharley@cmu.edu).
Website theme based on Scene Graph Representation and Learning.