Knowledge in 

Generative Models

@ ECCV 2024

A workshop on intrinsic computer vision knowledge in generative models.


📅  Date: 30 September 2024 Afternoon
📍  Location: Brown 2, MiCo Milano, Italy

Overview

In this workshop, we will discuss how knowledge about the visual world is represented in modern generative models for images, videos, and 3D assets. Recent advances in generative modeling have been successful in creating rich, diverse and increasingly convincing photorealistic and stylized images. The workshop aims to investigate how these models internally represent and process visual information, and whether they bring us closer to fulfillment of the known mantra “vision is inverse graphics”. We seek to understand how well generative models, such as GANs, auto-regressive models, and diffusion models, comprehend semantic constructs that are commonly used to convey visual understanding, such as object recognition, scene understanding, spatial awareness, intrinsic image decomposition, and so on. Can this understanding be leveraged to solve inverse (recognition) problems? Can it be improved to further enhance generative models’ abilities? Is something important still missing from how our large models represent the visual world?


Core Themes


Invited Speakers

 Northeastern University


David Forsyth


University of Illinois Urbana-Champaign

UC Berkeley

Phillip Isola

MIT

Nanyang Technological University


Spotlight Talks

We have stopped accepting new submissions. We'll contact authors who have already filled the interest form very soon.


We are collecting interest for poster presentations at our workshop. If you are interested in presenting your work, please fill out this Google form. We will follow up on the status of the poster presentations in the coming months.


📧  Contact: Anand Bhattad (bhattad@ttic.edu)

Organizers

Grace Luo
UC Berkeley

Shuang Li

University of Toronto & Vector Institute

If you find our workshop interesting, you might also like Generative Models for Computer Vision.