Content creation plays a crucial role in domains such as photography, videography, virtual reality, gaming, art, design, fashion, and advertising design. Recent progress in machine learning and AI has transformed hours of manual, painstaking content creation work into minutes or seconds of automated or interactive work. For instance, generative modeling approaches can produce photorealistic images of 2D and 3D items such as humans, landscapes, interior scenes, virtual environments, clothing, or even industrial designs. New large text, image, and video models that share latent spaces let us imaginatively describe scenes and have them realized automatically—with new multi-modal approaches able to generate consistent video and audio across long timeframes. Such approaches can also super-resolve and super-slomo videos, interpolate and extrapolate between photos and videos with intermediate novel views, decompose scene objects and appearance, and transfer styles to convincingly render and reinterpret content. Learned priors of images, videos, and 3D data can also be combined with explicit appearance and geometric constraints, perceptual understanding, or even functional and semantic constraints of objects. While often creating awe-inspiring artistic images, such techniques offer unique opportunities for generating diverse synthetic training data for downstream computer vision tasks, both in 2D, video, and 3D domains.
The AI for Content Creation workshop explores this exciting and fast-moving research area. We bring together invited speakers of world-class expertise in content creation, up-and-coming researchers, and authors of submitted workshop papers, to engage in a day filled with learning, discussion, and network building.
Welcome! -
Deqing Sun (Google)
Lingjie Liu (University of Pennsylvania)
Krishna Kumar Singh (Adobe)
Lu Jiang (ByteDance)
Jun-Yan Zhu (Carnegie Mellon University)
James Tompkin (Brown University)
Time CDT | |||
---|---|---|---|
08:45 | Welcome and introductions | 👋 | |
09:00 | Maneesh Agrawala (Stanford University) |
|
|
09:30 | Kai Zhang (Adobe) | ||
10:00 | Coffee break | ☕ | |
10:30 | Charles Herrmann (Google) | ||
11:00 | Mark Boss (Stability AI) | ||
11:30 | Poster session 1 - ExHall D #412-431
|
||
12:30 | Lunch break - ExHall C | 🥪 |
Time CDT | |||
---|---|---|---|
13:30 | Oral session + best paper announcement + best presentation competition
|
||
14:00 | Yutong Bai (UC Berkeley) |
|
|
14:30 | Nanxuan (Cherry) Zhao (Adobe) | ||
15:00 | Coffee break | ☕ | |
15:30 | Ishan Misra (+Rohit Girdhar) (Meta) |
|
|
16:00 | Panel discussion — Open Source in AI and the Creative Industry
|
🗣️ | |
17:00 | Poster session 2 - ExHall D #412-431
|