To develop machines that have the following capabilities.
Achieve deep understanding of scenes and events through joint parsing and cognitive reasoning about appearance, geometry, functions, physics, causality, intents and belief of agents, and use joint and long-ranged reasoning to fill the performance gap with human vision.
Represent visual knowledge in probabilistic compositional models across the spatial, temporal, and causal hierarchies augmented with rich relations, which are task-oriented, support efficient task-dependent inference from an agent’s perspective, and preserve uncertainties.
Acquire massive visual commonsense through web scale continuous lifelong learning from heterogeneous sources through weakly supervised HCI and dialogue with humans.
Understand human needs and values to interact with humans effectively and answer human queries about what, who, where, when, why and how in storylines through Turing tests.
Song-Chun Zhu
|
Martial Hebert
|
|
Abhinav Gupta
|
|
Nancy Kanwisher
|
|
Josh Tenenbaum
|
|
Fei-Fei Li
|
|
David Forsyth
|
|
Derek Hoiem
|
|
Brian Scholl
|
Philip Torr
|
|
Philippe Schyns
|
|
Andrew Glennerster
|
|
Ales Leonardis
|
Address: 425 Westwood Plaza, Los Angeles, CA 90095
Recommend: Uber, Super Shuttle, FlyAway
Click the upper-right button to view larger map, upper-left button to show all the places
Patrick Tam (patrick.tam@stat.ucla.edu): Hotel and Direction
Yixin Zhu (yixin.zhu@ucla.edu): Schedule