Research in our group focuses on two intimately connected research threads: Robotics and Embodied AI. We investigate the synergistic relations of perception and action in embodied agents and build intelligent algorithms that give rise to general-purpose robot autonomy.
In Robotics, we develop methods and mechanisms that enable autonomous robots to reason about the real world through their senses, to flexibly perform a wide range of tasks, and to adaptively learn new tasks. To deploy general-purpose robot autonomy in the wild, we have to deal with the variability and uncertainty of the unstructured environments. We address this challenge by closing the perception-action loop using robot perception and learning techniques.
In Embodied AI, we build computational frameworks of embodied agents. In these frameworks, perception arises from an embodied agent’s active, situated, and skillful interactions in the open world; and its ability to make sense of the world through the lenses of perception, in turn, facilitates intelligent behaviors.
Our work draws theories and methods from robotics, machine learning, and computer vision, along with inspirations from human cognition, neuroscience, and philosophy, to solve open problems at the forefront of Robotics and AI. We are always looking out for talented members to join our group.
Talks and Tutorials
You can learn more about my recent research from my talks and tutorials.
Open-Source Software & Data
I devote effort to making scientific research more reproducible and making knowledge accessible to a broader population. Open-sourcing research software and datasets is one of my key practices. You can find open-source code and data out of my research in the Publications page or on my GitHub. I highlight some public resources below:
robosuite: modular simulation framework and benchmark for robot learning
RoboTurk: large-scale crowdsourced teleoperation dataset for robotic imitation learning
SURREAL: distributed reinforcement learning framework and robot manipulation benchmark
AI2-THOR: open-source interactive environments for embodied AI
Visual Genome: visual knowledge base that connects structured image concepts to language