Research Questions

How can human demonstrations and interactions be used to bootstrap the learning process?

Writing code is an extraordinarily labor-intersive way to provide robots with human knowledge and usually requires highly-trained specialists. Learning from demonstration is an alternate paradigm for gaining human insight though faster, more natural means like task demonstrations and interactive corrections. However, such time-series data is often difficult to interpret, requiring the ability to segment activities and behaviors, understand context, and generalize from a small number of examples. How can demonstrations best be interpreted to leverage human insight into complex tasks? What kinds of demonstrations are most effective? How can robots take advantage of multiple types of cues like natural language, gestures, and gaze?

How can robots autonomously improve their understanding of the world through embodied interaction?

Human demonstrations and interactions can provide a good baseline of knowledge, but do not necessarily cater to a robot's specific internal representations, uncertainties, and capabilities. Ideally, robots should directly reason about these factors and autonomously collect data to improve modeling and control of their environment. How can techniques like active learning or interactive perception be used to experiment intelligently? How can reinforcement learning algorithms best utilize robot experiences? How can we utilize large data sets that already exist, such as the recorded experiences of other robots, or language and video data on the web?

How can robots learn from heterogenous, noisy interactions and still provide strong probabilistic guarantees of correctness and safety?

Robot learning from demonstration and interaction has seen much practical success, but often cannot provide strong performance guarantees. To be deployed in many real-world situations, such learners must be able to provide strong probabilistic guarantees of safe and correct performance, especially when working in proximity to humans. How can lifelong learning algorithms provide guarantees that a safety-critical task will be performed correctly, such as disposing of a hazardous material? Similarly, how can tasks like cleaning up a dinner table be continually optimized, while guaranteeing that no unsafe situations (such as collisions or spills) will occur with high probability? How can safety-aware algorithms be made sample efficient enough to work on real robotics problems?



People

Faculty

Scott Niekum

Scott Niekum

Director, Assistant Professor

Students

team img

Akanksha Saran

PhD Student

team img

Ajinkya Jain

PhD Student

team img

Yuchen Cui

PhD Student

team img

Daniel Brown

PhD Student

team img

Wonjoon Goo

PhD Student

team img

Reymundo A. Gutierrez

PhD Student

team img

Caleb Chuck

PhD Student

team img

Prasoon Goyal

PhD Student

Lab affiliates

team img

Josiah Hanna

PhD Student

team img

Taylor Kessler Faulkner

PhD Student

team img

Srinjoy Majumdar

Undergraduate student

Lab alumni

team img

Joel Iventosch

MS student