Preference Learning from Minimal Human Feedback for Interactive Autonomy
The lack of large robotics datasets is arguably the most important obstacle in front of robot learning and interactive autonomy. While large pretrained models and algorithms like reinforcement learning from human feedback (RLHF) led to breakthroughs in other domains like natural language processing and computer vision, robotics has not experienced such a significant breakthrough due to the excessive cost of collecting large datasets. In this talk, I will discuss techniques that enable us to train robots from very little human feedback. I will dive into reinforcement learning from human feedback and describe how active learning methods can enable us to make it more data-efficient. I will finally propose an alternative type of human feedback based on language corrections to further improve both data-efficiency and time-efficiency.

Erdem Bıyık, PhD
Assistant Professor of Computer Science | lead the Learning and Interactive Robot Autonomy | Lab LiraLab | University of Southern California
