[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Aligning Humans and Robots via Reinforcement Learning from Implicit Human Feedback

Created by
  • Haebom

Author

Suzie Kim, Hye-Bin Shin, Seong-Whan Lee

Outline

In this paper, we propose an implicit human feedback-based reinforcement learning (RLIHF) framework using electroencephalography (EEG) to overcome the limitations of conventional reinforcement learning (RL), which struggles to learn effective policies in sparse reward environments. We utilize error-related potentials (ErrPs) to provide continuous implicit feedback without explicit user intervention, and transform raw EEG signals into probabilistic reward components via a pre-trained decoder to enable effective policy learning even in sparse external reward environments. We evaluate the proposed method on obstacle avoidance and object manipulation tasks using a Kinova Gen2 robotic arm in a simulation environment based on the MuJoCo physics engine. We show that the agent trained with decoded EEG feedback achieves comparable performance to the agent trained with manually designed dense rewards. This demonstrates the potential of leveraging implicit neural feedback for scalable and human-centric reinforcement learning in interactive robotics.

Takeaways, Limitations

Takeaways:
A novel RLHF framework for solving the scarce reward problem is presented.
Enables natural interaction by leveraging implicit feedback without explicit user intervention
Achieving Effective Policy Learning via EEG-Based Implicit Feedback
Presenting the potential of scalable and human-centric reinforcement learning in the field of interactive robotics
Limitations:
Currently evaluated only in simulation environments, performance verification in real robot environments is required.
Further research is needed on the accuracy and generalization performance of EEG signal interpretation
Need to evaluate generalization performance for different tasks and users
The complexity and cost of EEG data collection and processing must be taken into consideration.
👍