[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Inverse Reinforcement Learning with Switching Rewards and History Dependency for Characterizing Animal Behaviors

Created by
  • Haebom

Author

Jingyang Ke, Feiyang Wu, Jiyi Wang, Jeffrey Markowitz, Anqi Wu

Outline

This paper points out the limitations of existing neuroscience decision-making research, which focuses on simplified behavioral tasks and explicit rewards, and deals only with repetitive and fixed behaviors of animals. In natural environments, animals often exhibit complex behaviors over long periods of time due to unobservable internal motivations. Time-varying inverse reinforcement learning (IRL) has been used to capture this, but it fails to consider that animal decisions are based not only on the current state but also on past history. In this paper, we present SWIRL (SWitching IRL), a novel framework that integrates time-varying and past-dependent reward functions. SWIRL models long-term action sequences as transitions between short-term decision processes, each governed by a unique reward function, thereby capturing how past decisions and environmental context shape behavior. We apply SWIRL to simulated and real animal behavior datasets and demonstrate that it quantitatively and qualitatively outperforms models without past dependence. This is the first IRL model that integrates past-dependent policies and rewards, advancing our understanding of complex and natural animal decision-making.

Takeaways, Limitations

Takeaways:
We present SWIRL, a novel IRL framework that integrates time-varying and past-dependent reward functions, enabling more accurate modeling of complex and natural animal decision-making.
Effectively analyze the impact of past behavior and environmental context on current decision-making.
Overcoming the limitations of existing IRL models and presenting new possibilities in analyzing natural behavioral data.
Validation of the model by application to real animal behavior data.
Limitations:
Lack of detailed explanation of parameter settings and optimization of the SWIRL model.
Further research is needed to determine generalizability to different types of animal behavioral data.
The computational complexity and scalability of the model need to be reviewed.
Difficulties in precisely defining and measuring intrinsic motivation.
👍