[공지사항]을 빙자한 안부와 근황 
Show more

Daily Arxiv

This is a page that curates AI-related papers published worldwide.
All content here is summarized using Google Gemini and operated on a non-profit basis.
Copyright for each paper belongs to the authors and their institutions; please make sure to credit the source when sharing.

Illuminating the Three Dogmas of Reinforcement Learning under Evolutionary Light

Created by
  • Haebom

Author

Mani Hamidi, Terrence W. Deacon

Outline

This paper proposes a conceptual revision of the three core principles of reinforcement learning (RL): the definition of the agent, the goal of learning, and the scope of the reward hypothesis. We reexamine these three principles from an evolutionary perspective, and first argue that evolutionary dynamics can operate in the brain throughout the life of an individual, in order to support the validity of RL as a model of biological learning. For the second principle (the goal of learning), we enrich the ‘adaptation rather than exploration’ perspective with evolutionary insights, and for the third principle (the scope of the reward hypothesis), we discuss scalar rewards and multi-objective problems through the analogy of evolutionary fitness. After discussing practical implications for exploration in RL, we address the absence of the first principle, the definition of the agent. We argue that the agent problem cannot be solved by the evolutionary paradigm alone, and suggest that the thermodynamics of nutrition and replication in the origin of life theory provide a promising foundation for understanding agents and resource-constrained reinforcement learning in biological systems.

Takeaways, Limitations

Takeaways:
By reinterpreting the fundamental principles of reinforcement learning from an evolutionary perspective, we provide a more biologically plausible and rich theoretical foundation.
It provides a new perspective on practical problems in RL, such as scalar reward vs. multi-objective problems, search strategies, etc.
We present a novel attempt to apply the origin of life theory to solving agent problems in RL.
Limitations:
The evolutionary paradigm alone does not completely solve the problem of defining agents.
There is a lack of experimental validation and concrete algorithmic presentation of the proposed evolutionary framework.
The specific mechanism for integrating origin-of-life theory and RL is unclear.
👍