Daily Arxiv

This page organizes papers related to artificial intelligence published around the world.
This page is summarized using Google Gemini and is operated on a non-profit basis.
The copyright of the paper belongs to the author and the relevant institution. When sharing, simply cite the source.

Addressing Moral Uncertainty using Large Language Models for Ethical Decision-Making

Created by
  • Haebom

Author

Rohit K. Dubey, Damian Dailisan, Sachit Mahajan

Outline

We present an ethical decision-making framework that improves pre-trained reinforcement learning (RL) models using a task-agnostic ethical layer. After initial training, the RL model undergoes ethical fine-tuning using feedback generated from a large-scale language model (LLM). The LLM assigns belief values to recommended actions in ethical decision-making processes, based on moral principles such as consequentialism, deontology, virtue, social justice, and care ethics. The ethical layer aggregates belief scores from multiple LLM-based moral perspectives using Belief Jensen-Shannon Divergence and Dempster-Shafer Theory to generate a probability score that serves as a shaping reward, guiding the agent to make choices aligned with a balanced ethical framework. This integrated learning framework helps RL agents navigate moral uncertainty in complex environments and make morally sound decisions across a variety of tasks. Testing multiple LLM variants and comparing them to other belief aggregation techniques demonstrates improved consistency and adaptability, and reduced reliance on handcrafted ethical rewards. This approach is particularly effective in dynamic scenarios where ethical issues arise unexpectedly, making it suitable for practical applications.

Takeaways, Limitations

Leveraging LLM to generate ethical training data and enhance the decision-making capabilities of RL agents by integrating diverse ethical perspectives.
Leveraging Belief Jensen-Shannon Divergence and Dempster-Shafer Theory to aggregate belief scores from multiple ethical perspectives, allowing agents to make decisions that align with a balanced ethical framework.
The effectiveness of the methodology is verified through comparison with various LLM variants and other belief aggregation techniques.
Effectively address ethical issues in dynamic scenarios, increasing practical applicability.
LLM bias may influence final ethical judgment.
The performance of the entire framework may be limited by the performance of the LLM.
May have limited ability to resolve complex ethical issues.
👍