Multi-Objective Reinforcement Learning for Designing Ethical Environments

Fecha de publicación

2025-01-22T09:02:54Z

2025-01-22T09:02:54Z

2021

Resumen

AI research is being challenged with ensuring that autonomous agents learn to behave ethically, namely in alignment with moral values. A common approach, founded on the exploitation of Reinforcement Learning techniques, is to design environments that incentivise agents to behave ethically. However, to the best of our knowledge, current approaches do not theoretically guarantee that an agent will learn to behave ethically. Here, we make headway along this direction by proposing a novel way of designing environments wherein it is formally guaranteed that an agent learns to behave ethically while pursuing its individual objectives. Our theoretical results develop within the formal framework of Multi-Objective Reinforcement Learning to ease the handling of an agent's individual and ethical objectives. As a further contribution, we leverage on our theoretical results to introduce an algorithm that automates the design of ethical environments.

Tipo de documento

Objeto de conferencia


Versión publicada

Lengua

Inglés

Publicado por

International Joint Conferences on Artificial Intelligence

Documentos relacionados

Reproducció del document disponible a: https://doi.org/10.24963/ijcai.2021/76

Comunicació a: 30th International Joint Conference on Artificial Intelligence (IJCAI 2021)

https://doi.org/10.24963/ijcai.2021/76

Citación recomendada

Esta citación se ha generado automáticamente.

Derechos

(c) IJCAI Organization, 2021

Este ítem aparece en la(s) siguiente(s) colección(ones)