Balancing multidimensional morality and progression

Evaluating the tradeoff for artificial agents playing text-based games

Bachelor Thesis (2023)
Author(s)

B. Şerbănescu (TU Delft - Electrical Engineering, Mathematics and Computer Science)

Contributor(s)

Pradeep Murukannaiah – Mentor (TU Delft - Interactive Intelligence)

Enrico Liscio – Mentor (TU Delft - Interactive Intelligence)

D. Mambelli – Graduation committee member (TU Delft - Interactive Intelligence)

Faculty
Electrical Engineering, Mathematics and Computer Science
Copyright
© 2023 Bianca Şerbănescu
More Info
expand_more
Publication Year
2023
Language
English
Copyright
© 2023 Bianca Şerbănescu
Graduation Date
03-07-2023
Awarding Institution
Delft University of Technology
Project
['CSE3000 Research Project']
Programme
['Computer Science and Engineering']
Faculty
Electrical Engineering, Mathematics and Computer Science
Reuse Rights

Other than for strictly personal use, it is not permitted to download, forward or distribute the text or part of it, without the consent of the author(s) and/or copyright holder(s), unless the work is under an open content license such as Creative Commons.

Abstract

Morality is a fundamental concept that guides humans in the decision-making process. Given the rise of large language models in society, it is necessary to ensure that they adhere to human principles, among which morality is of substantial importance. While research has been done regarding artificial agents behaving morally, current state of the art implementations consider morality to be linear, thus failing to capture its complexity and nuances. To account for this, a multidimensional representation of morality is proposed, each dimension corresponding to a different moral foundation. Then, the performance of three types of artificial agents tasked with choosing actions while playing text-based games is compared and analysed. One type of agent is implemented to only choose the most moral action, without aiming to win the games, another one prioritizes moral actions over game progression, and another strives to win the games while also playing morally. The latter outperforms the others in terms of game progression, while also taking few immoral actions. However, the agent prioritizing morality over progression performs only slightly worse while taking no immoral actions, proving that artificial agents can perform well while also behaving morally.

Files

Research_paper.pdf
(pdf | 0.235 Mb)
License info not available