research-papers
Beyond Preferences in AI Alignment
Original Paper: https://arxiv.org/abs/2408.16984 By: Tan Zhi-Xuan, Micah Carroll, Matija Franklin, Hal Ashton Abstract: The dominant practice of AI alignment assumes (1) that preferences are an adequate representation of human values (2) that human rationality can be understood in terms of maximizing the satisfaction of preferences