Skip to yearly menu bar Skip to main content


A Roadmap for Human-Agent Moral Alignment: Integrating Pre-defined Intrinsic Rewards and Learned Reward Models

Elizaveta Tennant · Stephen Hailes · Mirco Musolesi

Abstract

Chat is not available.