Skip to yearly menu bar Skip to main content


Poster

AlphaAlign: Incentivizing Safety Alignment with Extremely Simplified Reinforcement Learning

Yi Zhang · An Zhang · XiuYu Zhang · Leheng Sheng · Yuxin Chen · Zhenkai Liang · Xiang Wang

Abstract

Log in and register to view live content