Poster
in
Workshop: Self-Improving Foundation Models Without Human Supervision
AMPO: Active Multi Preference Optimization for Self-play Preference Selection
Taneesh Gupta · Rahul Madhavan · Xuchao Zhang · Chetan Bansal · Saravanakumar Rajmohan
Keywords: [ Group Contrastive Loss ] [ Self Play Preference Optimization ] [ Multi-Preference Optimization ] [ Active Learning ] [ RLHF ]
Multi-preference optimization improves DPO-style alignment beyond pairwise preferences, by contrasting entire sets of helpful and undesired responses, enabling richer training signals for large language models. During self-play alignment, these models often produce numerous candidate answers per query, making it computationally infeasible to include all of them in the training objective. We propose Active Multi-Preference Optimization (AMPO), which combines on-policy generation, a multi-preference group-contrastive loss, and active subset selection. Specifically, we score and embed large candidate pools of responses, then pick a small but informative subset—covering reward extremes and distinct semantic clusters—for preference optimization.The resulting contrastive-training scheme identifies not only the best and worst answers but also subtle, underexplored modes crucial for robust alignment. Theoretically, we provide guarantees of expected reward maximization using our active selection method. Empirically, AMPO achieves state-of-the-art results on AlpacaEval with Llama 8B, achieving a 52% win-rate over GPT-4o. We release our datasets (anonymously) at huggingface/MPO.