|
Sun 5:50 p.m. - 6:00 p.m.
|
Welcome and Opening Remarks
(
Opening Remarks
)
>
link
SlidesLive Video
|
Hua Shen · Joyce Chai · Yang Li · Antoine Bosselut
🔗
|
|
Sun 6:00 p.m. - 6:30 p.m.
|
Keynote 1
(
Invited Talk
)
>
SlidesLive Video
|
Been Kim
🔗
|
|
Sun 6:30 p.m. - 7:00 p.m.
|
Keynote 2
(
Invited Talk
)
>
SlidesLive Video
|
Frauke Kreuter
🔗
|
|
Sun 7:00 p.m. - 8:10 p.m.
|
Poster and Coffee break
(
Poster Session'
)
>
|
🔗
|
|
Sun 8:10 p.m. - 8:40 p.m.
|
Keynote 3
(
Invited Talk
)
>
SlidesLive Video
|
Hung-yi Lee
🔗
|
|
Sun 8:40 p.m. - 9:10 p.m.
|
Keynote 4
(
Invited Talk
)
>
SlidesLive Video
|
Brad Myers
🔗
|
|
Sun 9:10 p.m. - 10:10 p.m.
|
Lunch Break
(
Lunch Break
)
>
|
🔗
|
|
Sun 10:10 p.m. - 10:20 p.m.
|
CHI Oral 1: Policy Prototyping for LLMs: Pluralistic Alignment via Interactive and Collaborative Policymaking
(
Oral Presentation
)
>
link
|
Kevin Feng
🔗
|
|
Sun 10:20 p.m. - 10:30 p.m.
|
CHI Oral 2: Augmenting Image Annotation: A Human–LLM Collaborative Framework for Efficient Object Selection and Label Generation
(
Oral Presentation
)
>
link
|
HE ZHANG
🔗
|
|
Sun 10:30 p.m. - 11:00 p.m.
|
Keynote 5
(
Invited Talk
)
>
SlidesLive Video
|
Dan Bohus
🔗
|
|
Sun 11:00 p.m. - 11:30 p.m.
|
Keynote 6
(
Invited Talk
)
>
SlidesLive Video
|
Pavel Izmailov
🔗
|
|
Sun 11:30 p.m. - 11:40 p.m.
|
ICLR Oral 1: Scalably Solving Assistance Games
(
Oral Presentation
)
>
link
SlidesLive Video
|
Cassidy Laidlaw
🔗
|
|
Sun 11:40 p.m. - 11:50 p.m.
|
ICLR Oral 2: Preference Optimization For Concept Bottleneck Models
(
Oral
)
>
link
SlidesLive Video
|
Emiliano Penaloza
🔗
|
|
Sun 11:50 p.m. - 12:00 a.m.
|
ICLR Oral 3: Societal Impacts Research Requires Usage-Based Benchmarks for Creative Tasks
(
Oral Presentation
)
>
link
SlidesLive Video
|
Judy Shen
🔗
|
|
Mon 12:00 a.m. - 12:10 a.m.
|
ICLR Oral 4: InterFeedback: Unveiling Interactive Intelligence of Large Multimodal Models with Human Feedback
(
Oral Presentation
)
>
link
SlidesLive Video
|
Henry Zhao
🔗
|
|
Mon 12:10 a.m. - 1:00 a.m.
|
Poster and Coffee Break
(
Poster Session
)
>
|
🔗
|
|
Mon 1:00 a.m. - 1:10 a.m.
|
ICLR Oral 5: Representational Alignment Supports Effective Teaching
(
Oral Presentation
)
>
link
SlidesLive Video
|
Ilia Sucholutsky
🔗
|
|
Mon 1:10 a.m. - 1:20 a.m.
|
ICLR Oral 6: PARSE-Ego4D: Toward Bidirectionally Aligned Action Recommendations for Egocentric Videos
(
Oral Presentation
)
>
link
SlidesLive Video
|
Steven Abreu
🔗
|
|
Mon 1:20 a.m. - 1:30 a.m.
|
ICLR Oral 7: AI-enhanced semantic feature norms for 786 concepts
(
Oral Presentation
)
>
link
SlidesLive Video
|
Siddharth Suresh
🔗
|
|
Mon 1:30 a.m. - 1:40 a.m.
|
ICLR Oral 8: SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities
(
Oral Presentation
)
>
link
SlidesLive Video
|
Fengqing Jiang
🔗
|
|
Mon 1:40 a.m. - 2:10 a.m.
|
Keynote 7
(
Invited Talk
)
>
SlidesLive Video
|
Richard Ngo
🔗
|
|
Mon 2:10 a.m. - 2:50 a.m.
|
Panel Discussion with Experts
(
Panel Discussion
)
>
SlidesLive Video
|
Richard Ngo · Frauke Kreuter · Pavel Izmailov · Tammy Masterson
🔗
|
|
Mon 2:50 a.m. - 3:00 a.m.
|
Paper Award Announcement
(
Award Announcement
)
>
SlidesLive Video
|
Hua Shen · Yang Li · Joyce Chai · Antoine Bosselut
🔗
|
|
Mon 3:00 a.m. - 3:00 a.m.
|
Closing Remarks
(
Closing
)
>
link
SlidesLive Video
|
Hua Shen · Yang Li · Joyce Chai · Antoine Bosselut
🔗
|
|
-
|
PREFERENCE OPTIMIZATION FOR CONCEPT BOTTLENECK MODELS
(
Poster
)
>
link
|
Emiliano Penaloza · Tianyue Zhang · Laurent Charlin · Mateo Espinosa Zarlenga
🔗
|
|
-
|
PREFERENCE OPTIMIZATION FOR CONCEPT BOTTLENECK MODELS
(
Oral
)
>
link
|
🔗
|
|
-
|
Learning From Diverse Experts: Behavior Alignment Through Multi-Objective Inverse Reinforcement Learning
(
Poster
)
>
link
|
Jun-Jie Yang · Qian-You Zhang · Chia-Heng Hsu · Xi Liu · Ping-Chun Hsieh
🔗
|
|
-
|
Learning From Diverse Experts: Behavior Alignment Through Multi-Objective Inverse Reinforcement Learning
(
Oral
)
>
link
|
🔗
|
|
-
|
Probing Mechanical Reasoning in Large Vision Language Models
(
Poster
)
>
link
|
Haoran Sun · Yijiang Li · Qingying Gao · Haiyun Lyu · Dezhi Luo · Hokin Deng
🔗
|
|
-
|
Probing Mechanical Reasoning in Large Vision Language Models
(
Oral
)
>
link
|
🔗
|
|
-
|
The Human Visual System Can Inspire New Interaction Paradigms for LLMs
(
Poster
)
>
link
|
Diana Robinson · Neil Lawrence
🔗
|
|
-
|
The Human Visual System Can Inspire New Interaction Paradigms for LLMs
(
Oral
)
>
link
|
🔗
|
|
-
|
AI-enhanced semantic feature norms for 786 concepts
(
Poster
)
>
link
|
SIDDHARTH SURESH · Kushin Mukherjee · Tyler Giallanza · Xizheng Yu · Mia Patil · Jonathan Cohen · Timothy Rogers
🔗
|
|
-
|
Scalably Solving Assistance Games
(
Poster
)
>
link
|
Cassidy Laidlaw · Eli Bronstein · Timothy Guo · Dylan Feng · Lukas Berglund · Justin Svegliato · Stuart Russell · Anca Dragan
🔗
|
|
-
|
Patterns and Mechanisms of Contrastive Activation Engineering
(
Poster
)
>
link
|
Yixiong Hao · Ayush Panda · Stepan Shabalin · Sheikh Abdur Raheem Ali
🔗
|
|
-
|
Patterns and Mechanisms of Contrastive Activation Engineering
(
Oral
)
>
link
|
🔗
|
|
-
|
Value Alignment in the Global South: A Multidimensional Approach to Norm Elicitation in Indian Contexts
(
Poster
)
>
link
|
Atmadeep Ghoshal · Martim Brandao · Ruba Abu-Salma
🔗
|
|
-
|
Value Alignment in the Global South: A Multidimensional Approach to Norm Elicitation in Indian Contexts
(
Oral
)
>
link
|
🔗
|
|
-
|
Active Human Feedback Collection via Neural Contextual Dueling Bandits
(
Poster
)
>
link
|
Arun Verma · Xiaoqiang Lin · Zhongxiang Dai · Daniela Rus · Bryan Kian Hsiang Low
🔗
|
|
-
|
Active Human Feedback Collection via Neural Contextual Dueling Bandits
(
Oral
)
>
link
|
🔗
|
|
-
|
CTRL-Rec: Controlling Recommender Systems With Natural Language
(
Poster
)
>
link
|
Micah Carroll · Adeline Foote · Marcus Williams · Anca Dragan · W. Bradley Knox · Smitha Milli
🔗
|
|
-
|
CTRL-Rec: Controlling Recommender Systems With Natural Language
(
Oral
)
>
link
|
🔗
|
|
-
|
Vision Language Models See What You Want but not What You See
(
Poster
)
>
link
|
Qingying Gao · Yijiang Li · Haiyun Lyu · Haoran Sun · Dezhi Luo · Hokin Deng
🔗
|
|
-
|
Vision Language Models See What You Want but not What You See
(
Oral
)
>
link
|
🔗
|
|
-
|
Data-adaptive Safety Rules for Training Reward Models
(
Poster
)
>
link
|
Xiaomin Li · Mingye Gao · Zhiwei Zhang · Fan · Weiyu Li
🔗
|
|
-
|
Data-adaptive Safety Rules for Training Reward Models
(
Oral
)
>
link
|
🔗
|
|
-
|
D3PO: Preference-Based Alignment of Discrete Diffusion Models
(
Poster
)
>
link
|
Umberto Borso · Davide Paglieri · Jude Wells · Tim Rocktaeschel
🔗
|
|
-
|
D3PO: Preference-Based Alignment of Discrete Diffusion Models
(
Oral
)
>
link
|
🔗
|
|
-
|
SafeChain: Safety of Language Models with Long Chain-of-Thought Reasoning Capabilities
(
Poster
)
>
link
|
Fengqing Jiang · Zhangchen Xu · Yuetai Li · Luyao Niu · Zhen Xiang · Bo Li · Bill Yuchen Lin · Radha Poovendran
🔗
|
|
-
|
Symmetry-Breaking Augmentations for Ad Hoc Teamwork
(
Poster
)
>
link
|
Ravi Hammond · Dustin Craggs · Mingyu Guo · Jakob Foerster · Ian Reid
🔗
|
|
-
|
Symmetry-Breaking Augmentations for Ad Hoc Teamwork
(
Oral
)
>
link
|
🔗
|
|
-
|
A Sociotechnical Perspective on Aligning AI with Pluralistic Human Values
(
Poster
)
>
link
|
Dalia Ali · Aysenur Kocak · Michèle Wieland · Dora Zhao · Allison Koenecke · Orestis Papakyriakopoulos
🔗
|
|
-
|
A Sociotechnical Perspective on Aligning AI with Pluralistic Human Values
(
Oral
)
>
link
|
🔗
|
|
-
|
Mitigating Societal Cognitive Overload in the Age of AI: Challenges and Directions
(
Poster
)
>
link
|
Salem Lahlou
🔗
|
|
-
|
Mitigating Societal Cognitive Overload in the Age of AI: Challenges and Directions
(
Oral
)
>
link
|
🔗
|
|
-
|
Policy Prototyping for LLMs: Pluralistic Alignment via Interactive and Collaborative Policymaking
(
Poster
)
>
link
|
Kevin Feng · Inyoung Cheong · Quan Chen · Amy Zhang
🔗
|
|
-
|
A Benchmark for Scalable Oversight Mechanisms
(
Poster
)
>
link
|
Abhimanyu Pallavi Sudhir · Jackson Kaunismaa · Arjun Panickssery
🔗
|
|
-
|
A Benchmark for Scalable Oversight Mechanisms
(
Oral
)
>
link
|
🔗
|
|
-
|
From Intuition to Understanding: Using AI Peers to Overcome Physics Misconceptions
(
Poster
)
>
link
|
12 presenters
Ruben Weijers · Denton Wu · Hannah Betts · Yuxiang Guan · Vidya Sujaya · Kushal Dev · Reihaneh Rabbany · Jean-François Godbout · Kellin Pelrine · Tamara Jacod · William Delooze · Ying Wu
🔗
|
|
-
|
From Intuition to Understanding: Using AI Peers to Overcome Physics Misconceptions
(
Oral
)
>
link
|
🔗
|
|
-
|
Understanding (Un)Reliability of Steering Vectors in Language Models
(
Poster
)
>
link
|
Joschka Braun · Carsten Eickhoff · David Krueger · Seyed Ali Bahrainian · Dmitrii Krasheninnikov
🔗
|
|
-
|
Understanding (Un)Reliability of Steering Vectors in Language Models
(
Oral
)
>
link
|
🔗
|
|
-
|
Processing, Priming, Probing: Human Interventions for Explainability Alignment
(
Poster
)
>
link
|
Kenza Amara
🔗
|
|
-
|
Processing, Priming, Probing: Human Interventions for Explainability Alignment
(
Oral
)
>
link
|
🔗
|
|
-
|
Aligning LLMs with Domain Invariant Reward Models
(
Poster
)
>
link
|
David Wu · Sanjiban Choudhury
🔗
|
|
-
|
Aligning LLMs with Domain Invariant Reward Models
(
Oral
)
>
link
|
🔗
|
|
-
|
Drift: Efficient Implicit Personalization of Large Language Models
(
Poster
)
>
link
|
Minbeom Kim · Kang-il Lee · Seongho Joo · Hwaran Lee · Kyomin Jung
🔗
|
|
-
|
Drift: Efficient Implicit Personalization of Large Language Models
(
Oral
)
>
link
|
🔗
|
|
-
|
Bidirectional Alignment for Inclusive Narrative Generation
(
Poster
)
>
link
|
Ken Kawamura
🔗
|
|
-
|
Bidirectional Alignment for Inclusive Narrative Generation
(
Oral
)
>
link
|
🔗
|
|
-
|
Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment
(
Poster
)
>
link
|
Yifan Zhang · Ge Zhang · Yue Wu · Kangping Xu · Quanquan Gu
🔗
|
|
-
|
Beyond Bradley-Terry Models: A General Preference Model for Language Model Alignment
(
Oral
)
>
link
|
🔗
|
|
-
|
Exploring Persona-dependent LLM Alignment for the Moral Machine Experiment
(
Poster
)
>
link
|
Jiseon Kim · Jea Kwon · Luiz Felipe Vecchietti · Alice Oh · Meeyoung Cha
🔗
|
|
-
|
Exploring Persona-dependent LLM Alignment for the Moral Machine Experiment
(
Oral
)
>
link
|
🔗
|
|
-
|
Augmenting Image Annotation: A Human–LLM Collaborative Framework for Efficient Object Selection and Label Generation
(
Poster
)
>
link
|
HE ZHANG · Xinyi Fu · John Carroll
🔗
|
|
-
|
CoPL: Collaborative Preference Learning for Personalizing LLMs
(
Poster
)
>
link
|
Youngbin Choi · Seunghyuk Cho · Minjong Lee · MoonJeong Park · Yesong Ko · Jungseul Ok · Dongwoo Kim
🔗
|
|
-
|
CoPL: Collaborative Preference Learning for Personalizing LLMs
(
Oral
)
>
link
|
🔗
|
|
-
|
Broaden your SCOPE! Efficient Conversation Planning for LLMs using Semantic Space
(
Poster
)
>
link
|
Zhiliang Chen · Xinyuan Niu · Chuan Sheng Foo · Bryan Kian Hsiang Low
🔗
|
|
-
|
Broaden your SCOPE! Efficient Conversation Planning for LLMs using Semantic Space
(
Oral
)
>
link
|
🔗
|
|
-
|
Sycophancy Claims about Language Models: The Missing Human-in-the-Loop
(
Poster
)
>
link
|
Jan Batzner · Volker Stocker · Stefan Schmid · Gjergji Kasneci
🔗
|
|
-
|
Sycophancy Claims about Language Models: The Missing Human-in-the-Loop
(
Oral
)
>
link
|
🔗
|
|
-
|
The Alignment Trilemma: A Theoretical Perspective on Recursive Misalignment and Human-AI Adaptation Dynamics
(
Poster
)
>
link
|
Tarun Raheja · Nilay Pochhi
🔗
|
|
-
|
The Alignment Trilemma: A Theoretical Perspective on Recursive Misalignment and Human-AI Adaptation Dynamics
(
Oral
)
>
link
|
🔗
|
|
-
|
Superalignment with Dynamic Human Values
(
Poster
)
>
link
|
Florian Mai · David Kaczér · Nicholas Corrêa · Lucie Flek
🔗
|
|
-
|
Superalignment with Dynamic Human Values
(
Oral
)
>
link
|
🔗
|
|
-
|
InterFeedback: Unveiling Interactive Intelligence of Large Multimodal Models with Human Feedback
(
Poster
)
>
link
|
Henry Zhao · Wenqi Pei · Yifei Tao · Haiyang Mei · Mike Zheng Shou
🔗
|
|
-
|
Human Alignment: How Much We Adapt to LLMs?
(
Poster
)
>
link
|
Cazalet Tanguy · Ruben Janssens · Tony Belpaeme · Joni Dambre
🔗
|
|
-
|
Human Alignment: How Much We Adapt to LLMs?
(
Oral
)
>
link
|
🔗
|
|
-
|
OUTLIER-AWARE PREFERENCE OPTIMIZATION FOR LARGE LANGUAGE MODELS
(
Poster
)
>
link
|
Pragya Srivastava · Sai Nalli · Amit Jayant Deshpande · Amit Sharma
🔗
|
|
-
|
Towards LVLM-Aided Alignment of Task-Specific Vision Models
(
Poster
)
>
link
|
Alexander Koebler · Christian Greisinger · Jan Paulus · Ingo Thon · Florian Buettner
🔗
|
|
-
|
Towards LVLM-Aided Alignment of Task-Specific Vision Models
(
Oral
)
>
link
|
🔗
|
|
-
|
Can RLHF be More Efficient with Imperfect Reward Models? A Policy Coverage Perspective
(
Poster
)
>
link
|
Jiawei Huang · Bingcong Li · Christoph Dann · Niao He
🔗
|
|
-
|
Can RLHF be More Efficient with Imperfect Reward Models? A Policy Coverage Perspective
(
Oral
)
>
link
|
🔗
|
|
-
|
TRIG-Bench: A Benchmark for Text-Rich Image Grounding
(
Poster
)
>
link
|
Ming Li · Ruiyi Zhang · Jian Chen · Tianyi Zhou
🔗
|
|
-
|
TRIG-Bench: A Benchmark for Text-Rich Image Grounding
(
Oral
)
>
link
|
🔗
|
|
-
|
Envision Human-AI Perceptual Alignment from a Multimodal Interaction Perspective
(
Poster
)
>
link
|
Shu Zhong · Marianna Obrist
🔗
|
|
-
|
Envision Human-AI Perceptual Alignment from a Multimodal Interaction Perspective
(
Oral
)
>
link
|
🔗
|
|
-
|
Monitoring LLM Agents for Sequentially Contextual Harm
(
Poster
)
>
link
|
Chen Yueh-Han · Nitish Joshi · Yulin Chen · He He · Rico Angell
🔗
|
|
-
|
Monitoring LLM Agents for Sequentially Contextual Harm
(
Oral
)
>
link
|
🔗
|
|
-
|
Representational Difference Clustering
(
Poster
)
>
link
|
Neehar Kondapaneni · Emily Gu · Oisin Mac Aodha · Pietro Perona
🔗
|
|
-
|
Representational Difference Clustering
(
Oral
)
>
link
|
🔗
|
|
-
|
Position: Interpretability is a Bidirectional Communication Problem
(
Poster
)
>
link
|
Kola Ayonrinde
🔗
|
|
-
|
Position: Interpretability is a Bidirectional Communication Problem
(
Oral
)
>
link
|
🔗
|
|
-
|
PILAF: Optimal Human Preference Sampling for Reward Modeling
(
Poster
)
>
link
|
Yunzhen Feng · Ariel Kwiatkowski · Kunhao Zheng · Julia Kempe · Yaqi Duan
🔗
|
|
-
|
PILAF: Optimal Human Preference Sampling for Reward Modeling
(
Oral
)
>
link
|
🔗
|
|
-
|
AI Systematically Rewires the Flow of Ideas
(
Poster
)
>
link
|
Zhonghao He · Tianyi Qiu · Tao Lin · Moshe Glickman · Atoosa Kasirzadeh · John Wihbey · Max Kleiman-Weiner
🔗
|
|
-
|
AI Systematically Rewires the Flow of Ideas
(
Oral
)
>
link
|
🔗
|
|
-
|
PARSE-Ego4D: Toward Bidirectionally Aligned Action Recommendations for Egocentric Videos
(
Poster
)
>
link
|
Steven Abreu · Tiffany Do · Karan Ahuja · Eric Gonzalez · Lee Payne · Daniel McDuff · Mar Gonzalez-Franco
🔗
|
|
-
|
Online Learning and Equilibrium Computation with Ranking Feedback
(
Poster
)
>
link
|
Mingyang Liu · Yongshan Chen · Zhiyuan Fan · Gabriele Farina · Asuman Ozdaglar · Kaiqing Zhang
🔗
|
|
-
|
Online Learning and Equilibrium Computation with Ranking Feedback
(
Oral
)
>
link
|
🔗
|
|
-
|
Societal Alignment Frameworks Can Improve LLM Alignment
(
Poster
)
>
link
|
17 presenters
Karolina Stanczak · Nicholas Meade · Mehar Bhatia · Hattie Zhou · Konstantin Böttinger · Jeremy Barnes · Jason Stanley · Jessica Montgomery · Richard Zemel · Nicolas Papernot · Nicolas Chapados · Denis Therien · Timothy Lillicrap · Ana Marasovic · Sylvie Delacroix · Gillian Hadfield · Siva Reddy
🔗
|
|
-
|
Societal Alignment Frameworks Can Improve LLM Alignment
(
Oral
)
>
link
|
🔗
|
|
-
|
TraCeS: Trajectory Based Credit Assignment For Safe Reinforcement Learning
(
Poster
)
>
link
|
Siow Meng Low · Akshat Kumar
🔗
|
|
-
|
TraCeS: Trajectory Based Credit Assignment For Safe Reinforcement Learning
(
Oral
)
>
link
|
🔗
|
|
-
|
ValueMap: Mapping Crowdsourced Human Values to Computational Scores for Bi-directional Alignment
(
Poster
)
>
link
|
Priya DCosta · Rupkatha Hira
🔗
|
|
-
|
ValueMap: Mapping Crowdsourced Human Values to Computational Scores for Bi-directional Alignment
(
Oral
)
>
link
|
🔗
|
|
-
|
Observability of Latent States in Generative AI Models
(
Poster
)
>
link
|
Tian Yu Liu · Stefano Soatto · Matteo Marchi · Pratik A Chaudhari · Paulo Tabuada
🔗
|
|
-
|
Observability of Latent States in Generative AI Models
(
Oral
)
>
link
|
🔗
|
|
-
|
Rethinking Anti-Misinformation AI
(
Poster
)
>
link
|
Vidya Sujaya · Kellin Pelrine · Andreea Musulan · Reihaneh Rabbany
🔗
|
|
-
|
Rethinking Anti-Misinformation AI
(
Oral
)
>
link
|
🔗
|
|
-
|
Negotiative Alignment: An interactive approach to human-AI co-adaptation for clinical applications
(
Poster
)
>
link
|
Florence X Doo · Nikhil Shah · Pranav Kulkarni · Vishwa Parekh · Heng Huang
🔗
|
|
-
|
Negotiative Alignment: An interactive approach to human-AI co-adaptation for clinical applications
(
Oral
)
>
link
|
🔗
|
|
-
|
Societal Impacts Research Requires Usage-Based Benchmarks for Creative Tasks
(
Poster
)
>
link
|
Judy Shen · Carlos Guestrin
🔗
|
|
-
|
Inference-time Alignment in Continuous Space
(
Poster
)
>
link
|
Yige Yuan · Teng Xiao · Li Yunfan · Bingbing Xu · Shuchang Tao · Yunqi Qiu · Huawei Shen · Xueqi Cheng
🔗
|
|
-
|
Inference-time Alignment in Continuous Space
(
Oral
)
>
link
|
🔗
|
|
-
|
SWEPO: Simultaneous Weighted Preference Optimization for Group Contrastive Alignment
(
Poster
)
>
link
|
Taneesh Gupta · Rahul Madhavan · Xuchao Zhang · Chetan Bansal · Saravanakumar Rajmohan
🔗
|
|
-
|
SWEPO: Simultaneous Weighted Preference Optimization for Group Contrastive Alignment
(
Oral
)
>
link
|
🔗
|
|
-
|
Investigating Alignment Signals in Initial Token Representations
(
Poster
)
>
link
|
Carl Rosenblatt
🔗
|
|
-
|
Investigating Alignment Signals in Initial Token Representations
(
Oral
)
>
link
|
🔗
|
|
-
|
Rethinking AI cultural alignment
(
Poster
)
>
link
|
Michal Bravansky · Filip Trhlík · Fazl Barez
🔗
|
|
-
|
Rethinking AI cultural alignment
(
Oral
)
>
link
|
🔗
|
|
-
|
A Roadmap for Human-Agent Moral Alignment: Integrating Pre-defined Intrinsic Rewards and Learned Reward Models
(
Poster
)
>
link
|
Elizaveta Tennant · Stephen Hailes · Mirco Musolesi
🔗
|
|
-
|
A Roadmap for Human-Agent Moral Alignment: Integrating Pre-defined Intrinsic Rewards and Learned Reward Models
(
Oral
)
>
link
|
🔗
|
|
-
|
Representational Alignment Supports Effective Teaching
(
Poster
)
>
link
|
14 presenters
Ilia Sucholutsky · Katherine Collins · Maya Malaviya · Nori Jacoby · Weiyang Liu · Theodore Sumers · Michalis Korakakis · Umang Bhatt · Mark Ho · Joshua B Tenenbaum · Bradley Love · Zachary Pardos · Adrian Weller · Thomas L. Griffiths
🔗
|
|
-
|
We Shape AI, and Thereafter AI Shape Us: Humans Align with AI through Social Influences
(
Poster
)
>
link
|
Jingshu Li · Tianqi Song · Beichen Xue · Yi-Chieh Lee
🔗
|
|
-
|
We Shape AI, and Thereafter AI Shape Us: Humans Align with AI through Social Influences
(
Oral
)
>
link
|
🔗
|
|
-
|
The Lock-in Hypothesis: Stagnation by Algorithm
(
Poster
)
>
link
|
Tianyi Qiu · Zhonghao He · Tejasveer Chugh · Max Kleiman-Weiner
🔗
|
|
-
|
The Lock-in Hypothesis: Stagnation by Algorithm
(
Oral
)
>
link
|
🔗
|
|
-
|
Multi-Objective Probabilistic Preference Learning with Soft and Hard Bounds
(
Poster
)
>
link
|
Edward Chen · Sang Truong · Natalie Dullerud · Sanmi Koyejo · Carlos Guestrin
🔗
|
|
-
|
Multi-Objective Probabilistic Preference Learning with Soft and Hard Bounds
(
Oral
)
>
link
|
🔗
|
|
-
|
Cooperative Agency-Centered LLMs
(
Poster
)
>
link
|
Iyadunni J. Adenuga
🔗
|
|
-
|
Cooperative Agency-Centered LLMs
(
Oral
)
>
link
|
🔗
|
|
-
|
A Pilot Study of Weak-to-Strong Generalization in Safety, Toxicity, and Legal Reasoning
(
Poster
)
>
link
|
Ruimeng Ye · Yang Xiao · Bo Hui
🔗
|
|
-
|
A Pilot Study of Weak-to-Strong Generalization in Safety, Toxicity, and Legal Reasoning
(
Oral
)
>
link
|
🔗
|
|
-
|
Moral Alignment for LLM Agents
(
Poster
)
>
link
|
Elizaveta Tennant · Stephen Hailes · Mirco Musolesi
🔗
|
|
-
|
Moral Alignment for LLM Agents
(
Oral
)
>
link
|
🔗
|
|
-
|
Shared Similarity Between Humans and Chatbots: Exploring Human Willingness to Seek Social Support From Chatbots
(
Poster
)
>
link
|
Zicheng Zhu · Tianqi Song · Jefferson Lim · Chi-Lan Yang · Yi-Chieh Lee
🔗
|
|
-
|
Shared Similarity Between Humans and Chatbots: Exploring Human Willingness to Seek Social Support From Chatbots
(
Oral
)
>
link
|
🔗
|
|
-
|
Addressing and Visualizing Misalignments in Human Task-Solving Trajectories
(
Poster
)
>
link
|
Sejin Kim · Hosung Lee · Sundong Kim
🔗
|
|
-
|
Addressing and Visualizing Misalignments in Human Task-Solving Trajectories
(
Oral
)
>
link
|
🔗
|
|
-
|
Vision Language Models Know Law of Conservation without Understanding More-or-Less
(
Poster
)
>
link
|
Dezhi Luo · Haiyun Lyu · Qingying Gao · Haoran Sun · Yijiang Li · Hokin Deng
🔗
|
|
-
|
Order Independence With Finetuning
(
Poster
)
>
link
|
Katrina Brown · Reid McIlroy-Young
🔗
|