AAAI 2026

January 23, 2026

Singapore, Singapore

Would you like to see your presentation here, made available to a global audience of researchers?
Add your own presentation or have us affordably record your next conference.

Reward functions, learned or manually specified, are rarely perfect. Instead of accurately expressing human goals, these reward functions are often distorted by human beliefs about how best to achieve those goals. Specifically, these reward functions often express a combination of the human's terminal goals — those which are ends in themselves — and the human's instrumental goals — those which are means to an end. We formulate a simple example in which even slight conflation of instrumental and terminal goals results in severe misalignment: optimizing the misspecified reward function $\hat{r}$ results in poor performance when measured by the true reward function $r$. This example distills the essential properties of environments that make reinforcement learning highly sensitive to conflation of instrumental and terminal goals. We discuss how this issue can arise with a common approach to reward learning and how it can manifest in real environments.

Downloads

SlidesPaperTranscript English (automatic)

Next from AAAI 2026

GSAG-CDGAN: A Generalizable Small-Sample Attention-Guided GAN for Remote Sensing Change Detection (Student Abstract)
technical paper

GSAG-CDGAN: A Generalizable Small-Sample Attention-Guided GAN for Remote Sensing Change Detection (Student Abstract)

AAAI 2026

Lukun WangJiaming Pei
Jiaming Pei and 2 other authors

23 January 2026

Stay up to date with the latest Underline news!

Select topic of interest (you can select more than one)

PRESENTATIONS

  • All Presentations
  • For Librarians
  • Resource Center
  • Free Trial
Underline Science, Inc.
1216 Broadway, 2nd Floor, New York, NY 10001, USA

© 2025 Underline - All rights reserved