papers AI Learner
The Github is limit! Click to go to the new site.

A Roadmap for Robust End-to-End Alignment

2019-02-25
Lê Nguyên Hoang

Abstract

We analyze the AI alignment problem. This is the problem of aligning an AI’s objective function with human preferences. This problem has been argued to be critical to AI safety, especially in the long run. But it has also been argued that solving it robustly is extremely challenging, especially in highly complex environments like the Internet. It seems crucial to accelerate research in this direction. To this end, we propose a preliminary research program. Our roadmap aims to decompose alignment into numerous more tractable subproblems. Our hope is that this will help scholars, engineers and decision-makers to better grasp the upcoming difficulties, and to foresee how they can best contribute to the global effort.

Abstract (translated by Google)
URL

http://arxiv.org/abs/1809.01036

PDF

http://arxiv.org/pdf/1809.01036


Similar Posts

Comments