Associated people: Paul Christiano, Buck Shlegeris, Dario Amodei
Associated organizations: OpenAI
Iterated amplification (also called iterated distillation and amplification) aims to build a powerful aligned AGI by repeatedly invoking two steps: (1) amplification and (2) distillation.
One specific version of iterated amplification has been called “imitating expert reasoning” in the reward modeling paper (see also this comment).
Iterated amplification intends to build powerful AGI assistants that try to help humans.
The agenda does not intend to solve all problems, e.g. it doesn’t aim to solve philosophy (although to the extent that humans solving these problems, the AGI assistant would be able to help with them). See “Paul’s research agenda FAQ” § Goals and non-goals for more information.
No particular assumptions about AI timelines, as far as I know.
Iterated amplification is intended to be able to deal with the case of prosaic AGI, i.e. the case where humanity is able to build AGI without learning anything fundamentally new about the nature of intelligence. In other words, iterated amplification works to align scaled-up versions of current machine learning systems.
Iterated amplification has some “key hopes” it is based on:
Title | Publication date | Author | Publisher | Affected organizations | Affected people | Affected agendas | Notes |
---|---|---|---|---|---|---|---|
Challenges to Christiano’s capability amplification proposal | 2018-05-19 | Eliezer Yudkowsky | Machine Intelligence Research Institute | Paul Christiano | Iterated amplification |