Item | Value |
---|---|
Donations List Website (data still preliminary) | donor |
Organization | Title | Start date | End date | Employment type | Source | Notes |
---|---|---|---|---|---|---|
Center for Human-Compatible AI | 2019-01-01 | [1] | ||||
Center for Human-Compatible AI | Graduate Student | 2021-01-01 | graduate student | [2] |
Name | Creation date | Description |
---|---|---|
Clarifying some key hypotheses in AI alignment | 2019-08-15 | With Ben Cottier. A diagram collecting several hypotheses in AI alignment and their relationships to existing research agendas. |
Title | Publication date | Author | Publisher | Affected organizations | Affected people | Document scope | Cause area | Notes |
---|
Title | Publication date | Author | Publisher | Affected organizations | Affected people | Affected agendas | Notes |
---|---|---|---|---|---|---|---|
AI Alignment Podcast: An Overview of Technical AI Alignment with Rohin Shah (Part 2) | 2019-04-25 | Lucas Perry | Future of Life Institute | Rohin Shah, Dylan Hadfield-Menell, Gillian Hadfield | Embedded agency, Cooperative inverse reinforcement learning, inverse reinforcement learning, deep reinforcement learning from human preferences, recursive reward modeling, iterated amplification | Part two of a podcast episode that goes into detail about some technical approaches to AI alignment. | |
AI Alignment Podcast: An Overview of Technical AI Alignment with Rohin Shah (Part 1) | 2019-04-11 | Lucas Perry | Future of Life Institute | Rohin Shah | iterated amplification | Part one of an interview with Rohin Shah that goes covers some technical agendas for AI alignment. |
Showing at most 20 people who are most similar in terms of which organizations they have worked at.