Link to the paper - On the Feasibility of Learning, Rather than Assuming, Human Biases for Reward Inference
Link to the transcript
The Alignment Newsletter
Rohin's contributions to the AI alignment forum
Rohin's website
Patreon: https://www.patreon.com/axrpodcast
MATS: https://www.matsprogram.org
Note: I'm employed by MATS, but they're not paying me to make this video.