Brian and I discuss a range of topics related to his latest book, The Alignment Problem: Machine Learning and Human Values. The alignment problem asks how we can build AI that does what we want it to do, as opposed to building AI that will compromise our own values by accomplishing tasks that may be harmful or dangerous to us. Using some of the stories Brain relates in the book, we talk about:
Links:
Timestamps:
4:22 – Increased work on AI ethics
8:59 – The Alignment Problem overview
12:36 – Stories as important for intelligence
16:50 – What is the alignment problem
17:37 – Who works on the alignment problem?
25:22 – AI ethics degree?
29:03 – Human values
31:33 – AI alignment and evolution
37:10 – Knowing our own values?
46:27 – What have learned about ourselves?
58:51 – Interestingness
1:00:53 – Inverse RL for value alignment
1:04:50 – Current progress
1:10:08 – Developmental psychology
1:17:36 – Models as the danger
1:25:08 – How worried are the experts?
Show notes: Follow Melanie on Twitter: @MelMitchell1. Learn more about what she does at her homepage. Here is her New York Times Op-Ed about...
Image courtesy of Kendrick Kay: Brain art Show notes: Check out Kendrick’s lab website: CVN lab. Follow him on twitter: @cvnlab. The papers we...
[bctt tweet="Check out episode 6 of the Brain Inspired podcast: Deep learning, eyeballs, and brains" username="pgmid"] Mentioned in the show Ryan Poplin What is...