Catherine, Jess, and I use some of the ideas from their recent papers to discuss how different types of explanations in neuroscience and AI could be unified into explanations of intelligence, natural or artificial. Catherine has written about how models are related to the target system they are built to explain. She suggests both the model and the target system should be considered as instantiations of a specific kind of phenomenon, and explanation is a product of relating the model and the target system to that specific aspect they both share. Jess has suggested we shift our focus of explanation from objects – like a brain area or a deep learning model – to the shared class of phenomenon performed by those objects. Doing so may help bridge the gap between the different forms of explanation currently used in neuroscience and AI. We also discuss Henk de Regt’s conception of scientific understanding and its relation to explanation (they’re different!), and plenty more.
Timestamps:
0:00 – Intro
11:11 – Background and approaches
27:00 – Understanding distinct from explanation
36:00 – Explanations as programs (early explanation)
40:42 – Explaining classes of phenomena
52:05 – Constitutive (neuro) vs. etiological (AI) explanations
1:04:04 – Do nonphysical objects count for explanation?
1:10:51 – Advice for early philosopher/scientists
Grace’s websiteTwitter: @neurograce.Models of the Mind: How Physics, Engineering and Mathematics Have Shaped Our Understanding of the Brain.We talked about Grace’s work using convolutional...
Support the show to get full episodes and join the Discord community. Check out my free video series about what's missing in AI and...
Follow Julie on Twitter: @julie_grollier. Check out Julie’s website, with links to papers and lots of good summaries of things we discuss, like Memristors,...