Inductive risk, understanding, and opaque machine learning models

Emily Sullivan*

*Corresponding author for this work

Research output: Contribution to journalArticlepeer-review

Abstract

Under what conditions does machine learning (ML) model opacity inhibit the possibility of explaining and understanding phenomena? In this article, I argue that nonepistemic values give shape to the ML opacity problem even if we keep researcher interests fixed. Treating ML models as an instance of doing model-based science to explain and understand phenomena reveals that there is (i) an external opacity problem, where the presence of inductive risk imposes higher standards on externally validating models, and (ii) an internal opacity problem, where greater inductive risk demands a higher level of transparency regarding the inferences the model makes.
Original languageEnglish
Pages (from-to)1065-1074
Number of pages10
JournalPhilosophy of Science
Volume89
Issue number5
Early online date9 Jun 2022
DOIs
Publication statusPublished - Dec 2022

Fingerprint

Dive into the research topics of 'Inductive risk, understanding, and opaque machine learning models'. Together they form a unique fingerprint.

Cite this