Emanuele Ratti Ratti (University of Notre Dame), Ezequiel López Rubio (University of Málaga)
We argue that mechanistic models elaborated by machine learning cannot be explanatory by discussing the relation between mechanistic models, explanation and the notion of intelligibility of models. We show that the ability of biologists to understand the model that they work with (i.e., intelligibility) severely constrains their capacity of turning the model into an explanatory model. The more a mechanistic model is complex (i.e., it includes an increasing number of components), the less explanatory it will be. Since machine learning increase its performances when more components are added, then it generates models which are not intelligible, and hence not explanatory.