Robustness analysis versus reliable process reasoning Chiara Lisciandra Metascience ISSN 0815-0796 Metascience DOI 10.1007/s11016-014-9927-2 1 23 Your article is protected by copyright and all rights are held exclusively by Springer Science +Business Media Dordrecht. This e-offprint is for personal use only and shall not be selfarchived in electronic repositories. If you wish to self-archive your article, please use the accepted manuscript version for posting on your own website. You may further deposit the accepted manuscript version in any repository, provided it is only made publicly available 12 months after official publication or later and provided acknowledgement is given to the original source of publication and a link is inserted to the published article on Springer's website. The link must be accompanied by the following text: "The final publication is available at link.springer.com”. 1 23 Author's personal copy Metascience DOI 10.1007/s11016-014-9927-2 ESSAY REVIEW Robustness analysis versus reliable process reasoning Robert Hudson: Seeing things: The philosophy of reliable observation. Oxford: Oxford University Press, 2014, xii+274pp, £41.99, $58.50 HB Chiara Lisciandra Ó Springer Science+Business Media Dordrecht 2014 Robert Hudson’s book is a contribution to the recent debate on robustness analysis in scientific practice, with a specific focus on the empirical sciences. In this context, robustness analysis is defined as a way to increase the probability of a certain hypothesis by showing that the same result is obtained from several, alternative methods. The rationale underlying this practice is that it would be highly unlikely if different, independent means of observation provided the same wrong outcome.We do not believe in miracles; hence, the probability of the initial hypothesis being true increases if the same result occurs across conditions. Simple as this notion sounds, according to Hudson, the various attempts that have been offered in the literature to formalize it have proved to be unsuccessful. Hudson’s skeptical argument is formulated at the outset of the book and illustrated throughout by means of several episodes from the recent history of science, mainly from physics and biology. Meticulous reconstructions of the discovery of atoms and molecules, the controversy about the existence of the mesosome, and the debate around dark matter and dark energy are offered. The main argument that Hudson presents unfolds as follows. Imagine that a group of scientists, or different scientific teams, had several different experimental procedures at their disposal, all of uncertain reliability. Suppose that the observations from the various instrumental procedures were to converge on the same result. It might be argued that it is more probable that different, independent means of observation detect the same result if that result is true, than if that result is false. However, Hudson claims that if the reliability of the various different procedures is uncertain, then the epistemic advantage in switching from one procedure to the other is questionable. In such situations, scientists should first worry about making the instruments more reliable, rather than pointing out that the same result is consistent across conditions. C. Lisciandra (&) Social and Moral Philosophy, Department of Political and Economic Studies, University of Helsinki, PO Box 24, 00014 Helsinki, Finland e-mail: [email protected] 123 Author's personal copy Metascience There have been numerous cases in the history of science—some of which are presented by Hudson in the central chapters of the book—in which convergence on the same outcome turned out to be a blunder; progress was made once that better, more reliable instruments were devised and became available to scientists. Since robustness arguments have been used in support of both successful and unsuccessful results, relying on this strategy appears weak after all. Furthermore, intuitive as it sounds, the argument for the value of robustness analysis conceals fundamental difficulties. To start with, Hudson draws a distinction between robustness analysis as (a) a method of adopting various inferential machineries, whose error term is known, so as to aggregate their outcomes in such a way that the error entailed by each procedure cancels out across conditions (e.g., when the measurements of two balances with different error terms are compared with one another); and (b) as a method for discovering new physical entities or phenomena. In the first case, the reliability of different instruments is known and their independence is given. In the latter case, the reliability of the various instruments should be deduced from their convergent result. Hudson claims that neither case is immune to problems. In the first case, the nature of the difficulties is mainly pragmatic, insofar as the aggregation of results coming from various instruments is far from trivial. Also—but this is my addition— it could be asked why we should rely on results deriving from a less reliable procedure to confirm results deriving from a more reliable one. The less accurate procedure will give estimates that are more biased than the original one; thus, it is not clear how such practice can increase confirmation. In the second case, the problems are mainly epistemic and concern the significance of results given by various procedures whose reliability and independence cannot be easily assessed. In new frontiers of science especially, convergence on the same result (or the lack of it) is not always a knockdown argument in favor of (or against) the hypothesis under scrutiny. Suppose the various instruments did not detect the same effect. Should scientists question the initial result, or should they exclude those instruments that did not detect that result? The main problem, which echoes the experimenter’s regress argument, is that if the ability to capture a certain phenomenon is the criterion for establishing the reliability of an instrument, then negative results might be excluded at the outset. If so, the robust results will ultimately not stand as particularly significant. The previous issues are presented in the first chapter of the book, where Hudson reviews the most important philosophical arguments in support of robustness analysis, respectively, pragmatic, probabilistic, and epistemic arguments. He aims to rebut each of them and to replace robustness analysis with ‘a reliable reasoning process’, i.e., a procedure to assess the reliability of our scientific tools. In the following chapters, his strategy is to analyze whether robustness analysis in scientific practice has been successful. If it were so, then—even if the epistemological arguments have not been decisive so far—the results could speak for themselves. Notwithstanding the supporters of robustness analysis, however, Hudson argues that the main cases usually presented in the literature have been mistakenly interpreted as successful examples of robustness analysis. 123 Author's personal copy Metascience The first case that Hudson discusses is the debate on the existence of the mesosome, which engaged biologists in the sixties and the seventies of the last century. In the philosophical literature, the discovery that the mesosome was a byproduct of a specific experimental apparatus has been defended as a successful case of robustness analysis (see e.g., Culp 1994). According to the defenders of this view, it was because independent means of observation produced contrasting results that scientists concluded that the mesosome was in fact an experimental artifact. At the same time, they disregarded convergent results coming from practices whose independence could not be similarly assessed. To this interpretation, Hudson objects that the key factor in denying the existence of the mesosome was simply an acknowledgment that the procedures through which the mesosome was revealed were unreliable. It was only when it became clear that chemical fixation damages bacterial morphology, and new cryofixation techniques were developed, that the hypothesis about the mesosome was disproved. According to Hudson, robustness analysis did not play the role that was attributed to it, since the discovery of the mesosome was ultimately just the result of a reliable reasoning process. The episode of the mesosome is mainly known to philosophers working in the experimental robustness niche; but Hudson proceeds to address one of the most famous cases in the philosophy of science literature, namely Perrin’s discovery of atoms and molecules. The story usually goes that the consistency of the result through different and independent methods of measurement was decisive in ruling out the possibility of the result being the effect of one specific measurement tool. Once again, according to Hudson, the previous reconstruction is misleading. In support of his claim, Hudson calls the reader’s attention to many excerpts from Perrin’s works (Perrin 1910, 1916, 1926). He attempts to understand what was going on in the mind of the scientist, whether it was a form of robustness analysis or a different way of reasoning. Although Perrin himself admits his astonishment at the agreement of the results obtained by different physical processes, according to Hudson this was not what convinced the scientist that he had finally proved the atomic nature of matter. As Hudson reconstructs the story, it was one specific experimental apparatus, the experiment on Brownian motion with emulsions, that possessed for Perrin a higher degree of epistemic authority, which made it the standard by which to calibrate other methods. In other words, the key issue was not independent convergence on that result, but rather the fact that it was derived from the most reliable approach available. Finally, with the discussion on dark matter and dark energy, Hudson brings us to one of the most exciting enterprises in modern science. In the interest of space, I will mainly refer to the case of dark matter and its composition. The focus of Hudson’s argument is on three different research groups, based, respectively, in Italy, the United States, and France, each of which devised different procedures for their purpose. Hudson’s point is to show how rarely it happens in science that different groups rely on robust results in support of their respective arguments. Even when agreement is achieved between different laboratories, as in the case of some of the results of the dark matter project, scientists tend to base their arguments in favor of their own analysis on the virtues it has with respect to alternative ones; or on the ways in which it remedies the deficiencies of other approaches. Scientists usually do 123 Author's personal copy Metascience not refer to other groups’ results to strengthen their own position. The underlying rationale is that, as long as their instruments are reliable, they can trust their outcomes without the need to refer to the results derived from different methods, which would make their achievements superfluous or unnecessary. Both because scientific practice does not rely on robustness analysis and because this method is epistemically dubious, Hudson’s conclusion is to favor the skeptics. The last part of the book is dedicated to the consequences of the debate on robustness analysis for scientific realism. The aspect of the book I appreciated the most is that the arguments for and against robustness are embedded in actual scientific practice; at times, however, Hudson risks losing the reader in the details of historical reconstruction. Through the exposition of the argument, several questions may come to the mind. One of the main issues is that previous treatments of robustness analysis did not include the notion of reliability in the picture—and for good reason. It is indeed when we do not have information about the reliability of our instruments that convergence on the same outcome becomes relevant (Kuorikoski et al. 2010). The process that leads to scientific accomplishments is rarely the one where reliability is a given, or where a single experiment settles the conflicts between competing hypotheses once and for all. It is most likely the case that a process of continuous adjustment between the experiments we devise and the results we get from them is what fosters progress in science. If we rebut robustness analysis on the grounds that we are not justified in deriving conclusions unless a certain reliability of our tools is proved, then we should suspend judgment about the outcomes of our experiments across the board. Ironically enough, Hudson’s argument is developed and supported via different case studies: Is he himself relying on robustness analysis in support of his position? Would it not be enough, according to his view, to have one decisive case study to support his argument? To this objection, Hudson may reply that he is not supporting his argument via robustness analysis, as he is listing a number of cases to show that scientists do not apply this method in their work. In other words—he may say—he is simply reasoning by induction. The differences between induction and robustness, and between the confirmations we get from repeating the same experiment versus changing it, are discussed in one of the most interesting and thought-provoking sections of the book. On a different note, the definition of robustness analysis stated at the outset of the book strongly recalls the notion of coherence in epistemology. It would be a further contribution to the debate to clarify whether the two notions overlap or not. If they do, then a greater attention should have been given to the large body of literature dedicated to coherence in epistemology (Bovens and Hartmann 2000; Schupbach 2011). If they do not, it would have been interesting to see how they differ. Finally, given his focus on the experimental sciences, Hudson’s analysis takes no notice of a highly contentious area of the debate, which investigates robustness analysis as a method of testing our hypotheses precisely when empirical results are not available. Nonetheless, Seeing Things addresses a number of cutting-edge issues in the domain of experimental robustness, which makes this book an important contribution to our understanding of the notion of robustness. This is one of the first monographs to 123 Author's personal copy Metascience present different accounts of robustness analysis in a unified picture. It is particularly welcome in an area of research which is dispersed over various subfields; even more so because, despite his focus on experimental robustness, Hudson opens new perspectives on the topic, which will most likely have a bearing also on the non-empirical side of the debate. Acknowledgments With thanks to Alessandra Basso, Aki Lehtinen and Caterina Marchionni who took part in the reading group on Seeing Things at the University of Helsinki. References Bovens, L., and S. Hartmann. 2000. Coherence, belief expansion and Bayesian networks. In Proceedings of the 8th International Workshop on Non-Monotonic Reasoning, NMR’2000, eds. Baral C., Truszczynski M, Breckenridge, Colorado. Culp, S. 1994. Defending robustness: The bacterial mesosome as a test case. In PSA 1994: Proceedings of the 1994 biennial meeting of the philosophy of science association, vol. 1, ed. D. Hull, et al., 46–57. East Lansing, MI: Philosophy of Science Association. Kuorikoski, J., A. Lehtinen, and C. Marchionni. 2010. Economic modelling as robustness analysis. British Journal of Philosophy of Science 61: 541–567. Perrin, J. 1910. Brownian Movement and Molecular Reality. Trans. Soddy F. London: Taylor and Francis. Perrin, J. 1916. Atoms. Trans. Hammick D. L. London: Constable. Perrin, J. 1926. Discontinuous structure of matter. In Nobel lectures: Physics, 1922–1941, ed. Nobel Foundation, 138–164. New York: Elsevier. Schupbach, J. 2011. New hope for Shogenji’s coherence measure. British Journal of Philosophy of Science 62: 125–142. 123
© Copyright 2026 Paperzz