Timothy D Lyons - Peter Vickers - Contemporary Scientific Realism - The Challenge From The History of Science-Oxford University Press, USA (2021)
Timothy D Lyons - Peter Vickers - Contemporary Scientific Realism - The Challenge From The History of Science-Oxford University Press, USA (2021)
Contemporary Scientific
Realism
The Challenge from the History of Science
Edited by
T I M O T H Y D. LYO N S A N D P E T E R V IC K E R S
1
3
Oxford University Press is a department of the University of Oxford. It furthers
the University’s objective of excellence in research, scholarship, and education
by publishing worldwide. Oxford is a registered trade mark of Oxford University
Press in the UK and certain other countries.
DOI: 10.1093/oso/9780190946814.001.0001
1 3 5 7 9 8 6 4 2
Printed by Integrated Books International, United States of America
Contents
PA RT I H I S T O R IC A L C A SE S F O R T H E D E BAT E
PA RT I I C O N T E M P O R A RY S C I E N T I F IC R E A L I SM
Index 375
1
History and the Contemporary Scientific
Realism Debate
Timothy D. Lyons1 and Peter Vickers2
1Department of Philosophy
Durham University
[email protected]
The scientific realism debate began to take shape in the 1970s, and, with the publica-
tion of two key early ’80s texts challenging realism, Bas van Fraassen’s The Scientific
Image (1980) and Larry Laudan’s “A Confutation of Convergent Realism” (1981),
the framework for that debate was in place. It has since been a defining debate of
philosophy of science. As originally conceived, the scientific realism debate is one
characterized by dichotomous opposition: “realists” think that many/most of our
current best scientific theories reveal the truth about reality, including unobservable
reality (at least to a good approximation); and they tend to justify this view by the
“no miracles argument,” or by an inference to the best explanation, from the success
of scientific theories. Further, many realists claim that scientific realism is an empir-
ically testable position. “Antirealists,” by contrast, think that such a view is lacking in
epistemic care. In addition to discussions of the underdetermination of theories by
data and, less commonly, competing explanations for success, many antirealists—in
the spirit of Thomas Kuhn, Mary Hesse, and Larry Laudan—caution that the his-
tory of science teaches us that empirically successful theories, even the very best
scientific theories, of one age often do not stand up to the test of time.
The debate has come a long way since the 1970s and the solidification of its frame-
work in the ’80s. The noted dichotomy of “realism”/“antirealism” is no longer a
given, and increasingly “middle ground” positions have been explored. Case studies
of relevant episodes in the history of science show us the specific ways in which a
realist view may be tempered, but without necessarily collapsing into a full-blown
antirealist view. A central part of this is that self-proclaimed “realists,” as well as
“antirealists” and “instrumentalists,” are exploring historical cases in order to learn
Timothy D. Lyons and Peter Vickers, History and the Contemporary Scientific Realism Debate In: Contemporary Scientific
Realism. Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press 2021.
DOI: 10.1093/oso/9780190946814.003.0001
2 Contemporary Scientific Realism
the relevant lessons concerning precisely what has, and what hasn’t, been retained
across theory change. In recent decades one of the most important developments
has been the “divide et impera move,” introduced by Philip Kitcher (1993) and
Stathis Psillos (1999)—possibly inspired by Worrall (1989)—and increasingly
embraced by numerous other realists. According to this “selective” strategy, any re-
alist inclinations are directed toward only those theoretical elements that are really
doing the inferential work to generate the relevant successes (where those successes
are typically explanations and predictions of phenomena). Such a move is well
motivated in light of the realist call to explain specific empirical successes: those the-
oretical constituents that play no role, for instance, in the reasoning that led to the
successes, likewise play no role in explaining those successes. Beyond that, however,
and crucially, the divide et impera move allows for what appears to be a testable re-
alist position consistent with quite dramatic theory-change: the working parts of a
rejected theory may be somehow retained within a successor theory, even if the two
theories differ very significantly in a great many respects. Even the working parts
may be retained in a new (possibly approximating) form, such that the retention
is not obvious upon an initial look at a theoretical system but instead takes consid-
erable work to identify. This brings us to a new realist position, consistent with the
thought that many of our current best theories may one day be replaced. Realism
and antirealism are no longer quite so far apart, and this is progress.
A major stimulus for, and result of, this progress has been a better under-
standing of the history of science. Each new case study brings something new to
the debate, new lessons concerning the ways in which false theoretical ideas have
sometimes led to success, or the ways in which old theoretical ideas “live on,” in
a different form, in a successor theory. At one time the debate focused almost
exclusively on three famous historical cases: the caloric theory of heat, the phlo-
giston theory of combustion, and the aether theory of light. An abundance of lit-
erature was generated, and with good reason: these are extremely rich historical
cases, and there is no simple story to tell of the successes these theories enjoyed,
and the reasons they managed to be successful despite being very significantly
misguided (in light of current theory). But the history of science is a big place,
and it was never plausible that all the important lessons for the debate could be
drawn from just three cases. This is especially obvious when one factors in a “par-
ticularist” turn in philosophy of science where focus is directed toward partic-
ular theoretical systems. These days, many philosophers are reluctant to embrace
grand generalizations about “science” once sought by philosophers, taking those
generalizations to be a dream too far. Science works in many different ways, in
different fields and in different contexts. It follows that the realism debate ought
to be informed by a rich diversity of historical cases.
It is with this in mind that the present volume is put forward. In recent years
a flood of new case studies has entered the debate, and is just now being worked
History and the Contemporary Scientific Realism Debate 3
through. At the same time it is recognized that there are still many more cases
out there, waiting to be analyzed in a particular way sensitive to the concerns
of the realism debate. The present volume advertises this fact, introducing as it
does several new cases as bearing on the debate, or taking forward the discus-
sion of historical cases that have only very recently been introduced. At the same
time, the debate is hardly static, and as philosophical positions shift this affects
the very kind of historical cases that are likely to be relevant. Thus the work of
introducing and analyzing historical cases must proceed hand in hand with phil-
osophical analysis of the different positions and arguments in play. It is with this
in mind that we divide the present volume into two parts, covering “Historical
Cases for the Debate” and “Contemporary Scientific Realism,” each comprising
of seven chapters.
Many of the new historical cases are first and foremost challenges to the re-
alist position, in that they tell scientific stories that are apparently in tension with
even contemporary, nuanced realist claims. The volume kicks off in Chapter 2
with just such a case, courtesy of Dana Tulodziecki. For Tulodziecki, the miasma
theory of disease delivered very significant explanatory and predictive successes,
while being radically false by present lights. Further, even the parts of the theory
doing the work to bring about the successes were not at all retained in any suc-
cessor theory. Thus, Tulodziecki contends, the realist must accept that in this case
false theoretical ideas were instrumental in delivering successful explanations
and predictions of phenomena.
This theme continues in Chapter 3, with Jonathon Hricko’s study of the dis-
covery of boron. This time the theory in question is Lavoisier’s oxygen theory
of acidity. Hricko argues that the theory is not even approximately true, and yet
nevertheless enjoyed novel predictive success of the kind that has the power to
persuade. Just as Tulodziecki, Hricko argues that the realist’s divide et impera
strategy can’t help—the constituents of the theory doing the work to generate the
prediction cannot be interpreted as approximately true, by present lights.
We meet with a different story in Chapter 4, however. Keith Hutchison provides
a new case from the history of thermodynamics, concerning the successful pre-
diction that pressure lowers the freezing point of water. Hutchison argues that al-
though, at first, it seems that false theoretical ideas were instrumental in bringing
about a novel predictive success, there is no “miracle” here. It is argued that the
older Carnot theory and the newer Clausius theory are related in intricate ways,
such that in some respects they differ greatly, while in “certain restricted situations”
their differences are largely insignificant. Thus the realist may find this case useful
in her bid to show how careful we need to be when we draw antirealist morals from
the fact that a significantly false theory enjoyed novel predictive success.
Chapter 5 takes a slightly different approach, moving away from the narrow
case study. Stathis Psillos surveys a broad sweep of history ranging from
4 Contemporary Scientific Realism
Descartes through Newton and Einstein. We have here a “double case study,”
considering the Descartes–Newton relationship and the Newton–Einstein rela-
tionship. Psillos argues against those who see here examples of dramatic theory-
change, instead favoring a limited retentionism consistent with a modest realist
position. Crucially, Psillos argues, there are significant differences between the
Descartes–Newton relationship and the Newton–Einstein relationship, in line
with the restricted and contextual retention of theory predicted by a nuanced,
and epistemically modest, contemporary realist position.
Chapter 6—courtesy of Eric Scerri—introduces another new historical case,
that once again challenges the realist. This time the theory is John Nicholson’s
atomic theory of the early 20th century, which, Scerri argues, “was spectacu-
larly successful in accommodating as well as predicting some spectral lines in
the solar corona and in the nebula in Orion’s Belt.” The theory, however, is very
significantly false; as Scerri puts it, “almost everything that Nicholson proposed
was overturned.” Hence, this case is another useful lesson in the fact that quite
radically false scientific theories can achieve novel predictive success, and any
contemporary realist position needs to be sensitive to that.
Chapter 7 turns to theories of molecular structure at the turn of the 20th
century. Amanda Nichols and Myron Penner show how the “old” Blomstrand-
Jørgensen chain theory was able to correctly predict the number of ions that will
be dissociated when a molecule undergoes a precipitation reaction. While prima
facie a challenge to scientific realism, it is argued that this is a case where the di-
vide et impera strategy succeeds: the success-generating parts of the older theory
are retained within the successor, Werner’s coordination theory.
The final contribution to the “History” part of the volume—Chapter 8—
concerns molecular spectroscopy, focusing on developments in scientific
“knowledge” and understanding throughout the 20th century and right up to
the present day. Teru Miyake and George E. Smith take a different approach from
the kind of historical case study most commonly found in the realism debate.
Siding with van Fraassen on the view that Perrin’s determination of Avogadro’s
number, so commonly emphasized by realists, does not prove fruitful for re-
alism, and focusing on diatomic molecules, they emphasize instead the extraor-
dinary amount of evidence that has accumulated after Perrin and over the past
ninety years for various theoretical claims concerning such molecules. Taking
van Fraassen’s constructive empiricism as a foil, they indicate that, in this case,
so-called realists and antirealists may really differ very little when it comes to this
area of “scientific knowledge.”
The second part of the volume turns to more general issues and philosoph-
ical questions concerning the contemporary scientific realist positions. This
part kicks off in Chapter 9 with Mario Alai’s analysis of the divide et impera
realist strategy: he argues that certain historical cases no longer constitute
History and the Contemporary Scientific Realism Debate 5
The editors of this volume owe a great debt to the participants of all of these
events, with special thanks in particular to those who walked this path with us
a little further to produce the fourteen excellent chapters here presented. We are
also grateful to the unsung heroes, the many anonymous reviewers, who not
only helped us to select just which among the numerous papers submitted would
History and the Contemporary Scientific Realism Debate 7
be included in the volume but also provided thorough feedback to the authors,
helping to make each of the chapters that did make the cut even stronger. The
volume has been a labor of love; we hope that comes across to the reader.
References
Kitcher, P. (1993). The Advancement of Science, Oxford: Oxford University Press.
Laudan, L. (1981). “A Confutation of Convergent Realism.” Philosophy of Science
48: 19–49.
Psillos, S. (1999). Scientific Realism: How Science Tracks Truth, London: Routledge.
Van Fraassen, B. (1980). The Scientific Image, Oxford: Oxford University Press.
Worrall, J. (1989). “Structural Realism: The Best of Both Worlds?” Dialectica 43: 99–124.
PART I
HISTOR IC A L CASE S
F OR T HE DE BAT E
2
Theoretical Continuity, Approximate
Truth, and the Pessimistic Meta-Induction:
Revisiting the Miasma Theory
Dana Tulodziecki
Department of Philosophy
Purdue University
[email protected]
2.1 Introduction
The pessimistic meta-induction (PMI) targets the realist’s claim that a theory’s
(approximate) truth is the best explanation for its success. It attempts to do so
by undercutting the alleged connection between truth and success by arguing
that highly successful, yet wildly false theories are typical of the history of sci-
ence. There have been a number of prominent realist responses to the PMI, most
notably those of Worrall (1989), Kitcher (1993), and Psillos (1999). All of these
responses try to rehabilitate the connection between a theory’s (approximate)
truth and its success by attempting to show that there is some kind of continuity
between earlier and later theories, structural in the case of Worrall and theo-
retical/referential in the cases of Kitcher and Psillos, with other responses being
variations on one of these three basic themes.1
In this paper, I argue that the extant realist responses to the PMI are in-
adequate, since there are cases of theories that were both false and highly
successful (even by the realist’s own, more stringent criteria for success)
but that, nevertheless, do not exhibit any of the continuities that have been
1 It is worth pointing out that this is the case even for the (very) different structural realisms
that abound. In particular, even ontic structural realism, which differs substantially from
Worrall’s epistemic version, shares with Worrall’s approach an emphasis on mathematical con-
tinuities among successor theories (see, for example, Ladyman (1998) and French and Ladyman
(2011)).
Dana Tulodziecki, Theoretical Continuity, Approximate Truth, and the Pessimistic Meta- Induction: Revisiting the
Miasma Theory In: Contemporary Scientific Realism. Edited by: Timothy D. Lyons and Peter Vickers, Oxford University
Press. © Oxford University Press 2021. DOI: 10.1093/oso/9780190946814.003.0002
12 Historical Cases for the Debate
The PMI received its most sophisticated and explicit formulation in Laudan
(1981, 1984). The argument’s main target is the Explanationist Defense of
Realism, according to which the best explanation for the success of science is
the (approximate) truth of our scientific theories (see, for example, Boyd (1981,
1984, 1990)). Anti-realists employ the PMI to undercut this alleged connection
between truth and success by pointing to the (in their view) large number of
scientific theories that were discarded as false, yet regarded as highly successful.
Laudan, for example, provides a list of such theories that includes, among others,
the phlogiston theory of chemistry, the caloric theory of heat, and the theory
of circular inertia (1984: 121). However, since “our own scientific theories are
held to be as much subject to radical conceptual change as past theories” (Hesse
1976: 264), it follows, the anti-realist argues, that the success of our current the-
ories cannot legitimize belief in unobservable entities and mechanisms: just as
Revisiting the Miasma Theory 13
past theories ended up wrong about their postulates, we might well be wrong
about ours.2
Realist responses to the PMI typically come in several stages: first, realists try to
winnow down Laudan’s list by including only those theories that are genuinely suc-
cessful (Psillos 1999: chapter 5). While, according to Laudan, a theory is successful
“as long as it has worked reasonably well, that is, so long as it has functioned in a va-
riety of explanatory contexts, has led to several confirmed predictions, and has been
of broad explanatory scope” (1984: 110), Psillos argues that “the notion of empirical
success should be more rigorous than simply getting the facts right, or telling a story
that fits the facts. For any theory (and for that matter, any wild speculation) can be
made to fit the facts—and hence to be successful—by simply “writing” the right kind
of empirical consequences into it. The notion of empirical success that realists are
happy with is such that it includes the generation of novel predictions which are in
principle testable” (1999: 100). The specific type of novel prediction that Psillos has in
mind is so-called use-novel prediction: “the prediction P of a known fact is use-novel
relative to a theory T, if no information about this phenomenon was used in the con-
struction of the theory which predicted it” (101). Once success is understood in these
more stringent terms, Psillos claims, Laudan’s list is significantly reduced.
With the list so shortened, the next step of the realist maneuver is to argue that
those theories that remain on Laudan’s list ought to be regarded as (approximately)
true, since they don’t, in fact, involve the radical discontinuity with later theories
that Laudan suggests. Rather than being discarded wholesale during theory-change,
it is argued that important elements of discarded theories get retained: according to
Worrall (1989, 1994), theories’ mathematical structures are preserved, according to
Kitcher (1993) and Psillos (1996, 1999), those parts of past theories that were re-
sponsible for their success are. Because these components are preserved in later the-
ories, the argument goes, we ought to regard as approximately true those parts of
our current theories that are essentially involved in generating their successes, since
those are the parts that will carry over to future theories, just as essential elements
from earlier theories were carried over to our own. As I will show in section 2.4,
however, none of these strategies will work for the miasma theory: despite the fact
that the miasma theory made a number of use-novel predictions, and so counts as
genuinely successful by realist standards, none of the elements or structures that
were involved in those successes were retained by its successor.3
2 Laudan’s argument is usually construed as a reductio (see Psillos (1996, 1999)). For some recent
discussions about how to properly interpret the argument, see Lange (2002), Lewis (2001), and Saatsi
(2005).
3 The miasma case is especially significant in view of the markedly different analysis that Saatsi
and Vickers (2011) provide of Kirchhoff ’s theory of diffraction. Saatsi and Vickers diagnose a specific
kind of underdetermination at work in the Kirchhoff case and argue that “it should not be implausible
14 Historical Cases for the Debate
The most sophisticated version of the miasma theory saw its heyday in the
mid-1800s. The situation with respect to the various accounts of disease at that
time was fairly complicated, however, and so it is in some sense misleading to
speak of the miasma theory of disease, since there was a whole cluster of related
views that went under this label, rather than one easily identifiable position
(see, for example, Baldwin (1999), Eyler (2001), Hamlin (2009), Pelling (1978),
and Worboys (2000)). However, since all members of that cluster shared basic
assumptions about the existence and nature of miasma, I will disregard this
complication here, and treat them as one. According to this basic miasma view,
diseases were caused and transmitted by a variety of toxic odors (“miasmas”) that
themselves were the result of rotting organic matter produced by putrefaction or
decomposition. The resulting bad air would then act on individual constitutions
to cause one of several diseases (cholera, yellow fever, typhus, etc.), depending
on a number of more specific factors. Some of these were thought to be extra-
neous, such as weather, climate, and humidity, and would affect the nature of
the miasmas themselves; others were related directly to the potential victims and
thought to render them more or less susceptible to disease, such as their general
constitution, moral sense, age, and so on. Lastly, there were a variety of local
conditions that could exacerbate the course and severity of the disease, such as
overcrowding and bad ventilation.
Although the miasma theory is sometimes contrasted with so- called
“contagionist” views of disease (the view that diseases could be transmitted di-
rectly from individual to individual), this opposition is also misleading, for two
reasons: first, because both contagionists and anti-contagionists subscribed to
the basic miasmatic assumptions just described, with the debate not centering
on the existence or nature of miasmas, but, rather, on whether people them-
selves were capable of producing additional miasma-like effects and through
these “exhalations” directly give the disease to others. Second, although some
diseases were generally accepted as contagious (smallpox, for example), and
some were generally held to be non-contagious (malaria), most diseases (yellow
fever, cholera, typhoid fever, typhus, and so on) fell somewhere in between these
to anyone that given the enormous variation in the nature and methods of scientific theories across
the whole spectrum of ‘successful science’, some domains of enquiry can be more prone to this kind
of underdetermination than others” (44). In fact, they believe that “witnessing Kirchhoff ’s case, there
is every reason to expect that from a realist stance we can grasp the features of physics and mathe-
matics that contribute to such differences” (44). As a result, they take themselves to have “argued for
the prima facie plausibility” of a realist response that focuses on “showing how the field of theorizing
in question is idiosyncratic in relevant respects, so that Kirchhoff ’s curious case remains isolated and
doesn’t provide the anti-realist with grounds for projectable pessimism” (44). The miasma case, how-
ever, is not prone to any of the idiosyncrasies Saatsi and Vickers identify (or any other idiosyncrasies,
as far as I can tell).
Revisiting the Miasma Theory 15
two extremes. Pelling (1978: 9) appropriately terms these diseases the “doubtful
diseases”: instead of being straightforwardly contagionist or anti-contagionist
about them, people espoused so called “contingent contagionism” with respect to
them, holding that they could manifest as either contagious or non-contagious,
depending on the exact circumstances and, sometimes, even transform from one
into the other (see also Hamlin (2009)).
This version of the miasma theory was extremely successful with respect to a
number of different types of phenomena. Most famous are probably the sanitary
successes that it ultimately inspired, but it also managed to provide explanations
of disease phenomena that any theory of disease at the time had to accommo-
date. These included the fact that diseases were known to be highly seasonal, that
particular regions (especially marshy ones) were affected particularly harshly,
that specific locations (prisons, workhouses, etc.) were often known to suffer
worse than their immediate surroundings, that sickness and mortality rates in
urban centers were much worse than those in rural areas, and why particular
geographical regions/countries were struck much worse by disease than others.
The miasma theory managed to explain all of these through its claims that de-
composition and putrefaction of organic material was responsible for producing
miasmas. Diseases peaked when conditions for putrefaction were particularly
favorable: this was the reason why certain diseases were particularly bad during
periods of high temperature and in certain geographical regions (for example,
the many fevers in Africa), why urban centers were much more affected than
rural areas, and why even specific locations in otherwise more or less healthy
areas could be struck (sewage, refuse, and general “filth” would sit around in
badly ventilated areas). It should also be noted that miasma theorists were not
just making vague or simple-minded pronouncements about stenches produ-
cing toxic odors, but embraced very specific and often highly complex accounts
of how various materials and conditions gave rise to miasmas—Farr, for ex-
ample, drew in some detail on Liebig’s chemical explanations (1852, lxxx-lxxxiii;
see also Pelling (1978: chapters 3 and 4) and Tulodziecki (2016)). Moreover,
there were debates about exactly what sorts of materials were prime for potential
miasmas, such as debates about various sources of vegetable vs. animal matter,
and so on. Since, however, I cannot do justice to the details of these accounts and
their corresponding successes here, I will focus on a somewhat simpler, yet par-
ticularly striking, example of use-novel success, while merely noting that others
could be given.
The example in question is that of William Farr’s (then famous) elevation law
(1852). Farr (1807–83), although not himself a physician, was viewed as an au-
thority on infectious diseases in mid-1800s Britain. Among the positions he
held were that of Statistical Superintendent of the General Register Office and
member of the Committee of Scientific Inquiries. Through his various writings,
16 Historical Cases for the Debate
especially those on the various big British cholera epidemics, he established his
credentials as a medical authority.4
As we have seen, according to the miasma theory, any decomposing organic
material could in principle give rise to miasmas. However, it was thought that
the soil at low elevations, especially around riverbanks, was a particularly good
source for producing highly concentrated miasma, since such soil held plenty of
organic material and the conditions for putrefaction were particularly favorable.
It thus followed directly from the miasma theory that, if miasmas were really
produced in the manner described and responsible for disease, the concentra-
tion of noxious odors ought to be higher closer to such sources and dissipate
with increasing distance. Correspondingly, it was to be expected that mortality
and sickness rates would be higher in close proximity to sources of miasma, de-
clining as one moved away. Farr confirmed that this was the case, in a number of
different ways.
First, he found that “nearly 80 percent of the 53,000 registered cholera deaths
in 1849 occurred among four-tenth of the population living on one-seventh of
the land area” (Eyler 1979: 114), and, moreover, that, “cholera was three times
more fatal on the coast than in the interior of the country” (Farr 1852: lii).
Furthermore, he found that those deaths that did occur inland were in either
seaport districts or close to rivers, noting that “[c]holera reigned wherever it
found a dense population on the low alluvial soils of rivers” and that it “was al-
most invariably most fatal in the port or district lying lowest down the river”
(lii). Concerning coastal deaths, he found that the “low-lying towns on the coast
were all attacked by cholera,” while the high-lying coast towns “enjoyed as much
immunity as the inland towns” (liv). Further, he noted that mortality increased
and decreased relative to the size of the port, with smaller ports having lower
mortality. The Welsh town of Merthyr-Tydfil constituted an exception to this,
having a “naturally” favorable location, yet relatively high mortality. However,
Farr also noted that Merthyr-Tydfil had a reputation, with the Health of Towns’
Commissioners’ Report noting that “[f]rom the poorer class of the inhabitants,
who constitute the mass of the population, throwing all slops and refuse into
the nearest open gutter before their houses, from the impeded courses of such
channels, and the scarcity of privies, some parts of town are complete networks of
filth emitting noxious exhalations” (liv). In addition, much of the refuse was being
carried to the local riverbeds, with the result that “the stench is almost intolerable
in many places” (lv). In one area, close to the river, an “open, stinking, and nearly
stagnant gutter, into which the house refuse is as usual flung, moves slowly before
the doors” (lvi).
All of these phenomena were exactly what was to be expected on the mi-
asma theory: wherever there was disease, one ought to be able to trace it back
to miasmatic conditions and, similarly, wherever conditions particularly favor-
able to decomposition were to be found, disease ought to be rampant. If there
were exceptions to the general rule, towns such as Merthyr-Tydfil that were not
naturally vulnerable, one ought to be able to find alternative sources of miasma
without difficulty. Farr also proceeded to check these results against a variety of
data from different countries and concerning different diseases, and further con-
firmed what he had found. Moreover, the miasma theory did not merely accom-
modate these findings but, rather, all of the phenomena followed naturally from
the account.
In addition, while the miasma theory made predictions about what areas
ought to be affected by cholera and to what degrees, for example, none of these
predictions were confirmed until Farr analyzed the data from the General Board
of Health in the late 1840s (indeed, due to the fact that much of this data was not
collected until shortly before that time, it would have been impossible to confirm
these predictions until then). Thus, since it was not even clearly known to what
extent these predictions were borne out, they could not have played a role in
formulating the miasma theory in the first place and, so, ought to count as use-
novel. One might object that, use-novel or not, these predictions were simply too
vague to qualify a theory as genuinely successful in the realist sense. I will note,
however, that (i) the predictions were as specific as a theory of this type would
allow, (ii) the predictions were no more vague than Snow’s later predictions
about what ought to be expected on the (correct) assumption that cholera was
waterborne (see Snow 1855a, 1855b) and, so, if one regards Snow’s predictions as
successful, one ought to also regard the miasmatic predictions as successful, and
(iii) that miasmatists actually did a lot better than this (and, indeed, better than
Snow ever did) by providing some detailed and quantifiable results.
Farr’s elevation law is a particularly striking example of this and it is to this
that I will turn now. Farr’s law related cholera mortality to the elevation of the
soil. However, Farr did not just predict that there ought to be a relationship
between these two variables, but upon analyzing more than 300 pages of data,
he found that the “mortality from cholera is in the inverse ratio of the eleva-
tion” (Farr 1852: lxi). Farr grouped the various London districts by their alti-
tude above the Thames high water mark, in brackets of 20 feet (0–20, 20–40,
and so on) and was able to capture the exact relation between the decline of
cholera and increased soil elevation in the form of an equation. Specifically, he
found that:
The mortality from cholera on the ground under 20 feet high being represented
by 1, the relative mortality in each successive terrace [i.e. the terraces numbered
18 Historical Cases for the Debate
Figure 2.1 Farr, William. (1852). “Report on the Mortality of Cholera in England,
1848–49.” London: Printed by W. Clowes, for H.M.S.O., 1852, p. lxii.
“2,” “3,” etc.] is represented by ½ [for terrace 2], ⅓ [for terrace 3], ¼, ⅕, ⅙: or
the mortality on each successive elevation is ½, ⅔, ¾, ⅘, ⅚, &c. of the mortality
on the terrace immediately below it. (ibid.: lxiii; see Figure 2.1)
He then generalized this result: “Let e be any elevation within the observed
limits 0 and 350, and c be the average rate of mortality from cholera at that eleva-
tion; also let eʹ be any higher elevation, and cʹ the mortality at that higher eleva-
tion” (lxiii). Then, adding a as a constant, Farr found that the formula
c = c′ × (e′ + a)/(e + a)
Figure 2.2 Farr, William. (1852). “Report on the Mortality of Cholera in England,
1848–49.” London: Printed by W. Clowes, for H.M.S.O., 1852, p. lxiii
convincing they must have seemed, Langmuir (1961: 174) plotted Farr’s result
about cholera mortality in the various London subdistricts, grouped by elevation
(Figure 2.3). According to Langmuir, Farr had “found a confirmation that I be-
lieve would be impressive to any scientist at any time” (173).
Even more remarkably, it turned out that Farr’s predictions did not just hold
for the (sub-)districts of London, but were also confirmed by others in different
regions. For example, “William Duncan, Medical Officer of Health for Liverpool,
wrote that when he grouped the districts of his city by elevation as Farr had
done, that cholera mortality in the last epidemic obeyed Farr’s elevation law for
Liverpool as well” (Eyler 1979: 228).
Now, these predictions of Farr’s were certainly use-novel: Farr was predicting
new phenomena that were hitherto unknown and that were later borne out
by a variety of data from different regions, in different contexts, and from dif-
ferent times. Moreover, Farr’s law clearly could not have played a role in the
construction of the miasma theory since, first, it was obviously not even for-
mulated by then, but, second, even the data on which the law was based (the
statistics from the General Register Office, collected on Farr’s initiative) did not
exist and, indeed, in the case of Duncan in Liverpool, no one had even thought
about collecting the relevant information. In short: it followed from Farr’s law
that cholera mortality and soil elevation ought to exhibit a specific relation that
was then found to occur in the various sub-districts of London and various other
parts of the country.
20 Historical Cases for the Debate
Figure 2.3 Correlation of cholera mortality and elevation above the Thames
River, London, 1849. Langmuir, Alexander D. (1961). “Epidemiology of Airborne
Infection.” Bacteriological Reviews, 25(3), p. 174.
6 Doppelt (2007) argues that requiring novel predictions is not sufficient for genuine success; in-
stead, we ought to also look for various explanatory virtues, such as consilience, simplicity, or uni-
fying power. To make a detailed case for this would take us too far afield here, but it can be shown
that the miasma theory also exhibits some of these more traditional virtues. Certainly one might
argue that independent strands of evidence (epidemiological and pathological, for example) pro-
duced a degree of consilience and that the miasma theory unified a number of different phenomena
(regarding the various diseases) in a simple and elegant way by providing essentially the same kind of
explanation for a number of different afflictions.
Revisiting the Miasma Theory 21
7 Vickers’ discussion of possible refinements of the divide et impera move is worth mentioning
here, in particular his suggestion that, instead of the usual focus on working posits “it remains pos-
sible that we might develop a recipe for identifying certain idle posits” (2013: 209). Vickers develops
an account of how this might have gone in the case of Kirchhoff ’s theory of diffraction. However, it
is unclear how to extend Vickers’ conclusions from the Kirchhoff case to the miasma theory, and so
I will not discuss his views in any detail here. Note, however, that Vickers mentions further potential
cases (see also Lyons (2006), especially his discussion of Kepler’s Mysterium Cosmographicum).
22 Historical Cases for the Debate
smells don’t transmit disease—and neither were any of the laws.8 Specifically, not
only is there no analogue of Farr’s law in any of the modern disease theories, but
not even the phenomenon associated with it—the connection between cholera
mortality and soil elevation—was retained.
Incidentally, this last example is also the best bet as a candidate for Worrall’s
structure. Since Worrall emphasizes mathematical components, and there is
virtually no mathematical structure to be had in the miasma theory, the only
candidates for this view were of the kind put forward by Farr. One might object
that there were other kinds of statistics that were retained: claims, for example,
about the peaks, courses, and durations of epidemics. Moreover, these didn’t in-
volve or rely on miasma or its properties, and so one might think that these are
prime candidates for preservation. The problem with these, however, is that while
these are good candidates for preservation they are not good candidates for re-
alism, since they are all observational. True, they don’t depend on any theoretical
components of the miasma theory (or any other disease theory, for that matter),
but they don’t depend on any other theoretical account either. Rather, this is
merely empirical data—a constraint with which any viable theory of disease has
to work. Realists and anti-realists agree on these data sets, and there’s absolutely
no debate here about approximate truth or realism about unobservables—and
realism about observables was never the issue.
So, to sum up: the miasma theory made several use-novel predictions and so
counts as genuinely successful on the realist’s own terms. Further, it does not
exhibit any kind of realist continuity, neither theoretical, nor referential, nor
structural. There are no miasmas, the laws the theory gave rise to have been
abandoned, its mechanisms of transmission turned out not to exist, and the
properties that were ascribed to miasmas are not now ascribed to any of its etio-
logical successors.
2.5 Objections
One objection realists might raise against Farr is that his prediction was not
really use-novel, since it was based on empirical data and hence insufficiently
grounded in theoretical assumptions. Thus, the realist might argue, as Vickers
has done with respect to Meckel, that Farr “really reached his conclusion
not via his (false) theoretical ideas, but rather via his empirical knowledge”
(Vickers 2015).
8 One also cannot make a case for Psillos’s kind- constitutive properties (see Psillos
1999: chapter 12): miasma is akin in this respect to phlogiston, not the luminiferous ether, and an
analogous version of the argument that Psillos makes against phlogiston will also work for miasma.
Revisiting the Miasma Theory 23
First, note that it is true that once all the data is in, in some sense the original
theory that prompted the prediction can be discarded. However, just because this
is possible does not mean that there was no theoretical basis for the prediction in
the first place. In Farr’s case, there were strong theoretical underpinnings: both
his predictions and the empirical data he used were dependent on his disease
theory. This data was not just readily available for him; instead, he needed to col-
lect and generate it, which, given the extremely limited contemporary means at
his disposal, amounted to a gargantuan effort. Moreover, he did not just have to
gather the raw data, which was already extremely work-intensive but, in order
to make well-formed predictions, he also needed to organize and interpret this
data appropriately. Even here, his work was not straightforwardly empirical but
based on a number of theoretically generated assumptions, for example about
how to construct appropriate mortality rates, and so on. None of these efforts
made sense unless one subscribes to a disease theory according to which such
relationships are to be expected. Without his disease theory, there would never
have been any reason for Farr to gather any of the data he did and, indeed, on
other disease theories—such as Snow’s, say—there would not have been any
point in collecting it. Without Farr, quite possibly the relation between elevation
and cholera mortality would have never been discovered; certainly, without the
accompanying theory, it would have been meaningless.
One might still object that Farr’s prediction amounts to mere curve-fitting and
his data to “more of the same.”9 However, there is nothing “mere” about curve-
fitting in science. Curve-fitting is widespread, standard practice, and it is a mis-
take to think that it involves nothing besides empirical data. On the contrary,
most scientific curve-fitting has important theoretical elements: most obviously,
it has to be decided what the relevant parameters of the curve are, what family of
curves is appropriate, and so on. Further, picking a curve comes with predictions
about future data points, and here different curves will of course make different
predictions. As a result, finding that future data fits one’s curve confirms the idea
that one has hit on the (or at least a) right relation. Moreover, this data is “more
of the same” only if one has, in fact, hit upon a curve that works.10 Importantly
for this discussion, all sorts of scientific laws involve curve-fitting, so if realists
object to Farr’s law on these grounds, they ought to also object to other laws.11
12 Note that this is compatible with the view that in certain domains the theoretical plays a heavier
role than in others, where the empirical is a more common starting point. In fact, this does not strike
me as implausible. The general point about the interplay between the theoretical and the empirical in
the generation of scientific laws is a question of degree, not kind, and so I do not see how this could be
an objection to Farr without also being an objection to other laws.
Revisiting the Miasma Theory 25
13 In personal communication, Vickers has stated that this is, in fact, his preferred response to the
Farr case.
26 Historical Cases for the Debate
and decomposition and that disease can be transmitted through the air, both
central assumptions of the miasma theory.14 Without these assumptions, the
prediction about elevation and cholera mortality just disappears. Further, since
without these core theoretical assumptions, the prediction did not even make
sense, it is hard to see how Farr could somehow have made the elevation predic-
tion merely from other empirical knowledge. Both assumptions do theoretical
work for Farr and on taking out either one—the connection to decomposition or
beliefs about aerial transmission—there is no reason to expect the elevation rela-
tion. As a result, it is extremely hard to see how someone could have predicted it
without either of these components.15
It is perhaps noteworthy here that John Snow, famous for his view that cholera
was water-, not air-borne and on whose view there wasn’t a central connection
between cholera and putrefaction, did not predict the elevation relation. In fact,
not just did Snow not predict it, he did not even accommodate it, which was one
of the biggest criticisms voiced by his contemporaries (see, for example, Parkes
(1855)).16 Thus, not everybody came to Farr’s conclusion; notably, those who
had a different disease theory did not. Further, the other well-known empirical
knowledge had been around for a long time, yet it took Farr to predict a form
of the elevation hypothesis. Presumably, if the elevation relation had somehow
followed straightforwardly from this existing empirical knowledge, other people
would have predicted it long before Farr.17 So, to sum up: it is implausible to
think that Farr could have come to his conclusions about elevation and cholera
without the central theoretical assumptions of the miasma theory.
What about the other part of the objection, that Farr’s prediction was not as
risky as that of Poisson and, therefore, less significant? The first thing to note is
14 There are, of course, as with any other theory, some auxiliary assumptions involved, such as
assumptions about air dilution (that the air is cleaner the further away it is from miasmatic sources
and so on). An anonymous referee has suggested that different auxiliaries are needed for different
diseases and that the “miasma theory of cholera” is different from, say, the “miasma theory of ty-
phoid fever.” This, however, is to misunderstand the context of mid-19th century disease discussions.
Diseases, at the time, were not thought of as entities that would attack the body from the outside and
make it sick; instead, the prevalent view was a more physiological notion of disease, according to
which, in some sense, diseases originated in the victim. The ontological conception became influen-
tial later in the century, with the idea that different diseases were different species and with the advent
of early germ theories. For more detail, see Hamlin (1998: chapter 2).
15 I have previously argued that both assumptions ought to count as essential for realists. For more
argument for the claim that Farr could have come to the elevation conclusion just from existing em-
pirical knowledge, without miasmatic assumptions. Given the assumption that he somehow could
have done so, even if it is hard to see how, we might now worry that we are able to make this same ar-
gument in many other instances and, in particular, in instances that realists want to keep as examples
of genuine successes. Given enough ingenuity, perhaps we can always construct such a hypothetical
argument (see also Greg Frost-Arnold’s comment in response to Vickers (Frost-Arnold 2015)).
Revisiting the Miasma Theory 27
that even granting that Farr’s prediction was not as risky or significant as Poisson’s
does not mean that Farr’s prediction was not good. I have no qualms with the
claim that Poisson’s prediction was superior to Farr’s; however, it is not clear how
typical or frequent predictions like Poisson’s actually are and what percentage
of scientific predictions are Poisson-like.18 Indeed, it is entirely possible, if not
plausible, that such predictions, while exemplars of excellent predictions, are
somewhat atypical. Just as most scientists—even very good ones—are not Curies
or Einsteins, most predictions—even very good ones—are not like Poisson’s.
Thus, Farr’s prediction might have been very good, even if it was less good than
Poisson’s. And, thus, the mere fact that his prediction was less significant than
Poisson’s is not a reason to think that Farr’s prediction should not be viewed as an
instance of genuine success. But what about the claim that Farr’s prediction was
less risky? Even without a specific notion of riskiness it is clear that it would have
been bad for Farr if it had turned out that there was no relationship between ele-
vation and cholera mortality since this clearly followed from the miasma theory.
If no relation had been found, there would have been a need for another mias-
matic explanation (such as more sewage at high altitudes, for example) and, ab-
sent any such explanation, the absence of a relationship would have constituted a
strong argument against the theory.19,20
Still, a realist might argue, what matters is that what followed deductively from
the miasma theory was only the general fact that there ought to be some relation
between elevation and cholera mortality, but not the exact form of this relation
or even the fact that it was possible to capture it through a mathematical law.
This is true; however, once again, the fact that what followed logically from the
theory was only a general, not a specific prediction should not be held against
it, since this is all that could follow from such a theory, even under the very best
circumstances. In physical theories, laws and predictions, even if heavily based
on data, eventually usually have to be integrated into the theory mathematically,
18 It actually strikes me as an interesting project to look into the different types of prediction one
might find in different domains, what categories they might fall into, how frequent they are, and so
on. I would not find it surprising if different kinds of predictions were prevalent in different fields.
19 The same was true for other localities where high mortality was expected but not found, and for
localities where low mortality was expected, but it turned out to be high. Indeed, many of the con-
temporary discussions of the miasma theory and its competitors centered around just such issues.
20 What about the thought that Farr’s prediction did not have a low prior probability, but Poisson’s
did? I honestly do not know how to assign the priors in this case. It makes sense to assign Poisson’s
prediction a low prior probability but, in this case, the prediction was against the prevalent theory.
In Farr’s case, however, the prediction followed from a theory that was already the dominant di-
sease theory. It thus seems reasonable to think that its predictions, especially plausible ones like the
elevation relation, would not have been surprising to anyone. But what would the prior probability
have been if the miasma theory had been new? Certainly, given the humoral theory, according to
which diseases were an imbalance of individual humors, the probability of Farr’s prediction would
have been extremely low. What would the prior have been on Snow’s view, or some of the other
alternatives? Since we do not want to make the priors contingent on historical circumstances, it is
unclear to me how to determine what Vickers considers the crucial term, p(E/–T ).
28 Historical Cases for the Debate
21 This is not to say, of course, that Farr’s prediction, on its own, made the miasma theory suc-
cessful, or that a single prediction of this type could warrant realist commitment in the (approximate)
truth of a theory. Obviously, a theory’s success depends on many factors. The point is, rather, that the
best type of prediction a theory can make should play a large role in assessing that theory’s success
and that, if such predictions are discounted as a source of success for some class of theories, it is hard
to see how theories in that class could ever count as genuinely successful, even in principle.
Revisiting the Miasma Theory 29
realists will have to pay a cost that goes much beyond the specific example of Farr
and that now also includes, as we have seen, Snow’s water-hypothesis and con-
temporary disease etiologies. In fact, the cost might be even greater, since there
are plenty of theories whose predictions are not predominantly mathematical in
nature or concern population-levels. Realists might be willing to bite this bullet,
but the result would be a much narrower, perhaps domain-specific realism.
2.6 Conclusion
It follows from earlier sections that the case of the miasma theory supports the
PMI: it does so by supporting the view that there is no connection between suc-
cess and approximate truth.22 Regardless, realists need some kind of account of
what is going on in this case. Perhaps the most obvious option is to identify some
other kind of continuity at work in the miasma case and add it to the possibilities
presented earlier. In that event, perhaps the realist intuition that the (working
parts of the) predecessors to our current theories are approximately true can be
preserved.
However, this strategy is problematic: first, because it is not clear what pos-
sible candidates for retention are left after we have excluded theoretical entities,
mechanisms, laws, and structures. Second, even if a candidate could be found,
it’s unlikely that the resulting proposal could be generalized to the cases in the
already existing literature, and which already have other realist accounts. One
might not be worried by this and instead be tempted by a kind of pluralism about
retention: different types of preservation apply to different types of theories
under different circumstances. I am not unsympathetic to this proposal, but it
comes with one big worry for the realist, namely whether the resulting position
has any realism left in it. After all, if the proposal is that theories get better over
time, and increasing success brings with it increasing approximate truth, yet this
approximate truth consists of such different things in different cases, it’s not clear
what we should be realist about. We might still be committed to some vague idea
of progress, but we won’t be able to tell what that progress involves. In particular,
we cannot be realist about any particular element: we cannot be realist about the
parts that were involved in predictions, since, sometimes, those are not retained;
we cannot be realists about structures, since, sometimes, those are not retained;
we cannot be realists about theoretical entities or mechanisms, since, sometimes,
they are not retained. The same goes for a theory’s laws and, indeed, if realists
want to argue that the miasma theory is approximately true, they’ll have to keep
22 Note that of course it is not enough to establish the PMI. In order to achieve the latter, it also
needs to be shown that these kinds of cases are pervasive (see also Kitcher (1993: 138–139)).
30 Historical Cases for the Debate
adding to this list. But this means that, from the perspective of our current the-
ories, we simply have no idea which of their parts we ought to be realists about,
beyond a vague commitment to the view that something about them is probably
right, although we have not the slightest inkling what.
The alternative, of course, is to bite the bullet in the miasma case and accept
that this theory, although genuinely successful, was simply not one that was
approximately true. However, this is also a somewhat uncomfortable posi-
tion for the realist, quite independently of the fact that this case now supports
the PMI. The source of this discomfort is that we now have a case in which
we went from a not even approximately true theory directly to a successor
(the germ theory) that we do take be true—and not just approximately true,
but true simpliciter. This is uncomfortable, because the whole realist idea is
that there is a relation between increasing success over time and increasing
truth-content, and if it now turns out that discontinuities between theories
might sometimes be completely radical, then this does shed some doubt on
the “converging” part of convergent realism. Of course just because approxi-
mate truth might not be necessary for truth simpliciter does not mean that it
is not sufficient, and it doesn’t follow from the fact that (some) true theories
had no approximately true predecessors that approximate truth is not a reli-
able guide to truth as such.23 The point is rather that, if the convergent realist’s
story is right—if the right way to view the history and development of science
is as a succession of increasingly true theories that get better and better over
time—then we ought to expect the predecessors of theories that we now know
to be true to be approximately true. Finding true theories that do not conform
to this expectation does not invalidate the story—one case does not make a
pattern—but it is a strike against the plausibility of the general story that con-
vergent realists tell.
Acknowledgments
For helpful comments, many thanks to Tim Lyons, Marshall Porterfield, and Jan
Sprenger. Special thanks are due to Peter Vickers for numerous exchanges and
conversations about this case, and for generous written comments on a different
but related paper.
23 For that, we would need cases of approximately true predecessors that led to false successors,
and it is not clear how we would find these, since we also need true successors in order to assess these
claims in the first place.
Revisiting the Miasma Theory 31
References
Baldwin, P. (1999). Contagion and the State in Europe, 1830–1930. Cambridge: Cambridge
University Press.
Boyd, R. (1981). “Scientific Realism and Naturalistic Epistemology,” in P.D. Asquith and
T. Nickles (eds) PSA 1980, Vol. 2, East Lansing, MI: Philosophy of Science Association,
pp. 613–662.
Boyd, R. (1984). “The Current Status of the Realism Debate,” in J. Leplin (ed.) Scientific
Realism, Berkeley: University of California Press, pp. 41–82.
Boyd, R. (1990). “Realism, Approximate Truth and Philosophical Method,” in C.W.
Savage (ed.) Scientific Theories, Minnesota Studies in the Philosophy of Science, Vol. 14,
Minneapolis: University of Minnesota Press, pp. 355–391.
Doppelt, G. (2007). “Reconstructing Scientific Realism to Rebut the Pessimistic Meta-
Induction.” Philosophy of Science 74 (1): 96–118.
Eyler, J. M. (1979). Victorian Social Medicine: The Ideas and Methods of William Farr.
Baltimore/London: Johns Hopkins University Press.
Eyler, J. M. (2001). “The Changing Assessments of John Snow’s and William Farr’s Cholera
Studies.” Sozial und Präventivmedizin 46: 225–232.
Farr, W. (1852). Report on the Mortality of Cholera in England, 1848–49. London: W.
Clowes.
French, S., and Ladyman, J. (2011). “In Defence of Ontic Structural Realism,” in A.
Bokulich and P. Bokulich (eds.) Scientific Structuralism (Boston Studies in the Philosophy
of Science: Volume 281), Dordrecht: Springer, pp. 25–42.
Forster, M. R. (1999). “Curve-fitting problem.” In R. Audi (ed.) The Cambridge
Dictionary of Philosophy, Second Edition, Cambridge: Cambridge University Press,
pp. 197–198.
Forster, M.R. (2003). “Philosophy of the Quantitative Sciences: Unification, Curve Fitting,
and Cross Validation.” Manuscript. Available at http://philosophy.wisc.edu/forster/pa-
pers/Part1.pdf
Frost-Arnold, G. (2015). “Comment on Vickers (2015).” https://thebjps.typepad.com/
my-blog/2015/06/srpetervickers.html, last accessed 24 June 2018.
Hamlin, C. (1998). Public Health and Social Justice in the Age of Chadwick: Britain, 1800–
1854.Cambridge: Cambridge University Press.
Hamlin, C. (2009). Cholera: The Biography. New York: Oxford University Press.
Hesse, M. (1976). “Truth and the Growth of Scientific Knowledge.” Proceedings of the
Biennial Meeting of the Philosophy of Science Association 2: 261–280.
Kitcher, P. (1993). The Advancement of Science. New York: Oxford University Press
Ladyman, J. (1998). “What is Structural Realism?” Studies in History and Philosophy of
Science Part A 29 (3): 409–424.
Lange, M. (2002). “Baseball, Pessimistic Inductions and the Turnover Fallacy.” Analysis
62: 281–285.
Langmuir, A.D. (1961). “Epidemiology of airborne infection.” Bacteriological Reviews 25
(3): 173–181.
Langmuir, Alexander D. (1961). “Epidemiology of Airborne Infection.” Bacteriological
Reviews, 25(3): 174
Laudan, L. (1981). “A Confutation of Convergent Realism.” Philosophy of Science 48
(1):19–49
32 Historical Cases for the Debate
Laudan, L. (1984). Science and Values. Berkeley and Los Angeles: University of
California Press
Lewis, P. (2001). “Why the Pessimistic Induction Is a Fallacy.” Synthese 129: 371–380.
Lyons, T.D. (2006). “Scientific Realism and the Stratagema de Divide et Impera.” British
Journal for the Philosophy of Science 57 (3): 537–560.
Parkes E.A. (1855). “Mode of Communication of Cholera. By John Snow, M.D. Second
Edition.” British and Foreign Medico-Chirurgical Review 15: 456.
Pelling, M. (1978). Cholera, Fever and English Medicine: 1825–1865. Oxford: Oxford
University Press.
Psillos, S. (1996). “Scientific Realism and the ‘Pessimistic Induction.’” Philosophy of
Science 63 (3): 314.
Psillos, S. (1999). Scientific Realism: How Science Tracks Truth. London: Routledge
Saatsi, J. (2005). “On the Pessimistic Induction and Two Fallacies.” Philosophy of Science
72 (5): 1088–1098.
Saatsi, J., Vickers, P. (2011). “Miraculous Success? Inconsistency and Untruth in
Kirchhoff ’s Diffraction Theory.” British Journal for the Philosophy of Science 62
(1): 29–46.
Snow, J. (1855a). “On the comparative mortality of large towns and rural districts, and the
causes by which it is influenced.” Journal of Public Health, and Sanitary Review 1: 16–24.
Snow, J. (1855b). On the Mode of Communication of Cholera (second much enlarged ed.).
London: J. Churchill.
Tulodziecki, D. (2016). “From Zymes to Germs: Discarding the Realist/Anti-Realist
Framework,” in T. Sauer and R. Scholl (eds.) The Philosophy of Historical Case-Studies,
Boston Studies in Philosophy of Science, Switzerland: Springer, 265–284.
Tulodziecki, D. (2017). “Against Selective Realism(s).” Philosophy of Science 84
(5): 996–1007.
Vickers, P. (2013). “A Confrontation of Convergent Realism.” Philosophy of Science 80
(2): 189–211.
Vickers, P. (2015). “Contemporary Scientific Realism and the 1811 Gill Slit Prediction,”
Auxiliary Hypotheses: Blog of the British Journal for the Philosophy of Science, http://
thebjps.typepad.com/my-blog/2015/06/srpetervickers.html, last accessed 24
June 2018.
Worboys, M. (2000). Spreading Germs: Diseases, Theories, and Medical Practice in Britain,
1865–1900. Cambridge: Cambridge University Press.
Worrall, J. (1989). “Structural realism: The best of both worlds?” Dialectica 43
(1–2): 99–124.
Worrall, J. (1994). “How to Remain (Reasonably) Optimistic: Scientific Realism and
the” Luminiferous Ether.” PSA: Proceedings of the Biennial Meeting of the Philosophy of
Science Association 1994, 1: 334–342.
3
What Can the Discovery of Boron Tell Us
About the Scientific Realism Debate?
Jonathon Hricko
3.1 Introduction
In much of the recent work on the scientific realism debate,1 realists and their
anti-realist opponents have focused on theories with two significant features.
First of all, those theories purport to describe some kind of unobservable reality,
for example, unobservable entities like electrons or genes. Second, those theories
exhibit novel predictive success, i.e., they generate true predictions that scientists
did not use when constructing those theories. The clearest cases of such success
involve temporal novelty, where a prediction of some phenomenon is temporally
novel if that phenomenon was not known at the time the theory was constructed.
Other cases involve only use-novelty, where a prediction is use-novel, but not
temporally novel, if it predicts a phenomenon that was known at the time the
theory was constructed, but knowledge of that phenomenon was not used in the
construction of the theory.2 In this chapter, I’ll focus on an example of a tempo-
rally novel predictive success.
One of the central issues in the realism debate concerns whether the novel
predictive success of a theory constitutes a sufficient reason for thinking that (at
least some of) that theory’s claims regarding unobservables are (at least approxi-
mately) true. Realists of various stripes argue that some form of realism or other
provides the best explanation of why theories exhibit novel predictive success
(Musgrave 1988; Psillos 1999). This argument is the so-called no-miracles ar-
gument, the name of which comes from the realist’s contention that, without a
Jonathon Hricko, What Can the Discovery of Boron Tell Us about the Scientific Realism Debate? In: Contemporary
Scientific Realism. Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press
2021. DOI: 10.1093/oso/9780190946814.003.0003
34 Historical Cases for the Debate
3 Throughout the chapter, when I claim that past theories are/are not approximately true, these
claims should be understood as abbreviations for the conditional claim that, if our current theories
are approximately true, then these past theories are/are not approximately true. Otherwise, I might
be accused of presupposing realism about our current theories, and of thereby begging some central
questions in the realism debate.
4 See Lyons (2017, 3215) for a useful list of selective realist positions.
5 This point is especially clear within the context of Lyons’s pessimistic meta- modus tollens
(2002, 67).
Scientific Realism Debate 35
The false theory at the center of this example of novel predictive success is
Lavoisier’s oxygen theory of acidity. According to that theory, all acids contain
oxygen and a non-oxygen component that Lavoisier called the “acidifiable base”
or “radical” of the acid, two terms that he regarded as synonyms (1790 [1965],
65). Lavoisier was led to this theory by the fact that the acids that were most well
understood at the time were all shown to contain oxygen. Lavoisier holds that
36 Historical Cases for the Debate
different kinds of acids differ from one another in one of two ways. They may
contain different radicals, as phosphoric acid and sulfuric acid do—the radicals
in these acids are phosphorus and sulfur, respectively (66). And even if they con-
tain the same radical, they may still differ from one another by virtue of the fact
that they contain different amounts of oxygen, as sulfuric acid and sulfurous acid
do (66–68). Moreover, Lavoisier emphasizes that the base or radical of an acid
can be either a simple substance or a compound (115–116, 176–177). For ex-
ample, he claims that many vegetable acids have compound radicals composed
of carbon and hydrogen.
Oxygen is the substance that plays the most central role in Lavoisier’s theory of
acidity. It is the element that all acids share, and which “constitutes their acidity”
(65). Oxygen, for Lavoisier, is what gives acids their acidic properties, for ex-
ample, their sour taste. Lavoisier labels oxygen “the acidifiing [sic] principle” (65)
and in fact the very name “oxygen” that Lavoisier proposed for this substance
comes from the Greek words for “acid-generator” (51; Chang 2012b, 9). Though
to be sure, as Le Grand (1972, 11–12) has emphasized, Lavoisier admitted that
many compounds that contain oxygen are not acids, and his point was that ox-
ygen is a necessary condition for acidity, not a sufficient one.
Lavoisier’s theory is also a theory of the formation of acids. After recounting
a series of experiments in which acids are formed by the combustion of the
acidifiable bases phosphorus, sulfur, and carbon, Lavoisier writes: “I might
multiply these experiments, and show by a numerous succession of facts, that
all acids are formed by the combustion of certain substances” (1790 [1965],
64). Lavoisier’s view, then, is that an acidifiable base must be a combustible sub-
stance. In order to understand Lavoisier’s view regarding the formation of acids
by means of combustion, we must start with his claim that oxygen gas is a com-
pound of caloric and the base of oxygen gas (“oxygen base” for short, following
Chang (2011, 415)); and it is oxygen base which is the true acidifying prin-
ciple (Lavoisier 1790 [1965], 51–52). Lavoisier illustrates the combustion of an
acidifiable base and the subsequent formation of an acid in terms of the example
of phosphorus, the acidifiable base of phosphoric acid:
In other words, phosphorus effects the decomposition of oxygen gas into caloric
and oxygen base; oxygen base then combines with phosphorus to form an acid.
More generally, “[b]efore combustion can take place, it is necessary that the base
of oxygen gas should have greater affinity to the combustible body than it has to
caloric” (414–415). Hence, Lavoisier’s theory of acidity makes use of his theories
Scientific Realism Debate 37
of combustion and caloric as well. Lavoisier goes on to propose the term “oxy-
genation” to name the process by which combustible substances combine with
oxygen, in which case acidifiable bases are converted to acids by oxygenating
them (61–62).
As a whole, Lavoisier’s theory of acidity isn’t even approximately true.6 By the
early years of the 19th century, chemists already knew of two counterexamples to
Lavoisier’s theory, namely, prussic acid (hydrocyanic acid, HCN) and muriatic
acid (hydrochloric acid, HCl), neither of which contains oxygen, Lavoisier’s acid-
ifying principle. Additionally, by positing the muriatic radical as the non-oxygen
component of muriatic acid, Lavoisier posited a non-existent entity.7 Moreover,
if one or both of our current conceptions of acidity are approximately true, then
Lavoisier’s theory cannot be. According to the Brønsted-Lowry concept, acids
are proton donors. And according to the Lewis concept, acids are electron pair
acceptors. While these concepts diverge from one another in interesting ways,8
neither gives any sense to the claim that oxygen is the acidifying principle, which
is the central claim of Lavoisier’s theory.
It’s worth pressing this point regarding the falsity of Lavoisier’s theory a bit fur-
ther by considering the theories of combustion and caloric that Lavoisier made
use of in the context of his theory of acidity. One might think that Lavoisier’s
theory of combustion, at least, is approximately true. But it’s not obvious that it is.
Lavoisier’s theory requires the presence of oxygen gas for combustion. But by the
early 19th century, chemists knew of cases of combustion that occurred without
the presence of oxygen gas. Moreover, combustion, for Lavoisier, requires the de-
composition of oxygen gas into caloric and oxygen base (1790 [1965], 414–415).
The caloric theory of heat has been a problem case for realists at least since it
appeared on Laudan’s well-known list of false-but-successful theories (1981, 33).
Even if we confine our attention to the period in which Lavoisier was working,
this theory exhibited a number of successful explanations. As Chang makes clear,
these included explanations of:
the flow of heat toward equilibrium, the expansion of matter by heating, latent
heat in changes of state, the elasticity of gases and the fluidity of liquids, the heat
released and absorbed in chemical reactions, [and] combustion. (2003, 907)
6 See Chang (2012b, 8–10) for a good discussion of the problems with Lavoisier’s theories of
acidity, combustion, and caloric, which has informed the discussion in the remainder of this section.
7 See Hricko (2018) for a detailed discussion of the muriatic radical within the context of the re-
alism debate.
8 See Chang (2012a) for a good discussion of the ways in which these concepts diverge.
38 Historical Cases for the Debate
The novel predictions on which I’ll focus concern the composition of what
Lavoisier called boracic acid. Lavoisier writes:
As we’ll see shortly, there is a bit more to the predictions than what Lavoisier
says here.
Lavoisier’s novel predictions regarding boracic acid belong to a larger group
of similar novel predictions that he made on the basis of his theory of acidity. At
the time, there were a number of acids that chemists could neither produce from
simple substances nor decompose into simple substances. These acids included
boracic acid, muriatic acid, and fluoric acid (hydrofluoric acid, HF). Lavoisier
hypothesized that these acids are composed of oxygen and the boracic, muriatic,
and fluoric radicals, respectively. Regarding these radicals, Lavoisier writes:
the combinations of these substances [the radicals], either with each other, or
with the other combustible bodies, are hitherto entirely unknown . . . We only
know that these radicals are susceptible of oxygenation, and of forming the mu-
riatic, fluoric, and boracic acids. (209–210)
None of these radicals had been isolated, and in this passage, Lavoisier predicts
that they are combustible substances that form acids by means of oxygenation.
This prediction is a straightforward consequence of his theory of acidity, ac-
cording to which acids contain oxygen and a combustible radical. It’s also worth
emphasizing that Lavoisier’s theory predicts that the only components of these
three acids are oxygen and their respective radicals. The radical, for Lavoisier, is
the acid’s non-oxygen component, and cases in which an acid has more than one
Scientific Realism Debate 39
Importantly, these predictions are novel predictions. In fact, the novelty of these
predictions is temporal novelty, which arises from the fact that chemists at the
time had not yet succeeded in decomposing boracic acid into simpler substances
or producing it by means of simpler substances. Hence, the composition of
9 He thereby indicated his view that they are elements. In section 3.6.2, I discuss the prediction
that the boracic radical is an element and conclude that it is not relevant to the issue concerning
whether selective realists can accommodate false-but-successful theories.
40 Historical Cases for the Debate
boracic acid and the way in which it is formed were unknown and so couldn’t
have been written into the oxygen theory of acidity at the time Lavoisier
constructed it.
10 I discuss the details regarding the identification of boracic acid with boron trioxide in
section 3.6.1.
11 In section 3.6.4, I consider and respond to an objection to the identification of boron with the
boracic radical.
12 Given the impurity of the samples, one might question whether these chemists really confirmed
suspected that the substance he had isolated was a compound of oxygen and
“the true basis of the boracic acid,” which he conjectured was a metal and for
which he proposed the name “boracium” (1809, 84–85). Several years later,
he concluded that this conjecture had not been vindicated, and in light of
the analogy between the substance that he had isolated and carbon, he pro-
posed another name for what he calls “the basis of the boracic acid”: “boron”
(1812, 178). The terminology that these chemists used does not necessarily
imply a commitment to Lavoisier’s theory of acidity. In fact, at the time they
discovered boron, Davy was skeptical of Lavoisier’s theory of acidity while
Gay-Lussac and Thénard were committed to it.13 And by 1812, although
Davy was still using the term “basis” to refer to the non-oxygen component
of boracic acid, he had more decisive reasons for rejecting Lavoisier’s theory
since, in 1810, he had argued that the components of muriatic acid are hy-
drogen and chlorine (Davy 1811). However, Davy did previously claim that
“the combustible matter obtained from boracic acid, bears the same rela-
tion to that substance, as sulphur and phosphorus do to the sulphuric and
phosphoric acids” (1809, 82), which basically amounts to an identification
of boron with the boracic radical. Importantly, regardless of their attitude
toward Lavoisier’s theory of acidity, these three chemists isolated the sub-
stance that Lavoisier had hypothesized years earlier; in other words, boron
is the boracic radical.
Once they had isolated boron, Davy, Gay-Lussac, and Thénard also deter-
mined some of its properties, most notably, that it is a combustible substance
that combines with oxygen to form boracic acid. Both Davy (1809, 79, 82; 1812,
179) and Gay-Lussac and Thénard (1808, 172–173) describe the experiments
by which they converted the newly discovered substance into boracic acid by
means of combustion with oxygen. And by doing so, they also showed that the
components of boracic acid are boron and oxygen.
At this point, we can conclude that Lavoisier’s four novel predictions, which
I listed in section 3.3, were successful. The work of Davy, Gay-Lussac, and
Thénard showed that boracic acid, which we know as boron trioxide, contains
both oxygen and a combustible substance, namely, boron. Moreover, these
chemists confirmed Lavoisier’s prediction regarding the formation of boracic
acid, namely, that it forms by the combustion of the boracic radical (boron)
with oxygen. And by both producing and decomposing boracic acid, they
provided a convincing demonstration that its sole components are boron and
oxygen.
13 See Brooke (1980, 124, 158) for Davy’s skeptical attitude toward Lavoisier’s theory of acidity and
At this point, the upshot is that we can add Lavoisier’s oxygen theory of acidity to
the increasingly long list of theories that, although not even approximately true,
exhibited novel predictive success. As I discussed in section 3.1, such theories
pose a challenge to realism. My goal in this section is to consider two ways in
which a selective realist might respond to this challenge and argue that neither of
these ways is successful.
Before doing so, it’s worth going into a bit more detail regarding the chal-
lenge that the selective realist faces. The derivation of Lavoisier’s four successful
predictions involved a number of claims from his oxygen theory of acidity that,
by present lights, are not even approximately true. These claims include:
Lavoisier’s prediction that boracic acid contains oxygen made use of (1), (2),
and (4). His prediction that it contains a combustible substance made use of (1),
(3), and (4). His prediction that it contains only these two substances made use
of all of these claims, as did his prediction regarding the formation of boracic
acid in terms of the combustion of an acidifiable base with oxygen. If our cur-
rent theories are correct, then none of these four claims is even approximately
true. (1) is false in light of the rejection of the caloric theory of heat. Moreover,
we now consider oxygen gas to be an elementary substance, not a compound.
(2) is false, not just because there are acids that do not contain oxygen, but also
because we’ve rejected the idea that a chemical substance can be an acidifying
principle in Lavoisier’s sense. (3) is false because Lavoisier characterized the rad-
ical of an acid as its non-oxygen component, and so acids that lack oxygen also
lack radicals in Lavoisier’s sense. And (4) is false, not just because it involves the
caloric theory of heat, but also because there are acids that do not form by com-
bustion with oxygen because they do not contain oxygen.
The first selective realist response I’ll consider is to show, contrary to first
appearances, that the theoretical claims required for deriving Lavoisier’s
Scientific Realism Debate 43
successful predictions are in fact approximately true. One way the selective realist
might attempt to do so makes use of Vickers’s (2013, 199) distinction between
derivation external posits (DEPs) and derivation internal posits (DIPs), along
with his notion of the working part of a DIP.14 Realist commitment does not ex-
tend to DEPs, which merely inspire scientists to consider ideas; and it may not
even extend to all parts of a DIP. Realist commitment extends only to the working
part of a DIP, i.e., the part that “actually contributes to a given derivational step,”
and not to the other parts, which Vickers labels “surplus content” (201).
In order to evaluate this response, I’ll focus on one of Lavoisier’s predictions,
namely, the prediction that boracic acid contains oxygen. While this prediction
was in fact motivated by (1), (2), and (4), for the sake of argument, let’s grant
that (1) and (4) are DEPs, and only (2) is a DIP. It’s plausible that (2) has some
surplus content and that the working part of (2) is a more minimal claim that
suffices for deriving the prediction without mentioning oxygen base or an acidi-
fying principle. What might this more minimal claim be? The claim that all acids
contain oxygen is more minimal, and it suffices for deriving the prediction; but
it is not even approximately true since not all acids contain oxygen. The claim
that boracic acid contains oxygen is even more minimal, and also true; but it
doesn’t provide an explanation of this instance of novel predictive success since it
is identical to the prediction itself, which makes the derivation trivial. The claim
that many acids contain oxygen is more minimal, and also true; but it doesn’t
entail the prediction, and on its own, it doesn’t even make the prediction likely
since it’s consistent with the claim that many acids do not contain oxygen. The
challenge for the selective realist is to identify a working part that has two prop-
erties: it must be at least approximately true, and it must be sufficient for deriving
the prediction in a nontrivial way. The problem for the selective realist is that the
claims that have one property lack the other property.
Another consideration that makes this kind of selective realist response prob-
lematic concerns Lavoisier’s predictions regarding the other two undecomposed
acids: muriatic (hydrochloric) acid and fluoric (hydrofluoric) acid. The selective
realist must explain why Lavoisier’s prediction regarding the presence of oxygen
in boracic acid succeeded while his predictions regarding the presence of ox-
ygen in the muriatic and fluoric acids failed. And this explanation must be con-
sistent with the fact that Lavoisier’s predictions regarding the presence of oxygen
in these three acids were derived in exactly the same way, as I discussed in section
3.3. These predictions were motivated solely by Lavoisier’s theory of acidity and
by the fact that these three substances are acids. Explaining why one prediction
succeeded while the other two failed requires adopting a view that is historically
14 Vickers uses the term posit “as essentially synonymous with ‘proposition’ or ‘hypothesis’
inaccurate. This is the view that Lavoisier’s derivations of these three predictions
differ in some significant way, say, because he identified some property of boracic
acid that muriatic acid and fluoric acid lack and that truly indicates the presence
of oxygen in a compound. In short, for any candidate for the working part of a
DIP, the selective realist will have to explain, in a historically plausible way, why it
works when it comes to boracic acid and doesn’t work when it comes to the other
two acids. It’s worth emphasizing that the issue here is not simply about failed
predictions. It’s about whether, in light of those failed predictions (which were
derived in exactly the same way as the successful predictions), the working parts
in the derivation of the successful predictions can be regarded as approximately
true. My claim is that they cannot.
The second selective realist response is that the novel predictive success of
Lavoisier’s theory is not impressive enough to pose a challenge to realism. Vickers
(2013, 195–198) argues that successful novel predictions need to be sufficiently
impressive in order to pose a challenge to realism. For Vickers, the distinction
between impressive and unimpressive novel predictive successes is a matter of
degree. One way in which they differ in degree concerns “the degree to which the
prediction could be true just by luck (perhaps corresponding to the Bayesian’s
‘prior probability’)” (196). If the probability that Lavoisier’s predictions could be
true just by luck is sufficiently high, then we have a group of lucky guesses that
the realist can dismiss, as opposed to “miracles” that require a realist explanation.
Given the state of knowledge regarding acids in the years leading up to Davy’s
and Gay-Lussac and Thénard’s work on boracic acid, what can we say about the
prior probability of Lavoisier’s predictions being true? Many acids (e.g., sulfuric
acid and phosphoric acid) were known to contain oxygen and a combustible sub-
stance and to form via the combustion of that substance with oxygen. However,
there were problem cases. In 1787, Claude Louis Berthollet (1748–1822) dem-
onstrated that prussic acid (hydrocyanic acid, HCN) contains hydrogen, carbon,
and nitrogen, but not oxygen (1789, 38). Thomas Thomson (1773– 1852)
mentions another problem case: “Sulphurated hydrogen [i.e., hydrogen sul-
fide, H2S], for instance, possesses all the characters of an acid, yet it contains no
oxygen” (1802, 4). In light of these problem cases, assigning a very high prior
probability to Lavoisier’s predictions regarding boracic acid would have been
unwarranted. In that case, Lavoisier’s predictions were, at least to some degree,
impressive.
Vickers (2013, 196) discusses two other senses in which predictions can be
unimpressive; but neither of them applies to Lavoisier’s predictions regarding
boracic acid. First of all, Vickers argues that vague predictions are unimpres-
sive. His example is the prediction that the planet Venus is hot, which is suc-
cessful so long as Venus has a temperature greater than, say, 50°C. The problem
here seems to relate, not just to the vagueness of the term hot, but also to its
Scientific Realism Debate 45
imprecision, since there are many ways in which Venus could be hot. Lavoisier’s
predictions regarding boracic acid are, however, both less vague and more pre-
cise than this prediction. Secondly, Vickers argues that predictions may be un-
impressive if the match between the prediction and the experimental results is
not sufficiently close. But Davy’s and Gay-Lussac and Thénard’s results match
Lavoisier’s predictions quite well. It’s also worth mentioning that, for Vickers,
qualitative predictions (e.g., the Poisson white spot) can be very impressive, and
so the qualitative nature of Lavoisier’s predictions is no reason to dismiss them
as unimpressive.
I conclude that the novel predictive success of Lavoisier’s oxygen theory of
acidity poses a strong challenge to selective realism. The working parts of
Lavoisier’s theory are not even approximately true. And the predictive success
of that theory is sufficiently impressive that the selective realist cannot dismiss
Lavoisier’s predictions as likely to be true just by luck.
At this point, I’ve presented the main argument of the chapter. In the course
of doing so, I’ve also flagged and so far ignored a number of relevant issues, to
which I now turn.
In section 3.4, I claimed that when chemists working in the late 18th and early
19th centuries used the term boracic acid, they primarily had in mind the sub-
stance that we call boron trioxide (B2O3). Boron trioxide is the anhydride of boric
acid (H3BO3), and one might wonder why boracic acid didn’t refer solely to the
substance that we call boric acid. If it did, then some of Lavoisier’s predictions
(e.g., the prediction that oxygen and a combustible substance are the only
components of boracic acid) are false. Moreover, in order to conclude that the
work of Davy, Gay-Lussac, and Thénard demonstrated the success of Lavoisier’s
predictions, we need to be sure that these three chemists were experimenting
with the substance that was the subject of Lavoisier’s predictions. Since my ar-
gument depends on the claim that chemists used the term boracic acid to refer
primarily to boron trioxide, it’s important for me to defend this claim.
The use of the term boracic acid to refer primarily to boron trioxide was a
particular instance of a more general trend. A number of commentators have
observed that chemists at the time used the term acid to refer to two kinds of
substances that we distinguish today, namely, acids and their anhydrides (Laurent
46 Historical Cases for the Debate
1855, 44–45; Miller 1871, 314; Lowry 1915, 250). Knight (1992, 83) explains that
when chemists at the time used the term acid, they had in mind primarily the
anhydride, though they recognized that the presence of water was required for
anhydrides to manifest their acidic properties. Crosland (1973, 307) points out
that this water “was considered as no more an essential part of the acid than, say,
water of crystallization is a part of certain salts.” To take an example, Lavoisier
(1790 [1965], 69) uses the term carbonic acid to refer to what we could call
carbon dioxide (CO2), the anhydride of carbonic acid. And when he claims that
the components of carbonic acid are oxygen and carbon, he’s not neglecting hy-
drogen as a component, since he had in mind the anhydride, and since he recog-
nized that carbonic acid mixed with water does contain hydrogen.
Chemists at the time took the same attitude toward boric acid and its an-
hydride, boron trioxide. For instance, in his Dictionary of Chemistry, William
Nicholson (1753–1815) writes:
In a moderate heat this concrete acid melts with less intumescence than borax
itself, and runs into a clear glass, which is not volatile unless water be present.
This glass does not differ from the original acid, except in having lost its water
of crystallization. (1795, 18)
The concrete acid is what we would now call boric acid, while the clear glass
is its anhydride, which we would call boron trioxide. Lavoisier (1790 [1965]
244) distinguishes the combinations of boracic acid with other substances “in
the humid way” and “via sicca,” i.e., in the dry way, which corresponds to the dis-
tinction we draw between boric acid and boron trioxide today. Years later, Davy
treats the acid and the anhydride as two forms of the same substance:
Boracic acid, in its common form, is in combination with water; it then appears
as a series of thin white hexagonal scales; . . . By a long continuous white heat
the water is driven off from it, and a part of the acid sublimes; the remaining
acid is a transparent fixed glass, which rapidly attracts moisture from the air.
(1812, 180)
The white hexagonal scales are what we would call boric acid, while the trans-
parent fixed glass is what we would call boron trioxide.
Importantly, Lavoisier’s predictions regarding boracic acid are primarily
predictions regarding what we would call boron trioxide. Lavoisier was prima-
rily concerned with the composition of boracic acid without water (i.e., boron
trioxide); and he predicted that it is composed of oxygen and a combustible sub-
stance, namely, the boracic radical. He was also perfectly aware that boracic acid
with water (i.e., boric acid) contains hydrogen since it contains water.
Scientific Realism Debate 47
It is also the case that Davy, Gay-Lussac, and Thénard decomposed boron tri-
oxide as opposed to boric acid. To be sure, there is some disagreement about
this point in the literature, with some commentators claiming that it was boric
acid (e.g., Fontani, Costa, and Orna 2015, 30) and others claiming that it was
boron trioxide (e.g., Crosland 1978, 78). There are two good reasons for con-
cluding that these chemists decomposed boron trioxide. First of all, Gay-Lussac
and Thénard (1808, 170) describe the boracic acid they were working with as
“very pure and vitreous,” which looks like a description of boron trioxide as op-
posed to boric acid. Second, decomposing boron trioxide by means of alkaline
metals like potassium is a recognized method of producing impure samples of
boron. Chemists in the mid-20th century did attempt to substitute boric acid for
boron trioxide in such reactions, but these attempts were eventually abandoned
because of the risk of producing an explosive mixture (Zhigach and Stasinevich
1977, 216–217).
Since boracic acid referred primarily to boron trioxide, we can conclude that
Lavoisier’s predictions were successful and that the work of Davy, Gay-Lussac,
and Thénard confirmed those predictions.
In section 3.3, I briefly noted another prediction regarding the boracic, muri-
atic, and fluoric radicals that one finds in Lavoisier’s work, namely, that they are
elements. Lavoisier included the three radicals in his table of simple substances,
thereby indicating his view that they should be considered elements (1790
[1965], 175).
For Lavoisier, the terms element and simple substance are synonyms, and he
understands elements as substances that chemists cannot decompose into sim-
pler component substances (xxiv). As Scerri (2005, 129) puts it, elements, for
Lavoisier, are “observable simple substances that can be isolated,” i.e., from the
other substances that they combine with to form compounds, which requires
chemical decomposition. Lavoisier (1790 [1965], xxiv, 177) emphasizes that
“simple” should be understood relative to our present state of knowledge, so that
a substance that we regard as simple today may be shown to be a compound to-
morrow if chemists succeed in decomposing it. Until chemists are able to decom-
pose a substance, Lavoisier states that we should regard it as an element.
However, there’s a puzzle regarding the inclusion of the three radicals on
the table of simple substances. As Hendry (2005, 41–42) has noted, Lavoisier
applies his notion of element in an inconsistent way—since the boracic, muri-
atic, and fluoric acids had not yet been decomposed, they should appear on the
48 Historical Cases for the Debate
In section 3.4, I noted that Davy, Gay-Lussac, and Thénard isolated relatively
impure samples of amorphous boron.15 Given the impurity of the samples that
these three chemists isolated, did they really confirm Lavoisier’s predictions?
15 Methods for producing pure samples of boron were not developed until the 20th century
In short, the samples that they isolated were pure enough to confirm the novel
predictions from section 3.3, because those predictions don’t require that the
boracic radical is an absolutely pure chemical substance. Confirming the pre-
diction that the boracic radical is an element might require isolating a sample
with a greater degree of purity. But as I argued in section 3.6.2, this prediction is
not relevant to the issue at hand. That said, it’s worth going into a bit more detail
regarding the ways in which Davy, Gay-Lussac, and Thénard isolated as pure a
sample as they could, and how they thereby confirmed Lavoisier’s predictions. In
order to do so, I’ll make use of Chen’s (2016) account of experimental individu-
ation and go into a bit more detail regarding the experimental work on boron in
the early 19th century.
According to Chen, scientists experimentally individuate an entity or a sample
of an entity when their experiments satisfy three conditions: (1) they separate the
entity or sample from its surrounding environment, (2) they maintain the struc-
tural unity of the entity or sample, and (3) they manipulate the entity or sample
so as to investigate other aspects of nature.
Conditions (1) and (2) are the conditions most relevant to the isolation of
samples of chemical substances, and therefore to the confirmation of Lavoisier’s
prediction that boracic acid contains a combustible substance. According to
condition (1), chemists must separate the substances that they produce “from
their environments” (Chen 2016, 348), and “from the experimental instruments
that may have helped produce [them]” (365). And according to condition (2),
chemists must maintain the structural unity of these substances in the process.
In general, structural unity is the idea that “the components of an individual are
structured into a whole in some specific manner” (358). When it comes to chem-
ical substances, maintaining structural unity includes such things as ensuring
that a highly reactive substance that has been isolated doesn’t react with other
substances in its environment. These two conditions collectively yield an under-
standing of what it means to isolate a sample of a chemical substance. Moreover,
both of these conditions relate to the quality of the sample that chemists manage
to produce. We must ask whether chemists managed to produce and preserve
a sample that is both large enough and pure enough for them to determine the
properties of the substance.
To a large degree, Davy’s and Gay-Lussac and Thénard’s experiments satisfied
these two conditions. Davy’s initial attempts to obtain the substance by means of
decomposing boracic acid with a battery were unsatisfactory precisely because
they didn’t satisfy these conditions. Davy (1809, 76) writes that, via this method,
he “was never able to obtain it, except in very thin films,” and so “[i]t was not
possible to examine its properties minutely, or to determine its precise nature,
or whether it was the pure boracic basis.” By heating boron trioxide in a metal
tube with potassium, these chemists were able to produce more of the substance
50 Historical Cases for the Debate
in question. While Gay-Lussac and Thénard report using a copper tube for the
experiment (1808, 170), Davy reports repeating the experiment with a number
of different metal tubes, including tubes made of copper, brass, gold, and iron
(1809, 76–77). According to Davy, “[i]n all cases, the acid was decomposed, and
the products were scarcely different” (77). By showing that the results were the
same regardless of what kind of metal was used, Davy goes some length toward
showing that the substance had been separated from the instruments used to
produce it. Both Davy (1809, 78) and Gay-Lussac and Thénard (1808, 171) also
used water and muriatic acid to wash the mixture that resulted from this process
and thereby separate the boron they had produced from other substances like
potash and borate of potash. By doing so, they go some length toward showing
that the substance had been separated from its surrounding environment. In
sum, they were able to produce samples of the substance that were large enough
and pure enough for them to determine some of its properties (e.g., combusti-
bility) and conclude that it is a component of boracic acid.
Condition (3) is relevant to the confirmation of Lavoisier’s prediction re-
garding the formation of boracic acid. This condition concerns the “instru-
mental use” of an entity or sample of an entity “to investigate other phenomena
of nature” (Chen 2016, 358). Davy (1809, 79, 82; 1812, 179) and Gay-Lussac
and Thénard (1808, 172–173) used the samples of boron they had isolated to
form boracic acid by means of combustion with oxygen. Importantly, their
samples of boron were pure enough to confirm Lavoisier’s prediction. And in
the course of confirming this prediction, these three chemists used boron as
an instrument. However, it’s not clear that they were using boron to investigate
other phenomena of nature as opposed to continuing an investigation into the
same phenomena.
Experimental individuation is a matter of degree. Samples of chemical
substances can be more or less separated from their environments and from
the instruments that produced them, they can be more or less pure, and their
structural unity can be more or less maintained. The important point is that the
samples of boron that chemists produced in the early 19th century were pure
enough to confirm Lavoisier’s predictions.
In section 3.4, I claimed that boron is the boracic radical. This claim is crucial
to my argument since, if boron is not the boracic radical, we can’t conclude that
Lavoisier’s predictions regarding boracic acid are successful. Hence, this claim is
in need of some defense, especially since Chang (2012b, 54) has labeled all three
Scientific Realism Debate 51
3.7 Conclusion
Acknowledgments
This chapter branched off from a co-authored project with Ruey-Lin Chen,
and I’d like to single him out for special thanks. This chapter also benefited
from very helpful suggestions from Timothy Lyons. Thanks to the organizers
and participants of the Quo Vadis Selective Scientific Realism? conference at
Durham University in August 2017, especially Hasok Chang, Ludwig Fahrbach,
Amanda J. Nichols, Myron Penner, Jan Potters, Juha Saatsi, Yafeng Shan, and
Peter Vickers. Thanks to three anonymous referees for their helpful comments
and suggestions. Finally, thanks to the Ministry of Science and Technology in
Taiwan for supporting this work (MOST 106-2410-H-010-001).
References
Alai, Mario. 2017. “The Debates on Scientific Realism Today: Knowledge and Objectivity
in Science.” In Varieties of Scientific Realism: Objectivity and Truth in Science, edited by
Evandro Agazzi, 19–47. Cham: Springer.
Berthollet, Claude Louis. 1789. “Extrait d’un mémoire sur l’acide prussique.” Annales de
Chimie 1: 30–39.
Brooke, John Hedley. 1980. “Davy’s Chemical Outlook: The Acid Test.” In Science and
the Sons of Genius: Studies on Humphry Davy, edited by Sophie Forgan, 121–175.
London: Science Reviews.
Chang, Hasok. 2003. “Preservative Realism and Its Discontents: Revisiting Caloric.”
Philosophy of Science 70 (5): 902–912.
Chang, Hasok. 2011. “The Persistence of Epistemic Objects through Scientific Change.”
Erkenntnis 75 (3): 413–429.
Chang, Hasok. 2012a. “Acidity: The Persistence of the Everyday in the Scientific.”
Philosophy of Science 79 (5): 690–700.
Chang, Hasok. 2012b. Is Water H2O? Evidence, Realism and Pluralism. Dordrecht: Springer.
Chen, Ruey- Lin. 2016. “Experimental Realization of Individuality.” In Individuals
across the Sciences, edited by Alexandre Guay and Thomas Pradeu, 348– 370.
New York: Oxford University Press.
Crosland, Maurice P. 1973. “Lavoisier’s Theory of Acidity.” Isis 64 (3): 306–325.
Crosland, Maurice P. 1978. Gay-Lussac: Scientist and Bourgeois. Cambridge: Cambridge
University Press.
Davy, Humphry. 1808. “The Bakerian Lecture [for 1807]: On Some New Phenomena
of Chemical Changes Produced by Electricity, Particularly the Decomposition of the
Fixed Alkalies, and the Exhibition of the New Substances Which Constitute Their
Bases; and on the General Nature of Alkaline Bodies.” Philosophical Transactions of the
Royal Society of London 98: 1–44.
Davy, Humphry. 1809. “The Bakerian Lecture [for 1808]: An Account of Some New
Analytical Researches on the Nature of Certain Bodies, Particularly the Alkalies,
Phosphorus, Sulphur, Carbonaceous Matter, and the Acids Hitherto Undecompounded;
with Some General Observations on Chemical Theory.” Philosophical Transactions of
the Royal Society of London 99: 39–104.
54 Historical Cases for the Debate
Davy, Humphry. 1811. “The Bakerian Lecture [for 1810]: On Some of the Combinations
of Oxymuriatic Gas and Oxygene, and on the Chemical Relations of These Principles,
to Inflammable Bodies.” Philosophical Transactions of the Royal Society of London
101: 1–35.
Davy, Humphry. 1812. Elements of Chemical Philosophy. Philadelphia: Bradford and
Inskeep.
Fontani, Marco, Mariagrazia Costa, and Mary Virginia Orna. 2015. The Lost Elements: The
Periodic Table’s Shadow Side. Oxford: Oxford University Press.
Gay-Lussac, Joseph Louis, and Louis Jacques Thénard. 1808. “Sur la décomposition et la
recomposition de l’acide boracique.” Annales de Chimie 68: 169–174.
Hendry, Robin Findlay. 2005. “Lavoisier and Mendeleev on the Elements.” Foundations of
Chemistry 7 (1): 31–48.
Hricko, Jonathon. 2018. “Retail Realism, the Individuation of Theoretical Entities, and the
Case of the Muriatic Radical. In Individuation, Process, and Scientific Practices, edited
by Otávio Bueno, Ruey-Lin Chen, and Melinda B. Fagan, 259–278. New York: Oxford
University Press.
Kitcher, Philip. 1993. The Advancement of Science: Science without Legend, Objectivity
without Illusions. New York: Oxford University Press.
Knight, David M. 1992. Humphry Davy: Science and Power. Cambridge, MA: Blackwell.
Laudan, Larry. 1981. “A Confutation of Convergent Realism.” Philosophy of Science 48
(1): 19–49.
Laurent, Auguste. 1855. Chemical Method, Notation, Classification, and Nomenclature.
London: Harrison and Sons, St. Martin’s Lane.
Lavoisier, Antoine Laurent. 1790 [1965]. Elements of Chemistry. 1st ed. Translated by
Robert Kerr. New York: Dover. Translation of Traité élémentaire de chimie, 1789.
Le Grand, H. E. 1972. “Lavoisier’s Oxygen Theory of Acidity.” Annals of Science 29
(1): 1–18.
Lowry, Thomas Martin. 1915. Historical Introduction to Chemistry. London: Macmillan.
Lyons, Timothy D. 2002. “Scientific Realism and the Pessimistic Meta-Modus Tollens.” In
Recent Themes in the Philosophy of Science: Scientific Realism and Common Sense, ed-
ited by Steve Clarke and Timothy D. Lyons, 63–90. Dordecht: Kluwer.
Lyons, Timothy D. 2006. “Scientific Realism and the Stratagema de Divide et Impera.” The
British Journal for the Philosophy of Science 57 (3): 537–560.
Lyons, Timothy D. 2016. “Structural Realism versus Deployment Realism: A Comparative
Evaluation.” Studies in History and Philosophy of Science 59: 95–105.
Lyons, Timothy D. 2017. “Epistemic Selectivity, Historical Threats, and the Non-Epistemic
Tenets of Scientific Realism.” Synthese 194 (9): 3203–3219.
Miller, William Allen. 1871. Elements of Chemistry: Theoretical and Practical. Part II.
Inorganic Chemistry. 3rd London ed. New York: John Wiley & Son.
Musgrave, Alan. 1988. “The Ultimate Argument for Scientific Realism.” In Relativism and
Realism in Science, edited by Robert Nola, 229–252. Dordrecht: Kluwer.
Naslain, R. 1977. “Crystal Chemistry of Boron and of Some Boron- Rich Phases;
Preparation of Boron Modifications. In Boron and Refractory Borides, edited by Vlado
I. Matkovich, 139–202. Berlin: Springer-Verlag.
Nicholson, William. 1795. A Dictionary of Chemistry. Vol. 1. London: G. G. and
J. Robinson, Paternoster Row.
Psillos, Stathis. 1999. Scientific Realism: How Science Tracks Truth. London: Routledge.
Scientific Realism Debate 55
Scerri, Eric R. 2005. “Some Aspects of the Metaphysics of Chemistry and the Nature of
the Elements.” HYLE—International Journal for Philosophy of Chemistry 11: 127–145.
Stanford, P. Kyle. 2006. Exceeding Our Grasp: Science, History, and the Problem of
Unconceived Alternatives. New York: Oxford University Press.
Thomson, Thomas. 1802. A System of Chemistry. Vol. 2. Edinburgh: Bell & Bradfute, and
E. Balfour.
van Fraassen, Bas C. 1980. The Scientific Image. Oxford: Clarendon Press.
Vickers, Peter. 2013. “A Confrontation of Convergent Realism.” Philosophy of Science 80
(2): 189–211.
Zhigach, A. F., and D. C. Stasinevich. 1977. “Methods of Preparation of Amorphous
Boron.” In Boron and Refractory Borides, edited by Vlado I. Matkovich, 214–226.
Berlin: Springer-Verlag.
4
No Miracle after All
The Thomson Brothers’ Novel Prediction that Pressure
Lowers the Freezing Point of Water
Keith Hutchison
1 Rankine, ‘On an equation’ (1849) = Misc. Sci. Pap., pp.1–12. The prediction here was of the
form of the relationship between the temperature and saturated vapour pressure of a liquid/vapour
mixture, without a specification of precise numerical values, and this may well place it outside the
intended scope of the no-miracle argument.
2 For an overview, see: Hutchison, ‘Miracle or mystery,’ pp.106–7 and ‘Mayer’s hypothesis,’ esp.
pp.288–294.
Keith Hutchison, No Miracle after All In: Contemporary Scientific Realism. Edited by: Timothy D. Lyons and Peter Vickers,
Oxford University Press. © Oxford University Press 2021. DOI: 10.1093/oso/9780190946814.003.0004
No Miracle After All 57
3 Rankine called it the “thermodynamic function”, and denied (e.g. Misc. Sci. Pap., p.233) that it
had any asymmetric tendency to increase. For its first explicit identification, see pp.351–2 of his Misc.
Sci. Pap. For Clausius’s renaming, see the English trans. on pp.364–5 of Mechanical theory. A predic-
tion like this, that a particular state-function exists, is definitely subject to empirical verification, but
it is certainly atypical, so could perhaps be outside the scope of the no-miracle argument.
4 Hutchison, ‘Miracle or Mystery’ (2002). In that study, I did consider the dependence question,
but only briefly (on pp.109–11), but that discussion now feels too thin. As I recall, those remarks were
inserted at the request of a referee.
5 Carnot, Réflexions sur la puissance motrice du feu.
6 For a survey, see e.g., Cardwell, Watt to Clausius, esp. pp.186–260.
7 The whole process took place between 1847 and 1850. Though the prediction came out under
James’s name, William was heavily involved and may even have been the principal investigator. For
the chronology etc., see: Smith & Wise, Energy and empire, pp.294–99. For the original prediction,
see: J. Thomson, ‘Theoretical considerations’ and W. Thomson ‘The effect of pressure,’ reprinted
together on pp.156–69 of W. Thomson, Math. Phys. Pap., v.1. My quotations are from pp.165–6 of
William’s paper. Given that the sensitivity of the boiling point of water to pressure had long been
known, it is odd that the sensitivity of the freezing point was (apparently) quite unsuspected in 1848.
58 Historical Cases for the Debate
To preserve Carnot’s central claim, however, Clausius did need to invoke a new
premise—a second law—and with these decisive steps, modern thermody-
namics emerged. Thomson soon accepted Clausius’s understandings.
My concern here is with that slightly earlier prediction however. What was the
thinking that led James to conclude that external pressure lowered the freezing
point? How did he defend this conclusion? How did he estimate the magnitude
of the lowering? Did he really use the theory that was about to be discarded?
By answering these historical questions, I shall demystify the prediction, by
showing that though James’s argument uses suppositions rejected in modern
thermodynamics, and applies those falsehoods to a situation where the two the-
ories differ, the quantitative difference between the two accounts of James’s phe-
nomenon is only marginal. One says that heat is conserved here, the other says it
is not, but agrees that in this particular situation heat is nearly conserved. James’s
argument furthermore needed a principle that was central to both theories, but
at the time he wrote that principle was only supplied by the Carnot theory. So his
prediction did indeed need the precarious theory. Yet the critical inference did
not itself directly use those portions of the Carnot theory that were soon aban-
doned. The false components did however give credence to the true portions—
so mattered. But there is (seemingly) little miracle in using the truish sections
of a false theory to make a prediction. To demolish the no-miracle argument,
a bigger miracle is needed, one where a successful prediction hinges on bigger
deviations from the truth.
In discussing this episode, I deploy a very naive distinction between “true”
and “false” statements. This greatly simplifies my narrative, for it avoids compli-
cated hedging, circumlocutions, etc. The associated over-simplifications seem to
me to do no harm, because the nature of “scientific truth” is not the issue here
and does not need to be confronted. But my language might at times seem an-
noying to the fastidious reader.
The half-century beginning with 1775 saw a ten-fold increase in the effi-
ciency of steam engines, an improvement so extraordinary that the question
began to circulate whether there was any limit to this process.8 Attempts to
answer this question were generally of limited scope, focusing too closely
on familiar devices, but one was quite exceptional. In a masterpiece of ele-
gant analysis, Sadi Carnot introduced (in 1824) the notion of a “heat-engine,”
one characterised by the single fact that it produces mechanical work by
8 See, e.g. Fox, ‘Introduction etc.,’ pp.2–12 and Cardwell, Watt to Clausius, pp.153–81.
No Miracle After All 59
exchanging heat with its environment. The recent prevalence of the notion
that heat was a special fluid, caloric, suggested that there was, accordingly,
some sort of analogy between heat-engines and water-engines, given that the
latter operated through an exchange of water with their environment. This
analogy lies at the heart of Carnot’s discussion, but it also incorporates the
flaw that eventually led to the analysis being abandoned, the presumption that
heat (like water) was conserved. Yet it provided Carnot with his key insight,
an insight that survived the later demise of the 1824 analysis: just as a water-
engine requires two “reservoirs,” a high one to supply water and a low one to
receive “waste,” a heat-engine requires two heat-reservoirs, one at a high tem-
perature to supply heat, and one at a low temperature to receive heat. (Watt’s
separate condenser had made it abundantly clear that steam engines produced
waste-heat, and that, accordingly, their coal needed to be supplemented by a
source of cold, simply for the engine to function.)
The same analogy also provided Carnot with the idea of a heat-pump, and
that of a reversible engine, and he was then able to make a simple adjustment
to an argument that had already been applied to water-engines, to show that
if it was accepted that perpetual motion was impossible, no heat-engine could
be more efficient than one which was perfectly reversible (for given source
and sink levels). This is Carnot’s answer to the question challenging his con-
temporaries, and to enhance its bona fides, Carnot showed that his theory
was very fertile indeed. He produced some rough calculations of the maximal
efficiencies he had identified, rough only because too much of the requisite
data was lacking; and he produced a series of novel lemmata about the proper-
ties of fluids, typical of those deductions which are a standard feature of later
thermodynamics.9
But Carnot’s bold theory attracted little attention—until the late 1840s, after a
copy of his memoir had been discovered by William Thomson, then working
in Regnault’s laboratory in Paris (where the sort of data needed to complete
Carnot’s calculations was being systematically generated). Thomson was en-
thusiastic about the analysis. He soon published his own rehearsal of the theory,
with some of the core calculations systematized (via Regnault’s new data). But
before this came out, Thomson also showed that Carnot’s understanding pro-
vided a rationale for absolute thermometry, a temperature scale independent of
9 For the efficiency estimates, see Carnot, Réflexions, pp.94– 101. For the lemmata, see: ibid.,
pp.78–90; Fox, ‘Introduction etc.,’ pp.132–45; Cardwell, Watt to Clausius, pp.201–4.
60 Historical Cases for the Debate
the idiosyncrasies of any particular chemical substance. And he noticed that the
Stirling air-engine vividly illustrated Carnot’s claims: it used two heat reservoirs
and could be run in two different directions: as a heat-engine; or as a heat-
pump—where the transfer of heat was palpable.10
When a model Stirling engine was run as a pump, shifting heat between two
reservoirs at the same temperature, virtually no work was required to operate
the pump (just as Carnot’s theory would predict).11 It was this fact that led to
James’s prediction, because it meant that heat could easily be pumped from
beneath the piston of a cylinder containing a water-ice mixture at 0℃ to
(say) an icy lake at the same temperature (where the heat received would melt
some ice without changing the lake’s temperature). Yet if this was done, some
of the water in the cylinder would freeze, and in doing so it would expand
(for water colder than 4℃ expands when cooled). And it was well known that
the forces generated by freezing water are very great, for they often caused
serious damage. So quite a high force could be applied to the piston, without
preventing the expansion, and combining this force with the movement of
the piston would yield a supply of external work that exceeded (by far) the
negligible quantity of work consumed in pumping the heat to the lake. So a
perpetual motion would seem to be on offer. Or rather there had to be some
flaw in the reasoning here: some phenomenon, hitherto overlooked, must be
sabotaging the proposal.
Reflection quickly shows one such flaw: the argument that threatens a per-
petual motion presumes the temperature in the high-pressure cylinder to be
the same as that in the lake, i.e. that the temperatures of the two different
ice-water mixtures are identical, despite the different pressures at which the
phases change. (For it is this identity that enables the more or less work-f ree
pumping of the heat from the cylinder.) So if the temperature at which water
freezes changes with the pressure the argument collapses. Indeed, if the tem-
perature of the high-pressure cylinder were lower than that of the lake (acted
on by normal atmospheric pressure) work would be consumed in pumping
the heat “uphill” into the lake, and this work could suffice to prevent a per-
petual motion.
Such were the thoughts that led to James’s discovery, though his published an-
nouncement of the discovery includes a more systematic analysis, one extended
to include a quantitative estimate of the magnitude of the newly appreciated ef-
fect.12 Let us now sketch that argument.
We begin by observing that to understand the operation of any engine, we
obviously need to make a clear distinction between the fuel (materials which
take on permanently new forms in the course of the operation of the device)
and the engine proper (that which suffers no permanent change, the “working
substance” as it is often called, plus its mechanical container, etc.). Carnot
accommodates this fundamental distinction by having his engines operate in
cycles, their materials passing through a sequence of changes that bring them
back to their original configuration. To avoid wastage, these changes need (as
mentioned earlier) to be reversible; and in a thermal context, the simplest such
engine would seem to be one with a single working substance undergoing just
four changes, two to exchange heat with the two different reservoirs (each at the
same temperature as the engine materials13), and two to switch the temperature
of those materials with no exchange of heat. The materials need to be brought,
back and forth, between the pair of temperatures asserted by Carnot to be essen-
tial to a heat-engine.14
Adapting this idea of a “Carnot cycle” to his context, James considers an en-
gine composed of a cylinder containing a mixture of water and ice beneath a
piston. As the engine operates, the quantities of water and ice here vary, as ice
sometimes melts, and as water sometimes freezes, in association with the ener-
getic exchanges between the engine and its environment.
To see the details, imagine the cylinder to be thoroughly insulated, with a rel-
atively large weight on the piston, so that the water-ice mixture is at a highish
pressure, and in internal equilibrium at whatever temperature happens to be
the melting point of water for that pressure. Suppose now (as a preliminary) the
12 For the more informal discussion, see J. Thomson, ‘Theoretical considerations,’ pp.157–9; for
temperature as the ice-water mixture. Like Carnot before him (e.g. Réflexions, p.68n), James recog-
nized this problem, and avoided it. I, by contrast, create it here—by using a circumlocution that is
quite common in thermodynamics, which serves to abbreviate the wording, and which I presume
will be familiar to any concerned reader. Since he was writing at the very beginning of thermody-
namic analysis, James could not presume such familiarity and explains the logic more fully, overtly
imagining a small difference in temperature, but one that could be made arbitrarily small, perhaps
infinitesimal (etc.) Some such artifice lies behind many references to isothermal change in thermo-
dynamic argument, and is routinely deployed in descriptions of a Carnot cycle.
14 Thermocouples however operate without undergoing any change themselves, but electrical
engines were a rarity at the time of the Thomson brothers’ discovery, so I just ignore any minor chal-
lenge they might pose to my narrative.
62 Historical Cases for the Debate
insulation is removed, and the cylinder brought into thermal contact with a heat
reservoir at that same temperature, and that heat flows out of the cylinder and
into the reservoir, so some of the water in the cylinder freezes, and the resulting
expansion causes the piston to lift the weight, doing mechanical work on its
environment.
After a period of expansion, one chosen simply for pragmatic reasons, the
cylinder is insulated again, and the weight on the piston is slowly reduced. As
the pressure is reduced, the mixture in the cylinder expands, and work is done
on the environment, while water freezes, or ice melts, to bring the cylinder to a
new temperature, that which is the melting point of water at the new, reduced,
pressure. This is to be the first stage of our cycle. For stage 2, the insulation is
removed, and the cylinder is brought into contact with a second heat-reservoir,
one at the same temperature as the low-pressure cylinder. Heat is now injected
into the cylinder, melting some of the ice, so that the piston descends (as the
volume of its contents decreases) and work is done on the system. After a con-
venient time, the cylinder is insulated again, and stage 3 begins: the weight on the
piston is slowly increased, further compressing the mixture in the cylinder, with
yet more work being done by the environment. The stage terminates when the
weight on the piston reaches its original value. Along the way, some water will
freeze, or some ice will melt, and the cylinder will return to its original tempera-
ture. The final, fourth, stage then mimics the preliminary process described ear-
lier, beginning as thermal contact with the first heat-reservoir is reestablished.
Expansion resumes and heat flows to the reservoir, until the initial volume of
stage 1 (= the final volume of the preliminary stage) is reached. The cycle is now
complete, and can be repeated if desired etc.
This cycle constitutes a heat-engine. For it exchanges nothing but heat and
work with its environment. And it produces mechanical work: for each change
of volume takes place in two directions, and the compression that reverses an ex-
pansion takes place at a lower pressure than the corresponding expansion. So if
Carnot is right, heat must be flowing from a high-temperature reservoir to a low-
temperature one. But we know that in this case heat is flowing into the reservoir
during stage 4 and out of the reservoir during stage 2 (while no heat is flowing
during stages 1 and 3). So the high-pressure stage 4 must take place at a lower
temperature than the low-pressure stage 2. In other words, an increase in pres-
sure reduces the freezing point.
There is no doubt that James’s prediction here uses the Carnot theory, for it is
this that tells James heat must flow from a high temperature source to a low
No Miracle After All 63
temperature sink. But Carnot does more for James than simply assert this neces-
sity, it embeds the requirement in a sustained analysis which enhances its plau-
sibility. The fact that the Carnot theory requires such a thermal descent is not,
however, what makes it false. The theory is false because of something quite dif-
ferent, the fact that it declares the quantity of heat released to the sink identical to
that acquired from the source. Carnot’s discussion made much use of the heat-
conservation that this alleged identity illustrates,15 and the rationale for insisting
upon a descent of heat presumes conservation. But once that descent is accepted,
the prediction requires no further appeal to conservation. Indeed, James’s deri-
vation works just as well in post-1850 thermodynamics where the heat released
to the sink differs from that taken from the source. Cosmetic adjustments do, of
course, need to be made to articulations of the story; and a different context (like
that provided by Clausius), is needed to make the necessity of a low tempera-
ture sink plausible. So James may be thought of as taking a trustworthy deriva-
tion and placing it in bad company (via the alternative rationale for its central
premise). But no-one is challenged by a trustworthy argument making a suc-
cessful prediction. In other words, James’s prediction requires no miracle, and
the episode shows that a false theory really can—sometimes—make a successful
prediction. But it is a weak counter example to the no-miracle argument, for the
prediction proceeds via the true components of a false theory. To get a better
counter-example, it would be necessary to show something stronger: that James’s
deduction would collapse if denied some false component of the Carnot theory,
heat-conservation in particular. But it does not collapse. All that happens is that
its critical premise is left without a rationale. The Rankine story is far more pow-
erful, then, and does seem to be a real counter-example—if I got that story right.
15 The argument given by Carnot (at p.69 of Réflexions) to show that no engine can be more effi-
cient than a reversible uses the conservation of heat. See also: Carnot, Réflexions, pp.65, 76n, 100–
101; Fox, ‘Introduction etc.,’ pp.131, 150.
64 Historical Cases for the Debate
to consider how much this matters in the context of the no-miracle argument—
arguing that the differences are of no moment.
Like me, James gives two versions of the argument (see n. 12), one informal
(and focused on that key insight supplied by the Stirling engine) and the
other phrased, more systematically, in terms of a Carnot cycle. James however
introduces the Carnot cycle as a means of making his quantitative estimate, and
regards the effect itself as established by the more informal argument; while I use
that cycle to perform both these tasks (with the quantitative evaluation delayed
until later). Otherwise, our discussions of the cycle are very close, though that of
James seems more verbose, aimed at an audience with little exposure to thermo-
dynamic analysis (as, e.g. indicated in n. 13).
Our informal arguments however differ markedly, mainly because I cannot
follow the original argument in detail. There may (accordingly) be some serious
blemish in James’s prediction, a blemish hidden by my rephrasing.16 Our concern
here, however, is with the plausibility of the prediction, whether the conclusions
are reasonably justified by the premises (and their context), not whether a par-
ticular string of words adequately displays that connection. So a blemish in some
wording of the argument is of no moment, so long as some minor redrafting
sustains the conclusion—without introducing new premises. James may have
written his argument down hastily, and not done full justice to his thoughts, or
he may have gone to an opposite extreme, and penned numerous drafts, none of
which is canonical. In either extreme, the published argument does not tell the
full story, and the real argument for the prediction is that elaboration of the overt
wording that each of us privately prepares, as we absorb the message conveyed
to us by the written summary. So there are (I suggest) no grounds for concern
in the mere existence of small discrepancies between James’s words and my own
phrasing of the argument that supplies the prediction.
To estimate the magnitude of the effect he had discovered, i.e. how large a re-
duction (“−t”) in the freezing point is produced by a given increase (“[+]p”) in
pressure, James examines a concrete example of the Carnot cycle we introduced
to set out the qualitative argument.17 To reveal the relationship between p and t,
16 James does not, for example, evaluate the work done on the environment during the adiabatic
stages of his cycle, just treating the work done in these stages as negligibly small. I, by contrast, ex-
panded my description of the cycle slightly, in such a manner as would (hopefully) reduce any un-
certainty here. James’s brevity probably does not matter, but I am not sure. The issue hardly seems to
matter, since it is so easy to rephrase James’s argument.
17 James Thomson, loc. cit., n.12 = ‘Theoretical considerations,’ pp.160–4.
No Miracle After All 65
he uses three “known” quantities: (1) how much expansion is produced by the
freezing of water; (2) the latent heat required to melt ice at 0℃; and (3) the effi-
ciency of a Carnot engine that operates by absorbing heat at temperature 0℃ and
releasing heat at temperature −t℃. At that time, there was no novelty in the first
two of these data, while the third had recently been calculated by William, in his
systematic revision of Carnot’s fragmentary calculations.
The first of these items gives James the work done by his engine, as a multiple
of p—viz. “0.087 × p.” (James ignores the work done during the changes in tem-
perature, since the adiabatic stages of the cycle are so small, “infinitesimal” so
to speak.) The other two quantities give another, quite different, expression for
this same work, now as a multiple of t—viz. “4925 × 4.97 × t,” just the product
of William’s figure for the efficiency (“4.97”), the fuel consumed (“4925”) and
the magnitude of temperature drop (“t”). Equating these two figures, gives James
his final estimate, “t = 0.00000355p,” for the change in melting point produced
by an additional pressure p. Without specifying James’s units (viz. “cubic feet”;
“foot-pounds”; “pounds on a square foot”; “the quantity of heat required to raise
a pound of water from 0 to 1 degree centigrade”) these results are verging on the
meaningless, but our aim here is not to reveal the actual magnitude of the effect
he discovered, just to check that James’s calculation requires no appeal (direct or
indirect) to the conservation of heat.
We have clearly not established such an independence yet, for though the first
two items of data make no discernible appeal to conservation, the earlier cal-
culation of efficiency was made in the context of the Carnot theory, and used
data that had been prepared in a laboratory that did not embrace the possibility
that heat was created or destroyed. So application of this calculation to James’s
cycle could well have required some identification, overt or hidden, of the heat
supplied to some engine with that it released. And some such comparison had
indeed been required in the very calculations used by James, those performed
by William to indicate that his first “absolute” scale of temperatures was in good
agreement with familiar laboratory scales. When the Carnot theory was aban-
doned, William realised his calculations were suspect, so checked the agreement
he had claimed for the earlier scale, and found it no longer applied. He decided to
rescue the convenience it had promised, by developing a rather different absolute
scale, effectively the one in use today.18
We thus need to scrutinise James’s use of William’s earlier calculations. But
before doing so, it is worth observing that James’s method of calculating the
magnitude of the effect he had discovered is effectively a standard component
of modern thermodynamics textbooks, used to generate the so-called Clausius-
Clapeyron equation, relating the variation in the boiling-point of a liquid to
the external pressure. A version of this equation had been published by Émile
Clapeyron, in his important 1834 injection of Carnot’s discussion into a far more
mathematical, and especially graphical, idiom.19 The later survival of the calcula-
tion suggests that the result was not essentially tied to heat-conservation, but we
need here to confirm that suggestion—for on a strictly literal reading of James’s
account, a dependence of heat-conservation is evident.
James tells us that his estimate of the efficiency of his engine comes from “tables
deduced . . . from the experiments of Regnault” and published in his brother’s
major paper of 1849—to which James’s own discussion is attached as an appendix
(in the 1882 reprint). Strictly speaking this is not correct, as the coldest engine
listed in William’s table is one that operates just above 0℃, while James’s engine
operates just below that temperature. But William’s figure for the efficiency of a
perfect engine taking heat from 1℃ to 0℃ is 4.96, and James has presumably got
his 4.97 by extrapolation (something perfectly reasonable), or by access to data
not used in the final version of William’s table.
William had published those efficiency figures as the ones applicable to all
Carnot engines, so James was entitled to apply the figure to his own ice/water de-
vice, and that is the logic used by James here.20 Yet an appeal to heat-conservation
is lurking beneath the surface here, for (as observed in n. 15) the argument that
the figure at issue applied to all engines used the problematical presumption. But
a validation of this figure does not require a universal conservation of heat. It is
enough that heat be conserved in two particular engines, that used by William
to compute the efficiency and that used by James to evaluate the change in the
freezing point. And both these engines operate over a very small temperature
range, where the difference between the fuel heat and the waste heat is tiny. So the
claim that heat is conserved is effectively true in these particular circumstances,
and the suspect evaluation of the “infinitesimal” engine’s efficiency is in fact reli-
able, as William later checked.21
But there is also a risk, far less obvious, of a further appeal to heat-conservation
in William’s use of Regnault’s experimental data. For Regnault could well have
19 See the three equations on p.90 of Clapeyron, ‘Motive power.’ For modern versions, see
Pippard, Classical thermodynamics, p.54 (eq. 5.12) and Fermi, Thermodynamics, pp.63–69. Clausius
points out that James’s prediction is easily transferred to the new theory: see his 1850 remarks on
pp.80–3 of Mech. Theory.
20 The critical passage is on p.163 of Math. Phys Papers, v.1. William’s table is on p.140.
21 See n. 22.
No Miracle After All 67
presumed that heat was conserved in making his measurements, and this back-
ground assumption could have tainted his data. When, soon after these events,
William accepted that Joule had refuted the conservation of heat and endorsed
Clausius’s modification of the Carnot theory, William published a review of the
data behind James’s calculation and concluded that it had not been compromised
by any such presumption by Regnault.22
I see no reason to doubt William’s judgment here, all the more so because he
was certainly happy to admit error. For at the very same time, William reviewed
his earlier estimates of the efficiency of Carnot engines, concluding that these
were seriously in error, and publishing revised estimates (effectively those in use
today).23 This meant that the table cited by James was vitiated by the presumption
of heat-conservation made in its compilation, but it turned out that the errors got
smaller and smaller as the range of temperatures over which the engine operated
shrank. So the one figure used by James, the 4.97, was quite acceptable. Once
again, there is no dependence on the false components of Carnot’s account.
22 See W. Thomson, Math. Phys. Papers, v.1, pp.193–5. Chang, Inventing temperature, pp.76–
77, 86–7, 99–100 notes that Regnault was extremely cautious about allowing theory to taint his
measurements and (pp.174–5) that William Thomson found this caution a bit excessive. But Chang
notes too (77, n.48) that some of Regnault’s calorimetric measurements did require heat to be con-
served. Thomson had, of course, worked with Regnault’s laboratory, so his assessment here would be
especially reliable.
23 W. Thomson, Math. Phys. Papers, v.1, pp.189–91, 197–8.
68 Historical Cases for the Debate
References
Cardwell, Donald S. L. From Watt to Clausius: The rise of thermodynamics in the early
industrial age. (London: Heinemann, 1971).
Carnot, Sadi. Réflexions sur la puissance motrice du feu. [1824]. Critical edition, ed.
Robert Fox. (Paris: Vrin, 1978). My pages reference are however, to Fox’s English trans-
lation, Reflexions on the motive power of fire, (Manchester: Manchester University
Press, 1986).
No Miracle After All 69
5.1 Introduction
The pessimistic induction (PI) takes it that the evidence of the history of sci-
ence is an indictment for scientific realism: if past theories have been invariably
false, then it will be very unlikely that current theories will turn out to be true,
unless a sort of epistemic privilege is granted to current theories. But, the argu-
ment might go, what could possibly be the ground for this privilege? Assuming a
substantial difference between current theories and past (and abandoned) ones
would require that the methods used by current scientists are substantially dif-
ferent than those of the past scientists. But methodological shifts, the argument
would conclude, are rare, if they occur at all.
Yet, standards of evidence as well as conceptions of explanation have changed
as science has grown. If we look at these changes, we may draw different
conclusions regarding the bearing of the history of science on scientific realism.
Theories are not (nor have they been) invariably worthless because false. Some
have been outright falsehoods, while others hold in their domains under various
conditions of approximation. More importantly, some are such that important
parts of them have been retained in subsequent theories. Some, but not others,
have been licensed by more rigorous standards of success. Some, but not others,
have initiated research programs, which result in successive approximations to
truth. Hence, the history of science carries a mixed message, which needs to be
read with care.
Stathis Psillos, From the Evidence of History to the History of Evidence In: Contemporary Scientific Realism.
Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press 2021.
DOI: 10.1093/oso/9780190946814.003.0005
Descartes, Newton, and Beyond 71
In this chapter, I will look into two episodes in the history of science: the tran-
sition from the Cartesian natural philosophy to the Newtonian one, and then to
the Einsteinian science. These are complex episodes, and I can only offer sketches
here. But the point I want to drive home should still be visible: though the shift
from Descartes’s theory to Newtonian mechanics amounted to a wholesale re-
jection of Descartes’s theory (including his account of vortices as well as his
methodology), in the second shift, a great deal was retained; Newton’s theory
of universal gravitation gave rise to a research program which informed and
constrained Einstein’s theory. Newton’s theory was a lot more supported by the
evidence than Descartes’s and this made it imperative for its successor theory
to accommodate within it as much as possible of Newton’s theory: evidence for
Newton’s theory became evidence for Einstein’s. Besides, a point that these two
successive cases of theory-change will highlight is that the very judgment that
some theories are more supported by the evidence than others, and the concom-
itant judgement that some standards of evidence and explanation are better than
others, need not be made in a post hoc way by current philosophers of science.
Rather, they are made by the protagonists of these episodes themselves.
This double case study will motivate a rebranding of the divide et impera
strategy against PI that I introduced in my (1996) and developed in my (1999).
This rebranding urges shifting attention from the (crude) evidence of the his-
tory of science to the (refined) history of evidence for scientific theories. What’s
the connection between the two? As I will show in some detail before the
conclusions, to paraphrase Clausewitz, “moving from the evidence of history to
the history of evidence” is the continuation of divide et impera by other means.
The reader might wonder at the outset: are the lessons drawn from this double case
generalizable? Similar studies in optics and the theories of heat that I have performed
in the past (see my 1999, c hapter 6) suggest they are. But all these are theories of
physics and the reader might still wonder whether these lessons can be extended to
other areas of science. I will leave it open whether they can or cannot. The key point is
that it is sufficient to neutralize the force of the PI to show that some typical historical
cases, which have been used to feed epistemic pessimism about current science, do
not license this pessimism. On the contrary, they suggest that a more fine-tuned ac-
count of theory-change in science gives us reasons for epistemic optimism.
Here is the road-map. In section 5.2, I examine in some detail Descartes’s
account of method and explanation. I focus my attention on the role of
hypotheses in Cartesian science and the circumstances under which they
might be taken to be true. In section 5.3, I discuss Newton’s more stringent
account of scientific explanation and his critique of Descartes’s vortex theory
of planetary motions. I focus my attention on the fact that Newton’s initiated
a research program which aimed at the discovery of physical causes of the
72 Historical Cases for the Debate
systematic discrepancies between the predictions of the theory and the ac-
tual orbits of the planets; a program which put Newton’s theory to a series of
rigorous tests. In section 5.4, I discuss the retentionist pattern that Newton’s
theory of gravity set in motion, as this was exemplified in Einstein’s general
theory of relativity (GTR) and in particular in the explanation of the anoma-
lous perihelion of Mercury. In section 5.5, I compare the strategy suggested in
this paper with the earlier divide et impera strategy and show their substantial
similarity. Finally, I draw some general conclusions concerning the need to
shift our attention from the evidence of the history of science to the history of
evidence for scientific theories.
Descartes was the first to put forward a comprehensive theory of the world
based on entities invisible to the naked eye. All worldly phenomena, all mac-
roscopic objects and their behavior, are accounted for by the motion(s),
notably collisions, of invisible corpuscles, subject to general and universal
laws of nature. Within this framework, to offer a scientific explanation of a
worldly phenomenon X was to provide a particular configuration Y of matter
in motion, subject to laws, such that Y could cause X. Elsewhere, I have
explained the status and role of laws of nature in Cartesian natural philos-
ophy (cf. 2018b; forthcoming).1 The key relevant idea is that in Cartesian
physics, the possible empirical models of the world are restricted from above
by a priori principles, which capture the fundamental laws of motion, and
from below by experience. Between these two levels there are various me-
chanical hypotheses, which refer to configurations of matter in motion. As
Descartes explains in (III, 46) of the Principia, since it is a priori possible that
there are countless configurations of matter in motion that can underlie the
various natural phenomena, “unaided reason” is not able to figure out the
right configuration. Mechanical hypotheses are necessary, but experience
should be appealed to, in order to pick out the one in conformity with the
phenomena:
[W]e are now at liberty to assume anything we please [about the mechanical
configuration], provided that everything we shall deduce from it is {entirely} in
conformity with experience. (III, 46; 1982, 106)
1 There has been considerable debate about the status and role of laws of nature in Descartes. Some
notable contributions include Ott (2009), Hattab (2000), and Wilson (1999, c hapter 28).
Descartes, Newton, and Beyond 73
There are two sorts of hypotheses in Cartesian physics and his concomitant
hypothetico-deductive method: vertical and horizontal. “Vertical” I call the
hypotheses which posit invisible entities and their primary modes (shape and
size, i.e., bulk, and motion). These kinds of entities, being invisible, have to be
hypothesized. But there are “horizontal” hypotheses too. These concern mac-
roscopic (that is visible) entities. Being visible, these entities do not have to be
hypothesized; yet their structural relations have to be hypothesized because dif-
ferent structures yield the same appearances. The prime example is the “system
of the world”—the solar system. Here, there are competing, but empirically
equivalent, possible structures of the planetary motions.
In Principia (IV, 201; 1982, 283), Descartes states that sensible bodies are com-
posed of insensible particles. So insensible particles (of various bulks and speeds)
are everywhere: they fill (literally) space. A number of problems crop up, at this
juncture. How can there be epistemic access to these invisible entities? How is
knowledge of them possible? How can hypotheses be justified and accepted?
How can one of them be more preferable than another, given that there are more
than one which tally with the phenomena? All these are familiar questions. They
are at the heart of the current scientific realism debate. Let us see how Descartes
answers them.
Nor do I think that anyone who is using his reason will be prepared to deny
that it is far better to judge of things which occur in tiny bodies (which escape
our senses solely because of their smallness) on the model of those which our
senses perceive occurring in large bodies, than it is to devise I know not what
new things, having no similarity with those things which are observed, in order
to give an account of those things [in tiny bodies]. {E.g., prime matter, substan-
tial forms, and all that great array of qualities which many are accustomed to
assuming; each of which is more difficult to know than the things men claim to
explain by their means}. (IV, 201; 1982, 284)
[T]he matter of the heaven, in which the Planets are situated, unceasingly
revolves, like a vortex having the Sun as its center, and . . . those of its parts which
are close to the Sun move more quickly than those further away; and . . . all the
Planets (among which we {shall from now on} include the Earth) always remain
suspended among the same parts of this heavenly matter. (III, 30; 1982, 196)
some of these straws rotate about their own centers, and that those which
are closer to the center of the vortex which contains them complete their
circle more rapidly than those which are further away from it. (III, 30;
1982, 196)
Given the continuity thesis, we may transfer this mechanical model to the motion
of the planets and “imagine that all the same things happen to the Planets; and
this is all we need to explain all their remaining phenomena” (III, 30; 1982, 196).
Descartes, Newton, and Beyond 75
How about horizontal hypotheses? Here the chief rivalry was between the three
systems of the world: the Ptolemaic, the Tychonic, and the Copernican. The
bridge principle cannot be of help. But experience can help refute one of them,
viz., the Ptolemaic. As Descartes notes it is ruled out because it does not con-
form with observations, notably the phases of Venus (cf. III, 16; 1982, 89–90).
But when it comes to Copernicus vs. Tycho, Descartes notes that they “do not
differ if considered only as hypotheses”—they “explain all phenomena equally
well,” though Copernicus’s hypothesis is “simpler and clearer.” Hence, qua poten-
tial (hypothetical) explanations these two systems do not differ.
Descartes advocated a version of the Copernican theory because it’s “the sim-
plest and most useful of all; both for understanding the phenomena and for in-
quiring into their natural causes” (III, 19; 1982, 91). In his own version, he denied
the motion of the earth “more carefully than Copernicus and more truthfully
than Tycho” (III, 19; 1982, 91). To understand what goes on, we need to make a
short digression into Descartes’s theory of motion.
Notoriously, Descartes defines motion twice over. On the one hand, there is
the ordinary account of motion as “the action by which some body travels from one
place to another.” On the other hand, there is the “true” account of motion: “the
transference of one part of matter or of one body, from the vicinity of those bodies
immediately contiguous to it and considered as at rest, into the vicinity off some]
others” (Principia II, 24, 25; 1982, 50–51). According to this account, motion is
change of position relative to contiguous bodies that are considered at rest. But
the earth is carried over by a vortex on which it is, so to speak, pinned; hence, it
does not move (it is at rest) relative to its contiguous and at-rest bodies. Nor do
the other planets. The observed changes in their positions are wholly due to the
motion of the vortex which contains them.2
It is noteworthy that Tycho’s account is in conflict with Descartes’s theory of
motion: the Earth appears immobile, but it is not at rest. The reason for this is
based on a) the relativity of motion; and b) on Descartes’s claim that a body is
in motion only if it moves as a whole. As he observes in Principia (III, 38; 1982,
101–102), the relativity of motion implies that it could be either that the earth
rotates and the heavens with the stars are rest, or the other way around. But, on
the Tychonic hypothesis, condition b) implies that it is the earth which is actu-
ally in motion, since the earth moves as a whole body relative to the heavens (its
separation from the contiguous heavens taking place along the whole of the sur-
face of the earth), whereas the heavens with the stars in them are at rest, since the
separation takes place over a small fraction of their surface, viz., this part which
is contiguous with the earth. In fact, there is a deeper reason why Descartes
dismissed Tycho’s model of the System of the World. Given that the planets
(apart from the earth) revolve around the sun, and the sun revolves around the
earth, it is hard to see how this combination of motions could be modeled by
vortices. Hence, Descartes concluded that Tycho “asserted only verbally that the
Earth was at rest and in fact granted it more motion than had his predecessor”
(III, 18; 1982, 91).3
Despite his heliocentrism, Descartes noted that he accepted the Copernican
account not as true, “but only as a hypothesis,” which may be false. Hypotheses
need not be true to be useful. But can they be true? Or better, under what
circumstances, if any, should they be taken to be true?
3 For a more detailed account of Descartes’s theory of motion, cf. Garber (1992).
Descartes, Newton, and Beyond 77
for Descartes “as if true” seems to mean something like “true for all practical
purposes” (or, morally certain, as Descartes would put it). In the French transla-
tion of the Principia he added right after the claim that his hypothesis “will be as
useful to life as if it were true”: “because we will be able to use it in the same way
to dispose natural causes to produce the effects which we desire” (III, 44; 1982,
105). And again: “And I do not think it possible to devise any simpler, more intel-
ligible or more probable principles than these” (III, 47; 1982, 107).
Simplicity, intelligibility, and probability are marks of truth. Descartes adds
novel predictions, too; it is further support of the truth of a hypothesis, that is of
its being right about the causes of the phenomena, when it explains “not only the
effects that we were initially trying to explain but all these other phenomena that
we hadn’t even been thinking about” (III, 42; 1982, 104). On various occasions,
Descartes stresses that unification is a mark of truth too. In (III, 43; 1982, 104), he
notes emphatically that it “can scarcely be possible that the causes from which all
phenomena are clearly deduced are false.” He grounds this on the further claim
that if the principles are evident and the deductions of observations from them
“are in exact agreement with all natural phenomena,” it would “be an injustice
to God to believe that the causes of the effects which are in nature and which we
have thus discovered are false.”
The best explanation is known with moral certainty, i.e., certainty for all prac-
tical purposes. What is known with moral certainty could be false in the sense
that it would be in the absolute power of God to make things otherwise than
our best theory says they are. Descartes uses the metaphor of decoding a secret
message to make an analogy with the explanatory power of theories. The more
complex the message is the more unlikely it is that the code that was devised to
successfully decode it is a matter of coincidence. Similarly, the more a theory
reveals about the “fabric of the entire World,” the more unlikely it is that it is false
(cf. IV 205; 1982, 287).
4 For a thorough account of Newton’s criticism of Descartes’s theory of motion and of body, see
Katherine Brading (2012). Brading argues that both Descartes and Newton shared a law-constitutive
account of bodies but Newton developed it more systematically. See also Jalobeanu (2013).
5 For a detailed account of Newton’s criticism of Cartesian vortices, see Snow (1924) and Aiton
(1958a).
Descartes, Newton, and Beyond 79
But the problem with vortices appeared to be epistemological too: they were hy-
pothetical entities. In the preface to the second edition of the Principia, Roger
Cotes equated vortices with “occult causes,” being “utterly fictitious and com-
pletely imperceptible to the senses” (1729 [2016], 392).
But is the problem really that vortices, if they exist, are imperceptible? It seems
it is not. To see why it is not let us make a short digression to Huygens and Leibniz.
Mr Descartes has recognized, better than those that preceded him, that nothing
will be ever understood in physics except what can be made to depend on prin-
ciples that do not exceed the reach of our spirit, such as those that depend on
bodies, deprived of qualities, and their motions. (1–2)
Huygens posited a fluid matter that consists of very small parts in rapid motion
in all directions, and which fills the spherical space that includes all heavenly
bodies. Since there is no empty space, this fluid matter is more easily moved in
circular motion around the center, but not all parts of it move in the same di-
rection. As Huygens put it “it is not difficult now to explain how gravity is pro-
duced by this motion.” When the parts of the fluid matter encounter some bigger
bodies, like the planets: “these bodies [the planets] will necessarily be pushed
towards the center of motion, since they do not follow the rapid motion of the
aforementioned matter” (16). And he added:
This then is in all likelihood what the gravity of bodies truly consists of: we can
say that this is the endeavor that causes the fluid matter, which turns circularly
6 As Aiton (1958b) argues, it was only after the publication of Newton’s Principia that Cartesians
around the center of the Earth in all directions, to move away from the center
and to push in its place bodies that do not follow this motion. (16)
In fact, Huygens devised an experiment with bits of beeswax to show how this
movement toward the center can take place. At the same time, however, he had
no difficulty in granting that Newton’s law of gravity was essentially correct when
it comes to accounting for the planetary system, and especially Kepler’s laws. As
he put it:
I have nothing against Vis Centripeta, as Mr. Newton calls it, which causes the
planets to weigh (or gravitate) toward the Sun, and the Moon toward the Earth,
but here I remain in agreement without difficulty because not only do we know
through experience that there is such a manner of attraction or impulse in na-
ture, but also that it is explained by the laws of motion, as we have seen in what
I wrote above on gravity. (31)
Explaining the fact that gravity depends on the masses and diminishes with dis-
tance “in inverse proportion to the squares of the distances from the center” was,
for Huygens, a clear achievement of Newton’s theory, despite the fact that the me-
chanical cause of gravity remained unidentified. His major complaint was that he
did not agree with Newton that the law of gravity applies to the smallest parts of
bodies, because he thought that “the cause of such an attraction is not explicable
either by any principle of mechanics or by the laws of motion” (31).7
Leibniz in his Tentamen de motuum coelestium causis (1689), adopted a ver-
sion of the vortex theory despite his overall disagreement with Cartesian physics
and metaphysics. On this theory, a planet is carried by a “harmonic vortex,” the
layers of which moved with speeds inversely proportional to the distance from
the center of circulation, thereby satisfying Kepler’s area law. But the planet is
also subject to a radial motion (motus paracentricus), which is due to two forces
acting on the planet: the gravity exerted by the sun on the planet and the cen-
trifugal force arising from the circulation of the planet with the vortex. Given
all this, Leibniz was able to show that an elliptical orbit requires that the gravity
(which was conceived by Leibniz along the lines of the magnetic action) is in-
versely proportional to the distance, and then to deduce from this Kepler’s har-
monic law. In effect, Leibniz had posited two independent but superimposed
vortices for each planet, the vortex that causes gravity and the harmonic vortex
(cf. Aiton 1996, 12).
7 For further discussion of Huygen’s theory of gravity and his reaction to Newton’s, see Martins
(1994).
Descartes, Newton, and Beyond 81
This point, viz., that the problem with vortex theory was that it does not offer
a good explanation of gravity, is echoed by Roger Cotes: “For even if these
philosophers [the Cartesians] could account for the phenomena with the greatest
exactness on the basis of their [vortices] hypotheses, still they cannot be said to
have given us a true philosophy and to have found the true causes of the celestial
motions until they have demonstrated either that these causes really do exist or
at least that others do not exist” (Newton 1729 [2016], 393). But, as Newton has
shown, not only does the vortex hypothesis fail to account for the gravitational
phenomena; there is a rival theory—Newton’s own—that accounts for them.
In fact, Newton stressed repeatedly that gravity is a non-mechanical force;
hence it cannot be modeled by action within a mechanical medium. In the
General Scholium he noted that gravity does not operate “in proportion to the
quantity of the surfaces of the particles on which it acts (as mechanical causes are
wont to do) but in proportion to the quantity of solid matter” (1729 [2016], 943).
It transpires then that Newton’s theory is superior to Descartes’s as an ex-
planation. As Aiton (1958a, 1958b) has shown, most Cartesians succumbed to
Newton’s theory, and by the 1740s, the vortex theory had been fully abandoned.
But even before this, most Cartesians who advocated the vortex theory tried to
make it consistent with Newton’s.10 But there is more to it. Newton goes beyond
10 One of the last attempts to defend Descartes’s theory over Newton was by the Swiss mathema-
tician Johann Bernoulli who won the 1730 Académie Royale des Sciences prize for his essay on the
causes of elliptical orbits in a Cartesian vortex. He disagreed with two “bold suppositions” made by
Newton, which “shock the spirits accustomed to receive in Physics only incontestable and obvious
principles.” The first of these suppositions, according to Bernoulli, was “to attribute to the body a
virtue or attractive faculty” (1730, 6), while the second is the perfect void. But Bernoulli only gestured
Descartes, Newton, and Beyond 83
I have not as yet been able to deduce from phenomena the reason for these
properties of gravity, and I do not “feign” hypotheses. For whatever is not
deduced from the phenomena must be called a hypothesis; and hypotheses,
whether metaphysical or physical, or based on occult qualities, or mechanical,
have no place in experimental philosophy. In this experimental philosophy,
propositions are deduced from the phenomena and are made general by in-
duction. The impenetrability, mobility, and impetus of bodies, and the laws of
motion and the law of gravity have been found by this method. And it is enough
that gravity really exists and acts according to the laws that we have set forth
and is sufficient to explain all the motions of the heavenly bodies and of our sea.
(1729 [2016], 943)
In this, hypotheses of any sort are contrasted to propositions which “are deduced
from the phenomena and are made general by induction.” We shall see in the se-
quel how this claim may be understood. For the time being, the relevant point is
that Newton criticizes a certain account of explanation. But on what grounds? This
is explained by Cotes in his preface to the second edition of the Principia. Cotes
aims to highlight Newton’s methodological novelty, and he contrasts it with both
the Aristotelian way and the Cartesian way. He does not doubt that Cartesian me-
chanical philosophy is an improvement over the Aristotelian science. Nor does he
as to how Kepler’s laws could be accommodated in his system of vortices. For a detailed discussion,
see Aiton (1996, 17–18). See also Iltis (1973).
84 Historical Cases for the Debate
doubt that the Cartesians were right in aiming to explain the perceptible in terms
of the imperceptible. Indeed, Cotes praises the Cartesians for trying to show that
the “the variety of forms that is discerned in bodies all arises from certain very
simple and easily comprehensible attributes of the component particles” (Newton
1729 [2016], 385). Where did they go wrong then? In relying on speculative and
unfounded hypotheses about the properties of the invisible corpuscles.
If Cartesians are right on demonstration based on laws, but wrong to rely on
speculative and unfounded hypotheses, what is the alternative? It is to use the
phenomena as premises such that, together with the general laws of motion, fur-
ther laws are derived on their basis. What is thereby derived is not speculative
and unfounded. Indeed, that’s how Cotes describes Newton’s method:
Although they [the Newtonians] too hold that the causes of all things are to be
derived from the simplest possible principles, they assume nothing as a prin-
ciple that has not yet been thoroughly proved from phenomena. They do not
contrive hypotheses, nor do they admit them into natural science otherwise
than as questions whose truth may be discussed. (1729 [2016], 385)
The aim of explanation then changes: it is to use the general laws of motion
and phenomena in order to deduce further special laws, which have been estab-
lished in such a way that they cannot be denied while affirming the general laws
and the phenomena. In this sense, the special laws have been deduced from the
phenomena. They are then at least as certain as the phenomena and general laws
of motion. That’s how the law of gravity has been found.
We can then say that Newton’s experimental philosophy reverses the order of ex-
planation in science. Schematically put:
Descartes
General Laws & Hypotheses ➔ phenomena
Newton
General Laws & phenomena ➔ special laws
from data and observations. Hence, they are amenable to mathematical analysis.
For instance, phenomenon one is that the satellites of Jupiter (the circumjovial
planets), by radii drawn to the center of Jupiter, satisfy Kepler’s area law and
Kepler’s harmonic law. And phenomenon six is that the moon, by a radius drawn
to the center of the earth, satisfies Kepler’s area law.
It is then straightforward for Newton to apply Propositions 1, 2, and 4 (corol-
laries 6 and 7) of Book 1 of the Principia, which are proved for bodies qua math-
ematical points and forces qua mathematical functions, to the phenomena. As is
well known, Newton demonstrates the following two equivalences:
A body A moving in an ellipse (a conic section) around another body B which
is in a focus of the ellipse describes equal areas in equal times (that is, it satisfies
Kepler’s area law) if and only if it is urged by a centripetal force that is directed
toward body B.
A body A moving in an ellipse (a conic section) around another body B which
is in a focus of the ellipse has its periodic time T as the 3/2 power of its radius r
(that is, it satisfies Kepler’s harmonic law) if and only if the centripetal force that
urges it toward B is inversely proportional to the square of the distance r between
the two bodies.11
What is special about Newton’s account of explanation, as William Harper
and George Smith point out, is that he explains a phenomenon by relying on
equivalences such as those just described. These equivalences “make the phe-
nomenon measure a parameter of the theory which specifies its cause” (Harper
and Smith 1995, 147).12 In Book 3, the mathematically described centripetal
force of the foregoing equivalences is identified with gravity. In the scholium to
Proposition 5, Newton says: “Hitherto we have called ‘centripetal’ that force by
which celestial bodies are kept in their orbits. It is now established that this force
is gravity, and therefore we shall call it gravity from now on” (1729 [2016], 806).
Gravity is rendered a universal force by means of Rule III of Philosophy, which
is a rule of induction.13 This rule licenses the generalization of a quality, which is
found to all bodies on the basis of experiments, to all bodies whatsoever. Or, as
Newton put it:
11 In fact, Newton added a further proof of the inverse square law, one that as he says proves its
existence “with the greatest exactness” (1729 [2016], 802). The proof is based on the empirical fact
that “the aphelia [of the planets] are at rest.” As Newton argued, invoking proposition 45 of Book 1,
the slightest departure from the inverse square of the distance would “necessarily result in a notice-
able motion of the apsides in a single revolution and an immense such motion in many revolutions.”
This “systematic dependence,” as William Harper has put it, between the inverse square ratio and the
immobility of the aphelia is something that makes Newton’s account of gravity robust. Besides, as
Harper (2011, 42ff) has argued, looking for systematic dependencies between theoretical parameters
and the phenomena renders Newton’s method distinct from a simple hypothetico-deductive method.
12 For an elaboration of the idea of theory-mediated measurement, see Harper (2011).
13 For a useful discussion of Newton’s account of induction vis-à-vis Locke’s and Hume’s, see de
By reason of the deviation of the Sun from the center of gravity, the centrip-
etal force does not always tend to that immobile center, and hence the planets
neither move exactly in ellipses nor revolve twice in the same orbit. Each time
a planet revolves it traces a fresh orbit, as in the motion of the Moon, and each
orbit depends on the combined motions of all the planets, not to mention
the actions of all these on each other. . . . But to consider simultaneously all
these causes of motion and to define these motions by exact laws admitting of
easy calculation exceeds, if I am not mistaken, the force of any human mind.
(Herival 1965, 301)
Descartes, Newton, and Beyond 87
Kepler’s laws hold only approximately of the planets and their satellites mainly
because the actual motions are lot more complicated than the theory could cap-
ture. Was this something to give Newton pause? To cut a long story short, it was
certainly not. Not only Kepler’s laws hold quam proxime, but Newton himself
took pains to show that the equivalences he proved in Book I of the Principia
hold quam proxime too. This phrase (which occurs 139 times in the Principia)
literally means “most nearly to the highest degree possible”; it is probably best
translated “very, very nearly” or, as is more customary, “very nearly.”14
After proving Propositions 1 and 2 of Book I (which, as we have seen, assert
the equivalence: the centripetal force by which a body is drawn to an immo-
bile center of force is directed to this center iff that body satisfies Kepler’s area
law), Newton went on, in corollaries 2 and 3 of Proposition 3, to show that the
centripetal force is directed to the center of force quam proxime iff Kepler’s area
law is satisfied quam proxime. More generally, the theoretical equivalences that
Newton proved in Book I hold quam proxime too: p quam proxime iff q qua
proxime. Hence, given the laws of motion, Newton’s law of gravity is true at least
quam proxime (cf. Smith 2002, 156).
Still, Newton also showed the conditions under which the theoretical
equivalences would hold, not simply quam proxime, but exactly of the phe-
nomena. As he characteristically put it in Proposition 13 of Book III: “if the
sun were at rest and the remaining planets did not act upon one another, their
orbits would be elliptical, having the sun in their common focus, and they would
describe areas proportional to the times” (1729 [2016], 817–818). It follows,
as Smith put it, that “the true motions would be in exact accord with the phe-
nomena were it not for specific complicating factors” (2002, 157).
This interplay between exact, idealized motions and quam proxime ac-
tual motions renders Newton’s account of explanation substantially different
from Descartes’s. As Ducheyne put it, Newton’s target is not merely to explain
Keplerian motions but rather to show that, given the laws of motion, there is a
unique explanation of Keplerian motions (cf. 2012, 104). Given that Keplerian
motions are idealized motions, Newton’s target is to explain the actual deviations
from these motions.
As Smith has noted, Newton’s theory makes possible the following: On the as-
sumption that the theory is exact (that is on the basis of idealizations embodied
in the theory) there are systematic discrepancies between the predictions of the
theory and the observations. These discrepancies constitute “second-order phe-
nomena” which the theory, in the first instance, should explain by attributing
them, as far as possible, to physical causes. Hence, the theory makes possible
14 Ducheyne (2012, 82) offers a more literal translation as “as most closely as possible” or “utter-
most closely.”
88 Historical Cases for the Debate
15 For a detailed discussion of the method of successive approximations and the differences it
planetary motions and the actual ones call for an explanation in terms of specific
gravitational forces.
This kind of attitude toward the theory is embodied in Newton’s Rule IV.
Two independent fields insure that Newton’s law, even if not absolutely accurate,
represents real conditions with a far-reaching accuracy unmatched by hardly
any other law. In the astronomical domain, this law yields planetary motions
not only to the first approximation (Kepler’s laws); but even to the second ap-
proximation, the deviations of planetary motion due to perturbations by other
planets follow from Newton’s law so accurately that the observed perturbations
led to the prediction of the orbit and relative mass of a hitherto unknown planet
(Neptune). (1903, 86)
90 Historical Cases for the Debate
Indeed, M. Hall proved that the law previously examined by G. Green, which
replaces 1⁄r2 with 1⁄ r2+λ where λ stands for a small number, is sufficient to ex-
plain the anomalous perihelion motion of Mercury, if λ = 16 ⋅10–8. This figure
for λ would also give the right result for the observed anomalous perihelion
motion of Mars, though for Venus and Earth the consequence would be some-
what too large a perihelion motion. (1903, 92)
But this revision led to conflicts with other phenomena, such as the position of
the lunar apogee. Fixing the exponent of the inverse square law in such a way that
it accounted for the various phenomena was proved to be more difficult than
anticipated. An adequate account of the anomalous perihelion of Mercury had to
wait until a new theory emerged: Einstein’s GTR.
In November 18, 1915, Einstein published a paper in which, as he put it, he
found “an important confirmation of this most fundamental theory of relativity,
showing that it explains qualitatively and quantitatively the secular rotation of
the orbit of Mercury (in the sense of the orbital motion itself), which was dis-
covered by Leverrier and which amounts to 45 sec of arc per century” (quoted
in, Kennedy 2012, 170). He was, of course, referring to the GTR, which he had
presented a week earlier, in November 1915, in the same journal.16
Einstein took for granted that Newton’s theory (aided by standard perturba-
tion theory) successfully accounted for 531 arc-seconds per century of the total
575.1 arc-seconds per century precession of the perihelion of Mercury. In fact,
taking seriously the 43.4 arc-seconds per century as the observed discrepancy
that constituted an anomaly for Newton’s theory presupposed that the remaining
531 arc-seconds were accounted for by the gravitational effects on Mercury of all
other planets, as explained by Newton’s theory. This suggests that the evidence
for Newton’s theory was solid enough to allow taking this theory as the bench-
mark for the explanation of the 531 arc-seconds per century.
In fact, Einstein never doubted that he had to recover Newton’s theory of
gravity as a limiting case of his own theory. In his letter to Hilbert (November 18,
16 For a thorough and detailed study of Einstein’s account of the anomalous perihelion of Mercury
1915) he emphasized the role of the Newtonian limit in his formulation of the
theory of relativity. Stressing that he arrived at the field equations first, he noted
that three years ago
it was hard to recognize that these equations form a generalisation, and indeed
a simple and natural generalisation, of Newton’s law. It has just been in the last
few weeks that I succeeded in this (I sent you my first communication), whereas
3 years ago with my friend Grossmann I had already taken into consideration
the only possible generally covariant equations, which have now been shown to
be the correct ones. We had only heavy-heartedly distanced ourselves from it,
because it seemed to me that the physical discussion had shown their incom-
patibility with Newton’s law. (quoted in Renn et al. 2007, 908).
In his 1916 systematic presentation of GTR, Einstein (1916, 145) made clear
that starting from the field equations of gravitation “give us, in combination with
the equations of motion . . . to a first approximation Newton’s law of attraction,
and to a second approximation the explanation of the motion of the perihelion of
the planet Mercury . . . These facts must, in my opinion, be taken as convincing
proof of the correctness of the theory”.
How can this be, if Newton’s law is characteristically false? In other words,
how can recovering Newton’s law of gravity as an approximation be “convincing
proof ” for Einstein’s theory, if Newton’s theory has had no substantial truth-
content that could (and should) be recovered by (and retained in) Einstein’s
theory? What’s really interesting in Einstein’s case is that Einstein’s chief con-
ceptual innovation, the curvature of spacetime, provided evidence for Newton’s
account of gravity being in some sense a correct approximation to the actual grav-
itational phenomena. Conversely, as Smith has insisted, Newton’s law of gravity
provided evidence for Einstein’s own theory (and hence Einstein’s own account
of gravity) being more correct than Newton’s. This is because Einstein’s ac-
count identified “non-Newtonian sources of discrepancies, like the curvature of
space,” which provide “evidence for the previously identified Newtonian sources
presupposed by the discrepancies” (2014, 328). In other words, the evidence for
Newton’s account of gravity was solid enough to allow taking this account as the
benchmark for the explanation of the 531 arc-seconds per century.
In sum, some key theoretical components of Newton’s theory of gravity—the
law of attraction and the claim that the gravitational effects from the planets on
each other were a significant cause of the deviations from their predicted orbits—
were taken by Einstein to be broadly correct and explanatory (of at least part) of
the phenomena to be explained.
An objection to this retentionist argument could be that there is indeed reten-
tion in the transition from Newton’s theory of gravity to Einstein’s, but this is only
92 Historical Cases for the Debate
Before some conclusions are drawn, let me discuss the connection of the strategy
followed in the foregoing double case study with the move I introduced many
years ago (cf. 1996) and called the divide et impera strategy to address PI. I al-
ready noted in the introduction that the “moving from the evidence of history to
the history of evidence” strategy is, to paraphrase Clausewitz, the continuation of
divide et impera by other means. To see this let us recall the core motivation and
the core idea behind the divide-and-conquer strategy.
The key motivation was to refute PI by showing that the pessimistic con-
clusion that current theories are likely to be false because most past theo-
ries were false was too crude to capture the dynamics of theory change in
science. The problem with the conclusion of the PI is not that it is false. As
noted already, falsity is cheap; current theories are indeed likely to be false,
strictly speaking, if only because they employ idealizations and abstractions.
The problem with the conclusion of PI is that it obscures important senses in
which current theories are approximately true (better put: they have truth-
like components). More specifically, the conclusion of PI obscures the sub-
stantial continuity there is in theory-change. Now, the way I read PI was as
a reductio of the realist thesis that currently successful theories are approx-
imately true. PI, I argued, capitalizes on the claim that current theories are
inconsistent with past ones. So, according to PI, if we were to assume, for the
sake of the argument, that current theories are true, we would have to assume
that past theories were false. But given that those false past theories were
empirically successful (recall Larry Laudan’s list of past successful-yet-false
theories), truth cannot be the best explanation of the empirical success of
theories, as realists assume.
The key idea behind the divide et impera gambit was that the premise of PI
that needed rebutting was that past theories cannot be deemed truth-like be-
cause they are inconsistent with current theories. And the novelty of divide et
impera, if I may say so, was that it refocused the debate on individual past the-
ories and their specific empirical successes. If we focus on specific past theory-
led successes (e.g., the empirical successes of the caloric theory of heat, or of
Newton’s theory of gravity), the following questions suggest themselves: How
were these particular successes brought about? In particular, which theoretical
constituents of the theory essentially contributed to them? An important as-
sumption of the divide et impera strategy was that it is not, generally, the case that
no theoretical constituents contribute to a theory’s successes. By the same token,
it is not, generally, the case that all theoretical constituents contribute (or, con-
tribute equally) to the empirical successes of a theory. Theoretical constituents
that essentially contribute to successes are those that have an indispensable role
94 Historical Cases for the Debate
in their generation. They are those which really fuel the derivation of the predic-
tion. Hence, the theoretical task was two-fold:
If both aspects of this two-fold task were fulfilled, the conclusion would be that
the fact that our current best theories may well be replaced by others does not,
necessarily, undermine scientific realism. Clearly, we cannot get at the truth all at
once. But this does not imply that, as science grows, there is no accumulation of
significant truths about the deep structure of the world. The lesson of the divide
et impera strategy was that judgements from empirical support to approximate
truth should be more refined and cautious in that they should only commit us
to the theoretical constituents that do enjoy evidential support and contribute
to the empirical successes of the theory. Moreover, it was argued that these re-
fined judgements are made by the historical protagonists of the various cases.
Scientists themselves tend to identify the constituents that they think responsible
for the successes of their theories, and this is reflected in their attitude toward
their own theories.
I recently reviewed the debate that the divide et impera strategy has generated
(cf. Psillos 2018a). Hence, there is no point in repeating it here too. But given
the outline just presented, it transpires that the key thought behind the divide
et impera strategy was precisely that the evidence from the history of science
was too crude, undifferentiated, and misleading to indict realism. Besides, this
evidence portrayed theory-change in science as an all or nothing approach. The
divide et impera move had in it the thesis that the proper target was looking at
the history of evidential relations between the various successes of the theory
and the parts of the theory that fueled them. In this sense, the “moving from the
evidence of history to the history of evidence” strategy is a continuation of the
divide et impera strategy.
But what are the “other means” suggested by paraphrasing Clausewitz? These
are suggested by the dynamics of the theory-change we have discussed in the
double case study. What makes Newton’s strategy distinctive was a) the idea that
a law (or a theory, for that matter) holds quam proxime; and b) the idea that the
various discrepancies between the predictions of the theory and the observed
facts can become the source of a strategy for the development of theory and for
testing it more rigorously, until such time that there is reason to abandon the
theory. Hence, as for instance Harper and Smith (1995) have repeatedly claimed,
Newton significantly altered the criteria of rigorous testing of a theory. Now, a
Descartes, Newton, and Beyond 95
5.6 Conclusions
In the preface to the Treatise on Light (1690), Huygens summed up the Cartesian
approach to hypotheses as follows:
One finds in this subject a kind of demonstration which does not carry with it
so high a degree of certainty as that employed in geometry; and which differs
distinctly from the method employed by geometers in that they prove their
propositions by well-established and incontrovertible principles, while here
principles are tested by the inferences which are derivable from them. The na-
ture of the subject permits of no other treatment. It is possible, however, in this
way to establish a probability which is little short of certainty. This is the case
when the consequences of the assumed principles are in perfect accord with the
observed phenomena, and especially when these verifications are numerous;
but above all when one employs the hypothesis to predict new phenomena and
finds his expectations realized. (quoted in Harper 2011, 373)
We have seen that Newton replaced this conception of method with a more elab-
orate account of explanation, which, without neglecting the role of theoretical
virtues, took it that the explanation offered by the theory should be taken to be
quam proxime and should be further put to the test, while the theory is developed
96 Historical Cases for the Debate
Acknowledgments
Many thanks to Timothy Lyons and Peter Vickers for their encouragement and
useful comments. I am also grateful to an anonymous reader for insightful criti-
cism. Earlier versions of this paper were presented at the History of Science and
Contemporary Scientific Realism Conference, IUPUI, Indianapolis, in February
2016; the philosophy seminar of the University of Hannover, in May 2016; and
at the inaugural conference of the East European Network for Philosophy of
Science (EENPS), Sofia, June 2016. I would like to thank all those who asked
questions and made comments on these occasions. This paper would not have
been written without the generous intellectual help of Robert DiSalle, William
Harper, and Stavros Ioannidis.
References
Aiton, E. J. 1957. “The Vortex Theory of the Planetary Motions—I.” Annals of Science
13: 249–264.
19 For an elaborate account of how Descartes’s and Newton’s theories differ in the ways they repre-
Aiton, E. J. 1958a. “The Vortex Theory of the Planetary Motions—II.” Annals of Science
14: 132–147.
Aiton, E. J. 1958b. “The Vortex Theory of the Planetary Motions—III.” Annals of Science
14: 157–172.
Aiton, E. J. 1962. “The Celestial Mechanics of Leibniz in the Light of Newtonian Criticism.”
Annals of Science 18: 31–41.
Aiton, E. J. 1996. “The Vortex Theory in Competition with Newtonian Celestial Dynamics.”
In R. Taton and C. Wilson (eds.) The General History of Astronomy: Planetary
Astronomy from the Renaissance to the Rise of Astrophysics; vol 2B, The Eighteenth and
Nineteenth Centuries. Cambridge: Cambridge University Press, pp. 3–21.
Bernoulli. J. 1730. Nouvelles Pensees vers le Systeme de M. Descartes, et la Maniere d’en
Deduire les Orbites et les Aphelies des Planetes. Paris: Rue S. Jacques.
Brading, K. 2012. “Newton’s Law-Constitutive Approach to Bodies: A Response to
Descartes.” In A. Janiak and E. Schliesser (eds.) Interpreting Newton: Critical Essays.
Cambridge: Cambridge University Press, pp. 13–32.
Bussotti, P. 2015. The Complex Itinerary of Leibniz’s Planetary Theory. Dordrecht: Springer.
De Pierris, G. 2015. Ideas, Evidence, and Method: Hume’s Skepticism and Naturalism con-
cerning Knowledge and Causation. Oxford: Oxford University Press.
Descartes, R. 1982/1644. Principles of Philosophy. Trans. Valentine Rodger Miller and
Reese P. Miller. Dordrecht: D. Reidel Publishing Company.
Domski, M. 2013. “Mediating between Past and Present: Descartes, Newton, and
Contemporary Structural Realism.” In Mogens Laerke, Justin E. H. Smith, and Eric
Schliesser (eds.) Philosophy and Its History: Aims and Methods in the Study of Early
Modern of Philosophy. Oxford: Oxford University Press, pp. 278–300.
Ducheyne, S. 2012. The Main Business of Natural Philosophy: Isaac Newton's Natural-
Philosophical Methodology. Dordrecht: Springer.
Earman, J., & Janssen M. 1993. “Einstein’s Explanation of the Motion of Mercury’s
Perihelion.” In John Earman, Michel Janssen, and John D. Norton (eds.) The Attraction
of Gravitation: New Studies in the History of General Relativity. Boston: Birkhaeuser, pp.
129–172.
Einstein, A. 1916. “The Foundation of the General Theory of Relativity” reprinted in
A. Einstein The Principle of Relativity. Dover 1932.
Garber, D. 1992. Descartes’ Metaphysical Physics. Chicago: University of Chicago Press.
Harper, W. 2011. Isaac Newton’s Scientific Methodology. New York: Oxford
University Press.
Harper, W. & Smith, G. E.1995. “Newton’s New Way of Inquiry.” In Jarrett Leplin (ed.) The
Creation of Ideas in Physics. Dordrecht: Kluwer, pp. 113–166.
Hattab, H. 2000. “The Problem of Secondary Causation in Descartes: A Response to
DesChene.” Perspectives on Science 8: 93–118.
Herival, J. 1965. The Background to Newton’s Principia. Oxford: Clarendon Press.
Huygens, C. 1690/ 1997. Discourse on the Cause of Gravity. Karen Bailey (trans.).
Mimeograph.
Iltis, C. 1973. “The Decline of Cartesianism in Mechanics: The Leibnizian-Cartesian
Debates.” Isis 64: 356–373.
Jalobeanu, D. 2013. “The Nature of Body.” In Peter R. Anstey (ed.) The Oxford Handbook
of British Philosophy in the Seventeenth Century. Oxford: Oxford University Press, pp.
213–239.
Janiak, A. 2008. Newton as a Philosopher. Cambridge: Cambridge University Press.
98 Historical Cases for the Debate
6.1 Introduction
Let me begin by saying how grateful I am to Peter Vickers and Tim Lyons for
having invited me to two of their conferences.1 I can honestly say that I have
seldom participated in such interesting interdisciplinary settings and from
which I learned so much. The following article is based on a lecture that I gave
at one of these meetings on the work of the English mathematical physicist John
Nicholson. This particular episode in the history of physics has received little at-
tention in the scientific realism debate even though Nicholson’s theory had con-
siderable explanatory and predictive success. What makes the case all the more
remarkable is that by most standards almost everything that Nicholson proposed
was overturned.
My study is most closely connected with the research interests of Peter Vickers
who has published on the subject of inconsistent theories.2 To quote from the
brief directive that was given to speakers by the conference organizers: “The
following is among the key historical questions to be asked: To what extent did
theoretical constituents that are now rejected lead to significant predictive or
explanatory successes?” Many authors, some of whom are represented in this
1 I presented papers at the History of Chemistry and Scientific Realism meeting held in
Indianapolis between December 6 and 7, 2014, as well as the meeting Contemporary Scientific
Realism and the Challenge from the History of Science that also took place in Indianapolis.
2 P. Vickers, Understanding Inconsistent Science, Oxford, Oxford University Press, 2013.
Eric R. Scerri, How Was Nicholson’s Proto-Element Theory Able to Yield Explanatory as well as Predictive Success?
In: Contemporary Scientific Realism. Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford
University Press 2021. DOI: 10.1093/oso/9780190946814.003.0006
100 Historical Cases for the Debate
volume, appear to be somewhat puzzled by the fact that some inconsistent the-
ories were able to attract a good deal of success. However, in the approach that
I will be proposing, such features are seen to arise in a far more natural manner.
I will consider the work of Peter Vickers on quantum theory and quantum
mechanics as an example of the current research that has focused on the sci-
entific realism debate and the challenge from the history of science. In his
recent book Vickers considers two well-trodden examples from the history
of quantum theory. The first one consists of Bohr’s calculation of the spec-
trum of He+, a feat that brought admiration from no less a person than Albert
Einstein when it was first published. Several authors have claimed that Bohr’s
theory was in several respects inconsistent to the point of containing in-
ternal contradictions.3 And yet Bohr succeeded not only in calculating ex-
actly the energy of the hydrogen atom but in also assigning the electronic
configurations of many atoms in the periodic table. Even more dramatically,
Bohr gave a highly accurate calculation for the energy of the helium +1 ion.
The second episode began when Fowler criticized Bohr’s initially published
theory because it predicted an energy of precisely four times that of the hy-
drogen atom, or 4 Rydbergs, when in fact experiments revealed the energy of
He+ to be 4.00163 Rydbergs. In response, Bohr pointed out that he would take
account of the reduced mass of the electron in order to carry out a more accu-
rate treatment of the problem.4 On doing so he obtained an energy of 4.00160
Rydbergs, which is what led Einstein to say, “This is an enormous achieve-
ment. The theory of Bohr must then be right.”5
Vickers recounts a similar story concerning Arnold Sommerfeld.6 It emerges
that precisely the same formula is featured in Sommerfeld’s semi-classical
atomic theory as it is in the fully quantum mechanical theory that came later.
This equivalence occurs in spite of the fact that the existence of electron spin
and the wave nature of electrons, which play a prominent role in the fully
quantum mechanical theory, were completely unknown when Sommerfeld
published his account. Many commentators wonder how Sommerfeld could
have arrived at the correct formula without knowing about these aspects of
the more mature quantum mechanics, as opposed to the old quantum theory
within which he worked. For example, while Sommerfeld assumed definite
3 T. Bartelborth, “Is Bohr's Model of the Atom Inconsistent?” in P Weingartner and G. Schurz,
eds., Philosophy of the Natural Sciences, Proceedings of the 13th International Wittgenstein Symposium
: H, 1989, pp. 220–223.
4 Bohr’s revised calculation essentially consisted of considering the reduced mass of the hydrogen
atom rather than assuming that the nucleus was infinitely heavier than the electron.
5 http:// g alileo.phys.virginia.edu/ c lasses/ 2 52/ B ohr_ t o_ Waves/ B ohr_ t o_ Waves.
html#Mysterious%20Spectral%20Lines.
6 Vickers’s more recent views on the Sommerfeld question appear in P. Vickers, Disarming the
Ultimate Historical Challenge to Scientific Realism, British Journal for the Philosophy of Science, 2018.
Nicholson’s Proto-Element Theory 101
trajectories or orbits for electrons, the quantum mechanical account denies the
existence of orbits and prefers to speak of orbitals in which definite particle tra-
jectories are abandoned.
The response from somebody wishing to maintain a form of realism might
be to claim that something about Sommerfeld’s theory may have been correct,
and it is that part of the theory which accounts for the fact that the two math-
ematical formulas in question were identical. Those wanting to defend such a
position might claim that the core ideas were correct and were responsible for
Sommerfeld success, in spite of his lack of knowledge of electron spin, the wave
nature of electrons, and any other such later developments. Such a preservative
realist would presumably claim that some aspect of Sommerfeld’s theory was
latching on to “the truth.” There has been much discussion of this issue in the
philosophy of science literature and especially in connection with the realism
and anti-realism debate. Realists have tended to appeal to the slogan of divide
and conquer (divide et impera) that was first proposed, as far as I am aware, by
my old friend from graduate school in London, Stathis Psillos. Such an approach,
although perhaps not precisely under the same banner, has also been promoted
by Philip Kitcher and Larry Laudan.7 Meanwhile Carrier, Chang, Chakravartty,
and Lyons among others have expressed strong reservations about this strategy.8
It should be noted that these debates have usually been waged over what were
once very successful theories and scientific entities such as caloric, phlogiston,
or the ether.
As I will endeavor to show, it becomes more difficult to argue for some
form of “preservative realism” in cases like the one I am about to present,
concerning the mathematical physicist John Nicholson who was a contem-
porary of Niels Bohr at the time of the old quantum theory. In fact I shall
be arguing that it is rather difficult to find anything about Nicholson’s view
that has been preserved, except perhaps for the notion of the quantization of
angular momentum, which ironically did not play a prominent role in any
of the success that Nicholson’s theories achieved at the time when they were
first presented. What the defenders of preservative realism recommend in
7 P. Kitcher, The advancement of science, Science without legend, objectivity without illusions,
Oxford, Oxford University Press, 1993; L. Laudan, “A Confutation of Convergent Realism, Philosophy
of Science, 48, 19–48, (1981).
8 M. Carrier, “Experimental Success and the Revelation of Reality: The Miracle Argument
or Scientific Realism.” In Knowledge and the World: Challenges beyond the Science Wars, ed.
Martin Carrier, Johannes Roggenhofer, Günter Küppers, and Philippe Blanchard, 137– 61.
Berlin: Springer (2004); A. Chakravartty, A Metaphysics for Scientific Realism: Knowing the
Unobservable. Cambridge: Cambridge University Press (2007); H. Chang, “Preservative Realism
and Its Discontents: Revisiting Caloric.” Philosophy of Science 70, (5), 902–12 (2003); T.D. Lyons,
“Scientific Realism and the Pessimistic Meta-modus Tollens.” In Recent Themes in the Philosophy
of Science: Scientific Realism and Commonsense, ed. Steve Clarke and Timothy D. Lyons, 63–90.
Dordrecht: Kluwer, (2002).
102 Historical Cases for the Debate
the case of theories that featured the caloric or phlogiston, is that some parts
of the theories tracked the truth while others did not. However, this strategy
cannot be deployed in the case of Nicholson. Another strategy, that has been
recommended by John Worrall, has been the notion of structural realism.
Briefly put, this is the view that although the entities postulated by succes-
sive theories might change as history unfolds, the underlying mathematical
structure is seen to persist and to display continuity. I believe that this form
of strategy too will also fail in the case that I will be examining, all of which
leads me to say that my historical case has a good deal to contribute to the
general debate addressed in the present volume. I will now give a review of
the scientific work in atomic physics of John Nicholson before turning back
to the wider questions as to what one should make of the question of realism
and the development of scientific theories.
9 The mathematical tripos was a distinctive written examination of undergraduate students of the
University of Cambridge, consisting of a series of examination papers taken over a period of eight
days in Nicholson’s time. The examinations survive to this day although they have been reformed
in various ways. A. Warwick, Masters of Theory: Cambridge and the Rise of Mathematical Physics.
Chicago: The University of Chicago Press, 2003.
10 E.R. Scerri, The Periodic Table, Its Story and Its Significance, New York, Oxford University
Press, 2020.
Nicholson’s Proto-Element Theory 103
and Lockyer, Nicholson was an early proponent of the study of spectra for gaining
a deeper understanding of the physics of stars as well as the nature of terrestrial
elements.
The particular details of Nicholson’s proto-atoms were entirely original and
are represented in the form of a table (Figure 6.2). The first feature to notice is
a conspicuous absence of any one-electron atom.11 This is because Nicholson
believed that such a system would be unstable according to his electromagnetic
analysis.12
For Nicholson, the identity of any particular atom was governed by the
number of positive charges in the nucleus, regardless of the number of orbiting
electrons present in the atom. Nicholson may thus be said to have anticipated
the notion of atomic number that was later elaborated by van den Broek and
Moseley. Nicholson argued that a one-electron system could not be stable since
he believed this would produce a resultant acceleration towards the nucleus.
By contrast, Nicholson assumed that two or more electrons adopted equidis-
tant positions along a ring so that the vector sum of the central accelerations of
the orbiting electrons would be zero. The smallest atom therefore had to have at
least two electrons in a single ring around a doubly positive nucleus.13
By using his proto-atoms, Nicholson set himself the onerous task of calcu-
lating the atomic weights of all the elements and of explaining the unidenti-
fied spectral lines in some astronomical objects such as the Orion nebula and
the solar corona. It is one of the distinctive features of Nicholson’s work that
11 Nicholson rejected a one-electron atom because he believed that at least one more electron
was needed to balance the central acceleration of a lone electron. See p. 163 of McCormmach, “The
Atomic Theory of John William Nicholson,” Archives for History of Exact Sciences, 3, 160–184, 1966,
for a fuller account.
12 Nicholson’s list of proto-elements was extended to include two further members in 1914 when
he added proto-hydrogen with a single electron and archonium with six orbiting electrons.
13 By the year 1914, he had accepted the possibility of a one-electron atom. See H. Kragh, “Resisting
his interests ranged across physics, chemistry, and astrophysics and that he
placed great emphasis on astrophysical data above all other data forms.
14 Some of the proto-elements postulated by Nicolson had been invoked earlier by other authors.
For example, Mendeleev had predicted the existence of an element called coronium and Emmerson
had featured proto-fluorine in one of his periodic tables. B.K. Emmerson, “Helix Chemica, A Study
of the Periodic Elements,” American Chemical Journal, 45, 160–210, (1911).
15 It turns out that the name “plum pudding” was never used by Thomson nor any of his con-
temporaries. A.A. Martinez, Science Secrets, Pittsburgh, PA, University of Pittsburgh Press, 2011.
Because of the currency of the term I will continue to refer to it as such.
106 Historical Cases for the Debate
In order to see precisely how Nicholson envisaged his atom we consider his
expression for the mass of an atom, which he published between 1910 and 1911
in a series of articles16 on a theory of electrons in metals.17
2 e
2
m= 2
3 rc
In this expression m is the mass of an atom, e the charge on the nucleus, r the
radius of the electron’s orbit, and c the velocity of light. This expression can be
simplified to read
m ∝ e2 /r (i)
given the constancy of the velocity of light. Nicholson also assumed the positive
charge, ne, for any particular nucleus with n electrons would be proportional to
the volume of the atom.
ne ∝ V
m ∝ n2 e2 /r (ii)
ne ∝ V
or ne ∝ r 3 (since V ∝ r 3 ),
Substituting into (i) the nuclear mass would take the form
n2
m∝
n1/3
or
m ∝ n5/3
16 Nicholson’s theory of metals appears in, J.W. Nicholson, “On the Number of Electrons
Figure 6.4 Nicholson’s calculations and observed weights of the noble gases.
Note: Slightly modified table based on a report of Nicholson’s presentation.
Source: Nature, 87, 2189-501-501, 1911.
At this point Nicholson assigned the mass of 1.008 to his proto-atom of hy-
drogen,18 which allowed him to estimate the relative masses of the other proto-
atoms as shown in Figure 6.3.
From here Nicholson combined different numbers of these three particles
(omitting Cn) to try to obtain the weights of the known terrestrial elements
(Figure 6.4). For example, terrestrial helium could be expressed as,
He = Nu + Pf = 3.9896,
a value that compares very well with the weight of helium that was known at the
time, namely 3.99.19
18 This step seems a little odd given Nicholson’s statements to the effect that hydrogen the proto-
atom is not necessarily the same as terrestrial hydrogen. In using a mass of 1.008 he surely seems to be
equating the two.
19 The error amounts to approximately 0.3 of one percent. Moreover, Nicholson takes account of
the much smaller weight of electrons in his atoms. After making a correction for this effect he revises
the weight of helium to 3.9881 (or, to three significant figures, 3.99) in apparent perfect agreement
with the experimental value. Such was the staggering early success of Nicholson’s calculations. See
J.W. Nicholson, “A Structural Theory of the Chemical Elements,” Philosophical Magazine series 6, 22,
864–889, 871–872, (1911).
108 Historical Cases for the Debate
Figure 6.5 Nicholson’s composite atoms for the first 11 elements in the
periodic table.
Nicholson’s calculations of atomic weights were not confined to just the first
few elements as shown in the Figure 6.3. He was able to extend his accommo-
dation to all the elements up to and including the heaviest known at the time,
namely uranium, and to a very high degree of accuracy. For example, Figure
6.4 shows his calculations as well as the observed atomic weights for the noble
gases.20 Meanwhile Figure 6.5 shows the calculated and observed weights for the
first eleven elements in the periodic table.
Nicholson’s contemporaries initially reacted to his work in a positive but cau-
tious manner. For example, one commentator wrote,
The coincidence between the calculated and observed values is great, but the
general attitude of those present seemed to be one of judicial pause pending
the fuller presentation of the paper, stress being laid on the fact that any true
scheme must ultimately give a satisfactory account of the spectra.21
Nicholson promptly rose to this challenge and provided precisely such an ac-
count of the spectra of some astronomical bodies in his next publication.
This contribution involved the hypothetical proto-element nebulium, which
Nicholson took to have four electrons orbiting on a single ring around a cen-
tral positive nucleus with four positive charges. Like his other proto-elements,
20 Figure 6.4 did not appear in Nicholson’s own papers but in a 1911 article in Nature magazine as
part of a report on the annual conference of the British Association for the Advancement of Science
meeting at which Nicholson had presented some of his findings.
21 Anonymous, Nature, 501, (October 12, 1911).
Nicholson’s Proto-Element Theory 109
Nicholson did not believe that this element existed on the earth but only in the
nebulae that had long ago been discovered by astronomers, such as the one in the
constellation of Orion. Following a series of intricate mathematical arguments
building on Thomson’s model of the atom, Nicholson found that he could ac-
count for many of the lines in the nebular spectrum that had not been explained
by others who had only invoked lines associated with terrestrial hydrogen or
helium.
Nicholson’s feat could well be regarded as a numerological trick, given that it is
always possible to explain a set of known data points given enough doctoring of
any theory. In fact when it was first publicly proposed at a meeting of the British
Society for the Advancement of Science the reaction was indeed one of further
caution. A report that appeared in the magazine Nature stated that,
Scientists usually demand that a good theory should also make successful
predictions so as to avoid any suspicion that a theory may have been deliber-
ately rigged in order to agree with the experimental data.23 Surprisingly enough,
Nicholson’s theory was also able to make some genuine predictions. In addition
to providing a quantitative accommodation of many spectral lines that had not
previously been identified, Nicholson predicted several experimental facts that
were confirmed soon afterward.
Nicholson assumed that each spectral frequency could be identified with
the frequency of vibration of an electron in any particular ring of electrons.
Furthermore, he believed that these vibrations took place in a direction that
was perpendicular to the direction of circulation of the electrons around the
nucleus (Figure 6.6). The model that was eventually developed by Bohr in 1913
differed fundamentally, in that spectral frequencies are regarded as resulting
scientist may be important in deciding who prizes should be awarded to, but
does not matter in the broader question of how the scientific community gains a
better understanding of the world.
In this section the manner in which Nicholson was able to assign many unknown
lines in the spectrum of the Orion nebula will be examined. First I present a figure
containing the spectral lines that had been accounted for in terms of terrestrial hy-
drogen and helium (Figure 6.7). The dotted lines signify the spectral lines that
had not yet been assigned, or identified, in any way. This situation therefore pro-
vided Nicholson with another opportunity to test his theory of proto-atoms and
proto-elements.
As in many other features of Nicholson’s work his approach was rather simple.
He began by assuming that ratios of spectral frequencies correspond to ratios of
mechanical frequencies among the postulated electron motion.24 In mathemat-
ical terms he assumed
24 The detailed calculations can be found in Nicholson’s articles, J.W. Nicholson, “The Spectrum
of Nebulium,” Monthly Notices of the Royal Astronomical Society, (London), 72, 49–64, (1911); “A
Structural Theory of the Chemical Elements,” Philosophical Magazine, 6, (22), 864–89, (1911); “The
Constitution of the Solar Corona I, Protofluorine,” Monthly Notices of the Royal Astronomical Society
(London), 72, 139–50, (1911); “The Constitution of the Ring Nebula in Lyra,” Monthly Notices of the
112 Historical Cases for the Debate
where the f values emerged from his calculations, while one of the ν values
was obtained empirically from the spectra in question and the other one was
‘predicted’.
In his 1911 article “The Spectrum of Nebulium,” Nicholson also predicted the
existence of a new spectral line for the nebulae in question:
Now the case of k = −2 for the neutral atom has been seen to lead to another
line which will probably be very weak. Its wavelength should be 5006.9 ×
.86939 = 4352.9. It does not appear in Wright’s table.25
Remarkably enough this prediction was also soon confirmed. In a short note
in the same journal in the following year, 1912, Nicholson was able to report
that the spectral line had been found at a wavelength of 4353.3 Ångstroms
Royal Astronomical Society (London), 72, 176–77, (1912); “The Constitution of the Solar Corona II,
Protofluorine,” Monthly Notices of the Royal Astronomical Society, (London), 72, 1677–692, (1912);
“The Constitution of the Solar Corona III,” Monthly Notices of the Royal Astronomical Society
(London), 72, 729–39, (1912).
25 J.W. Nicholson, “The Spectrum of Nebulium,” Monthly Notices of the Royal Astronomical Society,
At the meeting of the Society of 1912 March the writer announced the dis-
covery of the new nebular line at λ4353 which had been predicted in his paper
on “The Spectrum of Nebulium.” A plate of the spectrum of the Orion nebula,
on which the line was found, and which had been taken with a long exposure
at the Lick Observatory in 1908 by Dr. W.H. Wright, was also exhibited. In the
meantime the line has been recorded again by Dr. Max Wolf, of Heidelberg,
who has, in a letter, given an account of its discovery, and this brief note gives a
record of some of the details of the observation.
The plate on which the line is shown was exposed at Heidelberg between
1912 January 20 and February 28, with an exposure of 40h 48m. The most
northern star of the Trapezium is in the center of the photographed region, and
the new line is visible fairly strongly, especially in the spectrum of the star and
on both sides.
The wave-length in the Orion nebula, obtained by plotting from an iron
curve, is 4353.9, which is of course, too large, as all the lines in this nebula are
shifted to greater wave-lengths, on account of the motion of the nebula. But the
correction is not so large as a tenth-metre.
The wave-length of the line on the Lick plate, as measured at the Cambridge
Observatory by Mr. Stratton, is 4353.3, the value calculated in the paper being
4352.9.26
26 J.W., Nicholson, “On the New Nebular Line at λ4353,” Monthly Notices of the Royal Astronomical
Nicholson next turned his attention to the spectrum of the solar corona
that had been much studied and that showed numerous lines that had
not yet been accounted for (Figure 6.9). In this study Nicholson was even
more successful than he had been with the spectrum of the Orion nebula,
Figure 6.10 Nicholson’s accommodation of 14 of the lines from Figure 6.9 using
proto-fluorine and ionized forms of this proto-atom.
the case of the proto-fluorine atom, Nicholson calculated the ratio of potential
energy to frequency to be approximately
In arriving at his result Nicholson had used the measured values of e and m,
the charge and mass of the electron. However, his method still did not provide
a means of estimating the radius of the electron, and he was forced to eliminate
this quantity from his equations, a problem that he would overcome a little later.
Nicholson proceeded to calculate the ratio of potential energy to frequency
in proto-fluorine ions with one or two fewer electrons and found 22 h and 18 h,
respectively. He noted that the three values for Pf, Pf+1, and Pf+2 were members of
a harmonic sequence,
By dividing each value by the number of electrons in the atom he found the
Planck units of angular momentum per electron to be,
Nicholson thus arrived at the general formula for the angular momentum of a
ring of n electrons as
1
2 (15 − n) n
This formula then allowed him to fix the values of the atomic radius in each case,
and since angular momentum did not change gradually, he took this to mean
that atomic radius would also be quantized.
Several authors have traced the manner in which Bohr picked up this hint
of quantizing angular momentum.27 This feature was not present in Bohr’s
initial atomic model, and he only incorporated it over a series of steps fol-
lowing a close study of Nicholson’s papers. Bohr also spent a good deal of
time trying to establish the connection between his own and Nicholson’s
atomic theory.
27 J. Heilbron, T.S. Kuhn, “The Genesis of Bohr’s Atom,” Historical Studies in the Physical Sciences,
1, 211–90, (1969).
Nicholson’s Proto-Element Theory 117
Nicholson’s calculated frequencies and the observed lines were “so close
and so numerous as to leave little doubt of the general correctness of the
theory . . . Nicholson’s theory stands alone as a first satisfactory theory of one
type of spectra.”28
Physics historian Abraham Pais saw the relationship between Bohr and
Nicholson sometime later in the following terms,
Bohr was not impressed by Nicholson when he met him in Cambridge in 1911
and much later said that most of Nicholson’s work was not very good. Be that as
it may, Bohr had taken note of his ideas on angular momentum, at a crucial mo-
ment for him . . . He also quoted him in his own paper on hydrogen. It is quite
probable that Nicholson’s work influenced him at that time.30
Returning to Heilbron:
The success of Nicholson’s atom bothered Bohr. Both models assumed a nu-
cleus, and both obeyed the quantum; yet Nicholson’s radiated—and with un-
precedented accuracy—while Bohr’s was, so to speak, spectroscopically mute.
28 McCormmach, “The Atomic Theory of John William Nicholson,” Archives for History of Exact
1991, p. 145.
118 Historical Cases for the Debate
By Christmas 1912, Bohr had worked out a compromise: his atoms related to the
ground state, when all the allowed energy had been radiated away; Nicholson’s
dealt with earlier stages in the binding. . . . Just how a Nicholson atom reached
its ground state Bohr never bothered to specify. He aimed merely to establish
the compatibility of the two models. The compromise with Nicholson was to
leave an important legacy to the definitive form of the theory.31
Later in the same paper Bohr proposed other formulations of his quantum
rule, including, with full acknowledgement of Nicholson’s priority, the quanti-
zation of the angular momentum.32
Now for one last commentator, Leon Rosenfeld, who reveals some further
aspects of how Nicholson has been regarded. In his introduction to a book by
Niels Bohr to celebrate the 50th anniversary of the 1913 theory of the hydrogen
atom, Rosenfeld writes:
The ratio of the frequencies of the two first modes happens to coincide with that
of two lines of the nebular spectra: this is enough for Nicholson to see in this
system a model of the neutral “nebulium” atom; and as luck would have it, the
frequency of the third mode, which he could then compute, also coincided with
that of another nebular line, which—to make things more dramatic—was not
known when he made the prediction in his first paper, but was actually found
somewhat later. . . .
From the mathematical point of view Nicholson’s discussion of the stability
conditions for the ring configurations and of their modes of oscillation is an
able and painstaking piece of work; but the way in which he tries to apply the
model . . . must strike one as unfortunate accidents.34
In the third paper, however, published in 1912, occurs the first mention of
Planck’s constant in connection with the angular momentum of the rotating
Century Physics, C. Weiner, ed., 40–108, Academic Press, New York, 1977, p. 69.
32 Ibid., p. 70.
33 M. Jammer, The Conceptual Development of Quantum Mechanics, New York, McGraw- Hill,
1966. p., 73.
34 L. Rosenfeld, in preface to N. Bohr, On the Constitution of Atoms and Molecules, New York, W.E.
electrons: again here there is no question of any physical argument, but just a
further display of numerology. . . .
Bohr did not learn of Nicholson’s investigations, as we shall see, before the end
of 1912, when he had already given his own ideas of atomic structure their fully
developed form.35
By contrast [with Nicholson] the thoroughness of Bohr’s single-handed attack
on the problem and the depth of his conception will appear still more impressive.36
There is clearly no “fence sitting” here to give Nicholson any benefit of the
doubt. Rosenfeld does not even believe that Nicholson’s apparent early success
was due to a cancellation of errors. But perhaps some aspects of Rosenfeld’s
life serve to explain some of his reaction. Rosenfeld was without doubt one of
Bohr’s leading supporters and also acted as a leading spokesperson for Bohr’s
Copenhagen interpretation of quantum mechanics for the last 30 or so years of
Bohr’s life. Rosenfeld is also known to have been an especially vitriolic and harsh
critic in spite of his apparently shy and retiring personality. His fellow Belgian
and one-time collaborator, the physical chemist Ilya Prigogine, described him
as a “paper tiger.”37 It is hardly surprising therefore that Rosenfeld championed
Bohr against any claims from anyone, such as Nicholson, who he regarded as
imposters, or anyone who might try to steal some of the thunder from Bohr.
The views of Rosenfeld can be contrasted this with those of Kragh, the con-
temporary historian and a Dane like Bohr,
But let us say, for the sake of argument, that Rosenfeld is right and that Nicholson’s
work was completely worthless. Even if this were true, I contend that Nicholson’s
publications contributed to Bohr’s developing his own atomic theory for the
simple reason that Nicholson served as his foil. In some places Bohr is quite dis-
missive of Nicholson’s work, such as when writing to his Swedish colleague Carl
Oseen, where he describes Nicholson’s work as “pretty crazy” while adding,39
35 Ibid., p. xiii.
36 Ibid., p. xiii–xiv.
37 Prigogine was born in Russia but emigrated to Belgium.
38 H. Kragh, Niels Bohr and the Quantum Atom: The Bohr Model of Atomic Structure 1913–1925,
gathers the impression that Bohr continued to hold this crucial view about Nicholson. I am grateful
for a reviewer for drawing my attention to this qualification.
120 Historical Cases for the Debate
I have also had discussion with Nicholson: He was extremely kind but with him
I shall hardly be able to agree about very much.40
In another passage from a letter to his brother Harald, Niels Bohr writes
After Bohr had published his three-part article, Nicholson continued to press
him in a number of further publications. If we must speak in terms of winners
and losers, Bohr would be regarded as a winner and Nicholson as a loser. But as
in all walks of life, it is not just about winning, but more about partaking. There
would be no athletic races for spectators to watch if the losers were not even to
40 Bohr to Oseen, December 1, 1911, cited in N. Bohr, L. Rosenfeld, E. Rüdinger, & F. Aaserud,
“Collected Works.” In Early Work (1905–1911), Vol. 1, J.R. Nielsen, ed., North-Holland Publishing,
Amsterdam, 1972, pp. 426–431.
41 N. Bohr to Rutherford, 1913, cited in L. Rosenfeld, preface to N. Bohr, On the Constitution of
In both instances where I have written “my,” Bohr had actually written “his”—which must surely
be typos.
Nicholson’s Proto-Element Theory 121
participate in the race. The very terms winner and loser are necessarily code-
pendent in the context of any scientific debate, as were the roles of Bohr and
Nicholson.
Now this picture that I have painted would seem to raise at least one obvious
objection. If all competing theories are allowed to bloom because there is no such
thing as a right or wrong theory, how would scientists ever know which theo-
ries to utilize and which ones to ignore? Indeed this is the kind of criticism that
was levelled against Feyerabend43 when he claimed that “anything goes” and was
promptly criticized by numerous authors.44 I think the answer to this question
can be found in evolutionary biology. Nature has the means of finding the best
way forward. Just as any physical trait with an evolutionary advantage eventu-
ally takes precedence, so the most productive theory will eventually be adopted
by more and more scientists in a gradual, or trial and error fashion. The theo-
ries that lead to the most progress will be those that garner the largest amount
of experimental support and which provide the most satisfactory explanations
of the facts. This entire process will not be rendered any the weaker even if one
acknowledges an anti-personality and anti–“right or wrong view” of the growth
of science.
More generally, I believe that the two aspects can coexist quite happily.
Scientists can, and regularly do, argue issues out to establish the superiority of
their own views, as well as their claims to priority. But the march of progress, to
use an old-fashioned term, does not care one iota about these human squabbles.
And it is the overall arc of progress that really matters, not whose egos are bruised
or who obtains the greater number of prizes and accolades.
Having examined the apparent successes of Nicholson’s theory we must still ask
how any of this was even possible given what we now know of his ideas. Here is a
brief list of what seems to be problematical in Nicholson’s scheme: First of all, the
proto-elements like nebulium that he postulated do not exist. Second, he identi-
fied mechanical frequencies of electrons with spectral frequencies and assumed
that these oscillations took place at right angles to the direction of electron cir-
culation. Third, Nicholson’s electrons were all in one single ring, unlike the sub-
sequent Bohr model in which they are distributed across different rings or shells.
simple assertion on whether or not the moon is made of blue cheese or anything
quite so specific.
I take it that one would never wish to claim that biological evolution is either
right or wrong, but rather that any biological developments are either suited to
their environment or not. Those developments that are suited to their biological
niches result in their being perpetuated in future generations while those that are
not simply wither away. So, I claim, it is with the development of science.
In passing, I should note some kinship with the views of the influential anti-
realist philosopher Van Fraassen, who supports an evolutionary view of the
progress of science when he writes,
For any scientific theory is born into a life of fierce competition, a jungle red in
tooth and claw. Only the successful theories survive—the ones which in fact
latched onto the actual regularities in nature.47
This view that has been criticized in particular by Kitcher,48 and Stanford,49 al-
though it would take me too far afield to comment on this exchange here.
My own brand of evolutionary epistemology also holds that the study of sci-
entific development should not be approached by concentrating on individual
discoverers nor through individual theories. I claim that science essentially
develops as one unified organism, while fully anticipating that this aspect will
meet the greatest resistance from critics. Needless to say, many people have real-
ized the societal/collective nature of science, and I cannot claim any originality
on that score. For example, there have been many programs such as the Strong
Program, Science Studies, and the Sociology of Science, which all take a more
holistic approach to studying how science develops. However, these programs
generally hold that social factors determine scientific discoveries. My interest lies
primarily in the actual science, while at the same time maintaining a radical form
of sociology that considers scientific society to be a single organism which is es-
sentially developing in a unified fashion.
I propose that scientific research is conducted by a tacit network of scholars,
and researchers, who frequently appear to be at odds with themselves, and fre-
quently are involved in bitter priority disputes, while unknowingly partaking in
the same overall process. For example, in the case of John Nicholson, and several
others that I have discussed in a recent book, the protagonists were seldom in
direct contact with the Bohrs, Paulis and other luminaries in the world of early
47 B. Van Fraassen, The Scientific Image, Oxford, Clarendon Press, 1980, p. 40.
48 P. Kitcher, The Advancement of Science: Science without Legend, Objectivity without Illusions,
New York, Oxford University Press, 1993.
49 K. Stanford, An Antirealist Explanation of the Success of Science, Philosophy of Science, 67, 266–
84, (2000).
124 Historical Cases for the Debate
20th century atomic physics. Nevertheless, one sees an entangled, organic devel-
opment in which ideas compete and collide with each other, even if one famed
individual is eventually associated with any particular scientific discovery or
episode.
I regard science as very much proceeding via modification by trial and error
rather than via “cold rationality.” My view is aligned with biological evolution
rather than with an enlightenment view of the supremacy of rationality and the
powers of pure deduction. I regard scientific development as being guided by
evolutionary forces, blind chance, and random mutation. When seen from a dis-
tance science appears to develop as one unified organism. Under a magnifying
glass there may be little sign of unity, so much so that some philosophers and
modern scholars have been driven to declare the dis-unity of science, a view that
I believe to be deeply mistaken.50
From the perspective that I propose, one can better appreciate how what a
realist regards as wrong theories, as in the case of most of John Nicholson’s scien-
tific output, can lead to progress. I do not need to explain wrong theories away in
my account. Nor do I believe that simultaneous or multiple discoveries deserve
to be explained away as being aberrations. Similarly, priority disputes, which are
so prevalent in science, can be seen as resulting from the denial of the unity of
science and an underlying body-scientific that I allude to.
Developments that might normally be regarded as wrong ideas frequently
produce progress, especially when picked up and modified by other scientists.
Nicholson’s notion of angular momentum quantization, which he developed in
the course of what has generally been regarded as an incorrect theory, was picked
up by Niels Bohr who used it to transform the understanding of atomic physics.
The image of science that I envisage resembles a tapestry of gradually evolving
ideas that undergo mutations and collectively yield progress. In some cases the
source of the mutation is random, just as mutations are random in the biolog-
ical world. This view stands in sharp contrast with the traditional notion of pur-
posive and well thought out ideas on the part of scientists.51
Consequently, the kinds of inconsistent theories that Peter Vickers and others
have examined no longer seem to be in such urgent need of explanation and
should not be regarded as being so puzzling. Indeed, such developments should
50 P. Galison, D. Stump (eds.), The Disunity of Science, Palo Alto, Stanford University Press, 1996.
51 I regard mutations in scientific concepts to occur on the level of individual scientists and
not primarily in entire social groups. My evolutionary epistemology thus also differs from Kuhn’s
in this respect since his unit of evolutionary change is the social group of scientists. See some in-
teresting discussions regarding Kuhn’s evolutionary account of scientific progress in B.A. Renzi,
“Kuhn’s Evolutionary Epistemology and Its Being Undermined by Inadequate Biological Concepts,”
Philosophy of Science, 76, 143–59, (2009); T.A.C. Raydon, P. Hoyningen-Huene, “Discussion: Kuhn’s
Evolutionary Analogy in The Structure of Scientific Revolutions and ‘The Road since Structure,’ ”
Philosophy of Science, 77, 468–76, (2010).
Nicholson’s Proto-Element Theory 125
be regarded as the rule rather than exceptions. I agree completely with Kuhn that
ideas and theories that survive should be regarded as facilitating progress and
not as tracking the truth.
As so many evolutionary epistemologists have urged, I regard knowledge as
being presumptive, partial, hypothetical, and fallible. If scientific progress is far
more organic than usually supposed, we can make better sense of why Nicholson
was able to contribute to the growth of science. None of the scientists I have
examined really knew what they were doing, in a sense.52 Their crude ideas de-
veloped in an evolutionary trial and error manner instead of in a strictly rational
manner, or so I propose.
Nicholson and many people like him contributed very significantly to the de-
velopment of science. Nicholson was not simply wrong, since he inadvertently
helped Bohr to begin to quantize angular momentum. Nicholson is as much part
of the history as Bohr is, and minor contributors matter as much as the well-
known ones. In anticipation of the following section, I believe that the early
Kuhn’s focus on revolutions may have served to diminish the importance of such
marginal figures.
Is an evolutionary view of the kind I propose compatible with the revolutions for
which Kuhn is so well known? First of all, as many authors have claimed, Kuhn’s
own work on historical episodes sometimes points to continuity rather than rev-
olution, such as in the case of quantum theory and the Copernican Revolution.53
Of course, Kuhn also acknowledges the role of normal science that in very
broad terms can be seen as upholding the role of minor historical contributors
to the growth of science. Nevertheless, as I see it, there was no sharp revolution
in the development of quantum theory only an evolution. Paradoxically, such
an evolution is easier to appreciate from the wider perspective of science as one
unified whole than from the perspective of individual contributors or individual
theories. Viewing theory-change as revolutionary, on the other hand, may mask
the essentially biological-like growth of science that I am defending.
Nevertheless, I am in complete agreement with Kuhn in supposing that sci-
ence does not drive toward some external truth. In this respect I am on the side
of anti-realism. I prefer to regard scientific process as driven from within, by
52 Scerri 2016.
53 T.S. Kuhn, Black-Body Radiation and the Quantum Discontinuity, 1894–1912, Chicago, Chicago
University Press, 1987; J. Heilbron, T.S. Kuhn, “The Genesis of the Bohr Atom,” Historical Studies in
the Physical Sciences, 1, 211–290 (1969).
126 Historical Cases for the Debate
evolutionary forces which look back to past science.54 This is not to say that the
world does not constrain our theorizing. Kuhn just wants us to see that the scope
of our theories is not determined by nature in advance of our inquiring about it.
Indeed, I find that I agree with Kuhn on a great number of issues, the main excep-
tion being the occurrence of scientific revolutions.
As scholars generally concur, Kuhn’s later work was aimed toward developing
an evolutionary epistemology.55 As Kuhn developed his epistemology of science,
he saw increasingly more similarities between biological evolution and scientific
change. Consequently as he developed his epistemology of science it became a
more thoroughly evolutionary epistemology of science. Wray makes the very
perceptive remark that while Kuhn was one of the key philosophers of science
who initiated the historical turn in the philosophy of science in the early 1960s,
he later came to adopt what he termed a historical perspective. According to Wray
this historical perspective, or developmental view, is nothing less than an evo-
lutionary perspective on science. In accord with Wray and others, Kuukkanen
(2013, 134) points out that the later Kuhn felt that his evolutionary image of sci-
ence did not get the amount of attention that it deserved. In the course of his last
interview, Kuhn deplored this situation while saying, “I would now argue very
strongly that the Darwinian metaphor at the end of the book [SSR] is right and
should have been taken more seriously than it was.”56
Just as evolution lacks a telos and is not driven towards a set goal in advance,
science is not aiming at a goal set by nature in advance. Kuhn continued to re-
gard his evolutionary view as important to the end of his life, or as Wray writes,
“Whatever else he changed he did not change this aspect.” Kuhn also claims
that in the history of astronomy, the earth-centered models held the field back
for many years. Similarly, he claimed that, truth-centered models of scientific
change were holding back philosophy of science. Somewhat grandiosely, Kuhn
notes a similarity between the reception of Darwin’s theory, and the reception of
his own theory his own view on the evolution of science, in that both meet the
greatest resistance on the point of elimination of teleology.
54 Kuhn prefers the phrase “pushed from behind.” I believe that “driven from within” better
conveys the way that scientific knowledge is being generated in an outward fashion rather than being
“pulled” toward the truth.
55 B. Wray, Kuhn’s Evolutionary Social Epistemology, Cambridge, UK: Cambridge University Press,
2011, 84.
56 K. Jouni-Matti, “Revolution as Evolution: The Concept of Evolution in Kuhn’s Philosophy.”
In Kuhn’s The Structure of Scientific Revolutions Revisited, Theodore Arabatzis, ed., 134–
152, Routledge, New York, 2013, p. 134. See T.S. Kuhn, J. Conant, & J. Haugeland, The Road
Since Structure: Philosophical Essays, 1970- 1993 (with an autobiographical interview), 2000,
Chicago: University of Chicago Press. Kuhn’s first thoughts on epistemology based on evolutionary
lines appear at the end of The Structure of Scientific Revolutions, 1962, Chicago: University of Chicago
Press, 169–172.
Nicholson’s Proto-Element Theory 127
Whereas Kuhn did not initially have an explanation for the success of sci-
ence, he later proposed that scientific specialization was the missing factor. Kuhn
argued that just as biological evolution leads to an increasing variety of species, so
the evolution of science leads to an increasing variety of scientific sub-disciplines
and specializations.57 For Kuhn science is a complex social activity, and the unit
of explanation is the group, not the individual scientists. This resonates with my
own view that I alluded to earlier that the growth of science is not successfully
tracked by considering individual scientists or individual theories. Kuhn asks
us to judge changes in theory from the perspective of the research community
rather than the individual scientists involved. Both early converts to a theory and
holdouts aid the community in making the rational choice between competing
theories. This, I suggest, is how the work of Nicholson should be viewed, namely
as a step toward the new theory in the context of Bohr and others.58
However, there is another respect in which I would want to qualify this view to
also reaffirm the importance of change on the level of individual scientists rather
than change predominantly within social groups. Kuhn’s evolutionary account
of the growth of science appears to be somewhat half-hearted. He does not say
very much on the mechanism of evolutionary change in scientific theories except
for his talk of specialization. Kuhn believes that the development of new sub-
disciplines such as the emergence of biochemistry from chemistry, for example,
is analogous to the evolution of new species which become incapable of mating
with members of the species from which they have evolved. Similarly Kuhn
holds that the members of the new scientific sub-discipline, which has branched
off from an older one, are unable to communicate with members of the mother
discipline. Kuhn even makes a virtue of this reconceptualized incommensura-
bility in suggesting that such isolation allows for a greater development within a
particular sub-discipline.
While authors like Renzi have criticized Kuhn’s evolutionary analogy, Raydon
and Hoyningen-Huene defend him, in suggesting that Kuhn did not intend
his analogy to be taken literally. Be that as it may, I do intend the evolutionary
analogy to be literal. Since the evolution of scientific theories is part of an under-
lying biological evolution of the human species I do not regard this suggestion to
be too far-fetched, even though these evolutionary processes may be occurring
at levels that are far removed from each other.
57 Out of physics and chemistry there emerges physical chemistry. Biology and chemistry give rise
merit in Kuhn’s work. I had been too stuck on the cartoon Kuhn (the best-seller Kuhn) who is sup-
posed to deny progress and who is often taken to be at the root of all evil, such as science wars and the
sociological turn—but only because I arrived at the idea of an evolutionary epistemology through my
own work in asking how a “wrong” theory can be so successful in so many cases.
128 Historical Cases for the Debate
In the case of the evolution of organisms, modern biology has taught us that
the underlying mechanism is one of random mutation on the level of errors that
occur in the copying of DNA sequences. The more accurate biological analogy
for the development of scientific theories would seem to be for the “mutation”
of ideas to occur in the minds of individual scientists such as Bohr or Nicholson.
The unit involved in the evolution of scientific theories would therefore be indi-
vidual scientists and not the social group that scientists belong to. Here then is
where I depart from Kuhn’s evolutionary epistemology. In my account the evolu-
tion of scientific theories takes place in a literally biological sense and is mainly
situated at the level of individuals rather than social groups.
Brad Wray has claimed that revolutions are essential to Kuhn because they
are incompatible with the view that scientific knowledge is cumulative and that
scientists are constantly marching ever closer to the truth. I must say that I dis-
agree with this view. Scientists may not be moving toward a fixed truth, but the
development may still be gradual and not revolutionary. After all, biological ev-
olution is not teleological but is generally thought to be gradual (pace Stephen
Jay Gould et al.). To me the main insight from Kuhn is his evolutionary episte-
mology not his notion of discontinuous theory-change.
For example, why consider the quantum revolution to have ended in 1912 as
Kuhn does? Surely an equally important revolution due to Bohr began around
this time of 1912. Or was the true revolution the coming of quantum mechanics à
la Heisenberg and Schrödinger in 1925–26 or maybe QED sometime later in the
59 A quick reply to this question is that Kuhn gives very few examples of what he considers to be
late 1940s or QCD in the 1970s? Or could it be that there are so many revolutions
that the very concept ceases to be helpful?
6.11 Conclusions
61 I am speaking mainly of popular science and science textbooks. Professional historians of sci-
7.1 Introduction
Amanda J. Nichols and Myron A. Penner, Selective Scientific Realism and Truth-Transfer in Theories of Molecular
Structure In: Contemporary Scientific Realism. Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press.
© Oxford University Press 2021. DOI: 10.1093/oso/9780190946814.003.0007
Selective Scientific Realism and Truth-Transfer 131
structures at the turn of the 20th century which resulted from the Blomstrand-
Jørgensen/Werner Debate about the structure of cobalt complexes. Both the
Blomstrand-Jørgensen chain theory and Werner’s coordination theory, which
ultimately replaced it, make predictions about the number of ions that will be
dissociated when the molecule undergoes a precipitation reaction. While chain
theory makes correct predictions in many cases, Werner’s coordination theory
correctly predicted the number of ions in cases where chain theory failed. If the
predictive success of chain theory depended on features that were later rejected
in coordination theory, we’d have a case which undermines SSR’s explanation of
truth transfer between failed theories and their successors. Conversely, the rise
of coordination theory supports SSR if the features of chain theory that resulted
in its predictive successes were also a part of coordination theory. It turns out the
latter is the case. We conclude by applying the lessons from the debate about the
structure of cobalt complexes to philosophical debate about scientific realism.
In brief, the chain theory utilized by Jørgensen involved three rules for mod-
eling the structure of inorganic compounds. Jørgensen applied this framework
with some experimental success: his models predicted certain outcomes of ion
dissociation which were confirmed in a series of precipitation experiments.
However, there were some experiments in which Jørgensen’s model failed to
accurately predict ion dissociation. During the same time period, Werner
was developing a geometric model for cobalt complexes much different from
Jørgensen’s. It turns out that Werner’s model was able to succeed experimentally
where Jørgensen’s had failed. This led, ultimately, to Blomstrand-Jørgensen chain
theory being replaced by Werner’s coordination theory.
Those who are skeptical about whether experimental success indicates the
truth of scientific theories might look at this historical episode as one which
justifies such skepticism. After all, in retrospect we can see that Jørgensen
achieved a degree of experimental success even though he was mistaken
about the structure of the cobalt complexes in question. Why should we
think that the experimental success of contemporary scientific theories
is in virtue of their truth, when experimental success obviously does not
guarantee truth?
However, we think that for this particular historical episode, SSR provides the
correct historical analysis. It turns out that the factors which led to Jørgensen’s
predictive success are also features of Werner’s coordination theory that ulti-
mately replaced Jørgensen’s chain theory. Jørgensen’s predictive success was
in virtue of him latching on to certain truths about how cobalt complexes are
structured. We further argue that the lessons from this historical case can illumi-
nate how two contemporary objections to realism—P. Kyle Stanford’s Problem
of Unconceived Alternatives and Timothy D. Lyons’ pessimistic modus tollens
argument—fall short as arguments against realism.
132 Historical Cases for the Debate
By the middle of the 19th century, chemists were routinely investigating both
physical and chemical properties of different compounds as well as identifying
ratios between elements in different compounds. Masses for different elements
were well established by this time, which meant that chemists could carefully
weigh samples before and after experiments involving a reaction and isolation
of a compound in order to determine the loss or gain of mass. An active area
of inquiry was molecular structure and trying to determine how atoms are ar-
ranged and connected to each other. Since J. J. Thomson didn’t discover electrons
until 1897, the modern theory of molecular bonds consisting of electrons had
not yet emerged.1 While chemists at this time did not have a theory explaining
how atoms connect, there was consensus that atoms connected in some way to
form molecules. Moreover, chemists were drawing connections from the phys-
ical and chemical properties to structural features of the molecule, primarily in
organic chemistry because of the work being done with carbon.2 It was discov-
ered that the way atoms were connected to one another seemed important, and
the arrangement that atoms have within space—their molecular geometry—
also seemed to affect chemical and physical properties. For example, Jacobus
Henricus van’t Hoff ’s work demonstrated that there are compounds that have
the same chemical formula with the same connectivity between atoms, but their
three-dimensional orientations differ, making the compounds behave differently
in polarized light. These optically active compounds introduced the idea of the
geometry of a molecule.3 Molecules were discovered to be not mere planar enti-
ties, but rather three-dimensional structures where each constituent atom’s ori-
entation in space can change the structure of the whole molecule.
Chemists had an early idea of valence, which can be understood as the af-
finity an atom has to combine with other atoms. Valence is expressed as a quan-
tity reflecting the number of bonds an atom may have. For instance, carbon is
said to have a valence of 4 (tetravalence) because it has four connecting points.
Hydrogen is univalent because it has one connecting point. Therefore, one
carbon atom will combine with four hydrogen atoms forming the compound
methane (CH4) (Figure 7.1). How elements bond to form compounds is a func-
tion of each constituent element’s valence, because not anything can bond to just
anything. Valence helps to predict molecular structure. Drawing on both these
valence patterns and the more developed branch of organic chemistry, Friedrich
August Kekulé predicted that the arrangement of atoms in molecules could be
C H
H
H
H H H H H H H H
H C C C C C C C C H
H H H H H H H H
chain.5 He thought that if a chloride came off the compound easily to react
with silver nitrate (and precipitate silver chloride as a product) it was because
the chloride was directly connected to the end of the ammonia chain, not the
metal.6
Sophus Mads Jørgensen continued the work with metal ammines in the
1890s, specifically with cobalt ammine complexes. A cobalt ammine complex
is a molecule made up of the metal cobalt, ammonia groups, and chlorides.
Jørgensen followed three bonding rules for structures when theorizing the
structure of these cobalt complexes. First, according to the valence rule,
bonding is constrained by an element’s valence. Thus, because the metal co-
balt is trivalent, a cobalt atom will have three connecting points. Second, ac-
cording to the dissociation rule (established from lab experiments involving
precipitation reactions with silver nitrate), chloride dissociation, that is, the
separation of chloride atoms from the metal ammine, will not occur if chloride
is directly bonded to metal. And third, according to the chain rule, the en-
tire structure of the compound would be chain-like, and thus, ammine chains
will form off a cobalt atom. While we are not claiming that Jørgensen explic-
itly formulated these rules in his model for cobalt complexes, it is clear from
his dependence on Blomstrand’s work and his own published research that
Jørgensen followed what we are calling the valence, dissociation, and chain
rules. Much of Jørgensen’s work centered on the compounds that are known
today as coordination complexes, and he was one of the first chemists to
provide experimental evidence that the cobalt ammine complexes were not
dimers, contrary to what Blomstrand hypothesized. Jørgensen’s 1890 paper
outlines this evidence and treats cobalt as trivalent, demonstrating that
Jørgensen is following the valence rule.7 Kauffman points out that in 1887
Jørgensen published his work using the precipitation experiment (discussed
in the next section), reasoning that if the chloride is directly bonded to a
metal atom, the chloride will not dissociate. This is what we have dubbed the
dissociation rule.8 Continuing the ideas originally proposed by Blomstrand,
Jørgensen thought of the ammines to be in a chain structure (the chain
rule), and Werner responds to these ideas directly in his paper proposing
his new coordination theory.9,10 This is relevant to what we later will argue
H Cl
H
N H H H H H H H H
H
Co N N N Cl
N
H
N H H H H
H
Cl
H
Applying these three rules results in the Blomstrand-Jørgensen chain theory of cobalt
ammine complexes, which includes models for structures of different compounds;
the accuracy of these proposed structures can be tested in precipitation experiments.
During a precipitation experiment, cobalt complex molecules split into ions (separated
charged particles). The experiment begins by mixing a cobalt complex solution with a
silver nitrate solution. A solid (the precipitate) forms by chloride separating from the
cobalt complex and bonding with the silver; the resulting precipitate is silver chloride.
Jørgensen’s three bonding rules make predictions with respect to the number of ions
that will result from a precipitate experiment.
An early example from Jørgensen is his postulated chain structure for co-
balt (III) hexammine chloride, further referred to as the hexammine complex.
Jørgensen’s structure of this cobalt complex, following the valence rule, consists of
a cobalt center with three groups connected to it (Figure 7.3). The three chloride
atoms are connected to ammine groups (NH3Cl), but none of the chloride atoms
are attached directly to the cobalt atom. Following the chain rule, the third group
is a chain of ammines with a chloride on the end. The dissociation rule predicted
that there should be four ions produced in the precipitation experiment: all three
chlorides will dissociate because they are each connected to an ammine group, not
the cobalt atom. Three chlorides plus the leftover part of the molecule would result
in four ions. That prediction was supported by measurement when four ions were
confirmed after the hexammine complex underwent a precipitation experiment
(Figure 7.4).
136 Historical Cases for the Debate
H Cl−
H
N H H H H H H H
H
H
Co+3
N N N N Cl−
H
N H H H H
H
H Cl−
cobalt ion, is attracting the negative parts, the chlorides. Ammines are neutral in charge, and their
attachment to the cobalt is not like the primary valence.
Selective Scientific Realism and Truth-Transfer 137
Cl–
H 3+
H H
H H
H N H
H N N
H
H Cl–
Co H
H N
N H
H N
H
H H
H
Cl–
3+
H
H H
H H
H N Cl–
H
H N N
H Cl–
H Co H
H N Cl–
N H
H N
H
H H
H
least one model must be mistaken. But note also, that both models at this stage
are making successful predictions—entailing results that are subsequently con-
firmed by experiment. And it’s this phenomenon—the phenomenon of incorrect
models making correct predictions—that seems to provide some prima facie jus-
tification for skepticism about a realist interpretation of current science. “Why,”
so the thinking goes, “should we be confident that the successful predictions of
current science are a marker of truth, if we have all these examples from history
of the predictive success of failed models?” In fact, one might extend this worry
about the shortcomings of current science to include the potential superiority of
“unconceived alternatives” to current theories.14 We’ll look at specific arguments
Cl H H H H H H
Co N N N Cl
Cl H H H
Cl H H H H H
H
Co1+ N N N Cl–
Cl H H H
against scientific realism in a later section of the paper and address these lines
of thinking more directly. However, it’s worth raising at this juncture in order to
remind ourselves what’s at stake, philosophically, by examining the 19th century
debate over the structure of cobalt complexes.
During this same time through 1899, Jørgensen and Werner worked
separately in synthesizing and characterizing other metal ammines. Like
experiments with hexammines, work with penta-and tetra-ammines yielded
experimental data that supported both Jørgensen’s chain theory and Werner’s
coordination theory. However, it was the triammines that forced the triumph
of one theory over the other.15 The metal triammine has a cobalt atom, three
ammine groups, and three chlorides. Jørgensen predicted a chain structure
for the triammine complex with a chain of three ammines: one chloride on
the end of the ammine group, and two chlorides directly attached to the co-
balt atom (Figure 7.7). Again following the dissociation rule, his structure
predicted two ions after precipitation: the chloride off the ammine chain
and then the remaining complex (Figure 7.8). Experimental work did not
H
H
Cl
Cl
H N
H Co H
H N
N H
H Cl
H
support Jørgensen’s chain structure. It turns out that zero ions precipitated
during the experiment.
However, Werner’s proposed structure for the triammine complex success-
fully predicted the number of ions for the precipitation experiment. Keeping
with the same two features of the hexammine model, Werner implemented an
octahedral, symmetric shape and secondary valence in his triammine model.
The cobalt atom is attached to three separate ammine groups and three separate
chlorides. The primary valence of cobalt is satisfied with the cobalt atom being
connected to three chlorides (Figure 7.9). Because the chlorides are directly at-
tached to the cobalt, they should not separate during precipitation. Therefore, no
ions are predicted, and precipitation experiments confirmed this prediction: no
ions were measured.
While we don’t have space for an in-depth discussion of the epistemic status of
models with respect to the physical systems being represented, a few explanatory
comments are in order. Margaret Morrison notes the “problem of inconsistent
models” arises when multiple inconsistent models provide useful information
about a physical system, and this problem undermines the degree to which one
should give a realist interpretation of a model—particularly if one lacks the
ability to tweak models in ways that would allow for empirical confirmation/
disconfirmation. Says Morrison, “In these contexts we usually have no way to
determine which of the many contradictory models is the more faithful repre-
sentation, especially if each is able to generate accurate predictions for a certain
class of phenomena.”16 This is exactly the situation during the early experimenta-
tion and modeling of cobalt complexes, when both Jørgensen’s chain theory and
Werner’s coordination complex model made predictions that were supported by
precipitation experiments.
However, note the conditions that Stathis Psillos presents concerning when a
model M can be interpreted as a realist depiction of a physical system X:
I think one can, in principle, take a realist stance towards particular models. For
although scientists do not start off with the assumption that a particular model
gives a literal description of a physical system X, there may be circumstances
in which a model M of X offers an adequate representation of X. These
circumstances relate to the accuracy with which a given model represents the
target system . . . . Amassing more evidence, such that novel correct predictions
for X derived from M, may be enough to show that M represents X correctly. In
sum, taking a realist attitude towards a particular model is a matter of having
evidence warranting the belief that this model gives us an accurate representa-
tion of an otherwise unknown physical system in all, or in most, causally rele-
vant respects.17
to discuss—concerning the practice of scientific research and the nature of theoretical justification.
For example, Day and Selbin (1962: 264) suggest “Werner had no theoretical justification for his two
types of valency.” Werner’s first paper (1893) proposing his theory outlines his justifications.
20 See Meissler and Tarr (2011: 323).
21 The widely-accepted idea that Jørgensen admitted defeat to Werner is debatable. Jørgensen
never formally stated defeat or acceptance of coordination theory, rather, it seems he ignored the
controversy after 1899. S.P.L. Sørensen reported in his eulogy that Jørgensen admitted defeat in a
conversation in 1907. See Kragh (1997).
22 See Day and Selbin (1962: 264).
Selective Scientific Realism and Truth-Transfer 141
In this section, we apply lessons from the scientific debate on the structure of
cobalt complexes to the philosophical debate about whether some version of sci-
entific realism is justified. Scientific realism reflects a certain cluster of intuitions
and beliefs about the nature and purpose of science. At the heart of the realist
sensibility is the notion that the success of scientific research and theory de-
velopment consists of a theory explaining the data and making accurate novel
predictions. Realists view the fit between data and theory as justifying an in-
ference to the truth of theoretical statements. As such, science is normed and
constrained by the mind-independent reality it seeks to map.
Although versions of scientific realism abound, we adopt as a starting point
Stathis Psillos’s elegant and comprehensive description of scientific realism as
a conjunction of three stances. According to Psillos, the realist adopts a meta-
physical stance that the world has a definite, mind-independent, natural kind-
like structure, a semantic stance that theories are truth-governed descriptions
in which putative referring terms do, in fact, refer, even to unobservable entities,
and an epistemological stance according to which mature and predictively suc-
cessful theories are held to be approximately true.23 Each stance has its philo-
sophical detractors.
Post-structuralists of various sorts will deny that the world has any accessible
mind-independent structure. Non-realist philosophers of science will deny that
successful science requires accepting that theoretical terms in scientific theories
are successful referring terms, and will further deny that predictive success is a
mark of, or best explained by, truth.
One of the most prominent arguments against the claim that predictive success
is a marker of truth comes from P. Kyle Stanford’s argument based on the possi-
bility of unconceived alternatives. According to Stanford,
[T]he most troubling historical pattern for scientific realism is the repeated
and demonstrable failure of scientists and scientific communities throughout
the historical record to even conceive of scientifically serious alternatives to the
theories they embraced on the strength of a given body of evidence, that were
23 See Psillos (1999: xix). More specifically our version of realism is best described as a type of di-
vide et impera realism; for the purpose of this paper we remain open to various options concerning
what, exactly, constitutes the truth-conducive constituents of scientific theories.
142 Historical Cases for the Debate
Stanford’s idea can be explicated as follows. Suppose that for some body of evi-
dence E, there are n number of theories being considered by the relevant scien-
tific community to account for E. Suppose further that through a combination of
testing and theoretical work, consensus emerges within the scientific commu-
nity about which of the candidate theories can be eliminated and about which
theory among the remaining candidates best explains E. The problem, says
Stanford, is that we have good reason based on the history of science to think
that the epistemic control condition for determining when eliminative inference
is reliable—namely, the condition that we must have all of the likely or plausible
alternative possibilities in view before proceeding to embrace the winner—is never
satisfied. Stanford continues:
26 Ibid.
144 Historical Cases for the Debate
beyond the scope of this paper to address all the nuances of Stanford’s project,
we feel that there are strong defenses that the realist can present in response to
Stanford. We further note that whether or not Stanford’s Problem of Unconceived
Alternatives is a successful objection to scientific realism will depend in large
measure on more fundamental epistemological assumptions, including whether
the logical possibility of an unconceived alternative does in fact constitute a
defeater for evidence one has in hand.
A charitable interpretation of Stanford’s philosophical objection to scientific
realism has an empirical consequence that seems implausible, a consequence
that is part of what we’re calling the No Final Theory Objection to the Problem
of Unconceived Alternatives. Suppose it’s always true, as Stanford maintains,
that the eliminative inference scientists perform at any given time is necessarily
constrained by the inability to consider unconceived alternative theories which
provide a better explanation for the data in question. This can be stated with the
following:
(1) For any current theory T that purports to explain some evidence E, there
is another theory T* such that:
(i) T* is unconceivable at present.
(ii) T and T* are inconsistent—they cannot both be true. And,
(iii) T* is a better explanation E than T.
But notice a consequence of (1) is that there can be no “final theory,” a term
that can be used to denote different things. Some, following Nobel Laureate
Steven Weinberg, use “final theory” to denote a final physical theory that
provides a unified explanation for gravity, quantum mechanics, and the strong
and weak nuclear forces.27 However, suppose, as suggested by Weinberg him-
self, possessing a final theory that provides a unified explanation of funda-
mental forces might still leave problems about consciousness and mind/body
interaction unsolved.28 In that case, having a final theory wouldn’t constitute
having a “theory of everything,” if “everything” included consciousness. For
ease of reference, we’ll use final theory to denote the type of unifying theory
that Weinberg has in view—nothing of consequence in our objection turns
on whether one has a wider scope in view for final theory. Thus, considering
candidates for a final theory as a substitution instance for the schema presented
in (1) it follows that:
(2) For any candidate for a final theory F, there will be a theory F* such that:
(i) F* is unconceivable at present.
(ii) F and F* are inconsistent—they cannot both be true. And,
(iii) F* is a better explanation than F.
Thus, if a final theory is one that provides a complete and accurate unification of
whatever fundamental forces there be, from (1) and (2) it follows that:
However, why should we think (3) is true? Why should epistemological consid-
erations based on the history of science tell us whether there is an accurate and
unified way of explaining how fundamental forces interact? Instead, it seems
that (1)–(3) serve as premises in a reductio against the Problem of Unconceived
Alternatives, for the negation of (3) seems much more plausible—asserting
(3) commits one to accepting a fundamental inconsistency between physical
reality and the logical possibility of, ultimately, representing physical reality
through comprehensive theory. If Stanford is right, not even an omniscient being
could have a final theory—which indicates that something is amiss in the initial
premises.
To be sure, the truth of a theory isn’t the only available explanation for a
theory’s predictive success. Suppose a theory T generates a testable observation
statement O and we do in fact observe that O. Observing O certainly doesn’t en-
tail that T is true. But would observing that O provide an epistemologically re-
spectable reason for thinking that T is true, or approximately true? A defeasible
reason, but a reason nonetheless?
Timothy D. Lyons doesn’t think so. In his 2017 “Epistemic selectivity, histor-
ical threats, and the non-epistemic tenets of scientific realism,” Lyons presents
two arguments against what he terms the realist meta-hypothesis:
Lyons’s first argument concludes that the meta-hypothesis is false; we’ve dubbed
it the False Basis Argument, because it employs the observation that successful
predictions are sometimes derived from constituents of theories that are not true.
In those cases, successful predictions are derived from theoretical bases that turn
out to be false, and the successful predictions made by the Blomstrand-Jørgensen
chain theory are examples of this.30 Here’s a slightly modified version of Lyons’s
argument:
1. If the realist meta-hypothesis were true, then all the constituents gen-
uinely deployed in the derivation of successful novel predictions are
approximately true.
2. However, we do find constituents genuinely deployed toward success that
cannot be approximately true.
3. Therefore, the realist meta-hypothesis is false.
30 One might think that the phenomenon of false theories making successful predictions parallels
what’s going on in Gettier cases in epistemology. In Gettier cases, we have examples of justified
true beliefs, but such that the truth of the belief is a “happy accident.” For example, the successful
predictions made by Jørgensen might be Gettier-like in that regard because it’s merely a “happy acci-
dent” that the false theory correctly predicted the number of ions that were dissociated in precipita-
tion experiments. However, the selective scientific realist wants to say that that is not so—rather, what
led to the predictive successes of Jørgensen was the fact that he followed accurate rules that survived
into the successor theory. Thus, while there’s a superficial similarity to Gettier cases (in the appear-
ance of truth being a function of “mere coincidence”), the similarity disappears on closer inspection.
Selective Scientific Realism and Truth-Transfer 147
chemical bonding explains the difference.36 Metal salts that would dissociate
in water are classified as ionic, where opposite charges (e.g. the positive sodium
and negative chloride) attract. Another type of bonding is classified as covalent,
where the electrons in the bond are shared (either equally or unequally), usu-
ally seen in bonding between two nonmetals. Transition metal complexes, like
the cobalt ammine complexes, have both types of bonding. Werner’s secondary
valence, the inner sphere around the cobalt, are covalent bonds whereas the pri-
mary valence (e.g. the chlorides in the outer sphere) are ionic bonds. Therefore,
the chlorides that will dissociate are the same type of bonds seen in other metal
salts: ionic. Ligand Field Theory further expands this theory,37 and one can say
that Werner’s coordination theory is approximately true, exhibiting novel pre-
dictive success, and has been preserved in present-day chemistry.
Jørgensen was deploying the valence rule, dissociation rule, and the chain rule
in order to, ultimately, make predictions about the number of ions that would
separate during precipitation experiments. In the psychological sense of deploy-
ment, he was genuinely deploying all three rules. But what turns out to be the
case is that the chain rule was an explanatory free-rider—it didn’t factor into the
actual causal story concerning the number of ions that separated during the pre-
cipitation experiments. Thus, this particular historical example doesn’t serve as
an example referenced in premise (2) of the False Basis Argument, for the rel-
evant sense of genuine deployment doesn’t apply. Moreover, as surveyed in the
previous paragraph, the successful working posits within Blomstrand-Jørgensen
chain theory survived into Werner’s model and through to the present day.
We have only looked at one case where it’s plausible to interpret the historical
sequence of events in a way that is friendly to scientific realism. In order to seri-
ously undermine (2) of the False Basis Argument, more work would need to be
done in order to look at the historical cases cited by anti-realists in support of the
pessimistic induction.38 However, we think that the transition from Blomstrand-
Jørgensen chain theory to Werner’s geometrical structures and the coordina-
tion theory that support it can be seen as the ever-increasing glory of science
at work. Blomstrand and Jørgensen were wrong, importantly wrong, about
some things but not everything. And Jørgensen did deploy a false constituent
in generating predictions which were then observed in experiments. However,
when faced with cases where false theoretical constituents lead to successful
predictions, it seems like one is faced with an interpretive choice: either there’s
a truth-conducive explanation lurking nearby that we don’t have clearly in view,
or there isn’t, in which case it turns out that the successful prediction happens to
39 We prefer “mere coincidence” as opposed to “miracle,” for presumably miracles, if any there be,
Why does Lyons think that we have more evidence for the ContraSRs than for
the realist meta- hypothesis? Avoiding language of confirmation/ disconfir-
mation, and opting instead for a hypothesis having “correlatively precise posi-
tive instances,” or “correlatively precise negative instances,” it seems like Lyons
supports (5) with something like the following No Instances Argument.
The controversial premises in the No Instances Argument are (8) and (11).
According to (8), we have a list of predictively successful theories that are not
approximately true. That is, we have a list of theories where false constituents
were genuinely deployed in deriving novel predictions. However, what our pre-
ceding discussion about genuine deployment in the Blomstrand-Jørgensen/
Werner case has shown, is that there’s a difference between a psychological sense
of genuine deployment and a causally robust sense of genuine deployment that
involves the working posits within a theory. Recall that in cases where it appears
that false constituents were genuinely deployed to generate predictive success,
one is faced with an interpretive choice between explaining the predictive suc-
cess in terms of approximate truth or mere coincidence. Either the predictive
success is accounted for because there’s a truth-conducive explanation lurking
nearby, or there isn’t (in which case the predictive success is a mere coincidence).
Certainly more investigative work would need to be done, along the lines of what
we’ve done with the story of cobalt complexes. However, if a long string of mere
coincidences is itself unlikely, than we have reason to think that the evidence for
(8) is not as strong as it’s taken by Lyons to be.41
What about (11)? In setting up the Weak Evidence Argument, Lyons states the
following:
He further states:
First, it is admittedly logical considerations that reveal that, on the one hand,
instances of the correlates of the ContraSRs can be secured without presup-
posing the truth of the ContraSRs, and, on the other, that such a feature is not
shared by the realist meta-hypothesis.43
Lyons’s worry here seems to be that there is no non-circular way of assessing pu-
tative positive correlates of the realist hypothesis which wouldn’t already presup-
pose the truth of the realist hypothesis. As such, those cases can’t really count as
evidence in favor of the hypothesis.
But can the realist hypothesis really be defeated so easily by a priori consid-
erations in epistemology? Here the realist meta-hypothesis is in a position sim-
ilar to analogous positions about the reliability of one’s cognitive faculties—sense
perception, say. Consider the following realist hypothesis about sense perception:
What possible evidence could one provide in support of the Sense Perception
Meta- Hypothesis that doesn’t assume the reliability of sense perception?
Not much. But here we’ll just point out that a number of philosophers, recog-
nizing the challenge of finding non-circular justification for our belief-forming
practices, have concluded that if there are any justified beliefs at all, then cer-
tain types of circularity are unavoidable and do not undermine justification. And
because it is extremely difficult to deny that there are any justified beliefs, scien-
tific or otherwise, the inevitable conclusion is that some types of circularity are
unavoidable but not such that they undermine justification.
For example, William Alston argues that our inability to give a non-circular
justification for the reliability of sense perception does not undermine the justi-
fied deliverances of sense perception.44 Michael Bergmann argues if one accepts
the foundationalist principle that some beliefs are justified non-inferentially,
one is thereby committed to approving of epistemically circular arguments;
Bergmann goes on to argue that because foundationalism is vastly superior to
competing accounts of the structure of knowledge, epistemic circularity can’t be
all bad.45 Psillos argues that “rule-circularity”—that is, arguing for some con-
clusion p via an inference rule R when p entails that R is reliable—is not an epis-
temologically vicious type of circularity.46 Thus, if there are benign instances of
circularity, like Alston, Bergman, and Psillos agree is the case, then it’s not the
case as Lyons asserts in (11), that there can be no correlatively precise positive
instances of the realist meta-hypothesis. True, realism will require interpreting
evidentially significant cases in ways that adopt some version of circularity, but
perhaps in a way that is not vicious circularity, particularly if the alternative to
realist interpretations of novel predictions requires appealing to a string of mere
coincidences.
In essence, what we are claiming is that there is lots of evidence for the realist
meta-hypothesis and lots of evidence against the ContraSRs if one understands
genuine deployment in a causally robust sense of working posits that lends itself
naturally to a realist interpretation of scientific practice, including practice that
involves theory change over time. Granted, there is a kind of circularity here.
But suppose, as Psillos, Alston, Bergmann, and others have argued, that some
instances of circularity are benign and are simply a feature of the epistemological
landscape, a landscape that includes the epistemological territories of scientific
practice and philosophy. In that case, more work would need to be done to es-
tablish that the type of circularity we are acknowledging is vicious, not benign.47
Moreover, interpreting cases of genuine deployment in ways that favor anti-
realism may also involve an appeal to circularity of the sort that requires some
question-begging assumptions about base-rates and the probability space of
detail involve internalism vs. externalism in epistemology. As Psillos (1999: 85) notes, those with ex-
ternalist intuitions will not be bothered by a type of rule-circularity that uses an inference rule R to
justify R. Moreover, externalists will be open to criteria for genuine deployment that reflect working
posits which actually carve nature’s joints, regardless of whether one is able to demonstrate successful
deployment according to internalist criteria.
Selective Scientific Realism and Truth-Transfer 153
theories from which the relevant base-rates are drawn. Peter Dicken (2013)
looks at ways in which the realism/anti-realism debate is subject to exempli-
fying instances of the base-rate fallacy.48 Dicken uses the common example of
circumstances involving medical diagnosis in order to illustrate the base-rate fal-
lacy and then proceeds to apply the statistical lessons learned to the realism/anti-
realism debate. In order to make the analogy explicit, let us define the ALPHA
condition as one where the facts on the ground make it very likely that the rele-
vant indicator conditions are present. Moreover, let the BETA condition be one
where the relevant indicator conditions are taken to indicate certain facts on the
ground.49 Here then, is the analogy:
Dicken illustrates the base-rate fallacy by noting that if, say, we’re looking at a
population of 100,000 and a base-rate of 0.1%, then the BETA condition is not a
strong indicator of the presence of disease. In that scenario, a positive test result
on a random subject yields only a 2% chance that the subject has the disease. This
is because even though the test has a 95% success rate, given the very low number
of actual disease carriers relative to the large population, the 5% rate of misdi-
agnosis means that the number of uninfected people misdiagnosed as having
the disease (i.e. 4995) is much, much higher than the number of infected people
accurately diagnosed as having the disease (i.e. 95). So, a positive test result on a
randomly drawn subject from this population has a 95/4995 (i.e. 2%) chance of
being accurate.
Applying the lesson to the realism/anti-realism debate, we get the following
two analogous conditions:
Yet while the threat of unconceived alternatives may well be more robust
than the threat of the standard pessimistic meta-induction, and while it may
indeed avoid the need for knowing the underlying base-rate of any arbitrary
scientific theory being true it seems that for the inference to be compelling
we must simply presuppose a different base-rate concerning the reliability of
scientists: maybe our historically unconceived alternatives are all heavily biased
towards researchers working in a very specific domain of inquiry; or maybe
while the history of science furnishes us with a great number of instances of
scientists failing to exhaust the relevant possibilities, this is due to the relatively
large number of scientific practitioners, rather than the likelihood of any arbi-
trarily successful researcher failing to consider another alternative theoretical
formulations.51
Are we at a stalemate with respect to worries about circularity more broadly, and
base-rate fallacies in particular? We think not, at least so far as defending SSR
through our particular historical example is concerned.
First, notice that one key determination in establishing the relevant base-rates
is determining the relevant “population” from which a candidate theory is being
assessed. There are many different ways to determine the relevant population
base for which one is attempting to determine base-rates. Given that there do not
seem to be clear and neutral criteria for determining base-rate thresholds in ways
that clearly privilege realism or anti-realism, it is unlikely for there to be such
clear and neutral criteria for determining the relevant population base.
Second, note that while the analogy between the medical diagnosis case and
the scientific realism case is helpful for illustrating the relevance of base-rates
to the realism/anti-realism debate, notice also that the two sorts of cases are
not completely analogous. With respect to medical diagnosis, the presence or
absence of disease is a binary, “on/off ” affair: one either has the disease or one
doesn’t. But with respect to the truth or approximate truth of any given theory
T, things are not cast in such binary terms—at least not in binary terms that the
scientific realist is obligated to accept. This is because, on the selective scientific
realist framework we are adopting, theories that are, strictly speaking, false when
taken in their entirety, can have true working posits. This further complicates the
task of determining the relevant population for determining base-rate threshold,
as determining base-rate isn’t as simple as determining which members of the
population “have or do not have the disease.”
Third, recall where we are in the dialectic with Lyons’s arguments against the
realist meta-hypothesis. The main focus of this section of our paper has been to
determine whether Lyons’s criticisms of realism undermine our interpretation
of the transition from the Blomstrand-Jørgensen model of cobalt complexes to
Werner’s coordination theory. We have argued that this episode of theory change
in 19th century chemistry can be plausibly interpreted using the tools of SSR,
noting that the features that lead to predictive success of the failed theory—its
working posits that were genuinely deployed in a causally robust sense—survived
and were part of the successful theory that replaced it. Lyons has identified a re-
alist meta-hypothesis that is required by realists—including the divide et impera
realism we are advocating in this paper:
We then examined two arguments by Lyons to the conclusion that the realist
meta-hypothesis is false.
According to the False Basis Argument, some theoretical constituents genu-
inely deployed in deriving successful novel predictions turn out to not be true—
a consequence that is inconsistent with the realist meta-hypothesis. However,
we argue that if we distinguish between psychological deployment and a more
156 Historical Cases for the Debate
7.6 Conclusion
Acknowledgments
This work was substantially improved by comments received when this paper
was presented at the Canadian Society for History and Philosophy of Science
2017 annual meeting, the 2017 Quo Vadis Selective Scientific Realism confer-
ence, the University of Notre Dame Centre for Philosophy of Religion, and the
Purdue University philosophy department colloquium. Thanks in particular for
detailed written comments by Tim Lyons and Peter Vickers, as well as by anon-
ymous reviewers. Also, work on this project was supported by a grant from the
Templeton Religion Trust.
References
Alston, William P., “Epistemic Circularity.” Philosophy and Phenomenological Research 47
[1] (Sep., 1986), 1–30.
Bergmann, Michael, “Epistemic Circularity: Malignant and Benign.” Philosophy and
Phenomenological Research 69 [3] (Nov., 2004), 709–727.
Day, M. Clyde, Jr., and Joel Selbin, Theoretical Inorganic Chemistry. Reinhold Publishing
Corporation (1962), 183.
Dicken, Peter, “Normativity, the Base- Rate Fallacy, and Some Problems for Retail
Realism.” Studies in History and Philosophy of Science 44 (2013), 563–570.
Douglas, Bodie E., and Darl H. McDaniel, Concepts and Models of Inorganic Chemistry.
Blaisdell Publishing Company (1965), 338.
Farber, Eduard, The Evolution of Chemistry. New York: The Ronald Press Company
(1952), 164–165.
Griffith, J.S., and L.E. Orgel. “Ligand-Field Theory.” Quarterly Reviews, Chemical Society
11 [4] (1957): 381–393.
Jørgensen, Sophus Mads, “Beiträge zur Chemie der Kobaltammoniakverbindungen: VII
I. Ueber die Luteokobaltsalze.” Journal für praktische Chemie 35 [2] (1887): 417, cited
in Kauffman.
Jørgensen, Sophus Mads, “Zur Constitution der Kobaltbasen.” Journal für praktische
Chemie 41 [2] (1890): 429, cited in Kauffman.
Kauffman, George B., Sophus Mads Jorgensen (1837–1914) A Chapter in Coordination
Chemistry History in Selected Readings in the History of Chemistry. Compiled by Aaron
J. Ihde and William F. Kieffer, editor, reprinted from the 1933–1963 volumes of Journal
of Chemical Education, Easton, Pennsylvania: Division of Chemical Education of the
American Chemical Society (1965).
Kragh, Helge, “S.M. Jørgensen and His Controversy with A. Werner: A Reconsideration.”
The British Journal for the History of Science 30 [2] (1997), 203–219.
158 Historical Cases for the Debate
Lyons, Timothy D. “Scientific Realism and the Strategema de Divide et Impera.” British
Journal for the Philosophy of Science 57 (2006), 537–560.
Lyons, Timothy D., “Structural Realism versus Deployment Realism: A Comparative
Evaluation.” Studies in History and Philosophy of Science 59 (2016), 95–105.
Lyons, Timothy D., “Epistemic Selectivity, Historical Threats, and the Non-Epistemic
Tenets of Scientific Realism.” Synthese 194 [9] (2017): 3203–3219.
Meissler, Gary L., and Donald A. Tarr, Inorganic Chemistry, 4th ed. Prentice Hall
(2011), 322.
Moore, F. J., A History of Chemistry, 3rd ed. McGraw-Hill Book Company, Inc. (1939),
294–295.
Morrison, Margaret. “One Phenomenon, Many Models: Inconsistency and
Complementarity,” Studies in History and Philosophy of Science 42 (2011): 342–351.
Psillos, Stathis. Scientific Realism: How Science Tracks Truth. Oxford: Oxford University
Press (1999).
Stanford, P. Kyle, Exceeding Our Grasp: Science, History, and the Problem of Unconceived
Alternatives. Oxford (2006).
Stanford, P. Kyle, “Unconceived Alternatives and the Strategy of Historical Ostension,”
in Juha Saatsi (ed), The Routledge Handbook of Scientific Realism. Routledge
(2018): 212–224.
Vickers Peter, “A Confrontation of Convergent Realism.” Philosophy of Science 80 [2]
(April 2013), 189–211.
Weinberg, Steven, Dreams of a Final Theory. Vintage Books (1994).
Weinberg, Steven, interviewed by John Horgan, https://blogs.scientificamerican.com/
cross-check/nobel-laureate-steven-weinberg-still-dreams-of-final-theory/.
Werner, Alfred, Zeitschrift für anorganische Chemie 3 (1893): 267–330, translated in
George B. Kauffman, Classics in Coordination Chemistry Part I: The Selected Papers of
Alfred Werner. New York: Dover Publications, Inc. (1968), 17.
8
Realism, Physical Meaningfulness,
and Molecular Spectroscopy
Teru Miyake1 and George E. Smith2
1Department of Philosophy
Tufts University
[email protected]
8.1 Introduction
1 See, for example, Wesley Salmon (1984), Deborah Mayo (1996), Peter Achinstein (2001), Stathis
Teru Miyake and George E. Smith, Realism, Physical Meaningfulness, and Molecular Spectroscopy In: Contemporary
Scientific Realism. Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press
2021. DOI: 10.1093/oso/9780190946814.003.0008
160 Historical Cases for the Debate
having only a fleeting existence. Absent any evidence that such microscopic parts
retain their integrity over a significant length of time, this result falls short of
establishing the existence of atoms or molecules.
An undeniably large epistemic gap exists, moreover, between having good
reason to think that the medium in which Brownian particles are suspended is
composed of microscopic parts in vigorous stochastic motion, and having know-
ledge about specific details of those parts. How much access to molecules was, in
fact, gained through Perrin’s experiments? According to physicists and chemists,
we now know that molecules have a great deal of structure, as we shall discuss
later. How much of this structure could have been inferred from measurements
of Avogadro’s number? The sizes of molecules could be calculated from a de-
termination of Avogadro’s number, but such calculations were based on the
assumption that molecules are spherical or at least have a distinct sphere of in-
fluence. As van Fraassen points out (2009, p. 8), dependence on such unrealistic
assumptions throws considerable doubt onto whether such calculations are truly
yielding information about molecules. Moreover, this shortcoming was expressly
recognized at the time. James Jeans adopts values of the molecular magnitudes
corresponding to those of Rutherford and Geiger in the 1916 edition of his The
Dynamical Theory of Gases and then concludes, “If, however, the molecules are
assumed as a first approximation to be elastic spheres, experiment leads to dis-
cordant results for the diameters of these spheres, shewing that the original as-
sumption is unjustifiable” (Jeans 1916, p. 9). Simply put, accordingly, whatever
grounds the various determinations of Avogadro’s number and the evidence for
fine-grained stochastic motion in the substrate supporting Brownian motion
had provided for molecular-kinetic theory, it had provided no information on
the sizes and shapes of molecules. In the 1921 edition of Theoretical Chemistry
from the Standpoint of Avogadro’s Rule and Thermodynamics, Walther Nernst was
still making a claim he made in the first edition of 1893: “At present scarcely an-
ything definite is known regarding either the nature of the forces which bind the
atoms together in the molecule and which hinder them from flying apart in con-
sequence of the heat of motion, or regarding their laws of action” (Nernst 1895,
p. 237; 1923, p. 327).
Some information can nevertheless be obtained about the structure of
molecules through kinetic theory. If equipartition of energy is assumed, then
kinetic energy should be distributed evenly through all the degrees of freedom
of a molecule. A spherical molecule has three degrees of freedom, while a mol-
ecule in the shape of a dumbbell can rotate along two axes, so it has two addi-
tional degrees of freedom. If the parts of the molecule are free to move relative to
each other, there will be additional degrees of freedom. The ratio of the specific
heat of a gas at constant pressure and the specific heat at constant volume, ac-
cording to kinetic theory, is a function of the number of degrees of freedom of
162 Historical Cases for the Debate
the molecular constituents of the gas. Measurements of the ratio of specific heats
might, then, provide some insight into the shapes of molecules. Unfortunately,
at the beginning of the twentieth century, measurements of the ratio of specific
heats of various gases could not be reconciled with what was then known about
molecules. In a lecture delivered to the Royal Institution in 1900, Lord Kelvin
named the specific heat anomaly one of two “Nineteenth Century Clouds over
the Dynamical Theory of Heat and Light” (Kelvin 1904, pp. 486–527).
How, then, did scientists first come to gain access to the structure and proper-
ties of atoms and molecules? Our view is that this was first made possible by the
development, in the late 1920s and after, of theories and techniques that enabled
theory-mediated access to the structure of atoms and molecules through spec-
troscopy. The analysis of spectra has a history that predates quantum mechanics
by more than a century, but it was the development of quantum mechanics
that first allowed physicists to make stable theory-mediated measurements
of parameters that they took to represent the real properties and structures of
molecules. In this paper, we thus aim to answer the following question:
(1) How did spectroscopists first come to have the ability to make stable
theory-mediated measurements of parameters that are taken to represent
the real properties and structure of molecules?
Anti-realists such as van Fraassen will grant that such concordant measurement
has been achieved but will deny that spectroscopists have thereby gained access
to the real structure and properties of molecules. For that reason, we shall ask the
further question:
(2) What evidence is there that when spectroscopists measure such parameters
they are gaining access to the real properties and structures of molecules—
in other words, that the parameters are physically meaningful and not just
artifacts of the particular way spectroscopists have chosen to represent
whatever lies beyond our senses?
pp. 129–174). Moreover, the values for the mean-free-paths obtained from vis-
cosity when extended to other phenomena—especially, to deviations from the
ideal gas law and refraction of light through gases—yielded values for the sizes
of the spheres that were not only discordant with one another, but also in vio-
lation of Avogadro’s rule (Nernst 1895, pp. 347–352; Meyer 1899, pp. 149–221,
299–348). Van Fraassen would call this a failure to ground the theory; we see it as
a representation of the molecular realm that never had clear claim to being phys-
ically meaningful. As this example suggests, the answer to the second question
is likely to be rather complicated, and so in this paper we do our best to give an
initial attempt to answer it. In particular, we believe that one of the keys to an-
swering this question consists in understanding a distinction that we will make
between representations that are physically meaningful and those that are not.
Molecular spectroscopy is a little-known topic among philosophers of science,
so we will take some care in explaining how it works. Section 8.2 of this paper
will be devoted to providing enough background on molecular spectroscopy to
be able to answer question (1)—that is, to explain how stable theory-mediated
measurement of parameters that are taken to represent properties of molecules
has been achieved. In order to keep things manageable, we will limit our discus-
sion to the molecular spectroscopy of diatomic molecules and will focus on the
early period of molecular spectroscopy, from the 1920s through around 1950. In
Section 8.3, we will consider ways in which one might attempt to answer ques-
tion (2). We will conclude that a proper answer to question (2) will require an
understanding of physical meaningfulness, which we will explain through an
examination of an example from celestial mechanics. A full answer to question
(2) will go beyond the scope of this paper, for it would require an examination of
the history of molecular spectroscopy in the period after 1950, but we will pro-
vide, in Section 8.4, a brief look at further work in molecular spectroscopy that
we believe supports our contention that spectroscopists indeed are gaining ac-
cess to physically meaningful parameters of molecules.
Molecular spectroscopy is a massive field. It grew rapidly from the late 1920s,
when quantum mechanics was first successfully applied to the analysis of spec-
troscopic data. This early period culminated with the publication of Gerhard
Herzberg’s landmark Molecular Spectra and Molecular Structure, published in
four volumes between 1939 and 1979. In this paper, we will mostly focus on re-
search in the field up through 1950, the year of publication of the second edi-
tion of Volume 1, on the spectra of diatomic molecules (Herzberg 1950). As an
164 Historical Cases for the Debate
indication of the amount of research that was being done at the time on dia-
tomic molecules, we note that the second edition of Volume 1 contains a table
of constants for diatomic molecules that goes on for 80 pages (pp. 501–581)
and lists 1574 references in its bibliography. Molecular spectroscopy has been
growing steadily ever since. Volume 4 of Herzberg, published in 1979, consists
solely of tables of constants for diatomic molecules. It is 716 pages long and
lists its references by molecule. The ordinary hydrogen molecule, 1H2, has 169
references, while 1H2H, a molecule consisting of one ordinary hydrogen atom
and one deuterium atom, has 51 references; 2H2, or molecular deuterium, has
60 references. There is now an online database of papers on diatomic molecules
called DiRef (Bernath and McLeod 2001), which covers the period after Volume
4; it contains 30,000 references for the period 1974–2000.
The molecular constants are parameters that can be determined from spec-
troscopic data, and they are taken to represent certain properties of molecules.
When a molecule undergoes a transition between two energy states, it absorbs or
releases a photon, and the frequency of the photon is taken to correspond to the
difference in energy levels between these two states. Thus, an examination of the
spectrum of light absorbed or released by a gas can provide information about
the energy states of the molecules in the gas. The interpretation of this spectrum
requires some knowledge about the possible energy states of a diatomic mole-
cule. A diatomic molecule at room temperature is taken to be a body that is not
merely in translational motion, but also undergoing vibrational and rotational
motion. The two nuclei in the molecule vibrate anharmonically in accord with
an asymmetric potential function that depends in large part upon the distance
between the nuclei. The shape of the potential function depends upon the ar-
rangement of the positive charges of the nuclei and the negative charges of the
electrons in the molecule. The rotation has the effect of centrifugally stretching
out the distance between the nuclei, but the effect is only significant at high levels
of rotation, so we shall omit discussion of it here. Energy can be emitted and
absorbed by the molecule through transitions between rotational, vibrational,
and electronic states. Energy is quantized, with the consequence that the spectral
lines correspond to a transition from one energy state to another, whether it be a
change in the rotational, vibrational, or electronic state of the molecule.
The energies of these states can be calculated for idealized systems by solving
the Schrödinger equation for such systems. For example, the simplest idealiza-
tion of a diatomic molecule is the rigid rotor, a system consisting of two point-
masses separated by a fixed distance, rotating about an axis perpendicular to the
line connecting the masses (Herzberg 1950, pp. 66–73). The Schrödinger equa-
tion for such a system has solutions for only certain energy values, customarily
given in units of wavenumber (cm–1) by
Realism and Molecular Spectroscopy 165
F ( J ) = BJ ( J + 1) , where B = h / (8π 2 cI ).
Here, J is the rotational quantum number, and can take the integral values 0, 1,
2, . . . , while h is Planck’s constant, c is the speed of light, and I is the moment of
inertia of the rigid rotor. Thus B, called the rotation constant, essentially gives
the inverse of the moment of inertia of the system, normalized and converted to
units of wavenumber. A further analysis of the selection rules for the quantum
number J shows that the spectrum of a rigid rotor should consist of a series of
evenly spaced lines, the first of which lies at wavenumber 2B, with the distance
between successive lines also being 2B. Thus, the value of B, and hence that of the
moment of inertia, of a rigid rotor, can be read directly off of its spectrum, pro-
vided that the spectrum has been interpreted correctly. Given the masses of the
rigid rotor, the distance between the two point-masses can be determined from
the moment of inertia.
A diatomic molecule is not, however, a rigid rotor—the distance between its
nuclei is constantly changing. If a diatomic molecule were, instead, a rotating
harmonic oscillator, it would produce a spectrum with a single peak that, at
higher resolution, contains peaks spaced uniformly corresponding to dif-
ferent rotational transitions. In fact, actual diatomic molecules are anharmonic
oscillators, and this gives rise to non-uniformities in the spectral lines. These
non-uniformities provide ways of determining constants characterizing both
the anharmonicity of the vibration and the effect of vibration on rotation, as
well as other effects. The values of these constants depend upon the potential
functions between the two nuclei, which in turn depend upon whether the
electrons in the molecule are in the ground state or excited states. In order to
keep things simple, we will only be considering the ground state—that is, a
state from which spontaneous transitions never occur—but no small part of
the evidence that the constants are characterizing a structure of the sort speci-
fied comes from comparing the values of the constants for different electronic
states.
With all of this in mind, the vibrational (G) and rotational (Fν) energy levels
for a diatomic molecule may be calculated theoretically, and are customarily
given by the following expressions (Bernath 2016, p. 225), again in units of
wavenumber (cm-1):
( ) (
Fν ( J ) = Bν J ( J + 1) − Dν J ( J + 1) + H J ( J + 1) + … )
2 3
166 Historical Cases for the Debate
The vibrational energy levels G(ν) are thus expressed as a series expansion in
ν + 1/2, where ν is the vibrational quantum number, while the rotational en-
ergy levels Fν(J) are expressed as a series expansion in J(J + 1), where J is the
rotational quantum number. The rotation constant Bν is the first coefficient in
the expansion for the rotational energy levels, and it depends on the vibrational
energy state as follows:
Bν = Be − α e (ν + 1 / 2) + γ e (ν + 1 / 2)2 + …
The constant Be, the first term in the expansion of Bν, is the value that the rota-
tion constant would have if the nuclei were not vibrating at all, in the manner
of a rigid rotor, with the distance between the nuclei being equal to the equi-
librium distance re, that is, the distance between the nuclei at the bottom of the
anharmonic potential. We will omit discussion of coefficients of higher-order
terms here. Spectroscopists have developed techniques for calculating the values
of these constants from spectra. In published tables of constants, the values given
are typically for the equilibrium values Be, αe, re, ωe, ωexe, ωeye. Note that these
constants are usually given in units of wavenumber, except for re.
Table 8.1 shows the values for some of these constants for a few diatomic
molecules. The three columns on the left show constants for different isotopic
forms of the hydrogen molecule. The first column shows a molecule consisting
of two ordinary hydrogen nuclei, the second column shows a diatomic mole-
cule consisting of an ordinary hydrogen and a deuterium nucleus, and the third
column shows a diatomic molecule consisting of two deuterium nuclei. The two
columns on the right show constants for two isotopic variants of the hydrogen
Source: From Herzberg (1950), except * = from Huber and Herzberg (1979), based on results from
1953 and after.
Realism and Molecular Spectroscopy 167
chloride molecule. There are complications that we are omitting for the moment,
but by 1950, spectroscopists had achieved stable theory-mediated measurement
of parameters that are taken to represent the structure and properties of a large
number of diatomic molecules. As already noted, Herzberg (1950) contains a
table that stretches over 80 pages and lists constants for a huge number of dia-
tomic molecules. These constants vary systematically from one molecule to an-
other, in strict correspondence to differences in their spectra.
than lines. High resolution is needed to resolve the structure. Because the struc-
ture in these bands corresponds to transitions between rotational and vibra-
tional energy states within different electronic states, they contain information
about the rotational and vibrational constants. The electronic spectra thus give
spectroscopists another independent way of determining both the rotational and
vibrational constants.
Further, Raman spectroscopy provides yet another independent way of cal-
culating the rotational and vibrational constants. Raman spectroscopy is done
by shining light of a particular wavelength, usually from a laser, at a sample and
examining the scattered light. When a photon with energy hν′ is incident on a
diatomic molecule, the molecule can go from a lower energy state E″ to a higher
one E′ by taking from the photon an amount of energy ΔE = E′ − E″, and then
170 Historical Cases for the Debate
scattering a photon with an energy hν′− ΔE. On the other hand, the molecule can
go from a higher to a lower state by imparting the energy ΔE to the photon, scat-
tering a photon with an energy hν′+ ΔE. The spectrum of the scattered light will
accordingly consist of a peak at the frequency of the incident light, with lines on
either side of the peak at frequencies given by ν′− (ΔE/h) and ν′ + (ΔE/h). Thus,
Raman spectroscopy provides another way of obtaining the differences between
the various energy levels of the diatomic molecule, from which the rotational
and vibrational constants can be determined. We note here that the spectra we
have previously mentioned involve the absorption and emission of light, whereas
Raman spectra involve scattering, an entirely different phenomenon. Raman
spectra appear in a different frequency range, and the theoretical link between
the observed lines in spectra and the molecular constants is different. In addi-
tion, there are other types of spectroscopy that were developed after the 1950s,
but we omit discussion of them here. Our point is that there is a massive amount
of redundancy in the determination of the molecular constants.
An anti-realist such as van Fraassen would nevertheless claim that redun-
dancy, or concordance in the values of the molecular constants that are measured
using various different kinds of spectroscopy, does not by itself show that these
molecular constants are actually giving us information about the real structure
of molecules. But evidence comes not merely from concordance in these values.
For each different method of determination of molecular constants, the molec-
ular structure is being accessed through a different causal pathway.
The different ways in which the molecular structure can be accessed ap-
pear as details in the molecular spectra that allow spectroscopists to decide
whether they are, indeed, accessing the molecular structure that they claim to be
accessing. Here is one example. The selection rules for rotational spectra imply
that there should be no pure rotation spectrum for homonuclear molecules,
but heteronuclear molecules should produce such a spectrum. This is be-
cause in order for there to be absorption or emission of a photon, there needs
to be a changing dipole moment. Homonuclear molecules always have a di-
pole moment of zero, so they cannot absorb or emit photons through rotation.
Spectroscopists have found that this is indeed the case—no pure rotational
spectra can be found for molecules such as H2, but such spectra can be found
for molecules such as HCl. In Raman spectroscopy, however, there is a different
causal pathway by which spectroscopists claim to be gaining access to the struc-
ture of molecules. The theory of Raman spectroscopy says that there ought to be
rotation lines even for molecules without a dipole moment, and this is indeed
what is found by spectroscopists—the Raman spectra of homonuclear molecules
yields rotational lines.
Spectra are extremely complicated, and they have many more features that
we have not mentioned. Spectra exhibit line splitting, shifts in the positions
Realism and Molecular Spectroscopy 171
of the lines, broadening of the lines, variations in intensity of the lines, and so
on. Without getting into specifics here, we claim that these details can often
be used to check whether the molecular constants really are measuring what
spectroscopists claim they are measuring, and, in fact, examination of those
details can yield further information about either the molecules themselves or
the causal pathways used to probe them.
With regard to the question of what reason there is to think that molec-
ular constants are giving us access to definite physical details of molecules,
and they are not just artifacts, we believe we can offer an impressive number
of different ways of checking whether the interpretation assigned to spectra by
spectroscopists is indeed correct. After having given all of these reasons, how-
ever, we nevertheless hesitate to answer this question affirmatively, because we
think there is still a problem along a different dimension. This is in trying to an-
swer the question of exactly what those molecular constants represent. We have
so far glossed over some details in our description of the molecular constants.
Take the rotation constant B, for example. If a molecule were a classical rigid
rotor, the distance between the nuclei would not vary, and the rotation constant
B could be interpreted straightforwardly as corresponding to the inverse of the
moment of inertia. But when we go beyond the rigid rotor to a vibrating rotator
with an anharmonic potential, the rotation constant is now dependent on the
vibrational quantum number, and its interpretation becomes more complicated.
Usually what is given in tables of constants is Be, which is what the value of the
rotation constant would be if the distance between the two nuclei at all times
was re, the distance between the two nuclei at the bottom of the potential. But
the molecule is never actually in a state with no vibration and hence never in a
state in which the distance re persists for any significant length of time.2 A way of
thinking about what is happening here is that the distance between the nuclei of
the molecule always fluctuates around the re value, and so the rotation constant,
which is dependent on this distance, must be fluctuating as well. So there is some
ground for saying that there is no feature of the actual molecule to which Be cor-
responds, even according to the theory that enables the measurement to be made
in the first place.
To take the point further, we said that the rotational, vibrational, and elec-
tronic energy states of a molecule can be separated. The separability is justified by
the Born-Oppenheimer approximation and the Born adiabatic approximation,
in which the assumption is made that the timescales involved in the motions
of the nuclei are much slower than the timescales involved in the motions of
electrons. The timescale at which electrons interact with one another and the
2 This is the case even in the lowest vibrational energy state, for which the energy is not zero but a
nuclei is of the order of 10–16 seconds. The period for vibrational motion is of the
order of 10–14 seconds, while the period for rotational motion is of the order of
10–11 seconds. Now, consider what this means. Each time the molecule rotates,
it has undergone of the order of a thousand vibrations, and since the rotation
constant depends on the distance between the nuclei, this must mean that the
rotation constant that is being measured can only be a time-averaged value—
it does not correspond to any static or instantaneous state of the molecule. In
other words, the molecular constants are a way of capturing a situation that is
enormously more complicated, with time-averaging smoothing or glossing over
fluctuations of much shorter periods.
Here, we claim that, nevertheless, it may be possible to argue that the mo-
lecular constants such as Be are physically meaningful, even though they do not
represent some actual state of a molecule. We believe that an analogy with ce-
lestial mechanics is helpful here. In celestial mechanics the orbits are character-
ized in terms of Keplerian parameters defining an ellipse and area swept out per
unit of time. These parameters define a counterfactual situation—they describe
the elliptical trajectory that would occur if there were only two bodies. To quote
Newton, “if the sun were at rest and the remaining planets did not act on one an-
other, their orbits would be elliptical, having the sun at a common focus, and they
would describe areas proportional to the times” (Newton 1999, p. 817f, emphasis
added). We contend that the Keplerian parameters, notwithstanding their coun-
terfactual status, are, even from a modern perspective, physically meaningful.
First, an examination of the 250-year history of celestial mechanics be-
tween the times of Newton and Einstein shows that deviations between
predictions made based on calculations from these orbital parameters, and ac-
tual observations of planetary motions, have been used to make inferences about
the existence and location of other bodies that might be influencing the motions
(Smith 2014). In other words, the Keplerian representation served as a research
instrument to expose physical sources of deviations from this representation.
The most famous example of this was the discovery of Neptune, but this one ex-
ample grossly understates how much was discovered about the motions in our
planetary system from this approach.
Second, at the time Newton began work on what became his Principia, he was
aware of four alternatives to the Keplerian representation that fit the observations
just as well as it did. Indeed, Newton was more familiar with both the planetary
theories, and the tables of their motions produced by Thomas Streete (1661),
Vincent Wing (1669), and Nicholaus Mercator (1676), than he was with any of
Kepler’s writings. These three employed three different alternatives to Kepler’s
rule that orbiting bodies sweep out equal areas in equal time, a rule which
Newton likely had learned of from reading Mercator’s systematic defense of his
approach over Kepler’s and Streete’s (Mercator 1676, pp. 144–164). A few pages
Realism and Molecular Spectroscopy 173
in the Principia after the counterfactual claim we just quoted, Newton went on to
show that the then most prominent violation of Kepler’s rule, namely the motion
of our moon, results from the action of solar gravity alternately accelerating and
decelerating the moon twice over the course of its monthly orbit, in the process
markedly distorting that orbit in a way theretofore unrecognized (Newton 1999,
pp. 840–848). In thus identifying a robust physical source for the deviation of the
moon from the area rule, Newton gave evidence supporting the counterfactual
claim that the moon would conform to the rule at least more closely than it does,
and the orbit would be closer to a Keplerian ellipse, were it not for the perturbing
effect of the sun.
This is but one of hundreds of examples from celestial mechanics in which the
Keplerian representation of orbital motion, counterfactual though it is, revealed
deviations that were then shown to have physically identifiable sources. The pre-
Newton alternatives to Kepler’s representation remind us that there can always
be many other comparably accurate representations, but not all of them would
yield deviations meeting the demand of having identifiable robust physical
sources. We thus distinguish between physically meaningful representations and
parameterizations and those that are not through the following criterion: physi-
cally meaningful representations and parameterizations yield deviations that have
physically identifiable sources.
3 The two specific papers to which we are referring are Todd S. Rose, Mark J. Rosker, and Ahmed
H. Zewail, “Femtosecond Real-Time Probing of Reactions. IV. The Reactions of Alkali Halides,”
Journal of Chemical Physics, vol. 91, 1989, pp. 7415–7436; and P. Cong, A. Mokhtari, and A. H.
Zewail, “Femtosecond Probing of Persistent Wave Packet Motion in Dissociative Reactions: Up to 40
ps,” Chemical Physics Letters 172, 1990, pp. 109–113.
176 Historical Cases for the Debate
function does make them a possible source for such support that needs to be
examined.
We have been emphasizing research after 1950 as the most instructive way
for comparing the physical meaningfulness of the diatomic molecular constants
with that of the Keplerian orbital parameters. Evidence supporting such a par-
allel between the two had emerged even before 1930. Table 8.1 shows the con-
sistency of the values of the molecular constants obtained from spectra with the
change of the isotopic mass of the hydrogen nucleus. The spectroscopic search
that resulted in the 1931 discovery of deuterium was prompted by the discovery
two years earlier of first one and then a second isotope of oxygen besides the
16O that had previously been used, under the assumption of its being unique, as
the basis for the system of atomic weights. Specifically, long-exposure absorp-
tion spectra of air had yielded an anomalous band spectrum of faint lines shifted
slightly in frequency from a long-established band spectrum. Analysis showed
that the shift corresponds to a 16O18O molecule under the assumption of a single
time-averaged potential function (Giaque and Johnston 1929a, 1929b). Still
more extended time exposure then revealed as well an even more rare 17O iso-
tope (Giaque and Johnston 1929c and 1929d). Band spectra of other molecules
involving oxygen, such as NO, subsequently independently confirmed both
(Naudé 1930), as did mass spectrometry (Aston 1932).
The parallel between the discovery of Neptune from anomalies in the motion
of Uranus and the discovery of 18O from an anomalous band in the spectrum
of oxygen extends beyond this. Neptune was widely heralded at the time as ev-
idence in support of Newtonian gravity. So too was the discovery of 18O for the
selection rules of the new quantum theory:
The evidence here involved isotopic shifts only in the time-averaged values of
the respective molecular constants, but some subtleties in these shifts were at
least attributed back then to short-time interactions between electrons and nu-
clei ignored in the time-averaging (Van Vleck 1936).
Let us return to the beginning of this paper. Realists have tried to respond to
van Fraassen by showing how, in the case of Perrin, concordant measurements of
parameters such as Avogadro’s number can be achieved and examining what this
concordance of measurements is telling us about the reality of molecules. The
measurements in question dated from 1908 to 1913. Whatever support for the re-
ality of molecules resulted from them at the time, if research over the subsequent
Realism and Molecular Spectroscopy 177
decades had failed to determine anything more about the structure and shape of
molecules beyond the spherical hypothesis, this would have rendered the con-
clusion that molecules are real to be of, at best, limited empirical significance.
In fact, however, the rotational and vibrational spectra of diatomic molecules,
and in some cases of polyatomic molecules as well, have yielded seemingly im-
pressive conclusions about detailed specifics of their structure and shape. We
say “seemingly” because the values of the constants characterizing their struc-
ture and shape involve a time-averaging that glosses over extremely complicated
short-term fluctuations in their structure. From the point of view of the realist-
instrumentalist debate, the question therefore arises whether these constants are
physically meaningful (in the sense that their values have causal implications)
or whether to the contrary they amount to nothing more than a heuristically
useful representation of the spectral data with no physical significance beyond
these data. We claim that the question of the physical meaningfulness of these
constants is now a far more appropriate focus for the realist-instrumentalist de-
bate than are the concordant measurements of the 1908–1913 period.
We can put this point in a slightly different way that might be more helpful.
We believe van Fraassen (2009) is right to emphasize that the primary aim of
researchers in microphysics in the early twentieth century was making stable
concordant measurement, or empirical grounding, to use van Fraassen’s term,
possible for microphysical parameters. Where we differ from van Fraassen is in
how we view the development of evidence after stable concordant measurement
has been achieved. In all the cases we know of from the history of the physical
sciences, scientists have continued to pursue discrepancies from concordant
measurements, no matter how small they may be, for these discrepancies have
always led to discoveries of new physical sources that had, in some cases, not
even been considered before. This methodology requires a distinction to be
made between representations that yield deviations that have physically identi-
fiable sources, and those that do not yield such deviations, which we term phys-
ical meaningfulness. We believe that a potentially important step forward in the
realist-instrumentalist debate can be achieved by focusing on the questions of
how exactly to characterize physical meaningfulness and how it might be pos-
sible to distinguish parameters that are physically meaningful from those that are
not, even as one must acknowledge that they do not correspond to any situation
that actually occurs.
Focusing the realism-instrumentalism debate on the question of the physical
meaningfulness of the parameters in the standard representation of diatomic
molecules has some virtues beyond those we have stressed earlier. As we have
provisionally drawn the distinction, whether a parameterization is physically
meaningful turns on how strong the evidence is in support of a specific phys-
ical source for each systematic discrepancy between theoretical calculation and
178 Historical Cases for the Debate
References
Achinstein, Peter. 2001. The Book of Evidence. Oxford: Oxford University Press.
Aston, F. W. 1932. “Mass-Spectra of Helium and Oxygen,” Nature, vol. 130, p. 21f.
Bernath, Peter F. 2016. Spectra of Atoms and Molecules, Third Edition. Oxford: Oxford
University Press.
Bernath, Peter F., and Sean McLeod. 2001. “DiRef, A Database of References Associated
with the Spectra of Diatomic Molecules,” Journal of Molecular Spectroscopy, vol. 207,
p. 287.
Bohr, Niels. 1913. “On the Constitution of Atoms and Molecules,” Philosophical Magazine
Series 6, vol. 26, pp. 1–25.
Boltwood, Bertram B., and Ernest Rutherford. 1911. “Production of Helium by Radium,”
Philosophical Magazine Series 6, vol. 22, pp. 586–604.
Born, Max, and Robert Oppenheimer. 1927. “On the Quantum Theory of Molecules,”
tr. H. Hettema, in H. Hettema, Quantum Chemistry: Classic Scientific Papers.
Singapore: World Scientific, 2000, pp. 1–24.
4 Part of the research for this paper was supported by the Ministry of Education, Singapore,
under its Academic Research Fund Tier 1, No. RG156/18, as well as the Radcliffe Institute for
Advanced Study.
Realism and Molecular Spectroscopy 179
Brown, John, and Alan Carrington. 2003. Rotational Spectroscopy of Diatomic Molecules.
Cambridge: Cambridge University Press.
Chalmers, Alan. 2009. The Scientist’s Atom and the Philosopher’s Stone: How Science
Succeeded and Philosophy Failed to Gain Knowledge of Atoms. Dordrecht: Springer.
Chalmers, Alan. 2011. “Drawing Philosophical Lessons from Perrin’s Experiments on
Brownian Motion: A Response to van Fraassen,” British Journal for the Philosophy of
Science, vol. 62, pp. 711–732.
Chaudesaigues, M. 1908. “Le mouvement brownien et le formule d’Einstein,” Comptes
Rendus, vol. 147, pp. 1044–1046.
Giauque, W. F., and H. L Johnston.1929a. “An Isotope of Oxygen, Mass 18,” Nature, vol.
123, p. 318.
Giauque, W. F., and H. L Johnston.1929b. “An Isotope of Oxygen, Mass 18. Interpretation
of the Atmospheric Absorption Bands,” Journal of the American Chemical Society, vol.
51, pp. 1436–1441.
Giauque, W. F., and H. L Johnston.1929c. “An Isotope of Oxygen of Mass 17 in the Earth’s
Atmosphere,” Nature, vol. 123, p. 831.
Giauque, W. F., and H. L Johnston.1929d. “An Isotope of Oxygen, Mass 17, in the Earth’s
Atmosphere,” Journal of the American Chemical Society, vol. 51, pp. 3528–3534.
Glymour, Clark. 1980. Theory and Evidence. Princeton: Princeton University Press.
Hecht, Charles E. 1990. Statistical Thermodynamics and Kinetic Theory. Mineola,
NY: Dover.
Herzberg, Gerhard. 1939. Molecular Spectra and Molecular Structure, I. Diatomic
Molecules. New York: Prentice-Hall.
Herzberg, Gerhard. 1950. Molecular Spectra and Molecular Structure: I. Spectra of
Diatomic Molecules, Second Edition. Princeton: D. Van Nostrand.
Herzberg, Gerhard. 1971. The Spectra and Structures of Free Simple Radicals: An
Introduction to Molecular Spectroscopy. Mineola, NY: Dover.
Huber, Klaus Peter and Gerhard Herzberg. 1979. Molecular Spectra and Molecular
Structure: IV Constants of Diatomic Molecules. New York: Van Nostrand Reinhold.
Jeans, James. 1916. The Dynamical Theory of Gases 2nd edition. Cambridge,
UK: Cambridge University Press.
Kelvin, Lord. 1904. Baltimore Lectures on Molecular Dynamics and the Wave Theory of
Light. Cambridge: Cambridge University Press.
Lefebvre-Brion, Hélène, and Robert W. Field. 2004. The Spectra and Dynamics of Diatomic
Molecules. Amsterdam: Elsevier Academic Press.
Mayo, Deborah G. 1996. Error and the Growth of Experimental Knowledge.
Chicago: University of Chicago Press.
Mercator, Nicholaus. 1676. Institutionum Astronomicarum Libri Duo, De Motu Astrorum
Communi & Proprio. London: Samuel Simpson.
Meyer, Oskar Emil. 1899. The Kinetic Theory of Gases. tr. Robert E. Baynes.
London: Longmans, Green; Die Kinetische Theorie der Gase. Breslau: Marschke &
Berendt, 1899.
Millikan, Robert A. 1911. “The Isolation of an Ion, A Precision Measurement of Its
Charge, and the Correction of Stokes’s Law,” The Physical Review, vol. 32, pp. 349–397.
Millikan, Robert A. 1913. “On the Elementary Electrical Charge and the Avogadro
Constant,” The Physical Review, Second Series, vol. II, pp. 109–143.
Naudé, S. Meiring. 1930. “The Isotopes of Nitrogen, Mass 15, and Oxygen, Mass 18 and
17, and their Abundances,” Physical Review, vol. 36, pp. 333–346.
180 Historical Cases for the Debate
Scientific realists use the “No Miracle Argument” (NMA): it would be a miracle
if theories were false, yet got right so many novel and risky predictions. Hence,
predictively successful theories are true. Of course, one could easily make up a
theory with completely false theoretical assumptions which predicted a phenom-
enon P (call it a FP-theory) if she knew P in advance and used it in framing the
theory. But how could she think of a FP-theory, without knowing P? Or knowing
P but without using it in building the theory? In fact, it is puzzling how one could
have built a FP-theory even if she used P inessentially: suppose Jill built a FP-
theory by knowing and using P, but she could have done without it, because, quite
independently of her, John built the same theory without using P. This I call Jill
using P inessentially, and it is something hard to explain, because it is under-
standable how the theory was built by Jill, but not by John (Lipton 1991, 166; Alai
2014c, 301).
So, how could one find a FP-theory without using P essentially? Well, one
could do this if P were very poor in content, i.e., very probable: any false theory
could predict the existence of some new planet somewhere in the universe. But
it took Newton’s theory to predict precisely the existence of a planet of such and
such mass with such and such orbit as Neptune. So, the riskier a prediction is (i.e.,
the more precise, rich in content, hence improbable it is), the less likely it is that it
has been made by a completely false theory.
It might be objected that since a falsity may entail a truth, even false theo-
ries might make novel predictions. Granted, in the Hyperuranion of possible
theories there are some with completely false theoretical assumptions which
Mario Alai, The Historical Challenge to Realism and Essential Deployment In: Contemporary Scientific Realism.
Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press 2021.
DOI: 10.1093/oso/9780190946814.003.0009
184 Contemporary Scientific Realism
(together with appropriate auxiliary assumptions) make true novel and improb-
able predictions (call them FNIP-theories). However, their rate is very small, for
it is a logical fact that the less probable a consequence is, the fewer the premises
entailing it. Hence, it is extremely improbable that FNIP-theories are found by
chance. But since scientists look for true theories, only by chance can they pick a
FNIP-theory. Therefore, it is extremely unlikely that scientists find a FNIP-theory.1
For instance, quantum electrodynamics predicted the magnetic moment of the
electron to be 1159652359 × 10–12, while experiments found 1159652410 × 10–12:
hence John Wright (2002, 143–144) figured that the probability to get such an
accuracy by chance, i.e., through a false theory, is as low as 5 × 10–8.
Here (following Alai 2014c) I will speak of “novel predictions” (NP) when the
predicted phenomenon is either unknown (historically novel) or not used (use-
novel) or not used essentially (functionally novel) by the theorist. So, it is prac-
tically impossible that a novel prediction which in addition is risky, i.e., a highly
improbable, was derived from a completely false theory.
Against the claim that success warrants truth, Laudan (1981) had objected that
many false theories in the history of science were successful, so realists replied
that however those theories didn’t have novel predictive success.2 Unfortunately,
it turns out that also many false theories had novel success (see the list in Lyons
2002, 70–72). Selective realists acknowledged the problem and argued that the
NMA does not warrant the complete truth of theories, but only their partial truth.
In particular, for Kitcher’s (1993) and Psillos’s (1999) “Deployment Realism”
(DR), we should be committed only to those particular hypotheses which were
deployed in deriving novel predictions: this is Psillos’s “divide et impera” move
against Laudan’s “reductio” (Psillos 1999, 102, 108, passim).
A rejoinder came from Timothy Lyons: first, he explained that Laudan’s orig-
inal objection was not “an articulation of the pessimistic meta-induction toward
the conclusion that our present theories are probably false,” but a modus tollens
refutation of the NMA (a “meta-modus tollens”) (Lyons 2002, 64–65; 2006, 557).
Secondly, he criticized DR by listing a number of false hypotheses which were ac-
tually deployed in the derivation of various novel risky predictions—including
Neptune’s discovery (2002; 79–83; 2006). Deployment realists, however, can an-
swer this new challenge by arguing that the NMA does not warrant the truth
of all hypotheses actually used in deriving novel predictions, but only of those
which were essential, i.e., strictly necessary to those predictions. False hypoth-
eses are no counterexample to the NMA if they were used de facto, but actually
superfluous, i.e., non-essential.
1 Alai 2014c, §1; 2014d, §5; 2016, §4. More on this objection at §9.7.1.
2 Musgrave (1985, 1988); Lipton (1993, 1994); Psillos (1999); Sankey (2001).
Historical Challenge to Realism and Essential Deployment 185
Thus, essentiality is crucial to two key claims of DR: (1) a predicted phenom-
enon is novel only if it was not used essentially in building the theory which
predicts it; and (2) a hypothesis is most probably true only if it was essential in
deriving a novel prediction. Here we shall focus on the latter claim.
According to Psillos (1999, 110) a hypothesis H is deployed essentially in
deriving a novel prediction NP if
(1) NP follows from H, together with some other hypotheses OHs and auxil-
iary assumptions AA, but not from OHs & AA alone;
(2) no other hypothesis H* is available which can do the same job as H, viz. is
(a) compatible with OHs and AA,
(b) non-ad hoc,
(c) potentially explanatory, and
(d) together with OHs and AA predicts NP.
Conditions (1) and (2) capture the idea that H is essential, i.e., strictly necessary,
when (1) OHs without H could not predict NP, and (2) no alternative hypothesis
is available which could substitute H in the derivation of NP.
didn’t understand his conditions for essentiality as a definition, but just as a crite-
rion: he writes “H indispensably contributes to the generation of P if . . . [(1) and
(2) are met],” not “if and only if . . . [(1) and (2) are met]” (1999, 110). Yet, Lyons’s
criticisms are at least prima facie well-founded, and must be taken seriously.
But Lyons also claims that his criticisms force deployment realists to
abandon the essentiality requirement altogether. Besides, he claims that the
“crucial” or “fundamental” insight of DR is that we should credit “those and
only those constituents that were genuinely responsible for, that actually led
scientists to, specific predictions” (2006, 543), no matter whether they were
indispensable or not. In fact, he argues that Psillos, after introducing his cri-
terion of essentiality (Psillos 1999, 109–110), “never mentions that rigid crite-
rion again”3 and fails to show that his case studies comply with it (Lyons 2009,
143). On the contrary, according to Lyons, Psillos often talks of commitment
to constituents which simply “generated,” or “brought about,” or were “respon-
sible for,” or “genuinely contributed to,” or “really fueled” (Psillos 1999, 108–
110) the success of the theory (Lyons 2006, 540). Moreover, Lyons says that
the same “fundamental insight” is also embraced, in some form or another, by
Kitcher, Niiniluoto, Leplin, and Sankey (538). Thus, he holds, DR should get
rid of the essentiality requirement and “return to the crucial insight” of actual
deployment (543).
At this point, however, Lyons argues that many assumptions which in his-
tory were actually employed and led to novel predictions are false. In fact, in
various papers (2002, 2006, 2016, 2017) he discusses a number of striking novel
predictions, toward each of which many false assumptions were employed (we
shall briefly review some of them later). Hence, he concludes the NMA does not
work even when restricted to particular hypotheses, and DR is false: “If we take
seriously the no miracles argument [i.e., if novel predictions derived from false
assumptions are a miracle] . . . we have witnessed numerous miracles” in history
(2006, 557).
However, in reply, a number of points should be noticed: first, by accepting
such purported “miracles,” one is left with the problem of explaining how they
are possible. If one wonders how scientist S made the novel risky prediction NP
which was then confirmed, a possible answer is that NP followed from true
assumptions in S’s theory, for all consequences of true assumptions are true. But
one cannot answer that NP followed from some false hypothesis in S’s theory,
because that would entail that S had luckily assumed a false hypothesis entailing
NP, and as mentioned, this is extremely improbable.4 Lyons (2002; 2003) and
others proposed alternative explanations, not involving the truth of the deployed
hypotheses, but I have argued elsewhere that they all fail (Alai 2012, 88–89;
2014a).
Secondly, as concerns which one is the “fundamental insight” of DR that
cannot be abandoned vs. what is merely accessory and should be dropped,
two questions must be distinguished: (I) the exegetic question whether Psillos
understands “deployment” as merely actual deployment, or as essential deploy-
ment; and (II) the theoretical question of how deployment should be understood
if we are to give DR its best currency in order to correctly assess its tenability.
As to (I), when Psillos claims that we should credit constituents that are “re-
sponsible” for success (1999, 108–109), Lyons understands “responsibility” as ac-
tual deployment, suggesting that this was the original core intuition of DR (2006,
540, 543). But this is incorrect. As Lyons himself acknowledges (2006, 539) on
the next page Psillos specifies that constituents to be credited are those “that
made essential contributions” to success and that “have an indispensable role” in
it, and right after this he spells out his conditions (1)–(2) (109–110). So, by re-
sponsibility Psillos clearly means essential responsibility.
Even apart from this, when an author says two different things, in the exegesis we
are allowed to discount one of them only if they are contradictory. But if they are not
contradictory we must keep both, and understand how they can be reconciled. In
this case, sometimes Psillos talks of responsibility or contribution, or deployment,
etc., without further qualifications, sometimes he introduces the essentiality or in-
dispensability qualification. But these two talks are not contradictory, and can be
easily reconciled, since the latter is just a specification of the former. In fact, when
talking plainly of responsibility or contribution, Psillos never adds “no matter
whether essential or not” or the like. Therefore, even his plain talk of responsibility
or contribution must be understood as implying “essential” or “indispensable.”
Even if in his case studies (1999, ch. 6) Psillos fails to explicitly refer back to
his criterion of essentiality and to explicitly show how it applies, he does use the
essentiality requirement. For instance, he argues that we need not be committed
to the caloric hypothesis, although it was actually deployed in Laplace’s predic-
tion of the speed of sound in air, because that prediction “did not depend on
this hypothesis” (121). Hence, he obviously would not give up his essentiality
requirement. Moreover, in the absence of any indication to the contrary, and
as this quotation suggests, we must assume that he understands the essentiality
requirement precisely in terms of his essentiality criterion.
Summing up, from the exegetic point of view, ignoring the essentiality re-
quirement would be misinterpreting Psillos’s position. But apart from this, since
here we must assess the tenability of DR, what matters more is not the exeget-
ical question, but (II) the theoretical question of how DR should be understood
in order to give it the most favorable interpretation. Now again, the answer is
188 Contemporary Scientific Realism
H does not merit realist commitment whenever H is doing work in the deri-
vation solely in virtue of the fact that it entails some other proposition which
itself is sufficient, when combined with the other assumptions in play, for the
relevant derivational step (2016, §3).
(2’) Th
ere is no other hypothesis H* which is proper part of H (hence weaker
than H) which together with OHs and AA entails NP.
Here in section 9.3 I explain (2’), spelling out the intuitive notion of indispen-
sability more precisely through Yablo’s (2014) concept of proper parthood, and
arguing that the troublesome requirements of Psillos’s criterion can be dropped
Historical Challenge to Realism and Essential Deployment 189
without loss, thus escaping Lyons’s criticisms; in section 9.4 I show how my crite-
rion immunizes DR from Lyons’s historical counterexamples; in section 9.5 I hold
that essentiality in this sense cannot be detected prospectively, and Vickers’s op-
timism on the identifiability of non-essentiality should be complemented by
some less optimistic considerations; but far from causing problems for realists,
this frees them from unreasonable obligations; in section 9.6 I explain in more
details why my criterion is enough, although apparently weaker than Psillos’s; in
section 9.7 I answer some actual and potential objections.
Intuitively, a hypothesis H* is a proper part of H iff its content is part of
the content of H but does not exhaust it. At first approximation and for most
purposes, a part of H is therefore a hypothesis entailed by H but not entailing it.
But this is not always enough: for instance consider
and
As (i) is entailed by (ii) without entailing it, (i) is a proper part of (ii). But
is implied by (ii) without entailing it, however it is not a (proper) part of it, for
it brings in a different subject matter. Therefore, just characterizing parthood
in terms of entailment will not always do for our purposes. Instead, we must
understand parthood as Yablo (2014):5 H* is a part of H iff it is entailed by H
and preserves its subject matter. The subject matter of a proposition consists
of two classes: that of its truthmakers (or reasons why it is true) and that of its
falsemakers (or reasons why it is false). A proposition H* entailed by H preserves
H’s subject matter, hence, is part of it, iff every truthmaker/falsemaker of H* is
entailed by a truthmaker/falsemaker of H. Moreover, H* is a proper part of H iff
every truthmaker/falsemaker of H* is entailed by a truthmaker/falsemaker of H,
but not vice versa. This explains the intuitive idea that (iii), although entailed by
(ii), is not part of it: because some truthmakers of (iii) (e.g., “it’s raining”) are not
entailed by any truthmaker of (ii).6
5 I thank Matteo Plebani for pointing this out to me. See Plebani (2017).
6 Although Yablo’s approach helps us to properly explain essentiality, it was not designed to this
purpose, but to answer the general question of “aboutness,” i.e., the subject-matter of sentences: why
and how sentences with the same truth-conditions say different things, like “here is a sofa” and “here
is the front of a currently existing sofa, and behind it is the back.” Thus, it offers a natural solution
to a wide variety of problems concerning agreement, orders, possibilities, priority, explanation,
190 Contemporary Scientific Realism
(H) inside each massive body there resides a demi-god, which attracts the
demi-gods dwelling in each other body by a force F = Gm1m2/r2.
H would predict the same novel phenomena as Newton’s actual theory, but we
wouldn’t need to believe it, because it is redundant, i.e., not essential to those
predictions. In fact, H consists in the following assumptions:
knowledge, partial truth, and confirmation. For instance, a white marble does not confirm (1) “All
crows are black” because (1) does not say the same as (2) “All non-black things are non-crows,” and
Yablo explains why (2014, pp. xiv, 7–8).
7 Whether they are or not can be decided in the same way as for H. In section 9.6 I shall notice that
different but compatible assumptions may all be essential for a given novel prediction.
Historical Challenge to Realism and Essential Deployment 191
most actual cases alternative hypotheses are not compatible with OHs and AA.
My condition (2’) escapes this problem, since it excludes, without exceptions, all
the alternative hypotheses which are part of H and can derive NP. In fact, in sec-
tion 9.4 I shall mention some historical examples of hypotheses which violated
condition (2’), confirming that it is not empty.
Yet, it might seem that here lurks a further difficult problem: as just said, all
the alternative hypotheses excluded by (2’) are compatible with OHs and AA,
since they are part of H. But how about those alternative hypotheses which are
not compatible with OHs and AA, because they are not part of H? Lyons stresses
that they exist in many historical cases, and if so, don’t they make the original
hypothesis H superfluous, i.e., non-essential? Why doesn’t my criterion exclude
such hypotheses?
In other words, Psillos’s condition (2) for essentiality is of medium strength: it
excludes all the hypotheses which can derive NP when joined with OHs and AA.
Lyons’s criticism is that it is too weak, and his implicit suggestion is that it should
be strengthened to exclude even the hypotheses which can derive NP in con-
junction with different collateral assumptions. My condition (2’), instead, is even
weaker than Psillos’s (for it excludes only a subset of the alternative hypotheses
excluded by (2), i.e., only those which are part of H), and this is why it escapes
Lyons’s first criticism: how can it be enough? In particular, shouldn’t we require
that H is unique, i.e., that there is no alternative hypothesis H’ incompatible
with H, which (no matter whether compatible and joined with OHs and AA,
or incompatible with them and joined with different hypotheses and auxiliary
assumptions) can derive NP? For if it existed, we couldn’t know which of them
is true.
I answer that we don’t need to explicitly exclude any other H’ except those
which are part of H (so, we don’t need Psillos’s condition (2) with its drawbacks,
nor its strengthening implicitly suggested by Lyons), because the required
uniqueness of H is already ensured by condition (2’) together with the risky char-
acter of the prediction NP. In fact, just as it happens for theories, it is practically
impossible to find a completely false hypothesis making risky novel predictions.
A false hypothesis H may yield a risky novel prediction NP, only if it is partly
true: i.e., if it has a true part H* from which NP can be derived. But in this case,
the essential role in the derivation of NP is played by H*, H being inessential by
my condition (2’).
This much we know a priori from the NMA (as spelled out in the five ini-
tial paragraphs of section 9.1), and anti-realists have not been able to spot any
inferential fallacy in that argument. However, they challenged it by apparent
counterexamples, i.e., by citing historical instances of false hypotheses which
were supposedly deployed essentially in deriving novel predictions. But in each
of those cases it can be shown that either (a) those hypotheses were not actually
192 Contemporary Scientific Realism
false or the predictions not actually true, or (b) the predictions were not actually
novel and risky, or (c) they were not deployed essentially (Alai 2014b).
Condition (2’) is designed precisely to exclude cases of type (c), and I shall
offer six examples in section 9.4 and three in section 9.5. Therefore, if H fulfills
(2’), it must be completely true. But if so, we don’t need in addition to exclude
that there is any incompatible assumption H’ from which NP could be essentially
derived, because any H’ incompatible with H would be (at least partly) false,
hence it could not (except by a miracle) be essential in deriving NP: H’ could
play a role in deriving NP only if it had a completely true part H’* sufficient to
derive NP, and obviously H’* would be compatible with H, since both would be
completely true.
This is a nutshell account of why my condition (2’) is enough to ensure the
required uniqueness of H, but I shall explain it in more detail in section 9.6. This
account, anyway, entails an empirical claim: in the history of science there have
never been a couple of incompatible hypotheses both essential in deriving the
same novel risky prediction. So, my proposal is empirically testable: it will be fal-
sified if any such couple is found, but confirmed if it is not.
Here is how my condition (2’) rules out as inessential five false assumptions
which according to Lyons were deployed to derive novel predictions (the first
three are more extensively discussed in Alai 2014b, §7). Three further examples
will be discussed in section 9.5. Yet another striking case is Arnold Sommerfeld’s
1916 prediction of the fine structure energy levels of hydrogen: for decades it was
thought to be derived from utterly false assumptions, and so considered “the ul-
timate historical challenge” to the NMA, for even physicists called Sommerfeld’s
success a “miracle.” But Vickers (2018) shows that the wrong assumptions were
not essential in deriving it, since the true part of Sommerfeld’s premises was
enough.
9.4.1 Dalton
Lyons argues that Dalton derived his true Law of Multiple Proportions (LMP)
from his false Principle of Simplicity:
PS: “Where two elements A and B form only one compound, its compound
atom contains 1 atom of A and 1 of B. If a second compound exists, its atoms
Historical Challenge to Realism and Essential Deployment 193
But this false PS has a true part: a weaker principle we might call “of Multiple
Quantities”:
LMP: The weights of one element that combine with a fixed weight of the other
are in a ratio of small whole numbers.
Therefore, although actually employed in deriving LMP, the false PS was not es-
sential, since its true part PMQ was sufficient to derive LMP.
9.4.2 Mendeleev
Mendeleev derived his predictions of new elements from his false Periodic Law:
Now, Q is a part of PL, true, and sufficient to derive Mendeleev’ predictions. So,
although actually employed, PL was not essential, while the essential assump-
tion, Q, was true.
9.4.3 Caloric
Lyons (2002, 80) points out that the caloric theory truly predicted that
However, these false claims were not essential in predicting ER. In fact, each of
them includes a true part not referring to caloric, from which ER could still be
derived, respectively
So, while the false (1)–(6) were used, they were not essential, since their parts
(1*)–(6*) are sufficient to derive ER (and true, as far as we know).
9.4.4 Kepler
(#1) the planets tend to rest, and they move only when forced to move;
(#2) the Sun has an Anima Motrix which spins it around its axis;
(#3) this spinning is transmitted to the planets through the Sun’s rays, which
push the planets in their orbits (thus the Sun exerts on the planets a di-
rective rotational force, not an attractive one);
(#4) the force of the Sun’s rays is inversely proportional to the distance from
the Sun, like the intensity of their light.
The model of the universe described by (#1)–(#4) is crazy in the light of our
mechanics (#1), metaphysics (#2), and physics (#3), but it was not in Kepler’s
time. Moreover, it was a natural abductive conclusion from two facts known at
that time:
(a) the planets move around the Sun approximately on the same plane and in
the same wise;
(b) the order of their velocities is inverse to that of their distances from
the Sun.9
9 This was already clear to Copernicus, who knew the times of the orbits and the relative distances
from the Sun, hence the relative lengths of the orbits. Kepler also knew that the times grow more than
the lengths, hence the velocity diminishes as the distance from the Sun increases. This suggests that
the planets be driven by a force emanating from the Sun, so varying inversely with the distance from
it (Koyré 1961, II, §1). Kepler thought that the light varies inversely with the distance from its source
(rather than with its square), hence he thought the Sun’s motrix force did the same.
196 Contemporary Scientific Realism
he had been more cautious in extrapolating from (a) and (b); but unfortunately
there were no technological analogies, like the sling, the wool winder, etc., to
suggest to him such a model. This true core is the following:
(#5) the solar system moves around the centre of the Sun as a coherent but
non-rigid disk;
(#6) its coherence is mainly due to a force exerted by the Sun on the planets
(Kepler thought it was a directive rotational force, and the only one in
play; while we know it is the attractive gravitational force, supplemented
by the planets’ own attractive forces and the effects of inertia);
(#7) the velocities of the planets are also proportional to the same force
(Kepler thought this happens because that force is responsible for their
motion, by overcoming their tendency to rest; instead we know this
happens because their motion is due to their tangential inertia, which
must be equal to that (gravitational) force, otherwise the planet would
either fly off on its tangent, or collapse on the Sun);
(#8) that force is in an inverse relation with the distance from the Sun (for
Kepler that relation was F=1/d, we know it is F=1/d2).
10 Some of which are mentioned within parentheses after (#6), (#7), and (#8) respectively.
Historical Challenge to Realism and Essential Deployment 197
enough to generate (P1)–(P3); therefore the model (#1)–(#4) was not essential
by condition (2’), hence the NMA does not commit to it.
For the sake of simplicity, in (#6), (#7), and (#8) I spoke of a force, as if Newton’s
gravitational theory were true; but, one could object, General Relativity shows
that there are no gravitational forces, and the planets’ motion is due to the curva-
ture of spacetime caused by the Sun. The objection is fair, but it only shows that
we must circumscribe more accurately the true core of Kepler’s model, which we
can do by generalizing (#6), (#7), and (#8) as follows:
(#5) the solar system moves around the centre of the Sun as a coherent but
not rigid disk;
(#6*) its coherence is due to an effect (i.e., for us, the curvature of spacetime)
which is mainly brought about by the Sun;
(#7*) the velocities of the planets are also proportional to the same effect;
(#8*) that effect is in an inverse relation with the distance from the Sun (hence
so are the planetary velocities).
This restricted core is shared by Kepler, Newton, and Einstein and is still
sufficient to derive Kepler’s novel predictions. But what if someday General
Relativity is replaced by a better theory and this by another one, etc.?11 If we
constantly weaken our hypotheses to keep them compatible with succes-
sive theories, won’t we reduce them to mere descriptions of observed phe-
nomena, and scientific realism to empiricism?12 I don’t think this pessimism is
warranted, because history shows that each successive theory T’, while giving
up part of the theoretical content of the earlier theory T, adds some new theo-
retical content. Even if in turn T’ is superseded by T” and some of its content
is given up, it is quite possible that (a) some of the original content of T is pre-
served even in T”, and (b) the theoretical content of T’ preserved in T” is even
larger than that of T preserved in T’.
For instance, in the case discussed here, the directive force has been replaced
by the attractive force, and this by a curvature of spacetime, but all of them are
unobservable entities or properties. Each theorist while discarding some theo-
retical assumptions adds some new ones (Newton adds inertia and gravitation
force, Einstein spacetime curvature). Tomorrow it might no longer be spacetime
curvature, but something else; but whatever it is, the Sun will have a major role in
it, and it will have an inverse relation with the distance from the Sun and a direct
relation with the velocities of the planets. Besides, each new model is probably
closer to the truth, because it has a better fit to the data and greater predictive
11 I deny that our current theories are completely true and will never be rejected: Alai (2017).
12 I owe this objection to an anonymous reviewer.
198 Contemporary Scientific Realism
power. Thus our theoretical knowledge (our theoretical true justified beliefs)
increases over time. While acknowledging that our beliefs are not yet “the whole
truth and nothing but the truth,” we can trust to be on the right track. I shall fur-
ther discuss similar worries in §§9.7.3–9.7.5.
9.4.5 Neptune
Finally, Lyons claims that the following false claims were used in the prediction
of Neptune (2006, 554):
(1) the sun is a divine being and/or the center of the universe (Kepler);
(2) the natural state of the planets is rest;
(3) a non-attractive emanation coming from the sun pushes the planets in
their paths;
(4) the planets have an inclination to rest, which resists the solar push, and
contributes to their slowing speed when more distant from the sun;
(5) the planets are pushed by a “directive” magnetic force;
(6) there exists only a single planet and a sun in the universe (Newton);
(7) each body possesses an innate force, which, without impediment,
propels it in a straight line infinitely;
(8) between any two bodies there exists an instantaneous action-at-a-
distance attractive force;
(9) the planet just beyond Uranus has a mass of 35.7 earth masses
(Leverrier)/50 earth masses (Adams);
(10) that planet has an eccentricity 0.10761 (Leverrier)/0.120615 (Adams);
(11) the longitude of that planet’s perihelion is 284°, 45’ (Leverrier)/299°, 11’
(Adams), etc.
However, (1)–(5) were no longer held in the 18th century. They were involved
in Neptune’s discovery only to the extent that they were involved in the dis-
covery of Kepler’s laws, which in turn were used by Newton. But I argued that
(1)–(5) were not involved essentially in Kepler’s discoveries. (9)–(11) concern
Neptune, so they could not be premises from which Neptune’s existence or lo-
cation was derived: they are only (partially wrong) parts of a global hypothesis
on the planet’s existence and behavior. (6) was practically true in this context,
given the negligible influence of the other planets. To my knowledge (7) was
never held by anybody; anyway, it entails the true claim that inertial motion con-
tinues indefinitely in a straight line. Finally, I just argued that (8), gravitational
force, was not essential in Newton’s predictions. So, even Neptune ceases to be a
counterexample to DR.
Historical Challenge to Realism and Essential Deployment 199
For deployment realists the components of discarded theories which were es-
sentially involved in novel predictions are most probably true; but Stanford
(2006), Votsis (2011), and Peters (2014) argue that it is not enough to identify
these components just retrospectively, i.e., as those preserved in current theories.
Rather, they must be identified prospectively, from the viewpoint of their con-
temporaries. In fact, deployment realists claim that the hypotheses of past the-
ories still preserved today are true because they were essential. Thus they resist
the Pessimistic Meta-Induction (PM-I), maintaining that both past and current
theories are at least partly true. But explaining that those hypotheses were es-
sential because they are preserved today, would be explaining that they are true
because they are preserved today, so begging the question of the truth of current
theories. In fact, since a hypothesis H is preserved today because we believe it is
true, saying that it is true because it is essential, and it is essential because it is pre-
served today, would be saying it is true because we believe that it is true.
Equally, deployment realists claim that the now rejected hypotheses deployed
in novel predictions were not essential, hence the NMA did not commit to their
truth. In this way they block Laudan and Lyons’s meta-modus tollens (M-MT).
But explaining that they were not essential because they are not preserved today
would be stipulating that the NMA warrants all and only the hypotheses ac-
cepted today. Thus the realist’s defense against both the PM-I and the M-MT
would be circular.
This is why it would seem desirable that the components of past theories
which are essential for their novel predictions are identified prospectively; more-
over, Votsis (2011), Peters (2014), Cordero (2017a, 2017b), and others claim that
such prospective identification is possible. On the opposite side, Stanford (2006,
167–180; 2009, 385–387) claims this is impossible, therefore the arguments for
DR are circular.13
I take an intermediate position: any time we believe a hypothesis is essential,
it may in fact be, but neither at that time nor later can we ever be certain that
it is. Yet, I argue that this does not make the arguments for DR circular. That
we cannot ever be certain is shown by many hypotheses which were firmly
believed by past scientists because they appeared to be essential to certain novel
predictions, but subsequently turned out to be inessential, in fact false.
For instance, Fresnel and Maxwell derived various novel predictions from the
hypothesis that
Fresnel and Maxwell could easily have seen that VM was a proper part of AV and
enough to derive those predictions. However, they didn’t consider VM (probably
they didn’t even think of it), no doubt because they presupposed that
and/or that
Hence, given their presuppositions, any vibrating medium was a material me-
dium composed of particles (i.e., either water, or air, or aether). Therefore from
their viewpoint, AV was required by their predictions as much as VM.
Again, Laplace predicted the speed of sound in air starting from the false
hypothesis that
Psillos (1999, 119–121) argues that H was not essential to Laplace’s prediction,
which could also be derived from a part of H, viz.
However, at that time adiabatic heating could only be explained as the disen-
gagement of caloric from ordinary matter, caused by mechanical compression
(Chang 2003, 904), for it was presupposed that
P1: gases can be heated without exchanges with the environment only if they
contain heat in a latent form,
and
But the material substance of heat was just caloric: therefore, given P1 and P2, H*
entailed H, hence H too had to be considered essential.
Generalizing, we might say that at any time in the history of science, when
a novel prediction NP is derived from an assumption H, there may be an as-
sumption H* which is proper part of H and sufficient—from a purely logical
viewpoint—to derive NP; however, given certain current explicit or implicit
presuppositions PRS, H may “appear to be conceptually or metaphysically
entailed by” H* (Vickers 2016, §4), hence scientists may falsely believe that H is
essential to NP. At a later time, however, the advancement of science may show
that the PRS are false, so scientists can (retrospectively) recognize that H was
inessential, after all. This recognition is facilitated if meanwhile experimental
or theoretical doubts have arisen about H, so that scientists started wondering
whether H was really necessary. Yet, this recognition is logically independent of
those doubts, as it follows already from the falsity of the PRS; hence, although
retrospective, it is not circular.
Vickers, however, suggests that although prospectively we cannot ever be
certain that H is essential to a prediction, in some cases we could (still prospec-
tively) recognize it is not essential, although “metaphysically or conceptually” re-
quired by other considerations. This may (i) suggest the realist to “restrict her
commitments to what is directly confirmed by the predictive successes” (i.e., H*),
and (ii) supply “a worthwhile heuristic,” i.e., show which commitments should be
abandoned first in case of empirical or theoretical refutations (i.e., H).
For instance, Bohr predicted (NP) the spectral lines of ionized helium by
assuming that
H: The electron orbits the nucleus only on certain specific orbital trajectories,
each characterized by a given quantized energy.
H turns out to be false, but NP could have been derived by the weaker
hypothesis that
H*: The electron can only have certain, specific, quantized energy states.
Moreover, says Vickers, Bohr could have seen (prospectively) that H was ines-
sential, because it entailed H*, which was enough to derive NP. But he still held
H because “it may have been inconceivable at the time to think that electrons
could have quantized electron energies without having associated quantized
orbital trajectories (cf. Stanford 2006, 171)”. Nonetheless, Bohr might have
202 Contemporary Scientific Realism
true, scientists grant that some of their assumptions are probably wrong, but only
future research will tell which ones.
Besides, the PM-I is probably right that no theory or component older
than 100 years or so was completely true (but not that none was at least partly
true). Cordero, Peters, and Votsis are also right that we know when a hypo-
thesis H is “true,” if this means “at least partly true”: but even if deployed in
novel predictions, H may have been inessential, and even if otherwise well
supported it can be partly false, and sooner or later replaced by another more
completely true.
The progress of research allows us to drop more and more false
presuppositions. Thus we may discover, retrospectively, that H was inessential,
for it had some proper part H* which (i) can be true even if H is false and (ii)
is sufficient to derive NP. That weaker H* may also be actively sought for, if H
encounters empirical or theoretical refutations, which suggest that it cannot pos-
sibly be essential (since essential components are completely true). In this case
(still retrospectively), one can recognize that H was inessential even before iden-
tifying its weaker substitute H*.
(γ) Neither the impossibility of recognizing essentiality, nor the difficulties in
prospectively recognizing non-essentiality are a problem for deployment realists,
because these recognitions are not necessary to resist the PM-I and the M-MT.
In fact, as soon as a risky novel prediction NP derived from H is confirmed, we
know (prospectively, and a fortiori retrospectively) that, short of miracles, there
is some truth in H. This is enough to refute the PM-I’s claims that both past and
present theories are completely false.
The M-MT, in turn, is always used retrospectively, therefore only a retro-
spective recognition of inessentiality is needed to resist it: if at time t hypo-
thesis H was firmly believed because considered essential to derive NP, but at
t’ it is shown to be false, the M-MT uses it as a counterexample to the NMA.
But realists can reply by two moves: (i) arguing that probably H was inessen-
tial, for only by a miracle could NP have been derived from a completely false
assumption; (ii) showing H was inessential, by identifying a proper part H*
sufficient to derive NP, and the false presuppositions PRS that at t prevented
us from distinguishing H from H*. Both moves are retrospective (at t’ or at a
later t”), yet sufficient to block the M-MT. Besides, move (ii) is independent
of the refutation of H, therefore the realist’s defense is not circular, as Stanford
claims.
(δ) Even when it is possible, the recognition of non-essentiality (especially
prospectively) is a task for scientists: philosophers just don’t have the necessary
expertise. The burden of scientific realists is arguing that certain criteria (e.g., es-
sential deployment) can justify our beliefs in the at least partial truth of theories
204 Contemporary Scientific Realism
H
H*
Figure 9.1
or hypotheses. But Fahrbach (2017) asks too much when he requires that they
also apply those criteria to actual research, so teaching practitioners which are
the working hypotheses and which are the idle parts in their theories, urging
changes, suggesting directions of research, etc.
Therefore, qua philosophers, realists need not be committed to any partic-
ular theory or hypothesis, not even to the best current ones: as argued by Smart
(1963, 36), this is a task for scientists. At most, they may argue that science is con-
vergent, hence, in general, current theories are probably more largely true than
past ones. Therefore, pace Stanford (2017), they need not be more conservative
than anti-realists. Actually, since they set for hypotheses a higher standard than
anti-realists—truth, rather than empirical adequacy or the like—for them it is
even more likely that any particular hypothesis fails to reach that standard, hence
must be substituted by a better one.
(ε) The discussion at (γ) shows that even when inessentiality can be recog-
nized prospectively, that is superfluous in resisting the PM-I and the M-MT. In
fact, the at least partial truth of H can be recognized prospectively, even if H is
not essential, and this is enough to block the PM-I. Besides, prospective recog-
nition of inessentiality is superfluous against the M-MT, because the latter is al-
ways used retrospectively.
Here I explain in more detail why my condition (2’) is enough and we can
do without the greater complexity of Psillos’s condition (2) or its possible
strengthening implicitly suggested by Lyons. When Psillos excludes any “other
hypothesis H*,” H* may be compatible or incompatible with H, and in different
mereological relations to it. The following cases are included:
(a) H* is proper part of H. This is precisely the kind of alternative hypothesis H*
excluded by my condition (2’) (Figure 9.1)
(b) H is a proper part of H* (Figure 9.2)
Historical Challenge to Realism and Essential Deployment 205
H* H
Figure 9.2
H = H*
Figure 9.3
H H*
Figure 9.4
Now, Psillos’s condition (2) excludes alternative hypotheses H* of all these five
kinds, and here the troubles arise, while my condition (2’) excludes only alter-
native hypotheses H* of kind (a), and I claim that this is enough for deploy-
ment realists. In cases (b), (d), and (e) H* has some content outside H; so, in
these cases H* might be incompatible with OHs and AA, and Lyons implicitly
suggests they should be excluded along with the cases in which H* is compat-
ible with OHs and AA. But we don’t need to worry about any of these cases,
and here is why:
206 Contemporary Scientific Realism
Figure 9.5
In case (b) H is weaker than and part of H*. Therefore the existence of H*
does not make H inessential, so there is no need to exclude it. In fact, there have
always been such unnecessarily rich alternatives to the true hypothesis (like e.g.
the ether theory with respect to the simple assumption of Maxwell’s equations, or
the aforementioned demi-gods theory with respect to Newton’s actual theory).
From a purely logical viewpoint, for any H there are infinite such H*, but they do
not make H inessential.
In case (c) H* coincides with H, so again it does not make H inessential and we
don’t need to exclude it.
In case (d) H and H* have no common content, and this may happen in
two ways:
(d1) H and H* are compatible, hence, they may be both true. In general it
is quite possible that a prediction is reached by independent inferential
routes starting from different hypotheses. This is just natural if we con-
ceive the world as a tightly connected whole, in which each phenom-
enon, at each scale, is causally linked to many other phenomena even
at different scales. For instance, Perrin famously derived the same value
for Avogadro’s number reasoning from tenets of chemistry, thermody-
namics, electrical theory, the theory of Brownian motions, and others.
In such cases, each hypothesis Hi deployed in a different inferential
route to NP is not “essential” in the sense that without Hi one could not
have predicted NP (for NP could have been predicted through some
different inferential routes involving some alternative hypothesis Hj).
However, Hi can be essential in the sense that NP could not have been
predicted along that particular inferential route without Hi, i.e., in the
sense that Hi cannot be substituted in the same inferential route by one
Historical Challenge to Realism and Essential Deployment 207
of its proper parts. This is precisely what my condition (2’) requires, and
it is all we need from a criterion of essentiality, for it is enough to war-
rant the complete truth of Hi (for we saw that, short of miracles, a hy-
pothesis deployed in predicting NP cannot be completely false, and it
can be partly false only if its false part is dispensable in the derivation).
Therefore we need not exclude case (d1), for in this case H is essential in
the required sense if it fulfills (2’). By the way, since H entails NP only in
conjunction with AA and OHs, also AA and OHs are essential if they
satisfy (2’).15
(d2) Otherwise, H and H* may have no common part and be incompatible.
One might think that we need to exclude this case: for if we had two in-
compatible hypotheses both predicting NP, we couldn’t know which of
them is true.16 However, as explained in my nutshell account at section
9.3, this case is already excluded by the risky character of NP. In fact, as
argued earlier, it is practically impossible to find a completely false hypo-
thesis entailing a risky novel prediction. But if H is essential in the sense
of (2’), it must be completely true. Therefore, if H* is incompatible with
it, we know that (i) H* is at least partly false and (ii) it cannot possibly
predict NP, for it contradicts an assumption (i.e., H) which is essential to
that prediction.17 Therefore, my condition (2’) (together with the risky
character of NP) already excludes that any hypothesis incompatible with
H also predicts NP.
Finally, case (e): H and H* partially overlap. Like in case (d), here too we must
distinguish:
15 Then we know they are true (unless NP was used in building them).
16 Besides, we would have a counterexample to the NMA, for there would be at least one false hy-
pothesis allowing us to derive a novel prediction.
17 Instead, if H is not essential in my sense, it consists of an essential part H and an inessential part
e
Hi. If H* also has a role in predicting NP, it may consist in an essential part H*e and an inessential
part H*i. Of course H*e cannot contradict He, but H*i may contradict Hi, in which case H and H* are
incompatible, yet have both a (non-essential) role in predicting NP, and are both partly true. Instead
He and H*e are two essential but compatible hypotheses as considered in case (d1).
208 Contemporary Scientific Realism
Since a falsity may entail a truth, even false theories might make novel
predictions.
Answer:
As argued at the beginning of section 9.1, in the Hyperuranion of possible
theories there are some with wholly false theoretical assumptions which make
true novel and improbable predictions (FNIP-theories). But their rate is exceed-
ingly small, hence it is extremely unlikely that scientists pick one of them (be-
cause scientists look for true theories, hence they could find FNIP-theories only
by chance). On the opposite, all true theories have true consequences, and those
sufficiently fecund have true novel consequences. Granted, true (and fecund)
theories are much fewer than false ones; however, they are not found by chance,
but on purpose and through reliable methods (White 2003; Alai 2016, 552–554;
2018, §III).
The author’s condition (2’) seems to make realism redundant for purely logical
reasons. A statement is derivable from a set of statements iff the content of that
statement is already included in that set of statements. The only proper part of
that set of statements that’s required to derive the statement is in fact the state-
ment itself. Thus, a novel prediction NP needs only NP to be entailed. But surely
we want to be realists about more things than the content corresponding to NP.18
Answer:
Most frequently and typically NP is not part of H, because, as explained by
Vickers19 and others, it is not entailed by H alone, but only in conjunction with
certain other hypotheses OHs of T, and various auxiliary assumptions AA.
Therefore (2’) does not allow us to dispense with H in favor of NP.
But the objection might become that since H&OHs&AA entails NP, we can
dispense with H&OHs&AA in favor of NP. This is reminiscent of Hempel’s
(1958) “theoretician’s dilemma”: theoretical hypotheses H are required to con-
nect the observable initial conditions IC to the observable final conditions FC,
and they are justified only if they succeed in this, i.e., if IC→H→FC. But then
why not drop H and just keep the empirical laws IC→FC? Hempel answered that
this would work from a purely logical point of view, but not from an epistemolog-
ical point of view: H does not predict only IC→FC, but many other phenomena
as well, most of which we would never have discovered without assuming H.
The answer here is similar: we would never have found NP, and many other
predictions, unless H&OHs&AA were true. In fact, NP is novel and risky, hence
it cannot have been found just by chance. Besides, the auxiliary assumptions
AA are the consequences of many independent theories T’, T”. . . Tn, typically
dealing with matters different from T. Therefore, it would be a miraculous co-
incidence if even one of them were completely false, yet the conjunction of their
consequences entailed NP. The only plausible explanation is that T, T’, T”. . . Tn
are all at least partially true (Alai 2014c, §4). Besides, each of T, T’, T”. . . Tn,
in conjunction with still different assumptions, issues many other predictions,
which are only explainable in the same way.
One might reply that guessing n true theories is even less likely than guessing
n false theories with one true joint consequence. This is right, but guessing n false
theories with one true, novel, and risky joint consequence is still too improb-
able to be a minimally plausible explanation of our frequent predictive successes.
So, how was it possible to derive NP and so many other novel predictions? The
only plausible explanation is that theories are not guessed, but found by a method
which is reliable in tracking truth. Scientific realists have a reasonable account
of why the scientific method is reliable, while anti-realists have never been suc-
cessful in providing an account of how predictive success might be achieved by
pure chance or “miracles” (Alai 2014a). Notice, here the truth-conduciveness of
the scientific method is not a petitio principii, but the conclusion of my inference
to the only plausible explanation in this paragraph (the second in this section).20
Answer:
Again, here H* is not part of H, hence its existence does not make H inessen-
tial. For instance, suppose the following:
Given that whatever follows from H also follows its Ramsey sentence (RSH), and
that H implies RSH, wouldn’t it follow that every hypothesis is dispensable in
favor of its Ramsey sentence?22
Answer:
No: in general, the essential part H* of H does not coincide with RSH, because
it contains both something more and something less than RSH. It must contain
something more because RSH would not allow to derive NP. For instance, as-
suming in Newton’s gravitation law (N) the terms “force” and “mass” are theoret-
ical and “distance” is observational, its Ramsey sentence is
( )
RS N : ∃x , ∃y x=Gy 1 y 2 /D2 .
But RSN is not enough to predict Neptune. We also need to say what kind of
properties or relations are x and y, how they behave, etc.: we need an expanded
Ramsey sentence like
( )
RS NEx : ∃x, ∃y ....x.... & ....y.... & x=Gy 1 y 2 /D2 ,
where ‘....x...’ and ‘...y...’ are various laws concerning x and y (which, in practice,
characterize them respectively as force and mass).
But typically the essential part H* contains also something less than RSH or
RSHEx. For instance, Newton’s Gravitation Law is partly false (since there exists
no gravitation force). Consequently, also RSN and RSNEx are partly false, hence,
inessential. Therefore, in general, the essential H* is different from RSH. It might
still be representable by an expanded Ramsey sentence RSH*Ex (a proper part of
RSHEx), but this wouldn’t be a problem for realism, because:
(i) Like any Ramsey sentence, RSH*Ex is not a merely empirical claim, it
states the existence of unobservable entities, even without naming them.
(ii) In general, RSHEx will have a very complex structure, hence it might be
practically impossible to write it out completely, so to dispense with H.
Besides, as argued in section 9.5, identifying the essential part of hypoth-
eses is very difficult, so even if we could spell out RSHEx, often we would
be unable to dispense with it in favor of its essential part RSH*Ex.
(iii) At any rate, my claim is that the NMA commits only to essential hypoth-
eses, not that only essential hypotheses can be true. So, if the essential
hypothesis turns out to be something like RSH*E, since Ramsey sentences
characterize only partially the entities to which they refer, realists may
hold that something stronger than RSH*E is true.
Answer:
As explained above, H** is neither part of H nor implied by it. Therefore H**
does not make H inessential.
Acknowledgments
23 Objection raised in discussion by John Worrall, referring to another topos from the debates on
confirmation.
Historical Challenge to Realism and Essential Deployment 213
References
Alai, M. (2012) “Levin and Ghins on the ‘No Miracle’ Argument and Naturalism,”
European Journal for Philosophy of Science vol. 2, n. 1, 85–110. http://rdcu.be/mTcs.
Alai, M. (2014a) “Why Antirealists Can’t Explain Success,” in F. Bacchini, S. Caputo
and M. Dell’Utri (eds.) Metaphysics and Ontology Without Myths, Newcastle upon
Tyne: Cambridge Scholars Publishing, pp. 48–66.
Alai, M. (2014b) “Defending Deployment Realism against Alleged Counterexamples,”
in G. Bonino, G. Jesson, J. Cumpa (eds.) Defending Realism. Ontological and
Epistemological Investigations, Boston-Berlin-Munich: De Gruyter, pp. 265–290.
Alai, M. (2014c) “Novel Predictions and the No Miracle Argument,” Erkenntnis vol. 79,
n. 2, 297–326. http://rdcu.be/mSra.
Alai, M. (2014d) “Explanatory Realism,” in E. Agazzi (ed.) Science, Metaphysics,
Religion: Proceedings of the Conference of the International Academy of Philosophy of
Science, Siroki Brijeg 24–24 July 2013 Milano, Franco Angeli, pp. 99–116.
Alai, M. (2016) “The No Miracle Argument and Strong Predictivism vs. Barnes,” in L.
Magnani and C. Casadio (eds.) Model-Based Reasoning in Science and Technology.
Logical, Epistemological, and Cognitive Issues, Cham: Springer, pp. 541–556.
Alai, M. (2017) “Resisting the Historical Objections to Realism: Is Doppelt’s a Viable
Solution?” Synthese vol. 194, n. 9, 3267–3290. http://rdcu.be/mSrG.
Alai, M. (2018) “How Deployment Realism Withstands Doppelt’s Criticisms,”
Spontaneous Generations vol. 9, n. 1, 122–135.
Carrier, M. (1991) “What is Wrong with the Miracle Argument?” Studies in History and
Philosophy of Science vol. 22, 23–36.
Chang, H. (2003) “Preservative Realism and Its Discontents: Revisiting Caloric,”
Philosophy of Science vol. 70, 902–912.
Clarke, Steve, and Timothy D. Lyons (eds.) (2002) Recent Themes in the Philosophy of
Science. Scientific Realism and Commonsense. Dordrecht: Kluwer.
Cordero, A. (2017a). Retention, truth-content and selective realism, in E. Agazzi (ed.)
Varieties of Scientific Realism, Cham: Springer, pp. 245–256.
Cordero, A. (2017b) “Making Content Selective Realism the Best Realist Game in Town,”
Quo Vadis Selective Scientific Realism? Conference, Durham, UK 5–7 August 2017.
214 Contemporary Scientific Realism
Fahrbach, L. (2017) “How Philosophy Could Save Science,” Quo Vadis Selective Scientific
Realism? Conference, Durham, UK 5–7 August 2017.
Hempel, Carl G. (1958) “The Theoretician’s Dilemma: A Study in the Logic of Theory
Construction,” Minnesota Studies in the Philosophy of Science vol. 2, 173–226.
Hoskin, M. (ed.) (1999) The Cambridge Concise History of Astronomy,
Cambridge: University Press.
Hudson, J. (1992) The History of Chemistry. London: Macmillan Press.
Kitcher, P. (1993) The Advancement of Science. New York: Oxford University Press.
Koyré, A. (1961) La révolution astronomique. Paris: Hermann.
Laudan, L. (1981) “A Confutation of Convergent Realism,” Philosophy of Science vol.
48, 19–49.
Lipton, P. (1991) Inference to the Best Explanation, London: Routledge.
Lipton, P. (1993) “VI—Is the Best Good Enough?,” Proceedings of the Aristotelian Society
vol. 93, 89–104.
Lipton, P. (1994) “Truth, Existence, and the Best Explanation,” in A. A. Derksen (ed.), The
Scientific Realism of Rom Harré. Tilburg: Tilburg University Press, pp. 89–11.
Lyons, T. D. (2002) “The Pessimistic Meta-Modus Tollens,” in S. Clarke and T. D. Lyons
(eds.) Recent Themes in the Philosophy Science. Australasian Studies in History and
Philosophy of Science, vol 17, Dordrecht: Springer, pp. 63–90. doi:10.1007/978-94-017-
2862-1_4.
Lyons, T. D. (2003) “Explaining the Success of a Scientific Theory,” Philosophy of Science
vol. 70, 891–901.
Lyons, T.D. (2006) “Scientific Realism and the Stratagema de Divide et Impera,” The
British Journal for the Philosophy of Science, vol. 57, 537–560.
Lyons, T.D. (2009) “Criteria for Attributing Predictive Responsibility in the Scientific
Realism debate: Deployment, Essentiality, Belief, Retention,” Human Affairs vol. 19,
138–152.
Lyons, T.D. (2016) “Structural Realism versus Deployment Realism: A Comparative
Evaluation,” Studies in History and Philosophy of Science Part A vol. 59, 95–105.
Lyons, T.D. (2017) “Epistemic Selectivity, Historical Threats, and the Non-Epistemic
Tenets of Scientific Realism,” Synthese vol. 194 n. 9, 3203–3219.
Musgrave, A. (1985) “Realism versus Constructive Empiricism,” in P. Churchland and C.
Hooker (eds.) Images of Science, Chicago, Chicago University Press, pp. 197–221.
Musgrave, A. (1988) “The Ultimate Argument for Scientific Realism,” in Robert Nola (ed.)
Relativism and Realism in Science, Dordrecht: Springer, pp. 229–252.
Peters, D. (2014) “What Elements of Successful Scientific Theories are the Correct Targets
for “Selective” Scientific Realism?” Philosophy of Science vol. 81, 377–397.
Plebani, M. (2017) “Aboutness for Impatients,” Academia.edu, https://www.academia.
edu/34295884/Aboutness_for_the_impatient.
Psillos, S. (1999) Scientific Realism: How Science Tracks Truth.
London-New York: Routledge.
Sankey, H. (2001) “Scientific Realism: An Elaboration And A Defense,” Theoria, vol.
98, 35–54.
Smart, J.J.C. (1963) Philosophy and Scientific Realism. London: Routledge.
Stanford, P. K. (2006). Exceeding Our Grasp: Science, History, and the Problem of
Unconceived Alternatives. Oxford: Oxford University Press.
Stanford, K. P. (2009) “Author’s Response,” in “Grasping at Realist Straws,” a Review
Symposium of Stanford (2006), Metascience vol. 18, 379–390.
Historical Challenge to Realism and Essential Deployment 215
10.1 Introduction
Here I propose a particular conception of both the current state of play and what
remains at stake in the ongoing debate concerning scientific realism. Recent
decades have witnessed considerable evolution in this debate, including a wel-
come moderation of both realist and instrumentalist positions in response to ev-
idence gathered from the historical record of scientific inquiry itself. In fact, I will
suggest that a set of commitments has gradually emerged that are now embraced
by many (though by no means all) who call themselves scientific realists and also
by many (though by no means all) who would instead characterize themselves as
instrumentalists or (perhaps less helpfully) “antirealists.” I will go on to suggest
that these commitments collectively constitute a shared “Middle Path” on which
many historically sophisticated realists and instrumentalists have already made
substantial progress together, perhaps without realizing how much closer their
respective views have thereby become to one another than either is to the clas-
sical forms of realism and instrumentalism whose labels and slogans they none-
theless conspicuously retain. I will then go on to suggest, however, that at least
one crucial disagreement still remains even between realists and instrumentalists
who walk this Middle Path together, and that this remaining difference actually
makes a difference to how we should go about conducting scientific inquiry itself.
Recognizing this difference poses a further challenge for instrumentalists both
on and off the Middle Path, however, and responding to it will illuminate a fur-
ther dimension of the realism debate itself: namely, the level(s) of abstraction or
generality at which we should seek and expect to find useful epistemic guidance
concerning scientific theories or beliefs. Along this dimension of the debate,
I suggest, Middle Path realists and instrumentalists are once again united, but
this time against both the shared presuppositions of their classical predecessors
on the one hand and radical or extreme forms of particularism on the other.
The first of the shared commitments that I suggest constitute a Middle Path be-
tween classical forms of both realism and instrumentalism is to what I’ve else-
where (Stanford 2015) called “Uniformitarianism,” in parallel with the great battle
in 19th century geology between Catastrophists and Uniformitarians concerning
the causes and pattern of changes to the Earth. Famously, Uniformitarians (like
Charles Lyell) argued that the broad topographic and geographic features of the
Earth were the product of natural causes like floods, volcanoes, and earthquakes
operating over immense stretches of time at roughly the same frequencies and
magnitudes at which we find them acting today. Their Catastrophist opponents
(like Georges Cuvier) held instead that such natural causes had operated in
the past with considerably greater frequency and/or magnitude than those we
now observe, on the order of the difference between a contemporary flood and
the great Noachian deluge reported in the Christian Bible. Catastrophists thus
held that the Earth had steadily quieted down over the course of its history, that
truly fundamental and wide-ranging changes to its topography and geography
are now confined to the distant past, and that contemporary natural causes will
further modify that topography and geography only in comparatively marginal
and limited ways. By contrast, Uniformitarians held that if given enough time
to operate present-day natural causes would continue to transform the existing
topography and geography of the Earth just as profoundly as it was transformed
in the past.
Likewise in the case of the realism dispute, Uniformitarians take the view that
the future of the scientific enterprise will continue to be characterized by the-
oretical revolutions and conceptual transformations just as profound and fun-
damental as those we have witnessed throughout the history of that enterprise.
By contrast, a textbook description of what I will call classical or Catastrophist
realism holds that “Our mature scientific theories, the ones used to underwrite
our scientific projects and experiments, are mostly correct” and “[w]hat errors
our mature theories contain are minor errors of detail” (Klee 1999 313–4). Such
Catastrophists take the view that truly profound and fundamental revisions to
our scientific understanding of the world are either largely or completely con-
fined to the past and that the future of scientific inquiry will not be characterized
by the sorts of fundamental revolutions or transformations by which theo-
ries like Newton’s mechanics, Dalton’s atomism, and Weismann’s theory of the
218 Contemporary Scientific Realism
allow not only that we can perfectly well articulate the competing claims of two
scientific theories about a single shared world using language that privileges
neither theory, but also that we can characterize the available evidence in sup-
port of each theoretical alternative in ways that are at least neutral between those
two competing theories (if not independent of any and all “theorizing” whatso-
ever) in reaching an impartial judgment that one is better or worse confirmed by
the existing evidence, or even that one or both theories are definitively refuted
(within the bounds of fallibilism) by that evidence. They similarly reject the fur-
ther consequence drawn by various interest-driven theorists of science that the
outcomes of such theoretical competitions are therefore typically determined
less by the available evidence in support of each theoretical alternative than by
the comparative power, standing, resourcefulness, and determination of the so-
cial groups who advance and defend them. Where Uniformitarianism assures
us that further profound and fundamental changes are still to come for our sci-
entific picture of the world, a further commitment to what we might naturally
call commensurability assures us that such changes will neither prevent us from
fully understanding the competing conceptions of nature thereby proposed nor
from impartially adjudicating the character and strength of the evidence in sup-
port of each of those competing conceptions. This is certainly not to suggest that
resolving such competitions is simple or straightforward, or that an individual
scientist’s own theoretical (and other) sympathies do not influence her evalua-
tion of the evidence, or even that such comparisons can always be convincingly
resolved by the particular body of evidence available at a given time. But it is
to claim that their ultimate resolution typically depends on the accumulation of
evidence that the relevant scientific community rightly sees as objectively and
impartially favoring one proposal over the other. Such a commitment to com-
mensurability would seem to similarly unite instrumentalists and realists on the
Middle Path together with one another and against such radically Kuhnian var-
ieties of instrumentalism or anti-realism.
Realists and instrumentalists who travel this Middle Path together are also
united against other classical varieties of instrumentalism or anti-realism by
their commitment to what I will call the Maddy/Wilson Principle, which I first
encountered when my colleague Penelope Maddy skeptically responded to my
own instrumentalist sympathies with this especially pithy formulation: “well,
nothing works by accident or for no reason.” That is, she seemed to regard instru-
mentalism as committed not simply to the view that our best scientific theories
are powerful cognitive instruments that guide our predictions, interventions,
and other forms of practical engagement with nature successfully, but also that it
is misguided or somehow illegitimate even to ask, much less try to discover or ex-
plain, how they manage to achieve that success when and where they do. Having
recognized the ability of our best theories to guide our practical engagement
220 Contemporary Scientific Realism
with nature successfully, such instrumentalism rejects even the demand for any
explanation of how or why they are able to do so.
When I reminded her of this conversation some years later, she was quick to
credit the influence of Mark Wilson’s Wandering Significance as having inspired
this particular expression of her disquiet. And indeed, Wilson does articulate
something very like this same concern in his own inimitable way:
Both Maddy and Wilson, then, endorse the broad principle that when a sci-
entific theory enjoys a track record of fine-grained and wide-ranging success
in guiding our practical engagement with the world, it typically does so in
virtue of some systematic connection or relationship between the description
of the world offered by that theory and how things actually stand in the world
itself.
It is no accident that Maddy and Wilson both regard instrumentalism as com-
mitted to the idea that there is simply nothing more to say or know about how
and why our instrumentally powerful scientific theories work as well as they do
when and where they do. Scientific instrumentalism has had a wide variety of
incarnations over the last several centuries, and some of the most influential
are indeed characterized by explicit or implicit versions of this commitment.
Perhaps most famously, some logical positivist and logical empiricist thinkers
held that the claims of our best scientific theories are not even assertions about
the world in the first place, nor, therefore, even candidates for truth or falsity,
but instead simply “inference tickets” allowing us to predict or infer some ob-
servable states of affairs from others. More recently, van Fraassen’s Constructive
Empiricism argues that it is illegitimate to demand anything more than the em-
pirical adequacy of a theory as an explanation of its success, insisting instead
“that the observable phenomena exhibit these regularities, because of which they
fit the theory, is merely a brute fact, and may or may not have an explanation
in terms of unobservable facts ‘behind the phenomena’ ” (1980 24). Thus, there
is indeed a long and distinguished philosophical tradition advocating forms of
instrumentalism that either reject or remain agnostic concerning the Maddy/
Wilson Principle, insisting instead that once we have recognized the instru-
mental utility of our scientific theories it is somehow misguided or illegitimate
even to ask how or why those theories manage to be so instrumentally useful
when and where they do.
Realism, Instrumentalism, Particularism 221
makes clear that there is no single such relationship holding between all earlier
successful theories and their contemporary successors: realists would need to
appeal to a wide and heterogeneous array of different systematic relationships or
“reasons” of this sort in explaining the diverse particular empirical successes of
the various instrumentally powerful, past scientific theories that have since been
abandoned or replaced.
I suggest that many realists and instrumentalists alike, especially those who
rely on evidence drawn from the historical record, have gradually come to em-
brace these shared commitments to Uniformitarianism, commensurability,
and the Maddy/Wilson Principle, perhaps without fully realizing how much
more they have thereby come to agree with one another than either of them
does with the more classical doctrines whose labels and slogans they nonethe-
less retain. Their broadly shared conception of the past, present, and future of
scientific inquiry anticipates many substantial continuities between future sci-
entific orthodoxy and that of the present day (just as there are between present
theoretical orthodoxy and that of the past), but also many substantial discon-
tinuities just as profound and significant as those we now find in past historical
episodes of fundamental revolution or upheaval. Moreover, I suggest that realists
and instrumentalists who travel this Middle Path have already made consider-
able progress in using further historical evidence to refine and elaborate that
broadly shared conception. Important recent work conducted by self-described
realists and antirealists alike, for example, has revealed that the mistaken (i.e.,
ultimately rejected) claims or components of successful past theories have often
played crucial and ineliminable roles in generating those theories’ empirical
successes (e.g., Lyons (2002, 2006, 2016, 2017), Saatsi and Vickers (2011); for
a fairly comprehensive listing, see Vickers (2013) and the references contained
therein). Likewise, Peter Vickers (2013, 2017) has recently argued that we can
at least identify posits (including some that do genuine work in generating a
theory’s successful implications) that are nonetheless inessential or eliminable
from those theories in a sense that renders them poor candidates for realist com-
mitment. Such increases in the nuance and sophistication of their broadly shared
vision of the past, present, and future of scientific inquiry should be recognizable
as important forms of progress by both realists and instrumentalists alike on the
Middle Path.
I do not mean to suggest, however, that these shared commitments of realists and
instrumentalists on the Middle Path leave no room for significant disagreement
Realism, Instrumentalism, Particularism 223
between them. As noted earlier, “selective” scientific realists are largely moti-
vated by the idea that the historical record itself (perhaps together with other
considerations) puts us in a position to reliably distinguish the particular elem-
ents, aspects, or features of our own best scientific theories (e.g., working posits,
core causal descriptions, structural claims, etc.) genuinely responsible for their
empirical successes from those that are instead “idle” or otherwise not required
for those successes (e.g., the electromagnetic ether). Even if we expect further
profound and dramatic changes in theoretical orthodoxy as scientific inquiry
proceeds into the future, Middle Path realists insist that we can nonetheless jus-
tifiably expect these privileged elements, aspects, or features to be retained and
ratified in some recognizable form throughout the further course of scientific in-
quiry itself. Note that without a commitment to at least this minimal form of sta-
bility or persistence, the selective realist loses not only her grounds for claiming
to have identified secure epistemic possessions on which we can safely rely as in-
quiry proceeds, but also her grounds for claiming that these particular elements,
aspects, or features of our best scientific theories were those genuinely respon-
sible for their various empirical successes (for which they have turned out not to
be essential or required after all).
By contrast, those on the Middle Path who embrace the “instrumentalist”
label certainly accept the Maddy/Wilson Principle’s insistence that there must be
such reasons for the successes of our best scientific theories, but they nonetheless
deny that we ourselves are generally in a position to know what those reasons are
in particular cases, or to reliably identify which particular elements or features
of our best scientific theories we should therefore expect to find persisting
throughout the course of further inquiry. Such instrumentalists will see the wide
variety of such reasons that (as we noted earlier) contemporary scientific ortho-
doxy would need to invoke in explaining the various successes of various past
scientific theories as piling up counterexamples rather than confirmation for the
idea that any single aspect, element, or feature of a successful scientific theory is
invariably or even just reliably preserved in its successors: sometimes equations
or claims about the “structure” of nature are preserved from one successful scien-
tific theory to its successors, but other times it is the fundamental entities posited
by that theory, or the “core causal descriptions” of those entities, or something
else altogether—no one of these forms of continuity (or corresponding versions
of selective realism itself) seems to capture even an especially wide range of cen-
tral historical examples.
Rather than seeking to adjudicate this remaining central point of disagree-
ment dividing even realists and instrumentalists traveling together on the Middle
Path, I will instead argue for its importance. It truly matters, I suggest, because
it actually makes a difference to how we should go about conducting further sci-
entific inquiry itself. To see why, recall first the case of classical, commonsense,
224 Contemporary Scientific Realism
which to compare them. But the instrumentalist shares any such strategic or in-
strumental reasons the realist may have for exploring or developing a theoretical
alternative conflicting with (some privileged part of) existing theoretical ortho-
doxy. The instrumentalist also has another that is ultimately far more impor-
tant: she believes that even more instrumentally powerful alternatives radically
distinct from contemporary scientific theories are actually out there still waiting
to be discovered.
Evaluating the promise, interest, or appeal of any particular theoretical pro-
posal is, of course, a complex and multi-dimensional affair, so the suggestion
here is certainly not that realists must always favor investing in the pursuit of
any theoretical proposal (about anything) consistent with existing orthodoxy
over the pursuit of any theoretical proposal (about anything) contradicting
that orthodoxy. Nor is there some threshold degree of theoretical conservatism
that realists must meet or exceed, either in general or in any particular case. The
point instead is that those who are realists regarding some particular scientific
theory (or privileged part thereof) should be prepared to treat the inconsistency
of a given alternative theoretical proposal with the central claims (or otherwise
privileged elements) of that theory as a reason to doubt that the alternative in
question is even a viable candidate for representing the truth about the do-
main of nature it seeks to describe. All else being equal, this in turn constitutes
a reason (for the realist) to discount or disfavor that alternative in competition
for funding or support against otherwise equally attractive alternatives that do
not similarly contradict (the relevant parts of) existing orthodoxy and therefore
remain more plausible avenues for successfully extending, expanding, sophisti-
cating, or supplementing our existing scientific conception of ourselves and the
world around us. But this same reason for increased skepticism remains in force
for the realist even when the alternatives being compared are anything but oth-
erwise equally attractive, that is, even when it is simply one among a wide range
of considerations bearing on the comparative attractions of investing in the pur-
suit of one theoretical proposal rather than another. Thus, holding all such fur-
ther considerations fixed in any particular case reveals that realists have reasons
instrumentalists lack for skepticism about just those theoretical proposals that
violate (the relevant parts of) existing theoretical orthodoxy.
Of course, the instrumentalist no less than the realist will need to make dif-
ficult choices about how to invest and distribute the scarce resources available
to support scientific inquiry itself. Many of the considerations she will weigh in
making such judgments or decisions will operate for her in precisely the same
way that they do for the realist: both, for example, will see alternative theoret-
ical proposals as more attractive or promising targets for investment the better
able they are to recover and/or explain whatever empirical consequences or
implications of existing theories have already been independently confirmed by
226 Contemporary Scientific Realism
Although realists may bristle at the suggestion that they defend a systematically
more theoretically conservative form of scientific inquiry than instrumentalists
do, this difference seems to generate an even more direct and immediate chal-
lenge to instrumentalism instead. After all, if such commitments to various parts
of contemporary theoretical orthodoxy are what enable realists to dismiss alter-
native theoretical possibilities out of hand when they conflict with too much of
what we think we already know about the world, it seems natural to worry that
the instrumentalist will, by contrast, wind up forced into an absurd permissive-
ness with respect to the alternative theoretical possibilities that she is prepared
to take seriously and/or consider potentially deserving investments of the time,
energy, money, and other scarce resources available for pursuing scientific in-
quiry itself. Surely even the instrumentalist thinks we should take a dim view
of investing scarce resources in finding and/or developing alternative theoret-
ical proposals contradicting claims like “fossils are the remains of once-living
organisms,” “many diseases are caused by bacterial or viral infections,” or “the
world around us is filled with microscopic organisms” as well as many others
Realism, Instrumentalism, Particularism 227
that seem put beyond serious question or reasonable doubt by the evidence we
already have. Although the instrumentalist remains open to the possibility that
future scientific communities may ultimately express these firmly established
facts using a very different theoretical vocabulary than our own,1 she should
nonetheless take a skeptical view of alternative theoretical proposals asserting
or implying the falsity of such claims (even as we ourselves express them). (Here
and throughout I assume that the theoretical alternatives in question simply con-
tradict, explicitly or implicitly, the relevant claims concerning fossils, infections,
microscopic organisms, etc., and do not offer convincing alternative explanations
of the evidence we presently take to support those claims.) But if the instru-
mentalist restricts her beliefs to only the empirical consequences or observable
implications of her best scientific theories, she might seem to lose any ground for
thinking that theoretical alternatives contradicting such claims should be taken
any less seriously or regarded with any more suspicion than those which instead
simply contradict, say, the far more speculative claims of our best scientific theo-
ries concerning the nature of dark matter or dark energy.
Recall, however, that the instrumentalist’s view is not that all claims of con-
temporary theoretical orthodoxy in science will ultimately be abandoned, but
instead simply that many central and fundamental claims will be and that we
are not generally in a position to predict in advance just which claims these will
be. What she rejects is the realist’s claim to have identified general or categorical
features of scientific theories and/or their supporting evidence from which their
approximate truth (or some analogue) can be reliably inferred. That is, the re-
alism debate itself has been most fundamentally concerned with whether there
is any general or categorical variety of empirical success or evidential support
that serves as a reliable indicator that a theory (or its privileged parts) will be
retained and ratified throughout the course of further inquiry. But realists and
instrumentalists alike can recognize exceptions to their general or generic ex-
pectations in particular cases based on evidence or other considerations specific
to the case in question. Many realists, for example, are prepared to make such an
exception in the case of quantum mechanics, whose empirical success is extraor-
dinary but whose very intelligibility to us as a description of how things actually
stand in some otherwise inaccessible domain of nature remains controversial.
Moreover, many realists are inclined to see the theory of evolution by natural se-
lection as extremely well confirmed despite the fact that the evidence supporting
that theory includes little in the way of the “novel predictive success” that they
argue is generally required to justify the claim that a theory (or its privileged
1 For example, when Joseph Priestly reported that after breathing “dephlogisticated air . . . my
breast felt peculiarly light and easy for some time afterwards” we judge that he made a true claim
about the effects of breathing oxygen using flawed or dated theoretical vocabulary, not a false or
empty claim about a substance that does not exist (see Kitcher 1993 100).
228 Contemporary Scientific Realism
parts) will persist throughout the course of further inquiry. While the instru-
mentalist holds instead that no generic characteristic or category of theories or
the evidence available in support of them entitles us to any more than the expec-
tation that the theory in question is a useful conceptual tool or instrument for
guiding our practical interaction with nature, she nonetheless remains just as
free as the realist to recognize particular cases as exceptions to this generic ex-
pectation on the basis of considerations specific to those cases. What she cannot
do is respond to the challenge by specifying generic epistemic characteristics or
categories that distinguish trustworthy from untrustworthy scientific beliefs, for
such generic characteristics and categories are just what she thinks we have yet to
identify successfully.
Accordingly, her judgment that a particular belief or claim is established be-
yond a reasonable doubt will have to be a function of her evaluation of the details
of the specific evidence she has in support of that particular belief. I have argued
at length elsewhere (Stanford 2010), for example, that the details of the evidence
we now have supporting the (once highly contentious) hypothesis that fossils
are the remains of previously living organisms should lead us to conclude that
this hypothesis is not merely a useful cognitive instrument but is in addition an
accurate description of how things stand in nature itself. Of particular impor-
tance in that case, I suggested, was an abundance of what I called “projective”
evidence in support of this hypothesis (especially from the field of experimental
taphonomy) in addition to merely eliminative and abductive forms of evidence.
In a similar fashion, it will be the details of the evidence available in particular
cases which convince the instrumentalist that any particular belief is established
beyond a reasonable doubt. (Indeed, this same response might also appeal to se-
lective or Middle Path realists, who face their own version of the problem insofar
as proposed theoretical alternatives might well preserve the “working posits,”
“structural claims,” or other privileged elements of contemporary theoretical or-
thodoxy while nonetheless disqualifying themselves from serious consideration
by contradicting claims like “many diseases are caused by bacterial infections” or
“fossils are the remains of once-living organisms.”)
Here an important strategic difference emerges between the realist’s and the
instrumentalist’s respective engagements with the historical record. The realist
begins from the attempt to explain the success of science and seeks to defend the
credibility of one particular explanation against potentially undermining histor-
ical counterevidence, in the process refining whatever criterion of epistemic se-
curity she proposes for identifying just which categories of scientific claims and
commitments she thinks can be trusted to persist throughout further scientific
inquiry. By contrast, the instrumentalist begins from the demonstrably serious
threat to the persistence of our scientific beliefs posed by the historical record
and tries to find ways to restrict our beliefs so as to obviate or sufficiently mitigate
Realism, Instrumentalism, Particularism 229
Since its inception in the writings of Smart (1963), Putnam (1975), van Fraassen
(1980), Boyd (1984), Laudan (1981), and others, the modern realism debate has
been predicated on the assumption that there is some point to ascending to the
230 Contemporary Scientific Realism
NOA’s attitude makes it wonder whether any theory of evidence is called for.
The result is to open up the question of whether in particular contexts the evi-
dence can reasonably be held to support belief (regardless of the character of the
objects of belief). Thus NOA, as such, has no specific ontological commitments.
It has only an attitude to recommend: namely, to look and see as openly as one
can what it is reasonable to believe in and then to go with the belief and com-
mitment that emerges. Different NOAers could, therefore, disagree about what
exists, just as different, knowledgeable scientists disagree. (1986b 176–7)
What binds these NOAers together, it seems, is simply their skepticism about
whether ascending to some more general or abstract philosophical level of in-
vestigation, analysis, or reflection on science itself will help them make any
progress in deciding what to believe about the world. Scraping away or refusing
to indulge in the universalizing interpretations offered by both realist and anti-
realist philosophers of science simply leaves us with the first order practices of
advancing, challenging, and defending particular scientific claims in particular
scientific contexts. “The general lesson,” Fine suggests, “is that, in the context of
science, adopting an attitude of belief has as warrant precisely that which science
itself grants, nothing more but certainly nothing less” (Fine 1986a 147).
In a similar fashion, Penelope Maddy sees the trouble at the root of both re-
alism and instrumentalism as their shared inclination to try to ascend to some
higher philosophical court of epistemic evaluation in which we leave behind the
ordinary sorts of evidence and evaluative standards that are characteristic of sci-
ence itself. She sees van Fraassen, for example, as conceding that the evidence in
favor of the existence of atoms is perfectly adequate and convincing for scientific
purposes but insisting that this does not settle whether or not we have sufficient
epistemological or philosophical grounds for holding such beliefs:
She suggests that realists like Boyd mistakenly take the bait here, trying to rise
to the challenge of defending our scientific beliefs in this higher, philosophical
court of inquiry where ordinary scientific evidence is disallowed, and thereby
being pushed “away from the details of the local debate over atoms and towards
global debates over such questions as whether or not the theoretical terms of ma-
ture scientific theories typically refer” (2001 46).
By contrast, Maddy’s “Second Philosopher” (2007) simply declines the invi-
tation to leave ordinary scientific evidence and methods of evaluation behind
and ascend to any such extrascientific or philosophical court of evaluation for
scientific beliefs. Indeed, she feels no temptation to follow realists and antirealists
down this shared rabbit hole unless and until someone can convincingly explain
what this further distinctively epistemological or philosophical inquiry is sup-
posed to achieve or accomplish and why the ordinary scientific evidence that
actually convinces her of the reality of atoms should be treated as inadequate or
irrelevant to that inquiry. She insists instead that scientific inquiry itself already
represents our best efforts to decide (for any purposes we do or should actually
care about) whether or not particular scientific claims are true and/or how much
confidence in any particular claim is warranted, and that such questions can only
be convincingly answered in the case-by-case or piecemeal manner that science
itself employs:
Although both Fine and Maddy deny that there are any useful answers to
questions about science at the level of generality at which philosophers have
traditionally sought them, they are both careful to leave room for at least
the bare possibility that we might secure claims of somewhat greater gen-
erality (regarding confirmation, explanation, and the like) “bottom up,” as
it were, by generalizing over and abstracting away from the details of par-
ticular investigations (Fine 1986a, 179–80; Maddy 2007 403n). But at least
some particularists are explicitly skeptical regarding this possibility. Magnus
and Callender, for example, have argued influentially that “profitable” re-
alism debates must be conducted at the “retail” level of individual, particular
234 Contemporary Scientific Realism
scientific inquiries rather than “wholesale.” And although they concede that
it might be “logically possible” to abstract away from or generalize over the
results of such retail investigations, they doubt that the resulting guidance
will be “either interesting or useful,” suggesting instead that “[w]e should pay
attention to particular cases for their own sake and not as proxies for some-
thing else” and that “the great hope for realism and anti-realism lies in re-
tail arguments that attend to the details of particular cases” (2004 336). And
Maddy herself has argued that when any such general epistemic guidance we
embrace conflicts with the more detailed, informed, and contextual judgments
of confirmation we make regarding particular cases, we typically prioritize
the normative force of the particular judgments over that of the more general
guidance. This is in part because such general guidance must itself remain suf-
ficiently vague, indeterminate, and open-ended to permit further specification
or adjustment in response to new developments: even if we generalized and
abstracted our way bottom-up from historical or other evidence to the con-
clusion that we should only believe in entities we can “detect,” for instance, she
suggests that we will subsequently adjust the boundaries of what we count as
“detection” (thereby including or excluding some new method) so as to pre-
serve the normative force of our reflective confirmational judgments about in-
dividual scientific cases (including those making use of the new method). In
this way, as Maddy says it, “the ordinary science, not the [general] criterion, is
doing the work” (2007 402).
Moreover, even many of those who do still seek “interesting and useful” ge-
neral epistemic guidance concerning our scientific beliefs have in recent decades
done so in ways that reflect growing particularist sympathies, suggesting that
such guidance may well consist in the accumulation and synthesis of principles
each of which applies to only a limited range of cases rather than the sort of fully
general criterion of epistemic security sought by their classical predecessors.
Richard Miller (1987), for example, argues that convincing realist commitments
must be grounded in “topic-specific truisms” whose force is restricted to partic-
ular contexts or domains of inquiry, while Jamin Asay (2019) suggests that re-
alism debates should be conducted at the level of particular sciences rather than
science as a whole. More generally, Larry Sklar argues that “by far the most inter-
esting issues will be found in the detail of how the global, skeptical claims . . . be-
come particularized and concrete when they appear as specific problems about
theories within the context of ongoing physical theorizing” (2000 78). And more
recently, Juha Saatsi has advocated a similarly modest particularism, arguing
first that it is “a manifestation of philosophical arrogance to think that as a re-
alist philosopher one commits oneself to providing a global recipe—largely inde-
pendently of the science steeped in relevant details—for revealing what aspects
Realism, Instrumentalism, Particularism 235
(including the novel predictive successes) of those theories. None of these claims
amounts to anything like the fully general criteria of epistemic security sought
by classical realists and antirealists, but each is nonetheless a clear example of
epistemic guidance intermediate in generality between those classical ambitions
and the radical particularist’s competing conviction that useful epistemic guid-
ance for science can only be found in the details of the evidence we have in sup-
port of particular scientific claims. Middle Path realists and instrumentalists
are thus once again united, this time charting a course between the fully general
commitments and expectations of their classical predecessors on the one hand
and the sharply opposed expectations of radical or extreme forms of particu-
larism on the other.
10.4 Conclusion
I’ve suggested here that the modern scientific realism debate is most cen-
trally concerned with whether or not we have yet discovered some particular
form(s) of empirical success or evidential support allowing us to reliably pre-
dict whether particular theories (or privileged parts thereof) will be retained
and ratified throughout the course of further scientific inquiry. As I’ve tried
to emphasize, realists and instrumentalists fundamentally disagree on the an-
swer to this question (and thus on the predictability of further changes in our
theoretical conception of the world around us) in ways that actually make
a difference to how we should pursue our further scientific investigation of
the world. But I have also argued that in recent decades many contempo-
rary realists and instrumentalists have come to share a set of fundamental
commitments that unite them more closely with one another than either is to
their classical predecessors, including Uniformitarianism, commensurability,
and the Maddy/Wilson principle, as well as to seeking useful guidance con-
cerning the epistemic security of our scientific theories and beliefs at a level
of abstraction and generality intermediate between that anticipated by radical
or extreme forms of particularism and the perfectly general criteria of epi-
stemic security envisioned by classical forms of realism and instrumentalism
alike. For such Middle Path realists and instrumentalists, arguably more im-
portant than any remaining point of debate or disagreement between them
is the shared epistemic project in which they are jointly engaged: realists and
instrumentalists alike on the Middle Path are already actively seeking to iden-
tify, evaluate, and refine candidate indicators of epistemic security for our sci-
entific beliefs, and both see the historical record of scientific inquiry itself as
the most important source of evidence we have available to us for pursuing
that joint project together.
Realism, Instrumentalism, Particularism 237
References
Asay, J. (2019) Going local: A defense of methodological localism about scientific realism.
Synthese 196: 587–609.
Bement, A. L., Jr. (2007) Important notice 130: Transformative research, National Science
Foundation, Office of the Director. https://www.nsf.gov/pubs/issuances/in130.pdf.
Boyd, R. (1984) The current status of scientific realism, in Jarrett Leplin (ed.) Scientific
Realism. Berkeley: University of California Press, 41–82.
Chakravartty, A. (2007) A Metaphysics for Scientific Realism: Knowing the Unobservable.
Cambridge: Cambridge University Press.
Egg, M. (2016) Expanding our grasp: Causal knowledge and the problem of unconceived
alternatives. British Journal for the Philosophy of Science 67: 115–141.
Fine, A. (1986a) The Shaky Game: Einstein, Realism, and the Quantum Theory.
Chicago: University of Chicago Press.
Fine, A. (1986b) Unnatural attitudes: Realist and instrumentalist attachments to science.
Mind 95: 149–179.
Fischer, R. (1930) The Genetical Theory of Natural Selection. Oxford: Clarendon Press.
Kitcher, P.S. (1993) The Advancement of Science. New York: Oxford University Press.
Klee, R. (1999) Scientific Inquiry: Readings in the Philosophy of Science. New York: Oxford
University Press.
Laudan, L. (1981) A confutation of convergent realism. Philosophy of Science 48: 19–48.
Lyons, T. (2002) Scientific realism and pessimistic meta-modus tollens, in S. Clarke and
T.D. Lyons (eds.) Recent Themes in the Philosophy of Science: Scientific Realism and
Commonsense. Dordrecht: Springer, 63–90.
Lyons, T. (2006) Scientific realism and the stratagema de divide et impera. British Journal
for the Philosophy of Science 57: 537–560.
Lyons, T. (2016) Structural realism versus deployment realism: A comparative evaluation.
Studies in History and Philosophy of Science Part A 59: 95–105.
Lyons, T. (2017) Epistemic selectivity, historical threats, and the non-epistemic tenets of
scientific realism. Synthese 194: 3203–3219.
Maddy, P. (2001) Naturalism: Friends and foes. Philosophical Perspectives 15: 37–67.
Maddy, P. (2007) Second Philosophy: A Naturalistic Method. Oxford: Oxford UP.
Magnus, P.D., and Callender, C. (2004) Realist ennui and the base rate fallacy. Philosophy
of Science 71: 320–338.
Miller, R. (1987) Fact and Method: Explanation, Confirmation, and Reality in the Natural
and the Social Sciences. Princeton: Princeton University Press.
Novick, A., and Scholl, R. (2020) Presume it not: True causes in the search for the basis of
heredity. British Journal for the Philosophy of Science 71: 59–86.
Peters, D. (2014) What elements of successful scientific theories are the correct targets for
“selective” scientific realism? Philosophy of Science 81: 377–397.
Putnam, H. (1975) Mathematics, Matter, and Method (Philosophical Papers Vol. 1).
London: Cambridge University Press.
Psillos, S. (1999) Scientific Realism: How Science Tracks Truth. London: Routledge.
Saatsi, J.T. (2017) Replacing recipe realism. Synthese 194: 3233–3244.
Saatsi, J.T., and Vickers, P. (2011) Miraculous success? Inconsistency and untruth in
Kirchhoff ’s diffraction theory. British Journal for the Philosophy of Science 62: 29–46.
Sklar, L. (2002) Theory and Truth: Philosophical Critique within Foundational Science.
Oxford: Clarendon Press.
238 Contemporary Scientific Realism
Smart, J.J.C. (1968) Between Science and Philosophy. New York: Random House.
Stanford, P.K. (2006) Exceeding Our Grasp: Science, History, and the Problem of
Unconceived Alternatives. New York: Oxford University Press.
Stanford, P.K. (2010) Getting real: The hypothesis of organic fossil origins. Modern
Schoolman 87: 219–243.
Stanford, P.K. (2015) Catastrophism, Uniformitarianism, and a Scientific Realism Debate
That Makes a Difference. Philosophy of Science 82: 867–878. doi:10.1086/683325
Van Fraassen, B.C. (1980) The Scientific Image. Oxford: Clarendon Press.
Vickers, P. (2013) A confrontation of convergent realism. Philosophy of Science
80: 189–211.
Vickers, P. (2017) Understanding the selective realist defence against the PMI. Synthese
194: 3221–3232.
Wilson, M. (2006) Wandering Significance: An Essay on Conceptual Behaviour.
Oxford: Clarendon Press.
Worrall, J. (1989) Structural realism: The best of both worlds? Dialectica 43: 99–124.
11
Structure not Selection
James Ladyman
Department of Philosophy
University of Bristol
[email protected]
11.1 Introduction
Structural realism is the best of both worlds, according to John Worrall (1989),
because it takes account of the most powerful argument against scientific re-
alism, as well as the motivations for it. The problem for scientific realism that
he takes most seriously is that posed by the actual historical record of scientific
theories in physics and chemistry, because it shows, all philosophical argument
aside, that not everything that is supposed by the highly empirically successful
theories of the past is real by the lights of current science. He was not troubled
by other arguments against scientific realism (such as the argument from the un-
derdetermination of theory by evidence), and the problem of theory change is
the only problem for standard scientific realism that Worrall sought to solve by
adopting structural realism. It is the problem that others, notably Stathis Psillos
(1999), seek to solve by what David Papineau dubbed “selective realism” (1996),
which involves analyzing case studies from the history of science to find a for-
mula that restricts the epistemic commitment of scientific realists to parts of
theories that will be retained. The idea is that close examination of the history
of science will reveal what it is about abandoned theoretical constituents that
distinguishes them from those that are retained, so that a selective commitment
to current theories can be then be applied in the confidence that the relevant on-
tology will not subsequently be abandoned. (Psillos’s formula is roughly “only
believe in the reference of the central theoretical terms of theories that play an
essential role in generating novel predictive success.”) While selective realism is
motivated and tested by studying the history of science, it is a form of realism
that involves criteria that can be applied to our best current theories to select in
advance what of their ontology will be preserved. If selective realism is under-
stood as selecting part of science which we can be confident will be retained on
James Ladyman, Structure not Selection In: Contemporary Scientific Realism. Edited by: Timothy D. Lyons and Peter
Vickers, Oxford University Press. © Oxford University Press 2021. DOI: 10.1093/oso/9780190946814.003.0011
240 Contemporary Scientific Realism
1 A referee claims that epistemic structural realism involves looking at our current best theories
and trying to identify the structure to which we then make a realist commitment. However, nobody
has ever actually done this with our best science in the attempt to tell what will be retained on theory
change.
Structure not Selection 241
kind of selective realism. The first is that it does not offer a criterion for selective
realist commitment to current theories (as argued in the next section), and the
second is that it addresses ontological problems other than the problem of theory
change.
Section 11.3 argues that the construction of a positive realist metaphysics that
can deal with the problem of theory change, must also take account of ontolog-
ical issues, most importantly the lack of a fundamental level and scale-relativity
(and explains how these problems are related to the problem of theory change).
Section 11.4 reviews the form of OSR involving the theory of real patterns de-
veloped by Ladyman and Ross, and shows how it addresses the ontological
problems for scientific realism, emphasizing the centrality of the idea of objec-
tive modal structure (where this is also retained on theory change even when
ontology changes radically). It is the latter that makes OSR a form of realism, and
differentiates it from van Fraassen’s structural empiricism, as well as allowing
OSR to avoid collapsing the distinction between abstract and concrete structure
(van Fraassen’s problem of “pure” structuralism, 2006), and making an account
of causation and other modal aspects of scientific knowledge possible. The paper
concludes with some remarks about the relationship between structural realism
and the historiography of science.
(i) There have been many empirically successful theories in the history of
science that have subsequently been rejected, and whose theoretical
terms do not refer according to our best current theories.
(ii) Our best current theories are no different in kind from those discarded
theories, and so we have no reason to think they will not ultimately be
replaced as well.
So, by induction we have positive reason to expect that our best current theories
will be replaced by new theories according to which some of the central theoret-
ical terms of our best current theories do not refer, and hence, we should not be-
lieve in the approximate truth or the successful reference of the theoretical terms
of our best current theories.
Both (i) and (ii) are dubious. The number of theories that can be listed in sup-
port of (i) is drastically reduced if the notion of empirical success is restricted
to making successful novel predictions (Psillos 1999). (ii) is undermined by the
unprecedented degree of quantitative accuracy and practical and predictive
success of current science. Furthermore, it may be argued that only since the
twentieth century have the physical sciences been fully integrated because of two
developments, namely the atomic valence theory of the periodic table and elec-
tronic theory of the chemical bond. Science is now very mature in the sense that
testing hypotheses at the cutting edge in one field relies upon background the-
ories from the basics of other sciences that are fully integrated with each other.
The common system of quantities and units, and many of the core background
theories used across all the sciences have been stable for a long time. There is no
precedent in history for the current extent of overlap and reinforcement among
2 Of course, they can be related by asking whether knowledge of the observable phenomena is
the limit of what is retained on theory change. Obviously observable phenomena are lost on theory
change, in so far as new theories are more empirically adequate than their predecessors.
Structure not Selection 243
theories in the natural sciences, nor for the commonality of statistical and other
methods.
However, there is another form of the problem of theory change that does not
have the form of an induction based on Laudan’s much-discussed list. Rather it
is an argument against the main argument for scientific realism namely the no
miracles argument. Laudan’s paper was also intended to show that the successful
reference of its theoretical terms is not a necessary condition for the novel pre-
dictive success of a theory (1981 45), and so there are counterexamples to the
no miracles argument. No attempt at producing a large inductive base need be
made; rather, one or two cases are argued to be counter-arguments to the re-
alist thesis that novel predictive success can only be explained by successful ref-
erence of key theoretical terms. As Psillos (1999 108) concedes, even if there are
only a couple of examples of false and non-referring, but mature and strongly
successful theories, then the “explanatory connection between empirical success
and truth-likeness is still undermined.” Hence, Laudan’s ultimate argument from
theory change against scientific realism is not really an induction of any kind, but
a reductio.
So there are examples of theories that were mature and had novel predictive suc-
cess but which are not approximately true.
Approximate truth and successful reference of central theoretical terms is not
a necessary condition for the novel-predictive success of scientific theories. So,
the no miracles argument is undermined since, if approximate truth and suc-
cessful reference are not available to be part of the explanation of some theories’
novel predictive success, there is no reason to think that the novel predictive suc-
cess of other theories has to be explained by realism (Ladyman and Ross 2007
84–85). Recall that the realist claims that novel predictive success is only intelli-
gible on the realist view so once it is established that it is possible where realism is
untenable then the realist’s argument is much less if at all compelling (assuming
it was to some extent compelling in the first place).
The ether theory of light and the caloric theory of heat seem to have enjoyed
novel predictive success by anyone’s standards.
If their central theoretical terms do not refer, the realist’s claim that approx-
imate truth explains empirical success will no longer be enough to establish
244 Contemporary Scientific Realism
realism, because we will need some other explanation for success of the caloric
and ether theories. If this will do for these theories then it ought to do for others
where we happened to have retained the central theoretical terms, and then
we do not need the realist’s preferred explanation that such theories are true
and successfully refer to unobservable entities. (Ladyman and Ross 2007 84, see
also Ladyman 2002)
Stathis Psillos (1999) first argues that these are the only two truly problematic
cases for the scientific realist. He then adopts his “divide and conquer strategy,”
which is to adopt a different solution for each though both are ways of denying
(II). In the case of the ether, he argues that according to his causal-descriptivist
theory of reference, the term “ether” refers, while in the case of caloric he argues
that there was never any reason why a realist of the time had to commit to it
being a material substance and that it was not after all a “central” theoretical term
in the sense that demands successful reference. Hence, he argues that we can be
confident that all the central theoretical terms of our current science do in fact
refer to things in the world.
Psillos argues that Fresnel’s use of the term ether referred to the electromag-
netic field and that continuity of reference being assured the problem posed
by that case is solved. This is contentious but anyway other cases are also prob-
lematic. For example, consider the relationship between classical and quantum
physics. Classical physics describes phenomena in diverse domains to very high
accuracy and made predictions of qualitatively new phenomena. It had a vast
amount of positive evidence in its favor. The rationalists and Kant had in var-
ious ways attempted to derive its principles a priori. However, it turned out to
be completely wrong for very fast relative velocities and for very small particles,
as was shown by relativistic and quantum physics respectively. This is surely the
problem of theory change in its most stark form, yet the problem is nothing to
do with the abandonment of theoretical terms, or difficulty in securing refer-
ence.3 Clearly the reference of the term “energy” at least overlaps in classical and
quantum theories because quantities in each coincide in the limit as the number
of particles increases. This does not solve the problem that the radical ontological
and metaphysical differences between the theories pose for the realist. Different
theories often use the same terms, to say very different things about what the
world is like. The question of which terms refer does not settle whether it is
3 A referee objects that this may be thought to be a less serious form of theory change because of
the precise mathematical relationships of approximation between the laws of the respective theories
but these can only be interpreted in terms of approximate truth if the ontological discontinuities in
question are not taken to be significant, while the laws are taken to have more than the purely empir-
ical content van Fraassen and other antirealists take them to have. This is just what ontic structural
realism recommends.
Structure not Selection 245
plausible to say that the metaphysics and ontology of the theory is even approxi-
mately true, because in many cases the same theoretical term is still used despite
radical changes.
For this reason, Worrall (1989) argues that the project of defending realism
by securing reference for abandoned theoretical terms is missing the point of the
problem as well as the key to its solution.4 Arguably “field” is just the new word
for what the ancients and early moderns termed “aether” or “ether.” Fields like
the ether permeate all of space and are posited to explain certain phenomena.
Anaxagoras is said to have introduced the idea of the ether to propel the planets,
and subsequently there were many ethers to account for various things, notably
the propagation of electricity and light through empty space. Fresnel’s optical
ether was supposed to be a solid and Maxwell’s field behaves very differently and
is said to be immaterial. However, from our current perspective the ether has
much more in common with the field than atoms do with the indivisible particles
of antiquity (Stein 1989).
Structural realism is not motivated by the need to secure reference across
theory change but by Heinz Post’s (1971) General Correspondence Principle, ac-
cording to which the well-confirmed laws of old theories are retained by their
successors as approximations within certain domains. Even comparing the well-
confirmed laws of theories that differ wildly in their metaphysical dimensions
such as quantum mechanics and classical mechanics, or general relativity and
Newtonian gravitation, the past theories are limiting cases of the successor the-
ories. This confirms Poincaré’s idea of physics as ruins built on ruins in the sense
that the old theories form the foundations of the new ones rather than the ground
being cleared before work begins. The key point is that more than the empirical
content and phenomenological laws of past theories is retained. For example, it
is not just Kepler’s laws and Galileo’s kinematical laws that are preserved as lim-
iting cases of general relativity, but Newton’s inverse square force law that unifies
and corrects those laws is retained in the form of the Poisson equation as a low-
energy limit of Einstein’s field equation. Similarly, the classical limit of quantum
physics gives rise to the approximate truth of the laws of Newtonian mechanics
for macroscopic bodies.5
The laws take mathematical form, and there are special cases, such as that of
Fresnel’s equations, where the very same equations are reinterpreted in terms
of different entities. However, of course this is not the norm, and in many
cases mathematical structure is lost and radically modified on theory change.
4 Whether or not Worrall is right there is less much emphasis on reference in the subsequent liter-
just the occurrent regularities between actual events. This is structural empiricism as defended by
van Fraassen.
246 Contemporary Scientific Realism
Structural realism does not require that all mathematical (or any other kind of)
structure is preserved on theory change. If it did it would be refuted by the fact
that the mathematical form of the physics of Maxwell is different from that of
the physics of Einstein. As French and Ladyman (2011) put it, “[t]he advocate of
OSR is not claiming that the structure of our current theories will be preserved
simpliciter but rather that the well-confirmed relations between the phenomena
will be preserved in at least approximate form and that the modal structure of the
theories that underlies them, and plays the appropriate explanatory role, will also
be preserved in approximate form” (32–33). However, this does not enable any
kind of selective realism because it says nothing at all about how the laws of the
theory are only approximate, and so does not enable the selection of what will be
retained in advance. French and Ladyman (2003) emphasized that they under-
stood OSR as the view that the mathematical structure of theories represents,
to some extent, the modal structure of the world. Hence, when the mathemat-
ical structure changes this often reflects the fact that some of how the world’s
modal structure was represented to be was wrong. For example, the modal struc-
ture of the world is different according to Special Relativity than it is according
to Newtonian mechanics because, for example, the velocity addition law of the
latter is not even approximately true in case of frames moving relative to each
other at velocities close to that of light, and this is reflected in the mathematical
change to the Lorentz transformations from the Galilean transformations.
The point of OSR in the context of theory change is that it inflates the ontolog-
ical importance of relational structure to take account of the fact that the ontolog-
ical status of the relevant entities may be very different in the different theories,
yet the relationship between the way they represent the modal structure of the
world can nonetheless be studied in depth by investigating the relevant mathe-
matical structures (even in the case of geocentrism and heliocentrism, as shown
by Saunders 1993). Ladyman (1997) argues that mathematical representation is
ineliminable in much of science and takes this to be key to OSR (and work by
French, Ross, and Wallace on OSR is also predicated on it). However, this does
not mean that this kind of structural realism only applies to mathematicised the-
ories, as shown by the following case.
Kuhn and Feyerabend popularized the view that theory change in the history
of science disrupts narratives of continuous progress and realism. The response
of many philosophers of science to this challenge, particularly in Popper’s school,
was to examine episodes in the history of science in detail to see if the histori-
cally inspired critiques were really supported by the evidence. Noretta Koertege
(1968) considered phlogiston theory as an example supporting the general cor-
respondence principle because the well-confirmed empirical regularities stated
in terms of phlogiston theory (such as that air saturated with phlogiston by com-
bustion does not support respiration) are true when translated into the language
Structure not Selection 247
of oxygen (de-oxygenated air does not support respiration). George Gale (1968)
argued that phlogiston theory was also a good explanatory theory that explained
the loss of weight of wood, coal, and ordinary substances when burnt, among
many other things. Furthermore, and crucially for structural realism, the fact
that combustion, respiration, and calcination of metals are all the same kind of
reaction and there is an inverse kind of reaction too is an example of a theoretical
relation that is retained from phlogiston theory. Even though there is no such
thing as phlogiston, the tables of affinity and antipathy of phlogistic chemistry
express real patterns (of which more in the next section) that we now express in
terms of reducing and oxidizing power.
Indeed as also discussed by Gerhard Schurz (2009), the terms “phlogistication”
and “dephlogistication,” meaning the assimilation and release of phlogiston re-
spectively when they were used in chemistry, can now be regarded as referring
to the processes of oxidation and reduction (where oxidation of X = the forma-
tion of an ionic bond with an electronegative substance and reduction of X is the
regaining of electrons). If the oxidizing agent is oxygen, and the oxidized com-
pound is a source of carbon then the product is carbon dioxide (“fixed air”), and
if the oxidizing agent is an acid, then hydrogen (“inflammable air”) is emitted, in
keeping with phlogiston theory. One could go further and allow that “phlogiston
rich” and “phlogiston deficient” refer too, namely to strongly electro-negative
and electro-positive molecules respectively. One could even argue that “phlo-
giston” refers to electrons in the outer orbital of an atom (as suggested by Andrew
Pyle, see Ladyman 2011). However, none of this changes the fact that phlogiston
theory is wrong and that there is nothing contained in all flammable materials
that leaves them on combustion, and nothing that metals lose when they become
calces. No form of selective realism could have told us in advance that phlogiston
would be abandoned, because it was essential to much empirical success.
The examples from mathematical physics cited earlier and the case of phlo-
giston theory show that even though the ontology of science may change quite
radically at the level of objects and properties (for example, there is no elastic
solid ether, no principle of combustion), there can be continuity of the modal
structure that is attributed to the world. OSR is not selective realism; rather, it is
the generic modification of standard scientific realism as exemplified by Psillos
(1999). The point is that “[t]heories, like Newtonian mechanics, can be literally
false as fundamental physics, but still capture important modal structure and re-
lations” (Ladyman and Ross 2007 118). Likewise important features of the modal
structure of the world can be represented in terms of phlogiston, as explained
earlier. When Worrall made his dialectical move in the realism-anti-realism
debate, philosophers of science had given a lot of attention to the problem of
the reference of theoretical terms, because it had been realized that they could
be discarded from even very successful science. In the light of this discussion
248 Contemporary Scientific Realism
it should be clear that the real problem is not the reference of theoretical terms
but that the notion of approximate truth is hard to apply to the ontology and
metaphysics of scientific theories. Atoms are simply not indivisible in meta-
physical terms, though they are of course in practical terms without a great deal
of very sophisticated engineering. OSR does not give an account of metaphys-
ical approximation, but it begins from the insight that more than the empirical
content of theories is preserved. The claim that modal structure is preserved is
compatible with radical ontological discontinuity, even in the extreme case of
phlogiston, which Psillos did not seek to redeem.
As mentioned in section 11.1, OSR was also introduced to address ontological
problems other than the problem of theory change. The next section outlines
several that are not discussed in the literature about selective realism. Among the
most important are the issues of fundamentality and scale-relativity.
The problems explained next are only specifically ontological problems for scien-
tific realism, because scientific realism involves ontological commitment on the
basis of science, but they are problems more generally for any philosophy of sci-
ence. There are a number of ontological problems for scientific realism that are
not discussed in the context of selective realism or epistemic structural realism.
Like the problem of theory change, these problems with standard scientific re-
alism arise from how science actually is.6
6 As mentioned earlier OSR is also supposed to take account of the problems with the standard
scientific realist interpretation of physics, in particular, to take account of the irreducibility of rela-
tional structure and the problems of identity and individuality in quantum mechanics and space-
time physics. These issues are not discussed in the present paper but are central for Ladyman (1998),
French and Ladyman (2003), Ladyman and Ross (2007), and French (2014). For a recent discussion
see Ladyman (2015, 2016).
7 Of course, many scientific realists are not fundamentalists and reject the idea that ontology
These problems are related. (a) suggests that a straightforward realist interpre-
tation of scientific theories and models is naïve. For example, the equator, the
orbitals of atoms, and temperature of a gas are all abstractions and idealizations
to a degree. (a) relates to (b) because something discrete may be represented
at a less fundamental level as something continuous (as with ordinary fluids),
and something heterogeneous at one scale may be idealized as something ho-
mogeneous at another, as with any polyatomic molecular substance. (b) relates
to (c) because the special sciences are often associated with restricted scales of
energy, length, and time. These problems must be faced by any scientific re-
alist. These problems are also related to the problem of theory change, because
past theories are also often those that we now regard as working at particular
scales and as describing the emergent entities that are now described by a part of
8 Entities like rivers and waves are arguably necessarily extended in time but nothing here depends
on that claim.
250 Contemporary Scientific Realism
physics that has the status of a special science compared to fundamental physics.
For example, atoms are now part of atomic physics, which is not fundamental
and describes reality at a restricted scale. The ontology of such theories remains
“effective,” in the sense that the entities it posits are part of empirically successful
descriptions and models.
Defenders of scientific realism such as Psillos (1999) and Anjan Chakravartty
(2017) are primarily concerned with epistemological issues rather than the
problems just outlined. However, in other contexts scientific realism is argued
to lead to the elimination of all but the most fundamental physical entities (see,
for example, Rosenberg 2011). In this way, scientific realism becomes anti-
naturalistic and cannot be claimed to take the ontological commitments of
science at face value. On the other hand, those who reject eliminativism or re-
ductionism need a version of realism that takes account of (a)–(c). The next sec-
tion reviews how the real patterns account of ontology incorporated into OSR by
Ladyman and Ross makes for a unified solution to the ontological problems for
scientific realism.
For French and Ladyman (2003) and in their later publications jointly and sep-
arately, OSR is to be understood as the claim that science represents the objec-
tive modal structure of the world. Ladyman and Ross (2007) and French (2014)
argue that incorporation of the idea that the world has a modal structure that is
described by science into OSR makes it a realist philosophy of science, and dis-
tinct from the structural empiricism of van Fraassen (2006). They also agree that
OSR so construed can accommodate the modal aspects of scientific knowledge
including causation, law, and symmetry.9 Modal structure is to be understood
roughly as an abstraction of causal and nomological structure. For example,
it is part of the modal structure of the world that metals expand when heated,
that mass is conserved in chemical reactions, and that heat cannot be con-
verted into work with perfect efficiency. These are all modal claims that support
counterfactuals and explanations, and they have all survived very different ideas
about the ontology in terms of which they are stated.
However, the forms of OSR that they defend also differ significantly espe-
cially in respect of how they relate to problems (a)–(c) of the last section. French
(2014) is an eliminativist about all ontology except the structures of physics, and
9 Berenstain and Ladyman (2012) also argue that the arguments for scientific realism are undercut
without it. Note that this makes OSR an immodest and to some extent speculative philosophical po-
sition, unlike epistemic structural realism which is quietist about metaphysics and very modest about
epistemology.
Structure not Selection 251
in this way avoids problems (b) and (c) by denying that anything that is not fun-
damental exists. This is a big price to pay for a naturalist because it departs so
radically from taking science at face value.10
Ladyman and Ross (2007) develop a version of OSR involving the theory of
real patterns in order to address (a)–(c) and develop a position that takes the
special sciences to give us irreducible knowledge of the modal structure of the
world.11 The commitment to objective modal structure is essential to making
the resulting naturalized metaphysics a form of realism, since the theory of real
patterns can also be developed as a form of instrumentalism. We take it that
all our scientific knowledge is about effective ontology in the sense of the pre-
vious section, and accordingly seek a unified way of thinking about the entities
of physics and the special sciences. We propose the idea of real patterns to this
end. Real patterns are the features that are used in everyday life and in science
to describe the world in terms of projectible regularities. As such they are not
confined to any ontological category and may be events, objects, processes, or
properties. For example, Mount Everest is a real pattern that features in a host
of projectible regularities about the world. To be genuine entities they need to
be non-redundant, so that, for example, the disjunction of real patterns is not a
new real pattern. Redundancy can be captured in different ways, the most ob-
vious being information theory. Real patterns compress information (though
at the cost of loss of precision). Much can be said about Everest that would be
much more costly to state otherwise. Describing the world without talking about
Mount Everest would be possible, but it would be harder. Real patterns simplify
the description of the world relative to some background already established lan-
guage (of more basic real patterns). Alternatively, redundancy can be understood
dynamically. Some coarse-grained variables feature in a much simpler dynamics
than that of the underlying variables; think of a rolling ball describable by the
position and momentum of its center of mass and its angular velocity, instead of
all the positions and momenta of its parts. However, some such coarse-grained
variables do not feature in any simpler dynamics and they are redundant and an-
ything defined in terms of them is not a real pattern.12
ephemeral. A rainbow is a real pattern but only with respect to vision perception. Real patterns can
also include events such as parties or revolutions which are of course notoriously difficult to individ-
uate. The theory of real patterns does not solve those problems of individuation but takes them to be
practical problems in some cases and pseudo-problems otherwise.
252 Contemporary Scientific Realism
The real patterns account of ontology is compatible with the scale relativity
of ontology, because it requires only that there be some regime in which the real
pattern in question is projectible (and non-redundant) for it to exist. Different
emergent structures are found at different scales, and accordingly there are real
patterns that exist at some scales and not at others. For example, Mount Everest
does not exist at atomic scale because there are no projectible regularities about
entities at that scale that involve Mount Everest. The real patterns account is com-
patible with taking the ontologies of the special sciences at face value because
insofar as the science in question captures the modal structure of the world, its
ontology involves real patterns just as much as that of physics. Hence, problems
(b) and (c) can be solved without adopting eliminativism about everything ex-
cept fundamental physics.
As Dennett (1991) makes clear in his original discussion, being a real pat-
tern is not an all or nothing matter. Mountains don’t have exact boundaries.
As many philosophers find mystifying, almost all events, objects, and proper-
ties commonly recognized as real involve some if not many kinds of vagueness.
The real patterns account accommodates (a) because it is not based on unreal-
istic ideas of a pristine scientific ontology to accompany exact and fundamental
laws. Representations can be more or less good at capturing the modal struc-
ture of the phenomena, and real patterns are indispensable to such successful
representations, even though approximation and idealization are essential to the
way the representation works.
The real patterns account fits well with the history of theory change in sci-
ence. The well-confirmed laws of past theories describe aspects of the modal
structure of the world, however, the ontologies in terms of which they are
stated are not apt for describing other aspects of the modal structure of the
world. In so far as past theories are empirically adequate, the entities they in-
volve are real patterns and it may be appropriate to continue to refer to them
in some regimes. For example, the force of gravity, the light waves of Fresnel,
the electric and magnetic fields of classical electrodynamics, and the heat
flows of Carnot are all real patterns and part of an effective ontology in the
respective regime. However, sometimes the real patterns in question are best
not described in terms of the old ontology, as with the case of phlogiston,
because it does not get enough of the modal structure right. However, even
in such cases we can say that there were some real patterns to the old theory,
in particular, the commonality of different forms of dephlogistication is a
real part of the modal structure of the world, as explained earlier. The real
patterns account allows that continuity of reference across theory change can
always be secured to some extent whenever there are real patterns that are
carried over as approximations.
Structure not Selection 253
Despite Kuhn’s subsequent clarifications of his own view of science (1977), his
Structure of Scientific Revolutions (1962) inspired schools of history and sociology
of science that demur explanations of theory change in terms of evidence or exper-
imental results (internalism). Rather, it became standard for explanations of theory
choice in science, in so far as they are offered, to appeal to economic, psychological,
and social (external) factors rather than the tribunal of experiment (a classic of the
genre being Shapin and Shaffer 1985). In the history of science, celebrated recent
studies have emphasized the rationality of the losers and the psychological and so-
cial influences on the victors, sometimes going as far as claiming that abandoned
theories should have been retained contrary to the theory-choice made by the scien-
tific community (for example, Chang (2012) argues that phlogiston theory should
have been retained despite its many failings; see Blumenthal and Ladyman (2017,
2018). In this and other cases, notably that of quantum mechanics (see Cushing
1995), it is claimed on the basis of the general considerations of underdetermination
that an alternative history could have delivered the same or more scientific success.
The orthodoxy in current historiography is that actors and their social
networks in the history of science should always be represented sympathetically,
and that their perspective be adopted in describing the relevant evidence and
theories. The opening lines of Harvard’s STS website describe the field as an “ap-
proach to historical and social studies of science, in which scientific facts were
seen as products of scientists’ socially conditioned investigations rather than as
objective representations of nature”.13 Realist historiography is often said to be
Whiggish and triumphalist. However, there is no reason why realists must be
Whiggish in the relevant sense and no reason why celebrating the success of sci-
ence is not compatible with recognizing the bias, error, and missteps that are also
part of its history, and the role of external factors where they are relevant.
While Whiggism is much derided it is less often clearly defined (see
Alvargonzález (2013) for a judicious and informative discussion). In political
history it is associated with an implausible teleology. In the history of science
it is similarly associated with the idea that the development of theories is an in-
evitable progression toward the truth. However, the idea that scientific meth-
odology should track the truth is compatible with a degree of contingency.
Furthermore, it is not unreasonable to think that there is a degree of inevitability
to some scientific discoveries, because so many people have come close to the
same ideas around the same time. For example, the inverse square law of gravity
may not have been known to Hooke, but he at least came near to it (Westfall
13 http://sts.hks.harvard.edu/about/whatissts.html
254 Contemporary Scientific Realism
1967). In any case, acknowledging the actuality of scientific progress does not
require believing it is inexorable, so the teleological element of Whiggism is not
required for a realist historiography of science.
Triumphalism is associated with judging past theories by the lights of pre-
sent ones and with dismissing the successes of abandoned theories and ridi-
culing the reasoning of the historical actors who defended them. However,
celebrating the success of theory-choice and development in particular cases,
and considering the history of science in the light of what we know now
(presentism), does not imply denigrating or ignoring the success of aban-
doned or rival theories or shallow judgment of historical actors. As discussed
earlier, Gale (1968) and Koertege (1969) are very clear about the successes
of phlogiston theory in the course of their explanations of why ultimately it
had to go. It is question-begging against realism to insist that the history of
science should be explained without reference to what we now know about
the world, and that internalist and presentist explanations of theory-choice
are ruled out. From the realist perspective, historiography that does not take
account of current scientific knowledge is likely to mislead us about the his-
tory of science and to beg the question against all forms of realism. What is so
compelling about the problem of theory change is that it works from within
realist historiography. Solving the problem requires being realistic about the
history of science and the limitations of scientific progress, while celebrating
its success.
Acknowledgments
References
Alvargonzález, D. (2013), “Is the History of Science Essentially Whiggish?” History of
Science 51 (1), pp. 85–99.
Berenstain, N, and Ladyman, J. (2012), “Ontic Structural Realism and Modality” in
Landry, E. & Rickles, D. (eds.) Structural Realism: Structure, Object and Causality, The
Western Ontario Series in Philosophy of Science 77, Dordrecht: Springer, pp. 149–168.
Blumenthal, G., and Ladyman J. (2017), “The Development of Problems within the
Phlogiston Theories, 1766–1791.” Foundations of Chemistry 19 (3), pp. 241–280.
Blumenthal, G., and Ladyman, J. (2018), “Theory Comparison and Choice in Chemistry,
1766–1791.” Foundations of Chemistry 20 (3), pp. 169–189.
Cartwright, N. (1983), How the Laws of Physics Lie, Oxford: Oxford University Press.
Structure not Selection 255
Department of Philosophy
Duke University
[email protected]
12.1 Introduction
C t = f ( Yt )
Jennifer S. Jhun, The Case of the Consumption Function In: Contemporary Scientific Realism.
Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press 2021.
DOI: 10.1093/oso/9780190946814.003.0012
258 Contemporary Scientific Realism
any particular things in the real world. Rather, they are methodological posits
that, in and of themselves, do not state truths; but they help the scientist un-
cover the truth about the way the world is. In particular, I’ll argue that such
posits help us discover scale-dependent structural features of the economy
that are real.
We begin in section 12.2 with a (very!) stylized reconstruction of the history
of the consumption function, starting with Keynes (1936) and ending with
Hall’s (1978) response to the so-called Lucas Critique. Hall’s insight was to ex-
plicitly incorporate agent expectations into the formulation of the consump-
tion function; in this way, he could avoid the criticism that agents would alter
their behavior in anticipation of policy changes and therefore change the very
structure of the economy that the modeler attempted to capture (rendering
the whole project moot). This explicit incorporation of intertemporal choice
is what Chao (2003), inspired by Worrall’s structural realism, identifies as real
structure.
However, a number of authors have expressed skepticism about its empirical ad-
equacy. It seems as if actual consumer behavior suggests that the Euler equation, or
the hypotheses it helps scaffold, generally don’t hold. Some more investigative work
must be done if we want to account for actual economic behavior. Yet, the Euler
equation approach persists as a cornerstone of consumption modeling. In order to
make sense of what role it is playing in economists’ reasoning strategies, section 12.3
considers the empirical challenges that the Euler equation faces—both by itself and
embedded in, for instance, the rational expectations permanent income hypothesis
proposed by Hall. The suggestion is that the consumption function—and indeed,
the underlying Euler equation itself—may not be the right object of the realist atti-
tude. We offer a proposal: the Euler equation partakes in a methodology that aims
to find out about real structure. That is, we can embed the Euler equation in a larger
realist project.
These considerations will help us identify a modest structural realism in
section 12.4 that respects that theoretical fixtures such as the Euler equation
in themselves do not represent real structure, and yet recognizes that they are
still used to uncover real structure. To hash out our position, we look to Lyons’s
(2005, 2011, 2017) axiological realism—which maintains that the aim of the-
ories is to find truth (whether or not we actually do, and whether or not we
are justified in thinking we do) in order to formulate an epistemic counterpart.
This epistemic counterpart will depart, however, fundamentally from Lyons’s
vision in that it takes a pragmatic, perspectival view of truth. Finally, in sec-
tion 12.5, we argue that this version of realism has the benefit of sidestepping
some of Reiss’s (2012) criticisms of standard realist defenses against charges of
instrumentalism.
The Case of the Consumption Function 259
C t = α + βYt
2 Technically, it is not Keynesian as in Keynes himself did not advocate for this (and certainly
not in the sense of the General Theory). People who considered themselves Keynesians presented
it in this way, but even for them it was a convenient tool rather than something to be unshakably
committed to.
3 Notably, in the long run it also looks linear through the origin. APC is more or less constant over
time, but not over the business cycle. That fact gave rise to an entire literature on the consumption
function from the 1940s to the 1970s. Friedman was working with Kuznets during this time, and the
original permanent income hypothesis can actually be found in a joint publication (1945).
4 He would document this long-run constancy again in a 1952 paper.
260 Contemporary Scientific Realism
the time series ran contrary to “Keynesian” theory, and instead (as Chao 2003
documents):5
(1) MPC is less than APC in budget data and short-run time-series data but
is equal to APC in the long run; (2) APS [average propensity to save] and APC
did not rise secularly, and (3) private demand increased sharply and APS was
sharply lower than in the interwar period level. (85)6
5 And here I put Keynesian in scare-quotes; Keynes himself in the second chapter of the General
Theory seems to gesture at something that looks like the permanent income hypothesis.
6 A number of other features cast doubt on the Keynesian formulation. For example, Molana
(1993) documents the following empirical challenges (assuming the consumption function takes the
form of Ct = α + βYt):
For instance: (i) Kuznets’ (1946) study illustrated that the APC did not contain a significant
trend, (ii) the cross section budget investigations by Brady and Friedman (1947) showed
that α was in fact positive with a tendency to shift upwards, (iii) Smithies’ (1945) time se-
ries study confirmed that the presence of a positive deterministic trend in a could not be
ruled out, and finally, (iv) estimates of β usually turned out to be lower than expected (see
Haavelmo, 1947) and the model persistently under predicted consumption. (338)
Cate (2013) further notes that the Keynesian account neglected the fact that consumption is not
only affected by changes in income but is also by wealth (572).
7 Modigliani’s life- cycle hypothesis of consumption is similarly based on forward- looking
expectations—this is why such models are now called permanent income-life cycle hypothesis
models. Both Keynes and Modigliani argued that consumers tend to smooth their consumption over
time, so consumption doesn’t depend so much on current income.
8 I note this because— as is unsurprising with any stylized history—this is not Friedman’s
formulation.
The Case of the Consumption Function 261
YtP = Yt −1 + q (YtP−1 − Yt −1 )
q is the fraction of last year’s error subtracted from Yt–1 in order to estimate this
period’s permanent income, Ytp . It is the amount by which my estimated in-
come differs from my estimate last year. The consumption function tells us how
permanent consumption depends on permanent income (rather than current
income), i.e.
C Pt = kYtP
Therefore,
C Pt = k (1 − q ) Yt −1 + kqYtP−1
9 And even this is somewhat overdramatic. These models were only just being used when Lucas
criticized them, so they didn’t at this time have a historical record of being the spectacular failures
they were made out to be.
262 Contemporary Scientific Realism
Robert Hall (1990) would try to get around this problem by explicitly incor-
porating optimization behavior into his formalism, locating structure in the un-
derlying decision-making mechanism.10
Aligning Friedman’s insights with Lucas’s, Hall proposed extending the per-
manent income hypothesis account to incorporate rational expectations, i.e.
assumptions about belief formation, directly into the model. The permanent in-
come hypothesis tells us that if permanent income were known (deterministic),
consumption would remain the same over time. Yet, it doesn’t—we have uncer-
tainty with respect to our income. Agents, according to rational expectations,
take into account all information in order to forecast the future. Supposing that
our agents are utility maximizers, the Euler equation encodes rational expecta-
tions as governing intertemporal choice. In the case of income uncertainty, it
manifests as the following first order condition on the agent’s decision-making
problem.
Ec t+1 = c t
c t+1 = c t + e t+1
That is, consumption tomorrow will be the same as today with some error.
Consumption follows a random walk, and any actual changes are unpredict-
able and random. If I’m forward looking, I’ll anticipate changes in my income by
smoothing consumption over time (but given that actual changes are random,
I do not anticipate substantial changes in income).
One can read the Lucas Critique as an urge to pursue a reductionist project, and
interpret the ubiquitous representative agent as a first-step maneuver in such a pro-
ject. As economics progresses, we will get ever closer to a model of the individual
agents in the economy rather than aggregates.11 So, as a first approximation, we
treat the consumption function as reflecting the behavior of a macroeconomic-
sized representative agent. Perhaps we can move next to (aggregate) heterogeneous
agents, and so on. According to this picture, aggregates are just a shorthand way of
talking about groups of individuals. Macroeconomics, thus, talks about real things
insofar as the entities in question are really inhabitants of a microeconomic on-
tology. Ultimately, what we really need to do is find and then put all household
consumption functions together in some appropriate way.
There’s something suspect about this particular way Lucas’s criticism has been
borne out. This is because it strikes me as an unpromising strategy to pursue a re-
ductionist project along these lines. If macroeconomic aggregates are reducible
to talk of microeconomic entities—the microfoundations project—as the Lucas
Critique seems to require, then it seems like macroeconomic features are not re-
ally autonomous from microeconomic ones. So, if we take the Lucas Critique se-
riously, then to be a realist about macroeconomics implies that we are realist only
insofar as there are microfoundations for our theory.
The difficulty is that there is often no straightforward relationship between the
microscale and the macroscale of the economy. In fact, formal theorems such
as the 1970s Sonneschein-Mantel-Debreu (SMD) theorem—also known as the
“anything goes” theorem—emphasize that there is no such correspondence be-
tween the configurations of individual agents and behavior at the macroscale.
In the context of excess demand functions, the SMD theorem itself states that
under standard assumptions on agents, no interesting macroscopic properties
are guaranteed to emerge from the population. That is, they don’t give rise to any
One of the most popular—if not the most popular—account of realism is struc-
tural realism. The paradigmatic defense of structural realism comes from
Worrall (1989), who argues that:
12
See Hart (1946).
13
Note that even aggregation is not uniformly problematic, nor is it always problematic. Nominal
GDP is an unproblematic aggregate that consists in just exact aggregation of the values of final goods.
The Case of the Consumption Function 265
indeed true or “approximately true” then it is no wonder that they get the phe-
nomena right. So it is plausible to conclude that presently accepted theories are
indeed “essentially” correct. (101)
Known as the No Miracles Argument, it insists that theories would not be as suc-
cessful as they are unless something had gone right—and this something, says
Worrall, is structure. One way of interpreting the position (and the usual way
to interpret Worrall) is in an epistemic sense, i.e. a claim about what we can have
knowledge of, where this is typically “the (preserved) mathematical structure of
our theories” (Morganti 2011, 1166).14
In what follows I mean structure quite loosely. For one, I am here interested in
causal structure, where structure consists of modal relationships that satisfy an
interventionist conception of causation á la Hoover (2001, 2011) or Woodward
(2005). Toggle one thing, and something else changes as a result. That relation-
ship must be (to some extent) invariant: “When a relationship is invariant under
at least some interventions . . . it is potentially usable in the sense that . . . if an
intervention on X were to occur, this would be a way of manipulating or con-
trolling the value of Y” (Woodward 2005, 16).15 Something that partakes in a
causal relationship is part of the causal structure of a system, and articulating a
bit of causal structure will involve articulating the relata in a difference-making
relationship (that is apt for counterfactual analysis). One important disclaimer,
however, is that the way I am thinking about structure will generalize from the
way that structure is typically understood by economists. Economists, at least
post– Lucas Critique, consider “deep parameters”— those that are invariant
under policy interventions—as structural features, and structure itself is defined
as a system of equations.16 Unlike the Cowles era economists, I am not using it
14 For the most part, the distinction doesn’t bear much on the suggestion I’m trying to make in this
paper, because (spoilers) the account I turn to is pragmatist in nature; for more on the distinction, see
Ladyman (1998). Ross’s (2008) ontic structural realism claims that “The basic objects of economic
theory are optimization problems” (741). Objects, on this view, are “heuristics, bookkeeping devices
that help investigators manipulate partial models of reality so as to stay focused on common regions
of measurement from one probe to the next.” What really matters, say, in the case of economic games,
are that they are “mathematical structures, networks of relationships whose relata are distinguished
as such by the mathematical representations of the structures in question—just like quarks and
bosons.”
15 One implication is that insofar as I am thinking about structure, I am really only concerned
with causal structure. I am here agnostic as to whether and what other kinds of structure there are.
I should note here that I am not using causation in the strictly Woodwardian sense either, who iden-
tifies causal relationships with a particular kind of relationship between variables. Despite this, causal
relationships we think of as holding in the world may very well on our account be the kinds of things
we try to capture with structural equations.
16 And even how and what exactly gets considered structure is going to go hand-in-hand with
the particular kind of methodology a school uses—for example, London School of Economics
style econometrics and Cowles Commission style econometric methodology proceed in different
directions, disagreeing on how theory and data bear on modeling. And the New Classical approach
that takes rational expectations as foundational, too, is relying upon a different methodology.
266 Contemporary Scientific Realism
[W]hen Worrall observes that in the history of science the mathematical equa-
tions of the new and old theories are invariant (isomorphism among theoretical
models), given that in the successful or mature scientific theories, the theoret-
ical model flourishingly represents a feature of the world (isomorphism be-
tween theoretical and empirical models), we can conclude that the represented
empirical models are also invariant under the same type of transformation as
the theoretical one . . .
In this interpretation the represented empirical model is invariant and there-
fore can be regarded as structure. (239)
. . . the research done to date has not supported the econometric restrictions im-
plied by the Euler equation approach. Nor has further investigation validated
the hypothesis that the response of consumption to income (henceforth, Y)
reflects only the usefulness of current Y in predicting future Y. Instead, research
typically finds “excess sensitivity” to current income. (467)
Flavin’s (1985) noteworthy study found that modeling income time series follows
an autoregressive–moving-average (ARMA) process, and also that consump-
tion exhibits excess sensitivity: “The empirical results indicate that the observed
sensitivity of consumption to current income is greater than is warranted by the
permanent income-life cycle hypothesis” (976). Because of the statistical discrep-
ancy between the null hypothesis and the data, she proposes liquidity constraints
on household consumption to explain why it is that consumption is excessively
sensitive to changes in income.17
There are ways to use posits like the random walk equation, or even the un-
derlying Euler equation, without assuming that there must be some straightfor-
ward correspondence (or failure as such) to something in the real world in order
to help us get a grip on the economy. Consider Campbell and Mankiw’s (1989,
1990, 1991) papers that aimed to show that rational expectations did not cor-
rectly capture macroeconomic aggregate behavior. Given the aggregate macro-
economic data, they “propose a simple, alternative characterization of the time
series data . . . [that] the data are best viewed as generated not by a single forward-
looking consumer but by two types of consumers” (1989, 185). About half of
these consumers behave like rational agents, but the other half follow a rule of
thumb, consuming their current income. (Both, as a note, optimize their utility
preferences.) This alternative hypothesis is “an economically important devia-
tion from the permanent income hypothesis” (187).
Campbell and Mankiw embed the permanent income hypothesis model
within a generalized model that allows for a fraction of income to accrue to
forward-looking agents and the rest to those who consume their permanent
17 Later in 1985 she notes that “The null hypothesis in this empirical literature typically consists of
the joint hypothesis that 1) agents’ expectations are formed rationally, 2) desired consumption is de-
termined by permanent income, and 3) capital markets are ‘perfect’ ” (117–118). There she diagnoses
excess sensitivity as a failure of the third assumption.
268 Contemporary Scientific Realism
They can directly test the permanent income hypothesis by setting λ to zero. And
in fact, their estimate of λ—that portion of the population that consumes cur-
rent income—is about 0.5 (195). The macroscopic phenomenon of smoothness
is explained by the fact that aggregate consumption “is a ‘diversified portfolio’ of
the consumption of two groups of agents” (210).
The point of this exercise, which again does not discard the original consump-
tion formalism associated with the random walk hypothesis, is not a straightfor-
ward case of starting with an idealized model and gradually de-idealizing it to
make it more realistic. If real behavior deviates from the idealized benchmark
(like one that assumes a single representative agent), however, we should seek
out what factors explain that deviation. The Euler construction is a tool that helps
us erect and navigate a larger, more complex framework, making space for po-
tentially relevant structural information (though not every and all pieces of in-
formation) that makes a difference to a system’s behavior. And for Mankiw and
Campbell in particular, the structural information about their population of in-
terest is how much of that income accrues to people who abide by rational ex-
pectations as opposed to those who consume their current income. They choose
to look at the distribution of agent kinds in the population. That is, theirs is a
project that tries to capture what we might call scale-dominant behavior.18 This
is behavior that is characteristic of a system at a particular scale, where particular
structural characteristics will be salient there.19
18 Wilson (2017), in the context of representative volume element (RVE) methods in physics,
points out that complex behavior can be managed by parsing it out into separate sub-models, each
“assigned the comparatively circumscribed duty of capturing only the central physical processes nor-
mally witnessed at its characteristic scale length . . . each localized model renders descriptive jus-
tice only to the dominant behaviors it normally encounters” (19). And indeed, as Batterman (2013)
emphasizes, “Many systems, say a steel girder, manifest radically different, dominant behaviors at
different length scales. At the scale of meters, we are interested in its bending properties, its buckling
strength, etc. At the scale of nanometers or smaller, it is composed of many atoms, and features of in-
terest include lattice properties, ionic bonding strengths, etc.” (255).
19 In order to check whether the results were robust, the authors carry out a battery of tests to
supplement their instrumental variables method. These include Monte Carlo methods to examine
the small-sample distribution of the test statistics in order to check for the small-sample bias. The
Monte Carlo experiment deploys their framework with adjustable parameters for population distri-
bution and aims to produce artificial data with similar structural—in this case statistical—properties
(such as that log-income and consumption are integrated processes and consumption and income
are cointegrated). In addition, they explore various generalizations of their model to see if there are
other candidate explanations.
The Case of the Consumption Function 269
Again, however (and as Chao points out), one might object that it is only the
random walk hypothesis that is in danger, rather than the Euler equation ap-
proach itself, as Chao notes: “Anomalies may reject the permanent income hy-
pothesis or the life-cycle hypothesis, but they only motivate modelers to modify
the Euler equations instead of abandoning them” (243). For example, instead of
discarding the Euler equation, discrepancies in behavior are used to modify it.20
One might suggest that this only indicates that one or more of the assumptions
going into the random walk hypothesis itself must be problematic, but it need
not also imply that the Euler equation (the intertemporal preference ordering)
is problematic. We have something of an underdetermination problem; the con-
sumption equation’s failure might be due to any number of things, such as wrong
assumptions about preferences (like that of constant relative risk aversion, which
is quite common).
But there are reasons to question the viability of the Euler equation itself.
Carroll (2001) questions whether we are able to estimate the (log-linearized)
Euler equations at all. Canzoneri et al (2007) compare the interest rates predicted
by the Euler equation against the money market interest rate to find out that they
are negatively correlated. Though they start out with preferences that satisfy con-
stant relative risk aversion, they do try out different utility functions. Estrella
and Fuhrer (1999) claim that “evidence that shows that some forward-looking
models from the recent literature may be less stable—more susceptible to the
Lucas critique—than their better-fitting backward-looking counterparts,” (4) a
count against the Euler approach. Blinder and Deaton (1985) have their “doubts
about the wisdom of modeling aggregate consumption as the interior solution
to a single individual’s optimization problem in adjacent periods, but in any case
think it fair to say that the research done to date has not supported the econo-
metric restrictions implied by the Euler equation approach” (467). There are at
least two reasons for this: corner solutions due to liquidity constraints and aggre-
gation problems. For instance, at least at the aggregate level, the Euler equation
seems suspicious.21 Attanasio (1999) finds that “The Euler equations for (non-
durable) consumption . . . are, for most specifications of preferences, non-linear.
As they refer to individual households, their aggregation is problematic” (781).
Elsewhere, Hvranek’s (2015) meta-analysis finds that on average, the estimated
20 For another interesting case study, see Zeldes (1989) and Runkle (1991), who estimated versions
of the Euler equation with different results. Attanasio (1999) reports that Zeldes (1989) “splits the
sample according to the wealth held and finds that the rate of growth of consumption is related with
the lagged level of income for the low wealth sample. The same result does not hold for the high
wealth sample. Zeldes interprets this result as evidence of binding liquidity constraints for a large
fraction of the population” (790). That is, the low-wealth sample violates the standard Euler equation
while the high-wealth sample does not. On the other hand, Runkle (1991) finds no such evidence.
21 Deaton (1992) specifies that “if tastes are the same, the Euler conditions will aggregate perfectly
if three conditions hold: (a) people live for ever, (b) felicity functions are quadratic (or the time in-
terval is very short), and (c) individuals know all the aggregate information” (167).
270 Contemporary Scientific Realism
elasticity of intertemporal substitution is zero—too low, given the way the Euler
equation is usually specified (consumption should be sensitive to variations in
the real interest rate).22,23
It’s doubtful that even individual agents optimize in this way, and when they
do it may be limited in scope—after all, agents are often thought to be “myopic”
in nature. (Of course, we could always restrict the behavior to a finite number
of periods, rather than an infinite horizon, which is entirely consistent with the
Euler approach.) Even if we were only considering individuals as the kinds of
agents that intertemporally optimize, aggregation problems notwithstanding,
this very well may be false except for fairly short time horizons. One substantial
task would then be to assess over what temporal scale intertemporal optimiza-
tion holds. So it seems a bit odd to say that the Euler equation is the product of
our uncovering some deep structure that is actually out there in the world.
Work since Hall (1978) has focused on extending the baseline Euler model
in various directions. For instance, one might incorporate habit-formation in
consumption behavior, which might explain the lagged behavior of consump-
tion in response to interest rate or follow Campbell and Mankiw and incorpo-
rate hand-to-mouth consumers. Throughout all this, as Chao notes, the Euler
construction still stands. Now, rather than taking the Euler construction itself
to be real, however, I’m more inclined to say that it helps find out what is real. I’d
argue that stipulating the Euler equation is a special case of a story about equilib-
rium reasoning that I have told elsewhere.24 Equilibrium conditions, constraints
on behavior, do not in themselves state truths about the way the world is—but
they do help yield interesting, informative insights about the world. That is, equi-
librium conditions are methodological tools rather than straightforward claims
about or corresponding to states of affairs. For instance, equilibrium conditions
like the Euler equation function more like the ceteris paribus laws we see com-
monly in economics. And while it may hold somewhere, and across some range
of circumstances, it need not hold universally.
Here’s what I mean. For example, we typically do not think “An increase in the
supply of a good, ceteris paribus, leads to an increase in price of that good” is in-
variant in the sense that it holds everywhere. We do, however, use it to figure out
in which pockets of the economy in which it does hold, so it will have to be in-
variant over some range of circumstances for some market over some time scale.
If our local widget economy is somewhat (causally) isolated from the rest of the
nationwide economy, we could apply ceteris paribus reasoning in order to think
about what would happen under certain kinds of shocks or interventions. I think
the Euler equation works in somewhat the same way—it’s not that it, itself, is
invariant—but rather, we use it to locate where relationships can be treated as
invariant, which means picking out a relevant scale of interest. When it fails,
we seek reasons as to why actual behavior diverges. And these relationships of
interest are often causal, in that they are apt for an interventionist difference-
making interpretation.
In particular, using the Euler equation means trying to see where, and
how well, it fits. To take an example, when Mumtaz and Surico (2011) ad-
just parameters to fit the data over time, what they are noting is where it is
and in what form that the Euler equation holds and doesn’t hold, as “the
parameters of the aggregate consumption Euler equation may vary over the
business cycle” (5). One of their self-imposed tasks is to show that “periods of
conditionally low (high) consumption correspond to periods of below-trend
(above-trend) consumption” (12). What this helps highlight is the partic-
ular scale of the system we should be paying attention to. So one thing that
Mumtaz and Surico’s account does is make salient particular time scales (and
salient behavioral properties at those scales) as relevant to the investigation.
For instance, they identify recession periods in the 1970s, 1981, 1992, and
2008 as periods during which agents tend to be more backward-looking and
consumption tends to be less sensitive to the real interest rate (13). Similarly,
we can use the Euler equation to investigate whether agents are subject to bor-
rowing constraints; strictly speaking this doesn’t mean that the Euler equation
should be thrown out, but that it applies within a certain scope. Rather than
merely pointing out that agents don’t optimize over time, we say something
about the extent to which they do. That is, I can tell an analogous story here
about the role of the Euler equation as I did with that of the random walk
hypothesis in the Campbell and Mankiw case. These equations are fixtures
that help us identify particular scales of interest and the behaviors that appear
dominant at those scales—scale-dominant behaviors.
Let me use an analogy from materials engineering to concretize what I mean
by scale-dominant behavior. Though this reductionist attitude is now less pop-
ular, one might think that the right way to determine the behavior of a system
is first to accumulate all the details at a micro-level. After doing so, we will be
able to infer or deduce information about what goes on at the macro-level. So a
bottom-up methodology would aim to achieve a picture of the whole system (be
it physical or social) first by gleaning information about its constituents and then
somehow putting all that information together.
But this approach is unsatisfactory in engineering contexts. For instance,
one everyday kind of problem we might worry about is how to manufacture
materials in order to construct sturdy buildings that can sustain some wear and
tear. Consider a steel beam:
272 Contemporary Scientific Realism
What’s also important in this case is structural detail at the mesoscale, where steel’s
granular structure becomes apparent—say at 30 nanometers. These boundaries
between grains affect the way molecular bonds break and reform when external
force is exerted on the beam. So, this structure that manifests at the mesoscale
contributes to the system’s macroscale behavior.
We can even tell a similar story about the permanent income hypothesis.
Campbell and Mankiw’s search for relevant population parameters (or Mumtaz
and Surico looking to see patterns of forward-looking behavior) is akin to
looking for details about grain in the steel beam. And like granularity, popula-
tion distribution is a mesoscale detail that we can treat as real as goings-on at
any other scale. But we had to do a little bit of work first—figuring out where the
Euler equation could plausibly fit in as successfully describing behavior. Rather
than describing the structure of the economy-at-large, the Euler consumption
equation might characterize the behavior of some members of the economy over
some (but not the infinite) time horizon. The particular configuration of these
members might be something that matters. Distributional features capture ad-
ditional relevant structure that affects the macro-behavior of a system; this is
additional information we need to understand this system’s behavior. This in-
formation also explains why it is that actual population behavior departs from
the idealized case with a homogeneous population where everybody consumes
current income.
Explaining the behavior of complex systems seems to require treating
features that appear as salient at particular scales as real parts of a real struc-
ture, because they are the kinds of things that figure into causal relationships.
But our treating economic structure as real is not the same as Worrall’s structural
realism; we are not realists about the random walk hypothesis, or even about
the Euler equation. The Euler equation plays a different role for us. It does not
figure as a theoretical posit that we understand as true, and thus is preserved over
time as the theory progresses. Rather, the Euler equation helps us orient our-
selves in a larger framework; it allows us to establish a vantage point from which
to look at a particular system of interest and get a grip on the relevant causal
relationships involved. We can think of investigation at a particular scale to be
investigation from a particular perspective of that system. To accommodate this,
we need to be perspectival realists, which, according to Hoover (2011), is “predi-
cated on the belief that models are used to assert true general claims about causal
The Case of the Consumption Function 273
The kind of realism I favor is modest; it doesn’t even claim that a theory marches
steadily toward a unique true theory even in the limit. All that it requires is that
at least sometimes, we’re justified in thinking that we are getting things right.28
I propose that the perspectival realism we endorse corresponds to a pragmatic
interpretation of Lyons’s axiological realism, capturing what is intuitively attrac-
tive about the view in addition to seamlessly providing an epistemic counterpart
to his position.
Lyons (2005) suggests that “science is interested, not in true statements, per
se, but a certain type of true statements, those that are manifested as true” (174).
These statements partake in theory complexes—groups of theories—and are
such that they have bearing on our world of experience; they can manifest as true.
25 That is, economic structures are structures relative to perspective (scale). Our account is also
different from another attractive account of realism, selective realism. First, perspectival realism gives
us something to make sense of theories synchronically—the attitude that we have at any given time
for a theory under consideration. On the other hand, selective realism is a diachronic account—it
explains something about a persistent element of a theory (or theories) over time.
26 It is within this larger framework that we can ask further questions such as: Is wealth a factor
that we should have included in our macroeconomic analysis? Is income inequality a structural fea-
ture that matters? And so on. See, for work along these lines, Carroll, Slacalek, and Tokuoka (2014).
27 But articulating multi-scale relationships is a tricky matter in itself. Bursten (2018) argues in
the case of nanoscience that articulating the relationship between scales may require conceptual
maneuvers that resist traditional categorization (e.g. reductionist ones). I would like to leave space,
for economics, to take advantage of these oddball conceptual maneuvers, though this may lead to
further complications for thinking about realism down the line.
28 Note here a sympathy with something like Mäki’s (2009) realistic realism.
274 Contemporary Scientific Realism
Lyons (2011) has called these statements, as a subclass of true statements, those
that are “experientially concretized” truths (XT statements): “true statements
whose truth is made to deductively impact, is deductively pushed to and enters
into, documented reports of specific experiences” (329).29 True statements are
candidates for—but not all are—XT statements, whose truth is experientially
concretized as true. And while actual XT statements are required to be true
in order to qualify as being so, we do not necessarily think ourselves to have
achieved truth per se. That is, despite the fact that that they inform the articula-
tion of or imply some claim that corresponds to an experience:
The emphasis here is on the pursuit of truth, rather than justification for believing
we have achieved it. That is, we can be committed to seeking truth and yet be ag-
nostic about whether we have achieved it or not.
Macroeconomics can be considered a complex of theories and models.
Statements such as “the proportion of income accrued to forward-looking agents
is λ” have implications for what we expect to see in the macroeconomic aggre-
gate. The Euler equations are such that (given some additional context) they
imply particular empirical consequences that may or may not hold.30
However, fixtures such as the Euler equations cannot be XT statements ac-
cording to Lyons’s account. And fundamental theoretical posits like the Euler
equation on its own are not even really treated as genuine candidates for truth.
Some may simply assume that they are outright false because they’re often con-
sidered idealizations. After all, the Euler equation approach is probably not a
good way of describing aggregate macroeconomic behavior. It is not even a good
way to think about individual behavior (people often simply fail at intertemporal
utility maximization, never mind that they are not infinitely lived agents
anyway). So in order to make Lyons’s insights relevant to economics, I propose
that we step away from the foundational conception of theory complexes as lying
on a bedrock of (manifest) truth from which we can infer other (manifest) truths
and think of theoretical posits as doing something other than serving as such
29 In his earlier work he tells us that a truth is manifest when “it has bearing on, makes a difference
to, the world of experience and is transmitted to tested predictions derived from the complex” (2005,
174). Such truth is “not manifested as true to us but in the consequences derived from the theory and
in the world of our experience” (177).
30 See, for this position, Hoover (2009).
The Case of the Consumption Function 275
ground-level manifest truths. The Euler equation has a shaping or framing role,
not a grounding one.
The Euler equation, in and of itself, is not literally true (nor, arguably, is it
false by itself either), so it cannot be an XT statement.31 But we use it to formu-
late claims that we do compare against empirical data to determine whether it
explains or describes our data well. And if we take Lyons’s account seriously, we’d
actually have what he calls an “evident XT-deficiency” where “it is evident that
non-XT statements are present in the complex” (2011, 330).
But it’s consistent to claim that the Euler equation doesn’t usually hold of any-
thing by itself and yet also claim that this approach is part of a methodology that
helps us find out what’s true about the world, by allowing us to anchor ourselves
in a particular economic context. Used wisely, I think such theoretical posits em-
body strategies that help us find manifest (or experientially concretized) truths—
truths that make a difference to our world of experience. The way it is doing so,
then, can’t be hashed out in terms of deductive informativeness (however way we
spell that out). Such non-XT statements figure as central to theory. For example,
Campbell and Mankiw do not discard the Euler equation approach from their
toolbox in lieu of a more realistic equation. Nor is it an idealization that simply
needs to be filled out with more detail in order to account for empirical data.
They use it, in fact, to construct a larger generalized framework around it. It re-
mains central to their reasoning, and such statements are never quite eradicated
from the science. But their primary role is not to deductively impute truth.32
Chao was right to emphasize the persistence of the Euler equation. I am
suggesting that the role of such equations is methodological rather than repre-
sentational; they serve to locate and articulate structure and serve as diagnostic
tools. They are fixtures that are central to analysis. When it comes to complex sys-
tems, we had to identify scale-dominant behavior in order to diagnose the ways
31 If anything, I would argue that the Euler equation has a role in shaping the complex of XT
statements, so that its role is prior to the question of whether or not a particular model represents
phenomena accurately. To put it in Lyons’s terms, we could understand the Euler equation as an in-
strument that even helps “remedy evident XT-deficiencies by increasing the number—and/or the
extent, degree, or exactitude of the experiential concretization—of XT statements” (2011, 330).
32 Lyons (2017) anticipates the kind of suggestion that I’m making.
[P]atently false posits can serve toward “bringing the truth” of high level posits “down” to
statements that describe experiences. Once novel predictions are derived, even from false
hypotheses, and new data is gathered, then, on this view, those mediating blatantly false
constituents previously deployed are eliminated in want of increasing the experientially
concretized truth of the accepted system. (3213)
So false posits can have a central role, making the distinction between our positions a subtle matter.
But for us the posit is never really eliminated. This is because (for instance) Mankiw and Campbell do
not attain their results on the basis of false assumptions; rather, the posit fully serves as a foundation
for the model they eventually articulate, so it is in effect built into the final product they offer. But
I think ultimately that the difference between our positions isn’t so much of a disagreement so much
as we are simply pointing at the nuanced guises that posits take in scientific reasoning.
276 Contemporary Scientific Realism
the actual behavior of a system deviates from the idealized model. We might start
an enquiry by examining whether the Euler equation fits our population-wide
data; when it fails to do so, we must figure out where else (if anywhere) it can have
explanatory or descriptive power—where those causal relationships stipulated
by the equation actually hold. The Euler equation is more like a navigational tool
(that helps us answer questions such as, where does it hold, and where does it
fail to hold? Why not?). And the more we are able to diagnose different kinds of
deviations from the ideal and able to tell when that benchmark is or isn’t the ap-
propriate one to use, the more our science progresses. In a way, economists too
are in the business of seeking and maximizing manifest truth.
Readers will note that this view, given its endorsement of Hoover’s perspec-
tival realism, also shares kinship with Chang’s (2016) pragmatic realism, where
the pragmatic refers to the practical aspect of activity:
With this notion of coherence in hand, Chang (2017) formulates his coherence
theory of reality as follows: “a putative entity should be considered real if it is
employed in a coherent epistemic activity that relies on its existence and its basic
properties (by which we identify it)” (15).
So for the economist, a coherent activity might be implementing a bit of policy,
which in turn requires conducting a bit of counterfactual analysis in a model.
Policymaking would require an economist to take seriously how behavior at the
macroscale might change in response to changes at other scales. I might want
to know how overall consumption changes when the underlying distribution of
agents changes. That is, exercises of counterfactual reasoning about complex sys-
tems behavior rely on the existence of a (multiscale) framework. While some-
thing like the Euler equation need not be considered real, it partakes in framing a
coherent epistemic project—this coherent multiscale framework where goings-
on at different scales must coordinate with one another. The scale-specific
structure that we discover in our investigation, insofar as it partakes in such a
framework, is what is real. Grasping the causal structure of the world is partly a
matter of trying to figure out how fundamental equations like the Euler equation
fit in and help articulate such a framework.
The Case of the Consumption Function 277
This modest realism is nothing new. What I am proposing here is that such
a realism supplies what’s missing in Lyons’s axiological account—the epistemic
oomph of what is experientially concretized as true. Experientially concret-
ized truth can be legitimate, plain old truth that we can have at hand—there’s
nothing that is truth per se in the pragmatic account apart from what is expe-
rientially concretized. Positing the Euler equation is the beginning of an enter-
prise that is open to the possibility that multiple perspectives of a system have to
be taken into account, and that the parcels of information we obtain from dif-
ferent perspectives must cooperate with one another. While how we place those
perspectives (and they very well may be separate models) into a coherent frame-
work is a question for another time, that complex is one—insofar as it allows us
to successfully conduct something like counterfactual analyses—that is apt for a
realist interpretation.
A final note. I have been deliberate in my noncommittal attitude to whether
or not the realism I care about is epistemic or ontic. This is because for the most
part, the distinction doesn’t bear much on the suggestion I’m trying to make.
Structural realism is, in both senses, a positive and a negative thesis. It is a posi-
tive thesis about what we can know, and what there is, and by parity about what
we do not know, and what there isn’t. My thesis is simply that there is structure
that we can be realists about, namely causal structure, and that the Euler equa-
tion is one tool that economists use to uncover it. And given the pragmatic na-
ture of our discussion, in that our knowledge of the economy involves how to
do things, the simple answer is: I am a realist both about what we can know and
about what there is at the same time, though I do not commit here to the neg-
ative thesis that that’s all we can know or that that’s all that there is. The realist
attitude outlined here simply tries to get right how economists actually go about
doing things, insofar as we conceive of economics as a policy science. This prag-
matic view emphasizes the coherence of the multi-scale causal structure, where
structure occupies a fairly fundamental place in economists’ thinking. At least,
given our investigation of economic methodology, we shouldn’t take as granted
the priority of (knowledge of) individuals over structure.
A few remarks before I conclude. I don’t consider the Euler equation itself to
be an approximation of the truth, nor something that is a pit stop for a model
that’s developing toward a future truer version of itself. The Euler equation aids
us in finding the truth, but in itself is not representational—so it is not an ap-
proximation of anything. If the Euler equation fails to “fit the data” it may be for
a number of different reasons; that the population is heterogeneous is a causally
278 Contemporary Scientific Realism
relevant feature about the structure of the population of interest. That is, the
Euler equation’s purpose is more diagnostic than representational.
Because this realism looks somewhat different from standard forms of re-
alism, which are grounded on correspondence views of truth, the defenses that
it can provide will look a bit different, too. In this section, we consider Reiss’s
(2012) criticism of realism; he argues that instrumentalism has at least three
advantages over it. I argue that we can sidestep those criticisms.
Reiss worries that “if we seek predictive success, we don’t necessarily want to
build causal models” (373). I do not claim anywhere that predictive success is the
mark of a successful model, nor does a model need be an accurate representation
in order to be true. Reiss maintains that “a model is not predictively successful
qua representing causal structure” (373–374). In fact, I agree! Reduced-form
models can be perfectly suitable for prediction in some cases. And this is because
a good reduced form model is, in fact, the collapsed form of a structural model.
So the relevant structure is incorporated into the reduced-form construction.33
And in order to use a model for diagnostic purposes, á la counterfactual rea-
soning, it ought to be causal.
Reiss’s larger worry is that “even when we are able to establish causality, to
make a causal model practically useful, additional facts about the represented
relations have to be discovered. But once we’ve discovered the additional facts, it
is irrelevant whether the relation at hand is causal or not” (374). That is, the fact
that a relationship is causal is irrelevant as to whether it is successful or not. Given
that diagnostic success for us is conceived in terms of possible (and possibly ac-
tual!) interventions, this seems implausible. If the structure of the economy is
something that counts as a difference-maker to its overall behavior, then I con-
sider that a causal contribution.
33 Every structural model gives rise to different reduced forms, and there might be observational
equivalence between some of them, but there remain legitimate questions about which one is a good
one. These questions are going to ask which ones have got the structure right (where structure is un-
derstood in my sense!).
The Case of the Consumption Function 279
The third criticism: “It’s better to do what one can than to chase rainbows” (375).
The final worry is that “building causal models has enormous informational
requirements” (375), and so a different heuristic that one might use is what is
known as Marschak’s Maxim. Marschak (1953) observed that for many questions
of policy analysis there is no need to identify fully specified models that are in-
variant to whole classes of policies. If the researcher instead focuses on particular
interventions, all that may be needed are combinations of subsets of the struc-
tural parameters—those required to identify the effect of the policy intervention.
There’s nothing stopping a structural realist of my stripe from abiding by
Marschak’s Maxim.34 Think of the analogy with materials. We don’t need all
the information about all the atoms in the steel beam—we do, however, need to
know about grain and how that granular structure affects macroscale behavior.35
The engineer’s version of Marschak’s Maxim would tell us to focus our attentions
at this scale. Something similar is the case in economics. It’s not necessary to
uncover all the individual preferences of individuals to get a sense of what con-
sumption in the aggregate looks like or all the causes that might be at play in
order to assess the effect of a particular disturbance to a system. We may need
to know something about the distribution of preferences, and finding this infor-
mation is a matter of finding the right perspective—that is, the appropriate scale.
And insofar as it is specific to scale and partakes in a multi-scale framework that
enables us to undertake counterfactual exercises, I call that structural and real.
12.6 Conclusion
I’ve argued thus far for a modest structural realism that accommodates at least
some of the ways that economists actually deploy idealization in their search for
the real causes underlying dynamic behavior. Making sense of these practices
necessitates our moving away from correspondence notions of truth to coherence
ones; aligning our realism with Lyons’s axiological realism cannot be achieved
otherwise. But a pragmatic spin on realism allows us to recognize the central
methodological role that idealizations have to play in obtaining (in particular
34 In Marschak’s 1953 Cowles paper, he actually describes what can be recognized as the Lucas
Critique. One way in which they differ is that Marschak is not concerned with expectations in the
way Lucas is. Marschak’s point, nonetheless, is that we need invariance—but the notion of being in-
variant is relative (as nothing is invariant to everything)—something that Lucas himself seems to
accept! This theme reaches back to Haavelmo and Frisch (see Aldrich 1989).
35 And the range of circumstances under which a particular granular configuration will persist; for
instance, with extremely high temperatures we can induce grain growth, which in turn affects macro-
scale properties by making steel more elastic.
280 Contemporary Scientific Realism
causal) knowledge. It is not the Euler equation that is real but the structure that
the Euler equation helps us uncover and use to explain economic behavior.
Acknowledgments
References
Aldrich, J. (1989). Autonomy. Oxford Economic Papers, 41(1), 15–34.
Attanasio O. P. (1999). Consumption. In, Taylor, J. B., & Woodford, M. (eds.) Handbook of
Macroeconomics, Vol. 1B. Amsterdam: Elsevier Sci, pp. 741–812.
Batterman, R.W. (2013). The Tyranny of Scales. In, The Oxford Handbook of Philosophy of
Physics. Oxford: Oxford University Press, pp. 255–286.
Blinder, A. S., Deaton, A., Hall, R. E., & Hubbard, R. G. (1985). The time series consump-
tion function revisited. Brookings Papers on Economic Activity, 1985(2), 465–521.
Brady, D. S., & Friedman, R. D. (1947). Savings and the income distribution. In, Studies
in Income and Wealth, Vol. 10. National Bureau of Economic Research, pp. 247–265.
Bronfenbrenner, M. (1948). The consumption function controversy. Southern Economic
Journal, 14(3), 304–320.
Bursten, J. (2018). Conceptual strategies and inter-theory relations: The case of nano-
scale cracks. Studies in History and Philosophy of Science Part B: Studies in History and
Philosophy of Modern Physics, 62, 158–165.
Campbell, J.Y. and Mankiw, N.G. (1991). The response of consumption to income: A
cross-country investigation. European Economic Review, 35(4), 723–756.
Campbell, J. Y., & Mankiw, N. G. (1990). Permanent income, current income, and con-
sumption. Journal of Business & Economic Statistics, 8(3), 265–279
Campbell, J. Y., & Mankiw, N. G. (1989). International evidence on the persistence of eco-
nomic fluctuations. Journal of Monetary Economics, 23(2), 319–333.
Canzoneri, M. B., Cumby, R. E., & Diba, B. T. (2007). Euler equations and money market
interest rates: A challenge for monetary policy models. Journal of Monetary Economics,
54(7), 1863–1881.
Carroll, C. D. (2001). Death to the log-linearized consumption Euler equation! (And
very poor health to the second-order approximation). Advances in Macroeconomics,
1(1), 1–38.
Carroll, C. D., Slacalek, J., & Tokuoka, K. (2014). The distribution of wealth and the
MPC: Implications of new European data. American Economic Review, 104(5), 107–11.
Cate, T. (ed.). (2013). An encyclopedia of Keynesian economics. Edward Elgar Publishing.
Chang, H. (2017). Operational Coherence as the Source of Truth. Proceedings of the
Aristotelian Society, 117(2), 103–122.
The Case of the Consumption Function 281
Hoover, K. (2011). Counterfactuals and Causal Structure. In, Illari, P. M., Russo, F., &
Williamson, J. (eds.) Causality in the Sciences. Oxford: Oxford University Press, pp.
338–360.
Hoover, K. (2009). Microfoundations and the Ontology of Macroeconomics. In,
Kincaid, H., & Ross, D. (eds.) Oxford Handbook of the Philosophy of Economic Science.
Oxford: Oxford University Press, pp. 386–409.
Hoover, K. (2001). Is Macroeconomics for real? In, Maki, U. (ed.) The Economic World
View: Studies in the Ontology of Economics. Cambridge: Cambridge University Press,
pp. 225–45.
Havranek, T. (2015). Measuring intertemporal substitution: The importance of method
choices and selective reporting. Journal of the European Economic Association, 13(6),
1180–1204.
Jhun, J. S. (forthcoming). Economics, equilibrium methods, and multi-scale modeling.
Erkenntnis. https://doi.org/10.1007/s10670-019-00113-6.
Jhun, J. S. (2018). What’s the point of ceteris paribus? or, how to understand supply and
demand curves. Philosophy of Science, 85(2), 271–292.
Keynes, J.M. (1936) The General Theory of Employment, Interest and Money.
Cambridge: Macmillan Cambridge University Press.
Kuznets, S. (1952). Proportion of capital formation to national product. The American
Economic Review, 42(2), 507–526.
Kuznets, S. (1942). Uses of national income in peace and war. New York: National Bureau
of Economic Research.
Ladyman, J. (1998). What is structural realism?. Studies in History and Philosophy of
Science, 29(3), 409–424.
Lucas, R. E. (1976). Econometric policy evaluation: A critique. In, Brunner, K., and
Meltzer, A. H. (eds.) The Phillips Curve and Labour Markets. Amsterdam: North
Holland, pp. 19–46.
Lyons, T. D. (2005). Toward a purely axiological scientific realism. Erkenntnis, 63(2),
167–204.
Lyons T.D. (2011). The problem of deep competitors and the pursuit of epistemically uto-
pian truths. Journal of General Philosophy of Science 42, 317–0338.
Lyons, T. D. (2017). Epistemic selectivity, historical threats, and the non-epistemic tenets
of scientific realism. Synthese, 194(9), 3203–3219.
Lucas, R. E., & Sargent T. J. (1979). After Keynesian macroeconomics. Federal Reserve
Bank of Minneapolis Quarterly Review, 3(2), 1–16.
Mäki, U. (2009). Realistic realism about unrealistic models. In, Kincaid, Harold, & Ross,
Don (eds.) The Oxford Handbook of Philosophy of Economics. New York: Oxford
University Press, pp. 68–98.
Mankiw, N. G., Rotemberg, J. J., & Summers, L. H. (1985). Intertemporal substitution in
macroeconomics. The Quarterly Journal of Economics, 100(1), 225–251.
Marschak, J. (1953). Economic measurements for policy and prediction. In, Hood, W.,
& Koopmans, T. (eds.) Studies in Econometric Method. New York and London: John
Wiley and Sons, 1–26.
Modigliani, G., & Brumberg, R. (1954). Utility analysis and the consumption function: An
interpretation of cross-section data. In Kurihara, K. K. (ed.) Post-Keynesian economics.
New Brunswick, NJ: Rutgers University Press, pp. 388–436.
Molana, H. (1993). The role of income in the consumption function: A review of on-going
developments. Scottish Journal of Political Economy, 40(3), 335–352.
The Case of the Consumption Function 283
University of Duesseldorf
[email protected]
13.1 Introduction
The aim of this paper is to analyze and assess a certain argument used by antirealists
in the scientific realism debate, the pessimistic meta-meta-induction (two “meta”s).
Let scientific realism be the view that our current successful theories are probably
approximately true. Antirealists challenge realism by presenting the pessimistic
meta-induction, PMI (one “meta”), according to which many theories in the past
of science were also successful for a while, but were refuted later on. Typically the
very first response by scientific realists to the PMI is that science has improved a
lot since the times of the past refuted theories and that these improvements block
the PMI and save realism.1 Antirealists then often reply that past realists could have
said the same thing, namely that science has improved a lot in the same manner as
realists claim for the recent past, but those improvements didn’t help past realists,
as the subsequent theory refutations show; hence, the recent improvements like-
wise don’t help current realists to block the PMI and to save realism, so the realists’
defense against the PMI fails. It is this argument by antirealists which is the focus of
this paper. I call it the pessimistic meta-meta-induction, PMMI.2 I will analyze the
PMMI and attempt to show that it does not succeed as a reply to the realist’s defense.
1 Stathis Psillos calls this response “the Privilege-for-current-theories strategy.” (Psillos 2018, Sect.
2.8). Mario Alai calls it the “discontinuity strategy”: “there are marked differences between past and
present science,” which imply that “pessimism about past theories can[not] be extended to current
ones.” (Alai 2017, p. 3271/67).
2 The term “pessimistic meta-meta-induction” is also used by Devitt (1991, p. 163) and by Psillos (2018,
Sect. 2.8). The PMMI is expounded by Wray (2013, 2018, Ch. 6) to whom this paper is a rejoinder, and Psillos
(2018). A related version directed at the realist defense based on the improvement of scientific methods is
criticized by Devitt (1991, p. 163/4). See also Doppelt (2014, p. 286), Müller (2015), Stanford (2015, p. 3),
Alai (2017, p. 3282), Vickers (2018, p. 49), Rowbottom (2019, p. 476), and Frost-Arnold (2018, p. 3).
Ludwig Fahrbach, We Think, They Thought In: Contemporary Scientific Realism. Edited by: Timothy D. Lyons and Peter
Vickers, Oxford University Press. © Oxford University Press 2021. DOI: 10.1093/oso/9780190946814.003.0013
We Think, They Thought 285
3 For a list of versions of (selective) realism see Lyons (2017, p. 3215). For an overview of structural
of predictions is here understood in the usual way, as use-novelty: The prediction is not used in the
construction of the theory.
5 In Section 13.11 I will hint at some reasons why I think scientists are right on this matter.
286 Contemporary Scientific Realism
is bothered by the revisionary nature of selective realism, as I am, then one will
find it worthwhile to explore alternative responses to the PMI.
The dialectic in which the PMMI arises has to be developed with some care
(Sections 13.2 to 13.5, and 13.7). In Section 13.2 I define scientific realism and
present the PMI. In Section 13.3 I analyze the PMI. Then I discuss the realist’s
defense against the PMI, which consists in claiming that science has improved
a lot since the times of the refuted theories, and formulate three assumptions
on which the defense is based (Section 13.4). Finally Section 13.5 introduces
the PMMI. I analyze it, and discuss some objections (Section 13.6). Then comes
the central part of the paper. I show how the assessment of the PMMI is related
to the assessment of the PMI, and present my objections to the PMMI (Section
13.7 to 13.9). Section 13.10 deals with the charge of ad hocness. In Section 13.11
I briefly discuss how the PMI and the PMMI fare, if it is assumed that there has
been a large increase in degrees of success in the recent past. A short conclusion
summarizes the argument against the PMMI.
The dialectic can be developed in a number of different ways. Obviously
I cannot discuss every possible branch of the dialectic, but have to make some
choices and assumptions. Some assumptions constitute, or rely on, strong
simplifications. For example, as will become apparent later on, I rely on a strongly
idealized picture of the history of science. The assumptions and simplifications
are always meant to serve the goal of understanding and assessing the PMMI.
Thus I won’t necessarily use the most prominent or plausible versions of the
positions and arguments in the realism debate, but ones that are useful and con-
venient for the purposes of this goal.
In this Section I set up the dialectic between the realist and the antirealist. Let
scientific realism be defined as the position that our current successful theories
are probably approximately true. (In the following I will omit “probably” and
“approximately.”) Examples of such theories are plate tectonics, the heliocen-
tric system, and the theory of evolution. There are many versions of scientific
realism, of course, but this definition is a fairly standard one. A theory enjoys
empirical success iff it has made sufficiently many sufficiently significant true
predictions and no important false predictions, in other words, iff it has passed
sufficiently many sufficiently severe tests and not failed any important tests. This
is not very precise, but will do for our purposes.
Scientific realists support their position with the No-Miracles Argument
(NMA), according to which the success of our current successful theories would
be a miracle if they were false (Putnam 1978, Smart 1963, p. 39). Although this
We Think, They Thought 287
argument is usually spelled out as an inference to the best explanation, I will only
work with the basic intuition of the NMA. Antirealists usually do not accept the
NMA, but considering challenges to the NMA other than the PMI would lead us
too far afield.
Antirealists attack realism by means of the PMI. I begin with a version of the
PMI that highlights the commonalities between the PMI and the PMMI, and
therefore looks a bit different from the ways it is usually presented. This version
starts with the observation that many realists in the past said the same thing as
today’s realists do: They also believed that their successful theories were true,
supporting their realism with arguments resembling the NMA. For example,
Clavius, Kepler, and Whewell defended realism about their theories by alluding
to the empirical success of their theories (Musgrave 1988, p. 229/30). However,
so the attack goes, many of their successful theories were later refuted. For ex-
ample, the Ptolemaic system endorsed by Clavius was refuted, and Kepler relied
on all sorts of later-abandoned assumptions about the sun and the planets (Lyons
2006). All theories just mentioned (heliocentric system, theory of evolution,
plate tectonics) had precursors that were successful to some degree and held
by many scientists to be true, but were later replaced.6 Thus, past realists were
proven wrong about their realism. It follows that today’s realists will probably
share the same fate as past realists, implying that realism about current successful
theories is likewise wrong. So, this is the PMI. In the next section we will see
that it is not essentially different from more familiar versions of the PMI. Wray
captures the spirit of the PMI nicely:
There is another objection to realism. Realism as just defined refers only to pre-
sent, but not to past successful theories. The reason is obviously that past suc-
cessful theories were often refuted, hence cannot be included in the scope of
realism. This invites the objection that this is not a good reason to restrict the
scope of realism to current successful theories. The realist has to offer an inde-
pendent reason for the restriction, or else the restriction is ad hoc, only serving
6 Other examples famously include the phlogiston theory, the caloric theory of heat, the ether
theory of light, and some more (see Laudan 1981 and the references in footnote 4). Note that the suc-
cessful refuted theories do not necessarily refer to unobservables, rather some refer chiefly or exclu-
sively to observables (Stanford 2006, Ch. 2, Fahrbach 2011, 2017, Sect. 3.3). For this reason the PMI is
also a threat to realism about theories of the latter kind.
288 Contemporary Scientific Realism
the purpose of making realism compatible with the existence of past refutations.7
This objection differs from the PMI since it does not use past refuted theories to
undermine realism directly, instead it only uses them to point out a weakness in
the definition of realism, namely an unprincipled restriction of the scope of re-
alism. I will discuss the charge of ad hocness in Section 13.10.
The PMI as just presented is an instance of a general argumentative strategy
on the antirealist’s side that has the following form: The target of each instance
of the strategy is some claim or piece of reasoning endorsed by the realist which
refers to current theories. The strategy consists in first observing that past realists
made, or could have made, the same claim (employed, or could have employed,
the same piece of reasoning) about past theories as the realist does about current
theories, and, second, pointing out that past realists were, or would have been,
proven wrong by later theory refutations or rejections. We are then invited to
conclude that the claim (or piece of reasoning) offered by today’s realist about
current theories is wrong as well. I will call any instance of this general strategy
an argument from the past. Different arguments from the past differ with respect
to which claim or piece of reasoning is inserted into the general form. The PMI
is obviously an argument from the past. It targets the realist claim that current
successful theories are true. Later we will encounter further arguments from the
past, in particular the PMMI.
I will now provide a detailed analysis of the PMI. This will pave the way for the
later analysis of the PMMI. The analysis consists of two remarks and a procedure
to simplify the PMI.
First remark. The PMI starts from the observation that many past realists said
the same thing as current realists do: they endorsed realism and supported it with
arguments like the NMA. The PMI goes on to state that past realists were proven
wrong by subsequent theory refutations and that today’s realists will share the
same fate as past realists. Now, past realists obviously didn’t say the very same
thing as current realists; rather they referred to different theories, namely past
successful theories. Furthermore, there is an inductive step at the end of the PMI
which carries us from the statement that past realists were wrong about past re-
alism to the conclusion that current realists are wrong about current realism.
This inductive step is obscured when the PMI states that past realists said “the
7 The charge of ad hocness has several variants, see Stanford (2006, p. 10), Devitt (2011, p. 292),
same thing” as present realists, and present realists “will share the same fate” as
past realists. The inductive step will be examined more closely in Section 13.7.
Second remark. The claim of the PMI that past realists were shown to be
wrong by subsequent theory refutations should not be taken to mean that past
realists were unjustified or irrational when they endorsed realism for their the-
ories. After all, the relevant refutations could not be known by past realists
at their respective times, because they occurred at respective later times.
Generally, the conditions which determine whether a position held by some
people at some time is justified or not should be knowable by those people
at that time, and should not depend on empirical information which only
becomes available later on.8 Hence, the PMI should not be understood to be
talking about the justifiedness of the assertions of past realists, but about the
truth or falsity of those assertions. We learn empirically, using information ac-
quired later on, that their assertions were false. Whether or not their assertions
where justified can be left open. The PMI then invites us to infer that current
realism is also false, and this is intended to imply that endorsing current re-
alism is unjustified.
Let us now simplify the PMI. First, referring to present and past realists and
their respective assertions makes the argument more vivid by turning it into a
story about the misfortunes of real people, but this is not essential to the argu-
ment. Rather we can focus on the contents of the assertions and just talk about
theories and the properties of theories such as success, truth, and falsity. Then
the PMI is the following inference:9
inferring therefrom that present realism is false, we can use past refutations to
undermine present realism directly. Then the PMI becomes:
This is a more familiar version of the PMI.10 I will use it from now on. Note
that the two simplifications mean that it is not essential to the PMI that we “see
ourselves as similar to the scientists of the past” or that we “recognize the sim-
ilarities between our predicament and the predicament of our predecessors”
(Wray 2013).
If both the PMI and the NMA are still operative at the end of the dialectic,
i.e., neither is taken to be defeated by some argument or other, then they have
to be balanced with each other to reach a final verdict on realism. To simplify
the discussion, I will assume that the PMI always trumps the NMA.11 Hence as
long as the PMI is operative during the dialectic, the NMA can be ignored. If
the PMI is not operative at the end of the dialectic, e.g., judged to be blocked
by the improvements of science, then the realist may invoke the NMA to sup-
port realism, but doing so still faces the charge of ad hocness, which I discuss in
Section 13.10.
Both simplifications may also be applied to other arguments from the past.
Recall that arguments from the past are instances of the following general
strategy: The aim is to undermine a certain target claim made by a current re-
alist about current theories by noting that past realists made (or could have
made) the same claim about past theories, but were (or would have been) shown
wrong by later theory refutations. The conclusion we are meant to draw is that
the target claim is wrong as well. The first simplification consists in abstracting
from present and past realists and their respective assertions, and just working
with the contents of the assertions, which refer to theories and their properties.
The second simplification cuts out the intermediate conclusion that the claim
about past theories was undermined by the ensuing refutations. Instead past
refutations (possibly combined with further information) are used to undermine
10 Current realism is the claim that current successful theories are probably true. Hence the ne-
gation of current realism, which is the conclusion of the PMI, can be taken to be the statement that
many current successful theories are false. Then the PMI is the inference from “Many past successful
theories were refuted” to “Many current successful theories are false.”
11 This assumption is actually quite strong. The NMA is arguably equivalent to, or reducible to, a
first-order argument about the support of theories by observation (see e.g., Bird 1998), while the PMI
is a second-order argument (on the meta-level), and one might hold that first-order arguments are
generally stronger than second-order arguments, or at least not that easily trumped by the latter. The
whole dialectic can be easily modified to accommodate a different judgement about the balancing of
the two arguments, although things get a bit more complicated.
We Think, They Thought 291
the present realist’s claim about present theories directly. Later we will apply both
simplifications to the PMMI.
Realists have developed a number of responses to the PMI. One response is se-
lective realism, on which I commented earlier. In this paper I focus on another
defense, the one that gives rise to the PMMI. This defense claims that science has
improved a lot since the times of the refuted theories, and these improvements
block the PMI. There are several versions of this defense, which differ with re-
spect to the precise nature of the invoked improvements. One version claims that
the methods of science have improved.12 Other versions hold that current theo-
ries are better than past theories in this or that respect, e.g., have fewer rival the-
ories, have more “explanatory success,” are more mature, or have more empirical
success.13 I will use the last version, involving the notion of empirical success.
Then the realist defense against the PMI can be taken to make two claims: First,
present theories are significantly more successful than past refuted theories, and
second, this difference in success between present and past theories blocks the
PMI. Both claims will be explained over the course of the paper.
The realist’s defense is based on three assumptions. I will adopt them without
much discussion, because they are quite plausible, and defending them is not
our topic here. The first assumption is that empirical success comes in degrees.
Earlier I defined a theory to be successful if it makes sufficiently many sufficiently
significant true predictions and only negligible false ones. Accordingly we can
understand the degree of success of a theory to be determined by the number
and quality of true predictions of the theory, in other words, by the number and
severity of the empirical tests the theory has passed. I use the notion of degree of
success because the realism debate is usually framed with the notion of success,
but one could just as well use other, closely related notions from confirmation
theory, such as the empirical support enjoyed by a theory from empirical evi-
dence, or some notion of confirmation of a theory by the empirical evidence.14
The notion of degree of success may be identified with some probabilistic notion
12 “[N]ot only are scientists learning more and more about the world, but also . . . they are learning
more and more about how to find out about the world; there is an improvement in methodology.”
(Devitt 1991 p. 163). See also Devitt (2007, p. 287, 2011, Sect. 3), Roush (2009), and Wray (2018,
Ch. 6).
13 For fewer rival theories (“attrition”) see Ruhmkorff (2013, 412), for “explanatory success” see
Doppelt (2007), for maturity see Vickers (2018, 49), for empirical success see, e.g., Bird (2007, p. 80),
Devitt (2011, p. 292), and Fahrbach (2017). It is an interesting question how the different versions are
related to each other, but I won’t discuss that here.
14 Compare Vickers (2013, Sect. 3) and Vickers (2019, Sect. 4).
292 Contemporary Scientific Realism
15 Strictly speaking, degrees of success are not ascribed to theories simpliciter, but to temporal
stages of theories.
16 If one thinks that novel predictive success comes in degrees (Vickers 2013, p. 196, 2016, Alai
2014, p. 298, 312), then one may construct a version of the realist defense against the PMI that relies
on degrees of novel predictive success. Such a defense also faces a PMMI-like argument.
We Think, They Thought 293
higher degree of success than phlogiston theory” are meaningful and justifiable.
The third assumption may be deemed especially plausible if the notion of de-
gree of success is identified with some probabilistic notion such as the likelihood
ratio. If one does not accept that theories from different specialties can be com-
pared with respect to degree of success, or only accepts comparisons inside some
discipline or subdiscipline of science, then one may pursue the whole discussion
at a more local level.
The increase in success over the history of science has obviously been far from
uniform. Different scientific areas started life at different times, and have been
developing at different rates. Hence, the theories accepted by scientists today
occupy a broad range of degrees of success. Some areas, especially in the nat-
ural sciences, have produced theories with enormous success, while many other
areas are stuck with theories at medium or lower levels of success. The realist
defense against the PMI as I am developing it here focuses on the theories with
the highest degrees of success. These are “our current best theories.” They consti-
tute the easiest kind of case for defending realism. Once secured, such a realism
may serve as a bridgehead for advancing to realist positions for harder cases, i.e.,
to develop appropriate (possibly weaker) realist positions for theories with me-
dium and lower degrees of success.
The realist’s defense against the PMI can now be taken to consist of two
claims: Our current best theories are significantly more successful than practi-
cally all refuted theories, and this difference in success blocks the PMI. The first
claim presupposes a way to delineate the set of “our current best theories,” a no-
tion of difference in success between theories, and an explanation of the term
“significant.” However these notions are understood, it is clear that the first
claim is a substantial empirical claim requiring support from the history of sci-
ence. I will return to all these issues later. The second claim (that the difference
in success blocks the PMI) means that the difference in success prevents past
refutations from undermining realism about our current best theories, i.e., the
inductive step of the PMI does not go through. Given the realist’s defense we can
finally formulate the PMMI.
Realists in the past could have reasoned in the same way as today’s realist does
in response to the PMI: “Our current best theories are significantly more suc-
cessful than past refuted theories, and this difference blocks the PMI.” But look
what subsequently happened, many of their theories were refuted. Hence, the
294 Contemporary Scientific Realism
reasoning of today’s realist fails, and the PMI is not blocked by the increase in
success.
Most people seem to find the PMMI to be intuitively quite compelling.17 However,
getting clear about its actual import is not so easy. To simplify the discussion let us
fix on a specific date at which the imagined realists of the past do their reasoning.
One plausible such time is the year 1900, at least for physics, because Newtonian
mechanics and the Newtonian theory of gravitation had been quite successful and
stable until around 1900, so that Max Planck famously got the advice, “in this field
[physics], almost everything is already discovered, and all that remains is to fill a
few unimportant holes” (Lightman 2005, p. 8). Then the imagined realists from
1900 don’t argue against our PMI, or PMItoday, but against a version of the PMI for
the year 1900, call it PMI1900. This leads to a second rendering of the PMMI:
Realists in 1900 could have reasoned in the same way as today’s realist does
in response to the PMItoday: “Our current best theories are significantly more
successful than past refuted theories, and this difference blocks the PMI1900.”
This reasoning by realists in 1900 would have been proven wrong by subse-
quent theory refutations. Hence the reasoning of today’s realist in response to
the PMItoday also fails and the PMItoday is not blocked by the increase in success.
The second rendering makes it clear that the realists of 1900 would not have
reasoned in exactly the same way as the realist does today, only in an analogous
way, namely at a lower level of success and targeting a different PMI. The PMMI
is obviously an argument from the past, i.e., an instance of the general antire-
alist strategy mentioned earlier: A piece of reasoning about current theories is
attacked by transferring it to the past and noting that the past version of the rea-
soning is undermined by subsequent theory changes.
The PMMI states that the reasoning of the imagined realists in 1900 “would
have been proven wrong” by subsequent refutations. This should not be taken
to mean that their reasoning would have been unjustified or irrational. The
refutations which proved their reasoning wrong occurred subsequently, hence
were not known in 1900. Therefore, the PMMI should not be understood as talking
about the justifiedness of the PMI1900, but about the reliability of the PMI1900.18 The
PMMI then invites us to infer that the PMItoday is likewise a reliable inference.19
17 Most references in footnote 3 seem to endorse the PMMI. Exceptions include Devitt and Psillos.
18 For our purposes we can define an inductive inference to be reliable, if its conclusion is true
given true premises. An inductive inference understood as a type is reliable, if the conclusions of its
instances are mostly true when its premises are true.
19 I assume that the statement that a given version of the PMI is not blocked by the increase in suc-
We may also describe the PMMI as follows: Using historical information acquired
after 1900, we learn empirically which kind of meta-inductive inference, pessi-
mistic or optimistic or neither, is reliable in 1900; we find that a pessimistic meta-
induction is reliable and project the finding to the present. In other words, the
PMMI uses historical information to calibrate the meta-inductive inference for
1900, and uses the result to calibrate the meta-inductive inference for today. Note
that the conclusion of the PMMI is intended to mean that the realist’s assertion
today, that the PMItoday is blocked by the increase in success, is unjustified.
Like the PMI, the PMMI can be simplified in two steps. First, let the
PMMI refer to theories and their properties rather than to imagined people
and their assertions. Then the premise of the PMMI marshals the historical
evidence and can be taken to state that the past of science exhibits a certain
pattern, namely 〈refutations before 1900, successes until 1900, refutations
after 1900〉 (Figure 13.1). This premise supports the intermediate conclu-
sion that the PMI1900 is a reliable inference (is not blocked by the increase in
success until 1900), where the premise of the PMI1900 refers to the first and
second component of the pattern, and the conclusion of the PMI1900 refers to
the third component of the pattern. The intermediate conclusion supports
the final conclusion that the PMItoday is a reliable inference (is not blocked by
the recent increase in success).
Second, we can remove the intermediate conclusion from the PMMI. Then the
PMMI looks like this:
The past of science exhibits the pattern 〈refutations before 1900, successes
until 1900, refutations after 1900〉.
The PMItoday is a reliable inference.
PMMI
PMI1900 PMItoday
refutations successes b1 refutations successes b2 refutations
l1 l2
1900 today
So, instead of using the pattern of the premise to first calibrate the meta-inductive
inference for 1900 and then using the result to calibrate the meta-inductive in-
ference for today, we use the pattern of the premise to calibrate directly the meta-
inductive inference for today. If this contraction of the PMMI is rejected, the
discussion that follows gets a bit more complicated, but yields essentially the same
conclusions.
Figure 13.1 invites a quick recapitulation of the whole discussion so far. The
realist is impressed by the successes of our current best theories (the second rec-
tangle from the right) and wants to infer realism about these theories. The antire-
alist is impressed by the previous refutations and recommends the projection of
theory failure to the future instead (the three rectangles to the right).20 The realist
replies that degrees of success have increased from past to present, blocking the
projection of theory failure. The antirealist responds that the whole situation al-
ready occurred in the past (the three rectangles to the left): First, past successes
also suggested realism about past theories; second, previous theory refutations
suggested the projection of theory failure instead; third, past realists could have
replied that the projection of theory failure is blocked by the increase in suc-
cess; and fourth, they would have been proven wrong by subsequent refutations,
showing that in 1900 the projection of theory failure would have been the right
thing to do. It follows (big curved arrow) that the projection of theory failure
goes through today. Unsurprisingly the big curved arrow will be a topic later.
20 We can assume that science will go on in the same way in the future as in the past. This implies
that the conclusion of the PMI that realism is false is equivalent to the statement that many current
best theories will be refuted in the future, and similarly for the PMI1900.
21 The principle against the double-counting of evidence may be seen as the complement of the
principle of total evidence, according to which every piece of evidence relevant for the truth of a state-
ment should contribute to our epistemic assessment of the statement. Together the two principles
We Think, They Thought 297
One might respond to the objection by pointing out that using empirical evi-
dence twice is legitimate in this case, because it is used for two different purposes.
The existence of refuted theories is used, first, for the premise of the PMI, and
second, to support via the PMMI that the inference of the PMI is reliable, and
these are two different purposes. However, one may retort that using empirical
information twice in this manner is still dubious. One should not use the same
empirical information both as a premise for an inference and to help determine
what to infer from this premise (to calibrate the inference). What is more, both
purposes serve the same final purpose, namely to assess the conclusion of the in-
ference, in our case the conclusion of the PMI that current realism is false, so it is
true after all that one piece of evidence is used twice for the same purpose.
I want to offer two considerations indicating that this kind of double-counting
may not be illegitimate in the case of the PMMI. First, the objection of double-
counting of evidence can also be directed against endorsing both the PMI and
antirealism. Take antirealism to be the inference from the success of current suc-
cessful theories to their falsity.22 Let the PMI be the inference from the existence
of successful-but-refuted theories to the conclusion that antirealism understood
as an inference is reliable. (To simplify things I use a somewhat stronger ver-
sion of the PMI here than in the rest of the paper.) Then the observations that
are used to show that current successful theories are indeed successful—which
is the premise of antirealism understood as an inference—are also used both to
show that the successful-but-refuted theories were successful and to refute those
theories—which is the premise of the PMI. Hence, these observations are used
both for the premise of antirealism and to determine via the PMI what to infer
from the premise of antirealism. They are used twice for the same final purpose,
namely to assess current successful theories (in the conclusion of antirealism un-
derstood as an inference). If we reject the PMMI for the reason that it uses the
same empirical evidence for both the premise of an inference and to determine
what to infer from the premise, then we should also reject the PMI for the same
reason. Given the intuitive plausibility of the PMI, this is a high price to pay.
Second, a version of the PMMI can be constructed that avoids the double-
counting of evidence. Divide the set of scientific areas randomly into two subsets
assert that when assessing a statement every relevant piece of evidence should count at least once and
no more than once.
22 This sounds like an unduly strong version of antirealism, but it is not, if one assumes ordinary
Bayesian probabilities for example. The priors of typical scientific theories are very low, very near to
zero, because scientific theories are typically quite general and have numerous non-negligible rivals.
Favorable observations eliminate many rivals, increasing the probabilities of the theories, but the
question is by how much. Antirealism can be taken to be the position that current successful theories
should still be expected to have many non-negligible rivals not eliminated by the observations, hence
their posteriors, although higher than the priors, are still near zero, hence our current successful the-
ories should be judged to be false.
298 Contemporary Scientific Realism
of equal size. Use the first subset to run the PMI, i.e., use the theory refutations in
the histories of the scientific areas of this subset for the premise of the PMI and
let the conclusion of the PMI refer to the current theories of this subset. Use the
second subset for the premise of the PMMI, i.e., to calibrate the meta-inductive
inference for the PMI. So, the premise of the PMMI refers to patterns such as
〈refutations before 1900, success until 1900, refutations after 1900〉 in the his-
tories of the second subset, and the conclusion of the PMMI states that the PMI,
which runs in the first subset, is reliable.
One may object that this solution comes at the price of weakening the prem-
ises of the PMI and the PMMI: the strength of both premises is halved, intuitively
speaking. However, this is not the case. The relevant quantity of both premises
is a frequency, the frequency of refutations in the two sets, i.e., the ratio of the
number of refuted theories in the respective periods of time to the number of
scientific areas. If the set of scientific areas is cut in half, the ratios will not change
much, because both the numerators (the number of refuted theories in the re-
spective periods of time) and the denominators (the number of scientific areas)
are cut in half. Hence, this proposal might work, and I will not hold the double-
counting of evidence against the PMMI. However, in order to keep the discus-
sion simple let us put dividing all scientific areas into two subsets right out of
our minds.
Before assessing the PMMI, we have to take a closer look at the assessment of the
PMI. This is the aim of this section. The aim is not to reach an actual assessment
of the PMI, e.g., to argue on the realist’s behalf that the PMI is blocked; I only
want to get clear about which issues need to be addressed when assessing the
PMI, without actually addressing them. At the end of the section we will see how
the PMMI arises in this context.
The PMI of 13.3 states that many past successful theories were later refuted,
therefore current realism is false. It is an induction: its premise refers to once-
successful theories and its conclusion refers to current successful theories. It uses
a binary notion of success where each theory at a given time is either successful
or is not.23 Then the induction proceeds inside the set of successful theories and
seems quite plausible. However, we now operate under the three assumptions
of the realist’s defense, in particular that empirical success is graded and that it
23 Most of the literature on the PMI starting with Laudan (1981) relies on a binary notion of (novel
predictive) success.
We Think, They Thought 299
has increased over the history of science. Given these assumptions, assessing the
PMI becomes much more difficult.
At this point I will make a fourth assumption. It states that there is a reason-
able definition of a global difference in success between past refuted and pre-
sent best theories such that this difference is non-zero. This assumption is quite
vague, but will do for our purposes.24 It is a substantial assumption about the
history of science, for which the realist has to provide empirical support, but the
realist defense is making a considerably stronger claim anyway, namely asserting
a “significant” difference in success between the two sets of theories, on which
more later. The antirealist may attack the fourth assumption, of course. She may
argue, for example, that a substantial number of refuted theories enjoyed degrees
of success comparable to the theories of the set of our current best theories, on
any reasonable definition of this set. This would imply, she could argue, that, cor-
rectly understood, the global difference in success between past refuted and cur-
rent best theories is zero, and the realist’s defense against the PMI fails. However,
this would amount to an objection to the realist’s defense that is different from
the PMMI, hence will not be pursued here.
The fourth assumption (that there is a non-zero global difference in success
between past refuted and present best theories) has the important consequence
that the inductive step of the PMI is an extrapolation. It extrapolates the occur-
rence of theory failure along degrees of success from past to present levels of
success. It is then not obvious that the extrapolation goes through, that theory
failure can be extrapolated in this way.
The realist may reject the extrapolation of theory failure along degrees of
success altogether. He may argue that from the existence of refuted theories at
a given level of success one can only infer something about the assessment of
theories at that level of success, not about any higher levels of success. This would
mean that the PMI is blocked by any non-zero global difference in success. The
antirealist may counter this move by arguing that theory failure can be extrapo-
lated along degrees of success, and this may then be the debate between the two
sides. However, I don’t want to pursue this branch of the dialectic; rather I will
assume that the realist grants that extrapolations of theory failure are possible
in principle. The question is only whether the extrapolation goes through in the
particular case at hand.
24 The global difference in success between past refuted and present best theories can be defined
in a number of ways. For example, start from today’s most successful theory overall, then descend
degress of success until sufficiently many theories for a set of “current best theories” have been
gathered. Determine the lowest degree of success of the theories in this set. Below that, determine
the highest degree of success above which there are only a negligible number of later refuted theories.
The difference between these two degrees of success may be taken to define the global difference in
success between past refuted and present best theories.
300 Contemporary Scientific Realism
[I]f realists are to blunt the threat of the Pessimistic Induction, they must iden-
tify some significant difference between today’s theories and past theories.
Without an argument to the effect that there is a fundamental difference be-
tween the theories we currently accept and the once successful theories we have
since rejected, we have little reason to believe that today’s theories will not end
up on the pile of ruins to which Poincaré drew attention. (2018, p. 93, emphasis
in original)
We have established that there are differences in success for which the extrapo-
lation of theory failure is highly implausible. However, we are interested in the
difference between past and present levels of success and today’s PMI. The ques-
tion is whether this difference is small, so as to allow the extrapolation of theory
failure to present levels of success, or big, so as to block the extrapolation. As the
discussion so far already suggests, this is a hard question. To answer it we have
to engage with difficult empirical and epistemological issues. We have to find
some way to actually measure or estimate the degrees of success of past refuted
and current successful theories, at least roughly. Then we have to find a suitable
delineation of the set of present best theories. Then we have to settle on a suitable
definition of global difference in success between past and present best theories,
and have to determine this difference. Finally we have to formulate and defend
a judgment concerning whether the global difference is small or big, whether it
allows or blocks the extrapolation of theory failure. It seems that we can only as-
sess the PMI, if we tackle these difficult issues.26
The realist defense against the PMI as presented here does not engage with the
difficult issues. It just claims that the global difference in success between past
and present theories is “significant” (meaning “big,” i.e., capable of blocking the
PMI), without telling us how the difference was determined, and why it is “signif-
icant.” Hence, the realist has some work to do here. He has to engage with the dif-
ficult epistemological and empirical issues to defend his claim that the difference
is “significant.” What about the antirealist? She wants the PMI to succeed. Hence,
she has to show that the difference in success between past and present theories
is small, i.e., does not block the extrapolation of theory failure. To show this she,
too, has to engage with the difficult epistemological and empirical issues. As long
as she does not do that, she, too, has some unfinished business—or so it seems.
It is at this point that the antirealist may offer the PMMI. The PMMI supports
that the PMI goes through, that is its conclusion, hence it promises to give
the antirealist an advantage, possibly decisive, over the realist in the debate
26 And I didn’t even mention the other factors that may be relevant for deciding whether theory
failure should be extrapolated to current levels of success, and the NMA which I assume for the sake
of simplicity to be trumped by the PMI.
302 Contemporary Scientific Realism
concerning the PMI. What is more, it avoids the difficult epistemological and
empirical issues of the PMI, because it looks at the past of science to establish
the reliability of the PMI. But I will now argue that despite first appearances, the
PMMI does not succeed in giving the antirealist an advantage in the debate.
I will offer two objections to the PMMI. The first concerns the inference, the second
the premise of the PMMI. The first objection focuses on the inductive step at the
end of the PMMI (the curved arrow in Figure 13.1). Given the fourth assumption,
that past refuted and present best theories are separated by a non-zero global dif-
ference in degrees of success, the inductive step of the PMMI is an extrapolation
from past to present levels of success. It extrapolates the pattern 〈refutations before
1900, successes until 1900, refutations after 1900〉 to the present to support the PMI.
Then the PMMI faces problems very similar to those faced by the PMI. It is not ob-
vious that the extrapolation goes through; it depends on the size of the difference
between past and present levels of success (as well as on other factors, which I ignore
here). Let us call the difference “small” if the extrapolation of the PMMI succeeds
and “big” otherwise. Once again the very existence of big differences can be backed
by considering future scenarios in which science will develop in a similar manner
in the next 100 or 1000 years as it did up to now, producing an extremely big in-
crease in success. For such an increase it is highly plausible that the extrapolation of
the PMMI fails. The question, though, is whether the increase in success so far has
been small or big. Once again this is a hard question. To find an answer, we have to
engage with difficult empirical and epistemological issues: we have to determine the
degrees of success of the involved theories, define and determine the global differ-
ence, and formulate and defend a judgment whether the global difference in success
is small or big, i.e., whether the extrapolation of the PMMI is plausible or not. Only
by engaging with these issues can we assess the PMMI, or so it seems.
The antirealist may evade the difficult issues of the PMMI in the same way
as she evaded those of the PMI: She may present a PMMMI. What then looms
is an infinite hierarchy PMI, PMMI, . . . PMnI, . . ., in which the assessment of
each PMnI faces difficult epistemological and empirical issues, which the anti-
realist evades by offering another argument from the past, the PMn+1I, which
invokes the past of science to establish the reliability of the PMnI. In this way the
antirealist avoids the difficult epistemological and empirical problems at every
level, but it is a bit unclear whether such an infinite hierarchy of arguments is
legitimate in principle.27 Anyway, it does not succeed for practical reasons: The
27 A related infinite hierarchy of arguments and counterarguments arises as follows: The realist
We Think, They Thought 303
premise of the PMI refers to the occurrence of past refutations. The premise of
the PMMI is more complicated, invoking patterns such as 〈refutations before
1900, successes until 1900, refutations after 1900〉. The premise of the PMMMI
refers to even more complicated patterns such as
(The conclusion of the PMMMI is the statement that the inference of the PMMI,
the curved arrow in Figure 13.1, goes through.) As n increases, the premises
of the PMnI become more and more complex, presupposing more and more
episodes of theory refutation. But the number of such episodes in the history of
science is finite. Hence, for some level N, the antirealist will not be able to rise to
the level N+1. It follows that she has to face the difficult empirical and epistemo-
logical issues at some level.28 She may choose the level, say level n, with n ≤ N, but
she cannot avoid the difficult empirical and epistemological issues altogether;
at some point she has to engage with them. What is more, it does not seem ad-
visable to go up the hierarchy of PMnI’s very far, because the complexity of the
premises grows exponentially as n increases, making it ever harder for the anti-
realist to show that the premise of the chosen PMnI is actually true (compare the
next section). So let us return to the assessment of the PMMI.
My first objection to the PMMI is then as follows. The PMMI promised to
help the antirealist to gain an advantage in the debate over the PMI. It seemed to
show that the extrapolation of the PMI goes through by invoking patterns such
as 〈refutations before 1900, successes until 1900, refutations after 1900〉, while
avoiding the difficult epistemological and empirical issues arising in the assess-
ment of the PMI. But we just saw that the inductive step of the PMMI is also
an extrapolation along degrees of success, whose assessment faces very similar
problems as the assessment of the extrapolation of the PMI. Thus, invoking the
PMMI just trades the problems of the PMI for very similar problems that are at
objects to each PMnI by arguing that it is blocked by the increase in success, and the antirealist
responds with the PMn+1I according to which realists in the past could have objected to the past ver-
sion of the PMnI in the same way, namely that it is blocked by the increase in success, but would have
been proven wrong by later theory refutations. This infinite hierarchy has the same problem as the
one in the text, namely the exponentially increasing complexity of the premises of the PMnIs.
28 Note that the antirealist only has to show for one level n that the PMnI succeeds, because the con-
clusion of the PMnI is that the PMn-1I succeeds, which has the conclusion that the PMn-2I succeeds,
and so on, until the PMMI shows that the PMI succeeds, which has the conclusion that realism is
false. In contrast, the realist has to argue that all PMnIs fail.
304 Contemporary Scientific Realism
least as hard as those of the PMI. It follows that the PMMI does not succeed in
helping the antirealist to gain an advantage in the debate over the PMI.
For the second objection, consider the PMMI in its unsimplified form. It starts
by claiming that people in the past, e.g., in 1900, could have reasoned in the same
way as today’s realist does. They, too, could have said: “Our current best theo-
ries are significantly more successful than past refuted theories, and this dif-
ference blocks the PMI1900.” This beginning of the PMMI presupposes that the
situation in 1900 was actually similar in relevant respects to the situation today.
That is a substantial claim, which needs empirical support from the history of
science. Likewise for the simplified PMMI, whose premise refers to the pattern
〈refutations before 1900, successes until 1900, refutations after 1900〉. It is a sub-
stantial empirical claim that the first and second component of the pattern are
sufficiently similar in relevant respects to the corresponding situations in the re-
cent past. More generally, the PMMI requires the premise that there were times
in the history of science in which the situation was sufficiently similar in relevant
respects to the situation today.29
Which respects are relevant for the similarity between the past and present
situations? I will examine two respects, the global difference in success between
the respective sets of best theories and the respective sets of refuted theories, and
the number of the respective scientific fields. Then the PMMI requires that there
were times in the history of science in which the global difference in success
and the number of relevant fields were both comparable to the corresponding
global difference and number in the recent past. Figure 13.1 makes clear what
I mean: the two rectangles dubbed “successes” should have roughly similar
lengths l1 and l2 (which represent the respective global differences in success)
and roughly similar breadths b1 and b2 (which represent the respective num-
bers of scientific fields with best theories). If, on the other hand, the lengths and
breadths of successes in the past were never similar to those today, for example, if
l1 was always much smaller than l2, and b1 was always much smaller than b2, then
the premise of the PMMI is never true.30
We have identified a substantial empirical assumption of the PMMI
about the existence of periods of time relevantly similar to the recent past.
29 I want to thank Ioannis Votsis for a number of challenging comments at this and other portions
of my argument.
30 If at some time l is smaller than l , but b is bigger than b , and some suitable function of l and
1 2 1 2 1
b1 such as their product l1∙b1 is roughly the same as that of l2 and b2, then the two situations may also
justifiably count as similar.
We Think, They Thought 305
Earlier, we took 1900 to be a suitable time, for the reason that central the-
ories of physics had been quite successful and stable for a while up until
around 1900. But the realist may argue that this is only then-f undamental
physics; the simultaneous stability of theories in recent times has been much
broader, encompassing the best theories of a far bigger number of scien-
tific fields (mostly from the natural sciences) than just then-fundamental
physics. The antirealist either has to show that theory stability occurred on
a much broader scale in 1900 than just then-f undamental physics, or she has
to present some other historical periods at which the increase in success and
the breadth of stability was comparable to today. Once again we find that
the antirealist, if she wants to hold on to the PMMI, has some unfinished
business.
It is worth noting that both objections may also be directed against other
arguments from the past. Recall that arguments from the past aim to undermine
a target claim made by a realist about current theories by noting that past realists
could have made the same claim about past theories, but would have been
proven wrong by later theory refutations. One objection consists in the challenge
to show that there actually were situations in the past of science that were suffi-
ciently similar in relevant respects to the situation today. The antirealist cannot
just claim that this is so, but has to provide adequate evidence that such situations
really occurred in the past. The other objection is that there is one respect in
which all past situations are dissimilar to the present situation, namely with re-
spect to degrees of success. If the fourth assumption (about a non-zero global dif-
ference between past and current best theories) is correct, then the inductive step
of any argument from the past will be an extrapolation from past lower to current
higher degrees of success. Whether the extrapolation is plausible will depend on
several factors, such as the global difference in success between past and present.
The antirealist has to show that the factors are such as to allow the extrapolation
of theory failure.
of past refutations can be used to attack realism about current theories. Only the
seemingly arbitrary restriction of realism to current theories is at issue.
The realist may offer a similar response to the charge of ad hocness as to the
PMI: our current best theories are more successful than past refuted theories,
and this difference in success provides an independent reason to treat past theo-
ries differently from our current best theories in the definition of realism. To this
move Stanford responds: “More success gives no reason to believe that we have
now crossed over some kind of threshold . . . such that these predictive powers
[of scientific theories] are now finally substantial enough [to accept them as
true]” (2009, 384 footnote, emphasis added). At this point we need to reintro-
duce the NMA intuition, as it is generally taken to be the source of support for
realism. Then it is plausible that the NMA intuition gets stronger as degrees of
success increase: the higher the success of a theory, the bigger the “miracle” if the
theory were false, and the higher the probability that the theory is true. It follows
that Stanford’s claim that more success gives us no reason to believe that we have
crossed over the threshold for truth is false. Rather, more success gives us some
reason.31 The question is, only, whether the reason is strong enough to support
realism as defined above.
At other places Stanford’s wording is a bit more cautious. In his (2015), he says
“little reason”:
Of course there are always important differences between each successive gen-
eration of theories (including our own) and their historical predecessors in a
given domain of scientific inquiry, but there seems little reason to think that
such differences are sufficiently categorical to warrant the conviction that con-
temporary scientific theories have now finally managed to more-or-less sort
things out at last, given that the same inference as applied to earlier theories,
predicated on the salient advances and advantages of those theories over their
predecessors, has turned out to be so repeatedly and reliably mistaken. (p. 3,
emphasis added)
“Little reason” is more cautious than “no reason,” but it is still clearly meant to
imply “not enough reason.” To support this claim, Stanford merely offers an ar-
gument from the past (starting with “given that” in the quote), which runs into
the problems noted earlier.32 He does not offer any first order reasons concerning
31 Even if the NMA intuition is not granted, more empirical success of a theory means that the
theory has passed more tests, which generally implies, on standard accounts of confirmation such
as Bayesianism, that the theory is incrementally confirmed. Hence, more generally, such additional
success offers some reason to think that we have crossed the threshold for truth.
32 It is not clear whether one should interpret Stanford’s argument from the past as the PMI, the
PMMI, or a further argument from the past concerning only ad hocness. Maybe it is a combination of
these. In any case, it runs into the problems noted earlier.
We Think, They Thought 307
the quantity and quality of the empirical evidence, and the strength of the empir-
ical support for our current best theories. Hence, concerning first order reasons,
his claim is as unprincipled as the opposite claim of the realist that present levels
of success suffice for the NMA to support realism. Whether either claim is justi-
fied, or no claim at all can be justified, surely depends on how big the increase in
success from past to present has actually been. If it has been small, then Stanford’s
claim may be justified and the charge of ad hocness may be plausible, but if it has
been big, then not. Consider once again the future scenario in which degrees of
success keep growing for another 1000 years at the same rate as they have done so
far. In such a scenario the NMA would surely give us more than “little reason” to
believe that we have crossed over the threshold for truth.
But the question is, of course, whether current levels of success are high
enough for the NMA to establish realism. The realist answers this question in
the affirmative, but according to the charge of ad hocness the affirmative answer
requires a justification, which the realist has not yet given. To find a justification,
the realist must engage with difficult empirical and epistemological issues: he has
to determine and compare the degrees of success of past and present best the-
ories and formulate and defend judgments based on the NMA intuition about
the plausibility of restricting realism to present levels of success at the expense of
past levels of success. So, the realist finds himself in a situation not unlike with re-
spect to the PMI as detailed earlier. In the next section I will hint at one possible
way how the realist may proceed in this situation.33
We saw at several points that the debate can only progress if we engage with diffi-
cult empirical and epistemological issues, namely compare the degrees of success
of past and present theories and formulate and defend judgments about the plau-
sibility of the PMI, the PMMI, and the charge of ad hocness.
Actually some such work has already been done.34 It strongly suggests that our
current best theories have enjoyed a very large increase in degree of success in
the recent past and swim in a sea of evidence today. One reason is, very roughly,
that the amount of scientific research in general has been growing at very high
rates in the recent history of science, which has led to huge improvements in
33 Alternatively, the realist may search for some deep epistemic difference between past refuted
and current best theories. However, I don’t think that there is such a magic bullet, which scientists
are unaware of, which is deeply buried in scientific practice, requiring the analytic digging powers of
philosophers for its unearthing. To my mind, there is just ordinary empirical evidence and ordinary
empirical success, but with large differences between different theories and different times, as I indi-
cate in the next section.
34 See, for example, Doppelt (2007, 2014), Park (2011), Mizrahi (2013), and Fahrbach (2017).
308 Contemporary Scientific Realism
computing power and the amount and quality of empirical evidence. This has
translated into ever-increasing degrees of success for the best theories. While
acquiring their tremendous success, practically all of them have been entirely
stable. There are no convincing examples of theories that enjoyed comparable
successes and were later refuted. At several points in the discussion we imagined
scenarios in which science develops in the same manner in the next few centu-
ries as it has done so far. The work just mentioned shows that the state of science
today is not so far from what we imagined in those scenarios. (Thus, Figure 13.1
offers a totally distorted view of the history of science.) Much more work needs
to be done to develop this approach further. Here I will just assume that it is gen-
erally correct, and confine myself to noting the implications for the assessment of
the PMI, the PMMI, and the charge of ad hocness.
First, the large increase in success and simultaneous stability of our current best
theories implies that the extrapolation of theory failure from past to present levels of
success is highly implausible. Our current best theories are far more successful than
any past refuted theories; hence the existence of the latter should have no bearing on
our assessments of the former. So the difference in success indeed blocks the PMI.
Second, the large difference between past and present levels of success also
blocks the PMMI. It prevents the extrapolation of patterns like 〈refutations be-
fore 1900, successes until 1900, refutations after 1900〉 to the present. What is
more, it is highly doubtful that the premise of the PMMI can be made true. To
make it true we have to find times in the history of science in which the situa-
tion was sufficiently similar to the recent past, in which the then-best theories
were similar in number and made comparable gains in success as in the recent
past. However, the recent past is quite unprecedented in this regard. No other
period comes anywhere close. (Again the picture of the history of science as
presented by Figure 13.1 is much distorted.) Thus, the PMMI does not succeed
in supporting the PMI and undermining the realist’s defense against the PMI.35
Third, the tremendous success of our current best theories makes the appli-
cation of the full NMA intuition to these theories very plausible: their success
would indeed be a miracle if they were false. By contrast, the success of past theo-
ries, whether refuted later on or not, does not suffice to trigger the full NMA intu-
ition: their success was far more moderate, and it was not such a big miracle when
they occasionally turned out to be false. So, the NMA fully supports the infer-
ence to truth for our current best theories, but does so only rather weakly for past
35 Similar remarks apply to the PMnIs with n > 2. Note that, the bigger and broader the recent
increase in success, the more the PMI faces competition from an optimistic meta-induction, OMI,
which uses the recent stability of our current best theories to project the stability of these theories
into the future, and likewise for higher-level optimistic meta-inductions, OMnIs.
We Think, They Thought 309
theories.36 It follows that the realist’s restriction of realism to current and future
highest levels of success is not ad hoc but supported by good reasons. What is
more, the big increase in success in recent times means that the realist need not
commit himself to a precise threshold beyond which success suffices for truth,
but can be very vague in this matter. All he needs to make plausible is that the
threshold is somewhere in the large interval of degrees of success between past
and present, and the NMA is able to do that. Thus, some of the difficult empirical
and epistemological issues turn out to be not so difficult after all, but are readily
solvable, because the recent increase in success has been so enormous.
13.12 Conclusion
36 Most of past refuted theories were not completely false, but contained a substantial amount of
truth (judged from today), as selective realists have shown in numerous case studies. Furthermore
many theories have been completely stable for centuries now, such as the heliocentric system, oxygen
chemistry, the reduction of temperature to mean kinetic energy, and so on (for more examples see
Bird 2007, 73).
310 Contemporary Scientific Realism
very similar problems as the PMI. It also involves an extrapolation from past to
present levels of success, which gives rise to empirical and epistemological is-
sues very similar to those of the PMI: determining the degrees of success of the
involved theories and reaching and justifying judgments concerning the plau-
sibility of the extrapolation. So, the PMMI trades one set of problems, those of
the PMI, for another set of very similar problems (as well as some additional
problems). Therefore, the PMMI does not succeed in helping the antirealist to
gain an advantage in the debate over the PMI. The upshot is that both realists and
antirealists can only make progress in the debate if they engage with the difficult
empirical and epistemological issues.
Acknowledgments
For discussion and support I would like to thank Claus Beisbart, Christopher
von Bülow, Matthias Egg, Christian Feldbacher, Berna Kilinc, Tim Lyons, Sam
Ruhmkorff, Yafeng Shan, Corina Strößner, Paul Thorn, Peter Vickers, Ioannis Votsis,
Brad Wray, two anonymous referees, and audiences in Lausanne and Durham.
References
Alai, M. (2014). “Novel Predictions and the No Miracle Argument.” Erkenntnis, 79(2),
297–326.
Alai, M. (2017). “Resisting the Historical Objections to Realism: Is Doppelt’s a Viable
Solution?” Synthese, 194(9), 3267–3290.
Bird, A. (1998). Philosophy of Science. Montreal & Kingston: McGill- Queen’s
University Press.
Bird, A. (2007). “What is Scientific Progress?” Noûs, 41(1), 64–89.
Devitt, M. (1991). Realism and Truth. 2nd edn. Oxford: Basil Blackwell.
Devitt, M. (2007). “Scientific Realism,” in: Frank Jackson and Michael Smith (eds) The
Oxford Handbook of Contemporary Philosophy. Oxford: Oxford University Press,
767–791.
Devitt, M. (2011). “Are Unconceived Alternatives a Problem for Scientific Realism?”
Journal for the General Philosophy of Science, 42(2), 285–293.
Doppelt, G. (2007). “Reconstructing Scientific Realism to Rebut the Pessimistic Meta-
Induction.” Philosophy of Science, 74, 96–118.
Doppelt, G. (2014). “Best Theory Scientific Realism.” European Journal for Philosophy of
Science, 4(2), 271–291.
Fahrbach, L. (2011). “How the Growth of Science Ended Theory Change.” Synthese,
180(2), 139–155
Fahrbach, L. (2017). “Scientific Revolutions and the Explosion of Scientific Evidence.”
Synthese, 194(12), 5039–5072.
Frigg, R., & Votsis, I. (2011). “Everything You Always Wanted to Know about Structural
Realism but Were Afraid to Ask.” European Journal for Philosophy of Science, 1(2),
227–276.
We Think, They Thought 311
1Department of Philosophy
University of Salzburg
[email protected]
2Department of Mathematics
Politecnico di Milano
[email protected]
14.1 Introduction
Patricia Palacios and Giovanni Valente, The Paradox of Infinite Limits In: Contemporary Scientific Realism.
Edited by: Timothy D. Lyons and Peter Vickers, Oxford University Press. © Oxford University Press 2021.
DOI: 10.1093/oso/9780190946814.003.0014
The Paradox of Infinite Limits 313
1 It is important to point out that one can also “construct” infinite systems without taking any
limits such as with the assumption of an infinitely long cylinder or two dimensional systems.
Although such examples deserve attention and have been discussed in the philosophical literature
(e.g. Earman 2017; Shech 2018), we will restrict our analysis to infinite idealizations resulting from
the use of infinite limits.
2 For completeness, let us clarify that in the literature one can find at least three different versions
of scientific realism: Explanatorianism, Entity realism and Structural realism (Chakravartty 2017).
In this paper we consider a general characterization of scientific realism, leaving the question of how
these particular versions can deal with infinite idealizations for future work. However, we point out
that the Paradox of Infinite Limits is especially problematic for Explanatorianism (Kitcher 1993;
Psillos 1999), which recommends a realist commitment with respect to those parts of our best theo-
ries that are indispensable to explaining empirical success.
3 See also (Fletcher, Palacios, Ruetsche, and Shech 2019) for a recent collection of papers on infi-
infinite idealizations they can actually lead one to the Paradox of Infinite Limits,
depending on whether the idealization is regarded as essential for the explana-
tion of the physical phenomenon under investigation. We then argue that, even
in the case of essential idealizations, there are ways of coping with the alleged
incompatibility between infinite idealizations and scientific realism, which ulti-
mately rely on empirical considerations.
We organize the paper as follows. In Section 14.2, we distinguish between
idealizations, approximations, and abstractions broadly constructed. This dis-
tinction is partially based on Norton’s (2012) distinction between idealizations
and approximations and on Godfrey- Smith’s (2009) distinction between
idealizations and abstractions. In the following section, we formulate the general
Paradox of Infinite Limits and we explain in greater detail the sense in which it
raises a challenge for scientific realism. In particular, in sub-section 14.3.3 we
state the condition of Empirical Correctness in precise terms, and then we pro-
ceed to analyze the various possible understandings of the use of infinite limits
in physics. Specifically, in Section 14.4 we develop the concept of approxima-
tion and argue that it is not problematic from a scientific realist perspective by
means of concrete examples: in fact, the paradox does not arise in the case of
approximations without idealizations, whereas it can be readily disarmed in the
case of idealizations yielding approximations. In Section 14.5, we address the use
of infinite limits as essential idealizations, explaining why that would lead the
scientific realist to a paradox. Yet, by focusing on the controversial example of
first-order phase transitions, we show that there is available a procedure to dis-
pense the infinite idealization “on the way to the limit,” thereby avoiding a con-
tradiction with the claim that real target systems undergoing the phenomenon
to be explained are finite. Finally, in the last section, 14.6, we address the case of
continuous phase transitions to illustrate the use of infinite limits as abstractions,
and we also explain why these limits would not raise any conflict with scientific
realism.
ways in which real systems are represented in scientific theories, one can distin-
guish at least between three different strategies labeled by philosophers of sci-
ence as idealizations, approximations, and abstractions.
Although there is no consensus about the nature of scientific idealizations,
most philosophers agree that idealizations involve a misrepresentation of a real
system (i.e. the target system) driven by pragmatic concerns, such as mathe-
matical tractability. Recent accounts provide a more precise characterization of
idealizations by arguing that they always refer to imaginary systems considered
to be analogues of the real-world systems of interest. For example, Godfrey-
Smith (2009, p. 47) says “I will treat [idealizations] as equivalent to imagining the
existence of a fictional thing that is similar to the real-world object we are inter-
ested in.” In a similar vein, Norton (2012, p. 209) defines an idealization as “a real
or fictitious system, distinct from the target system, some of whose properties
provide an inexact description of some aspects of the target system.” According
to these authors, there will be an idealization only when there is reference to a
novel (fictional) system, which has properties that resemble the properties of
real-world systems. In other words, when the misrepresentation of the proper-
ties of a target system coincide with the exact properties of a fictional system.
An example of idealizations is the frictionless plane just mentioned. This fic-
tional system was firstly introduced by Galileo to derive the equations of motion
of an object moving down an inclined plane. Although no such planes exist in
reality, they have proven to be extremely useful to predict the behavior of real
world systems. Why can these fictional systems explain the behavior of con-
crete systems observed in the world, and how can we justify their use from a
scientific realist perspective? The standard justification for idealizations of this
kind is that they provide approximately true descriptions of real world systems,
where approximation to truth is simply understood as a relation of similarity be-
tween the properties of fictional systems and the properties of real world systems
(Godfrey-Smith 2009). It is also believed that these idealizations are dispensable,
in the sense that it is possible in principle to de-idealize the theory by systemati-
cally eliminating distortions and by adding back to the model details of concrete
systems (McMullin 1985).4
This characterization of idealizations puts us in position to distinguish
them from other strategies of theory construction such as approximations and
abstractions. For instance, according to Norton (2012), approximations are in-
exact descriptions of certain properties of the target system, which are given in
terms of propositions expressed in the language of a theory, that do not need to
correspond to the relevant properties of some other fictional system. In fact, the
4 This kind of idealization corresponds to what has been called “Galilean idealization” (e.g.
use of approximations does not require one to make reference to any new system
different from the target systems. In this sense, one can say that approximations
involve distortions or misrepresentations of the target system but, differently
from the case of idealizations, these distortions are merely propositional. For
Norton (2012), idealizations can be demoted to approximations by discarding
the idealizing system and extracting the inexact description, but the inverse
promotion will not always succeed. However, we will see that there seem to be
cases of essential idealizations in which idealizations cannot be easily demoted
to approximations.
In recent years, the topic of essential idealizations has generated a great deal
of excitement among philosophers of science. In particular, it has been argued
that infinite idealizations arising from mathematical limits that are ineliminable
cannot provide approximations of realistic systems, because the latter exhibit be-
havior that is qualitatively different from the behavior exhibited by the idealized
infinite systems (e.g. Batterman 2005). We will examine some possible cases of
ineliminable idealizations in Section 14.5, but before doing so, let us mention
some examples of approximations and abstractions.
Let us recall that an approximation is an inexact description of the target
system that does not (necessarily) involve the introduction of a fictional system.
This means, there is no appeal to fictional system in which the inexact properties
of the target system are true. A good example of an approximation without ideal-
ization also mentioned by Norton (2014) is the case of a mass falling in a weakly
resisting medium. As we know, the speed v of a falling mass at time t is given by:
dv / dt = g − kv ,
where g is the gravitational acceleration and k the friction coefficient. The speed
of the mass at time t, as it starts from rest, is given in terms of the Taylor expan-
sion series by:
If we assume that the friction coefficient is small, we can approximate the pre-
vious expression by taking only the first term of the series expansion:
v(t )= gt
When we use this expression to describe the behavior of a real mass falling in a
resisting medium, we do not intend to give a literal description of the situation,
The Paradox of Infinite Limits 317
but rather to give a good approximation of it. This strategy allows us to simplify
problems that may be otherwise intractable. It is important to note, however, that
one can promote this approximation to an idealization by introducing a fictional
system in which a body falls under the same gravity in a vacuum, so that the fall
is described exactly by v(t) = gt (Norton 2014).
So understood, idealizations can also be distinguished from abstractions.
According to GodfreySmith (2009), an abstraction is the act of “leaving things
out while still giving a literally true description of the target system” (p. 48). In
contrast to idealizations, abstractions do not intend to state claims that are lit-
erally false and do not make reference to fictional systems.5 Instead, they in-
volve the omission of a truth by leaving out features considered to be irrelevant.
Abstractions also differ from approximations in that the former do not involve
propositional misrepresentations of the target system whereas the latter do.
We will argue next that all the forms of inaccurate representations mentioned in
the previous section, i.e. idealizations, approximation, and abstractions, can be
made compatible with the more relaxed realist framework that accepts the con-
tent of scientific theories to be at least approximately true. However, in order to
arrive at that conclusion, we need to offer a more precise definition of the notion
of “approximate truth.” Chakravartty (2010) distinguishes between three kinds
of approaches for approximate truth in the standard philosophical literature: the
verisimilitude approach, the possible-world approach, and the type hierarchy
approach. The verisimilitude approach, elaborated by Popper (1972), consists
in comparing the true and false consequences of different theories. In the pos-
sible word approach, which is meant to be an improvement of the verisimilitude
approach, the truth-likeness is calculated by means of a function that measures
a mathematical “distance” between the actual world and the possible worlds in
which the theory is strictly correct (Tichy 1976; Oddie 1986) so that one can gen-
erate an ordering of theories with respect to truth-likeness. Finally, in the type hi-
erarchy approach, truth-likeness is calculated in terms of similarity relationships
between nodes that represent concepts or things in the word (Aronson 1990).
As Chakravartty (2010) points out, the problem that all these approaches have
in common is that they do not pay attention to the different ways in which scien-
tific representations give inaccurate account of the target systems. This is an im-
portant limitation of these approaches, because the notion of approximate truth
is conferred to the ball. The scientific model that predicts where the cannonball
will land may include a number of distortions with respect to other properties
like the gravitational force, which is generally assumed to have the same magni-
tude and direction at all points of the trajectory. However, such a model does not
involve misrepresentation with respect to the properties that do not make a dif-
ference for the occurrence of the phenomena. In fact, the model is simply silent
about them, and hence it does not say anything false as regards these irrelevant
properties (Jones 2005). Although this strategy does not involve misrepresenta-
tion of certain properties it does give an inexact description of the target system
because it leaves out factors that are irrelevant for the behavior under considera-
tion. One then needs an appropriate notion of approximate truth in this case too.
The intended notion will differ from the one involved in cases of idealizations and
approximations in that there is no misrepresentation of properties. Chakravartty
(2010, p. 39) suggests an articulation of the notion of approximate truth qua
abstractions connected with the notion of comprehensiveness: “The greater the
number of factors built into the representation [i.e. the more comprehensive is
the description], the greater its approximate truth.” In so far as one can quantify
these factors, there will be a precise notion of approximate truth applying also to
the case of abstractions. It is important to note that sometimes abstractions may
be even crucial or essential for the explanation of certain classes of behavior.6 For
instance, if we want to explain the behavior of cannonballs in general, we should
not include details that are specific of a certain fire. This is possible just because
abstractions generally aid in the explanation of certain phenomena by enabling
us to account for some common behavior that is generated among disparate sys-
tems (see also Weisberg 2007).
Mathematical limits are vastly used in physics. For instance, in the statistical me-
chanical theory of phase transitions, it is assumed that the number of particles as
well as the volume of the system goes to infinity; similarly, in the ergodic theory
of equilibrium and in the explanation of reversibility in thermodynamics it is
assumed that time goes to infinity. Appealing to infinite limits has the technical
advantage that they render the formal account of physical phenomena more trac-
table. Furthermore, in some cases, like in the examples we just mentioned, it even
6 This account of abstraction is consistent with Cartwright’s (1994, p. 187) view, according to
which an abstraction is a mental operation, where we “strip away—in our imagination—all that is
irrelevant to the concerns of the moment to focus on some single property or set of properties, as if
they were separate.”
320 Contemporary Scientific Realism
The use of infinite limits has been frequently equated with the use of an infi-
nite idealization. However, we want to stress a distinction between different uses
of infinite limits as approximations, idealizations, and abstractions. In order to
draw such a distinction, we begin by setting up the formal framework within
which our discussion is cast. Let Sn represent a system characterized by some
physical parameter n, which may take on discrete or continuous values: in par-
ticular, it could denote the number N of molecules constituting a gas system, so
that the parameter takes on values in the natural numbers ; or, it could denote
the time t during which a physical process unfolds, so that the parameter takes
on values in the real numbers . As the variable n increases, one defines the fol-
lowing sequence of systems
{ f1 , f 2 , ..., fn } n ∈,
The Paradox of Infinite Limits 321
where the notation fn indicates the relevant property possessed by each finite
system Sn, so that fn := f S n . Here, some care should be taken when evaluating
the infinite limit n → ∞ . In fact, as Butterfield (2011) suggested, there is a crucial
difference between “the limit of a sequence of functions” and “what is true at that
limit”: that is, respectively,
• (i) the limit f∞ of the sequence {f1, f2, ..., fn}n of functions, and
• (ii) the function fS associated with the limit system S∞.
∞
More to the point, if the parameter n takes on values on the natural num-
bers, (i) obtains when one adjoints infinity to the set , and hence the
limit f ∞ := limn→∞ fn is defined as the function being the last element of
this sequence with n ∈ ∪ {∞}. In this case, there is no reference to the in-
finite system S∞, so that using the limit in this sense does not lead to an ide-
alization. To the contrary, (ii) depends exactly on how the limit system is
constructed. Indeed, the limit function fS represents the relevant quantity
∞
possessed by the infinite system S∞: as such, it tells us just what is true at the
limit. It should be emphasized, though, that its existence is contingent upon
the type of convergence one adopts: this point will be useful for our discus-
sion in Section 14.5.
The proposed distinction becomes less abstract if one casts it in terms of
values of quantities. In fact, supplying numerical values is right what enables us
to directly check whether or not the expected results turn out to be empirically
correct. So, given that the putative function f takes on values v(f), one should
consider yet another sequence, namely the sequence of values
Of course, the actual value of each function fn depends on the state sn in which the
system Sn is: in fact, there is also a sequence of states {s1, s2, ..., sn} implicitly under-
stood along with the sequence of systems.
As above, one needs to distinguish between (i) the limit of the sequence of
values of the function f, i.e. limn→∞ v ( fn ) , and (ii) the value v(fS ) of the natural
∞
limit function computed when the limit system S∞ is in the limit state s∞.
Before addressing the issue whether, and how, it is possible to dispense from
the infinity system in the explanation of some physical phenomenon, let us
conclude this section by noting that the formal setting we have just presented
enables us to sharply distinguish between different ways to characterize the
use of infinite limits in physics. The proposed taxonomy identifies three main
types, that is:
322 Contemporary Scientific Realism
Beside these cases, one can add to this taxonomy the use of infinite limits as
abstractions, which does not directly follow from such a scheme: in fact, as we
will see in greater detail in Section 14.6, these are cases in which the variable
n does not represent any physical parameter of the target system. For example,
the parameter can represent the number of times that one has to apply a trans-
formation that successively coarse grains the system. Here the appeal to an in-
finite limit is merely instrumental in that it allows us to find fixed points in a
topological space.
In the rest of the paper, we evaluate how each type of infinite limit fares against
the so-called Paradox of Infinite Limits and its consequences for scientific re-
alism, which we present next.
• (I) Finiteness of Real Systems: If Sn represents a real system, then the variable
n corresponding to some physical parameter cannot take on infinite values.
• (II) Indispensability of the Limit System: The explanation of the phenom-
enon P can only be given by means of claims about an infinite system S∞
constructed in the limit n → ∞.
• (III) Enhanced Indispensability Argument (EIA): If a claim plays an indis-
pensable role in the explanation of a phenomenon P we ought to believe in
its existence.
The ostensive problem can be further articulated and made more precise
when dealing with the description of particular phenomena, as we will see in
greater detail for the Paradox of Reversible Processes and the Paradox of Phase
Transitions in Sections 14.4 and 14.5, respectively. But the tension between these
statements captures the core idea of the paradox. Intuitively, the fact that real
systems are finite in the sense expressed by statement (I) can be understood as
The Paradox of Infinite Limits 323
7 See (Colyvan 2001), (Baker 2005; Baker 2009) and (Baron 2016) for a careful discussion of the
indispensability arguments.
324 Contemporary Scientific Realism
theories are (at least approximately) true because otherwise their success would
be a miracle. Underlying this argument is the inference-to-the-best-explanation
(Boyd 1983). By using a similar reasoning, if the best explanation available of a
certain phenomenon P involves assuming an infinite idealization and there is no
known way to de-idealize the model without losing explanatory power, then we
should commit to the existence of such an idealization for the same reasons that
ground scientific realism. This a particularly important for Explanationanists
(Kitcher 1993; Psillos 1999), since they recommend holding a realist attitude to-
ward entities that are indispensable to the explanation of certain phenomena.8
Therefore, we claim, a more satisfactory solution to the Paradox of Infinite
Limits comes from the rejection of statement (II), rather than statement
(III). To make our point, we need a criterion of empirical correctness that we
introduce next.
Before spelling out our intended criterion for correctness, it should be stressed
that the concept of explanatory indispensability heavily depends on what one
means by scientific explanation, which is a huge and much debated topic in
philosophy of science.9 For our purposes, we restrict ourselves to what we con-
ceive as a minimal requirement for a good explanation of some physical phe-
nomenon P, namely that one recovers empirically correct results. Arguably,
this yields a necessary condition in that, if an account fails to agree, at least
approximatively, with the observed data then it cannot be said to explain P
(whether it yields also a sufficient condition is less straightforward to estab-
lish, but our argument here does not really need that much). The relevant data
are given by the values of physical quantities of interest, like energy, position,
momentum, spin, etc., corresponding to properties of the physical system in-
volved in the phenomenon. The proposed condition for empirical correctness
is as follows:
Let us explain the content of this definition. Note that one cannot expect
that empirical data can be sharply determined, in that observations always
involve some margin of error. Hence, our condition is formulated in such a
way to allow for degrees of inexactness: in fact, it only requires that the value
of f be approximately equal, rather than exactly equal, to the data D, that
is v( fn )≈ D for some given n. Deciding the degrees of inexactness that one
may tolerate is ultimately a pragmatic matter, and that is why the number ε
is left unfixed in the definition: yet, by keeping ε very small, one ensures that
the results produced by system Sn are empirically correct to a pretty good
approximation.
Infinite limits can be used in the explanation of the physical phenomenon
P when empirical correctness is satisfied by the infinite system S∞ for some
physically relevant function f. Accordingly, one has v( f s∞ ) ≈ D, meaning that
the value at the limit, namely (ii) the function fS associated with the limit
∞
system S∞, is the same, or at least approximatively the same, as the observed
data D. More to the point, infinite idealizations are regarded as indispensable
to the explanation of P insomuch as no finite system Sn can yield empirically
correct results for f. That is, even though the parameter n grows while still
remaining less than ∞, the values taken on by fn fail to provide an approxi-
mation of the observed data for some sufficiently small ε. To put it in tech-
nical terms, this alleged indispensability of the infinite idealization typically
manifests itself in the cases in which the limit is singular, and hence (i) the
limit limn→∞ v ( fn ) of the sequence of values does not coincide with (ii) the
value v(fS ) yielded by the limit system. (Butterfield (2011) actually makes
∞
a similar point when discussing the emergence of certain properties at the
limit).
Armed with this stated condition of empirical correctness, we now proceed
to show how infinite idealizations can be actually dispensed to the explanation
of physical phenomena that seem to require one to take the limit n → ∞, thereby
allowing one to give up statement (II) in the Paradox of Infinite Limits. In fact,
such a condition grounds the use of limits as approximations of the properties
of real target systems. As we argue later, that is the basis to demonstrate that
cases of idealizations yielding approximations as well as the more controversial
cases of essential idealizations do not pose a threat to scientific realism.
The Paradox of Infinite Limits 327
may still wonder why one should appeal to a mathematical limit in the first place.
In other words, one may ask, how can one justify the use of an infinite limit to
describe the target system? In order to answer this question, Butterfield (2011)
puts forward what he calls a Straightforward Justification, which is based on two
features that limits enjoy, namely mathematical convenience and empirical ad-
equacy. Regarding the first feature, taking the limit often enables one to ignore
some degrees of freedom that complicate the calculations, and so infinite sys-
tems, when they exist, turn out to be more tractable than finite systems for which
the parameter n is very large.10 As for empirical adequacy, arguably that assures
that the values obtained at the limit are close enough to the real values. Hence,
while the empirically correct results are given by the values of the function
computed for the actual n0, namely the values of the real target system, taking
the limit for n growing to infinity still proves adequate within some acceptable
margin of approximation. Thus, based upon these two desirable features pro-
posed by Butterfield, one has both pragmatic and empirical reasons to justify
the use of the infinite limit. That reinforces our claim that one does not need to
commit to the reality of the infinite idealization S∞, contrary to what is entailed
by statements (II) and (III) in the Paradox of Infinite Limits.
The other scenario, namely the case in which taking the infinite limit n → ∞
yields an approximation without idealization, is more straightforward to deal
with. On the basis of our taxonomy, it occurs when the limit (i) rather than (ii)
yields empirically correct results or (ii) is not well defined at all for some phys-
ically significant function f: as a consequence, the infinite system S∞ does not
satisfy empirical correctness. It follows that it cannot be even used to explain the
relevant phenomenon. Indeed, an approximation without idealization should
be understood as a misrepresentation of the target system whose properties are
given an inexact description in terms of the limit f∞ of the sequence {f1, f2, ...., fn}
of functions representing the relevant physical quantities, rather than in terms
of the properties of a fictional limit system. Accordingly, it would not make
any sense to reify the infinite idealization, which means that statement (II) in
the Paradox of Infinite Limits does not hold, and hence there cannot arise any
ensuing problem for scientific realism. Very recently, Norton (2012) recognized
that sometimes infinite limits are used precisely as approximations without
idealizations. For him, there are two sufficient conditions under which an ap-
proximation cannot be promoted to the status of idealization, which are closely
related to the formal conditions we stated earlier: that is, (1) the limit system
10 In addition, Palacios (2018) points out that it does not suffice to show that the behavior that
arises in the limit arises already for a large value of the parameter n, but one also needs to show that it
arises for realistic values of n0. This latter requirement ought to be empirically grounded and is meant
to assure that the lower level theory that results from a limiting operation is empirically correct and
therefore capable of describing realistic behavior.
The Paradox of Infinite Limits 329
does not exist in the sense that it is paradoxical, and (2) the limit system has
properties that are inadequate for the idealization in that they do not match with
the properties of the finite target system. The following quotation illustrates his
proposed criterion for the failure of idealizations:
Norton’s goal is to show that, on the basis of these two conditions, some infinite
limits that are regarded as idealizations in the literature do not deserve to be ele-
vated to such a status and should in fact be demoted to mere approximations, as
it happens in many examples in which mathematical limits are employed in ther-
modynamics and statistical mechanics.11 In his view, an example in which an in-
finite limit does not give rise to a suitable idealization due to condition (2) is the
account of phase transitions in the thermodynamical limit within classical sta-
tistical mechanics that we will address in the next section. Another controversial
case that Norton claims to be an instance of failure of idealization due to condi-
tion (1) is the paradox of thermodynamically reversible processes arising in the
infinite-time limit, which has recently drawn attention in the philosophical lit-
erature (Norton 2014; Norton 2016; Valente 2019). Let us discuss this case next.
Reversible processes are conceived as sequences of equilibrium states through
which a thermodynamical system progressively passes in the course of time.
They are typically constructed by means of the infinite-time limit, and as such
they are interpreted as processes taking place infinitely slowly. They owe their
name to the fact that, ideally, they could be traversed in both directions of time.
However, reversible processes are fictitious processes: indeed, thermodynam-
ical processes just take a finite amount of time t to complete, even if they pro-
ceed quasi-statically, and they can only unfold in one temporal direction but not
in the reversed one. So, at best, taking the infinite-time limit t → ∞ can yield
approximations of real processes. The question that interests us is whether or not
the limit processes can also be intended as infinite idealizations. According to
Norton, reversible processes do not deserve to be elevated to such a status since
they display contradictory properties. More to the point, he submits that they are
11 To be sure, Norton concedes that, under certain circumstances, approximations can actually
be promoted to idealizations. For instance, in his (2014) paper Norton presents the case of the law of
ideal gases in thermodynamics as an example of an approximation which gives rise to an ideal system
that can as well serve as an idealization. In fact, an ideal gas system is defined as constituted by a large
number of non-interacting, spatially localized particles, and as such it can describe, at least approxi-
mately, the behavior of very rarefied real gases under appropriate circumstances.
330 Contemporary Scientific Realism
plagued by a paradox: for, on the one hand, (I) thermodynamical processes are
driven by a non-equilibrium imbalance of driving forces, which is necessary in
order to enact the transition of the system from one state to another; on the other
hand, though, (II) a reversible process amounts to a sequence of equilibrium
states, for which there must be no imbalance of driving forces. By connecting the
driving forces enacting the process with its duration in time, Norton’s Paradox
of Reversible Processes becomes somewhat similar to the general Paradox of
Infinite Limits. For instance, when a fixed quantity Q of heat is exchanged be-
tween two bodies in thermal contact, the driving forces are given by the tempera-
ture difference ∆T between the bodies. So, if the latter is equal to zero, the process
of heat transfer cannot take place. Now, by assuming Fourier law Q = − k ∆T ∆t
(with k being the heat transfer coefficient), the amount of heat that passes from
the hotter to the colder body is also proportional to the time ∆t that the process
would take to complete. As a consequence, if one tries to construct a reversible
process by letting time go to infinity, the temperature difference must vanish, as
prescribed by statement (II), but then no heat would be exchanged between the
two bodies, thereby raising a contradiction with statement (I). Based on such a
paradox, Norton argues that reversible processes fail to be idealizations.
However, Valente (2019) objected to this conclusion, by offering a way to cir-
cumvent the alleged contradiction. As he pointed out, Norton’s paradox arises
due to the misconception that reversible processes should be treated as ac-
tual thermodynamical processes. Instead, they ought to be regarded as mere
mathematical constructions that are introduced to apply infinitesimal calculus
to thermodynamics. In fact, formally, they simply correspond to continuous
curves in the space of equilibrium states of a thermal system: along these curves
one can calculate the exact integrals of state-functions, such as entropy, and
then compute the values of other physical quantities of interest, like heat and
work. For this reason, as it was observed by the mathematical physicist Tatjana
Ehrenfest-Afanassjewa (1956) in her book on the foundations of thermody-
namics, reversible processes should better be called “quasi-processes,” so as to
avoid the misunderstanding that they would correspond to actual processes,
which could even be reversed. Accordingly, while statement (II) holds by def-
inition, it would be a mistake to ascribe to reversible processes the properties
required by statement (I): that is, contrary to what happens during real ther-
modynamical processes, there cannot be any non-equilibrium imbalance of
forces moving a thermal system from one state of equilibrium to another along
the continuous curve representing a quasi-process. As a result, the ostensive
paradox disappears, and hence one may as well regard reversible processes as
idealizations, in accordance with Norton’s own criterion. Notice that here, dif-
ferently from the Paradox of Infinite Limits, in order to disarm the apparent
The Paradox of Infinite Limits 331
contradiction we need to drop statement (I) rather than statement (II): the
reason is that what is to be explained in this case is just the idealized object
constructed in the limit. The connection with real thermodynamical processes
can then be given thanks to the notion of approximations. To illustrate this idea,
let us again refer to Ehrenfest-Afanassjewa’s own work. After warning that,
strictly speaking, infinitely slow processes cannot exist (cfr. p. 11), she went on
to argue as follows:
E = γ m0 c 2
p = γ m0 v
Arguably, when the Lorentz factor goes to 1, one can recover the values of the
classical counterparts of these quantities defined, respectively, as E = m0 c 2 and
p = m0 v . Since the Lorentz factor takes on the form
332 Contemporary Scientific Realism
1
γ (v ) = ,
1 − (v / c )
2
2
v
one can make this expression go to 1 in the limit → 0, where c is the speed of
c
light and v the velocity of a moving body in a given inertial frame. There are dif-
ferent possible ways of interpreting this limit. First, one can interpret it as taking
the limit of a sequence of systems in which the velocity v goes to zero. The limit
will then give rise to an idealization, namely to a fictional system in which the
value of the velocity for all bodies is null in the given inertial frame. Nonetheless,
understanding the limit in this sense has the problem that, e.g. the momentum
p, which depends exactly on the velocity of the bodies, will be always zero. This
naturally means that the quantities defined in the limit system will not provide
a good approximation for the behavior of restless objects, in which momentum
is different from zero. Therefore, the purported idealization would not serve to
describe the behavior of moving objects. Likewise, a similar problem arises if
one interprets the Newtonian limit as taking the limit of a sequence of systems in
which c goes to infinity. In this case, the limit system will be an imaginary system
in which the speed of light is infinite. In this system, quantities such as the ki-
netic energy that are defined as a function of the speed of light c will also go to
infinity, thereby failing to give an approximation of the actual kinetic energy of
real moving bodies, which is instead finite.
On the contrary, a much more suitable interpretation of the Newtonian limit
2
v
c → 0 should not be given in terms of fictional systems, but rather as a mere
approximation of the velocity of bodies that are moving slowly compared to the
speed of light. This can be seen more clearly by noticing that the Lorentz factor γ
can be expanded into a Taylor series:
2n 2 4 6
1 ∞
v n 2k − 1 1 v 3 v 5 v
γ (v ) = = ∑ ∏ = 1 + 2 c + 8 c + 16 c + ...
1 − (v / c )
2
n=0 c k = 1 2k
Accordingly, if one considers only the first term of this Taylor expansion, one will
recover the exact values of the classical quantities of interest, and this will consti-
tute just an approximation of the velocity v of the objects for which v << c. Taking
more terms into account will give results that are more accurate from the rela-
tivistic point of view, but that depart from the classical values. It is important to
The Paradox of Infinite Limits 333
The claim that there are essential idealizations has been put forward by some
authors, especially Batterman (e.g. 2001, 2005, 2011), in reference to singular
limits, whereby empirically adequate results are obtained just for the infinite
system and not for finite systems. In terms of Butterfield’s distinction, this appar-
ently mysterious case occurs when, given a function f representing some physical
quantity, (i) the limit of the sequence of values and (ii) what is true at the limit for
n → ∞ differ from each other, but only (ii) is empirically adequate (these cases
correspond to essential idealizations according to the taxonomy presented in
Section 14.3.1). So, if the variable n growing to infinity represents some physical
parameter, it means that the limit system would not yield an approximation of the
relevant property of the target systems, which are instead finite. An infinite ide-
alization is then deemed as essential in that, arguably, it is only the limit system
S∞ that allows one to explain the physical phenomenon for which the function
f is relevant. So, if one further assumes EIA, then one falls into the Paradox of
Infinite Limits, hence threatening scientific realism. In this section we present a
334 Contemporary Scientific Realism
possible solution to this paradox that casts doubt on the “essential character” of
the idealization, which is partially based on Butterfield’s (2011) results.
Phase transitions are sudden transformations of a thermal system from one state
into another, occurring for instance when some material changes from solid to
liquid state due to an increase of temperature. That is a much debated case study
where, according to indispensabilists like Batterman, there arises an essential
idealization when taking the so-called thermodynamic limit. Informally, the ar-
gument goes as follows. According to thermodynamics, which deals with the be-
havior of thermal systems from a macroscopic point of view, phase transitions
occur when the function representing the derivatives of the free energy is discon-
tinuous. Allegedly, such a discontinuity matches with the observed data, since
sudden transitions from one phase to another appear to take place abruptly. Yet,
when attempting to recover the same phenomena within statistical mechanics,
which describes thermal systems at the microscopic level as being composed by
a very large number N of molecules, one faces a technical impossibility: that is,
if N is finite, the function representing the derivative of the free energy remains
continuous no matter how large N is. Instead, one can recover the sought-after
discontinuity by taking the thermodynamical limit, which prescribes that both
the number of molecules N and the volume V of the system go to infinity while
keeping its density fixed (Goldenfeld 1992). This motivates the persistent atti-
tude among physics, such as Kadanoff (2009), to emphasize the importance of
infinite systems to explain the phenomenon of phase transitions:
Phase transitions cannot occur in finite systems, phase transitions are solely a
property of infinite systems. (p. 7)
transitions. Let us stress that this is just a special case of the general Paradox of
Infinite Limits. In fact, Callender’s conditions can be explicitly connected with
the three statements presented in Section 14.3.1. For, condition (2) is tanta-
mount to statement (i) expressing the basic desideratum for scientific realism
that real systems are finite, whereas condition (4) assures that the phenomenon
to be explained, namely phase transitions, occurs for such systems, which is
captured by statement (ii). On the other hand, condition (3) requires empirical
correctness to be satisfied for the partition function being discontinuous: yet, as
discussed earlier, if one tries to explain the phenomenon within the framework
of statistical mechanics, in accordance with condition (1), then one is bound
to reify the thermodynamical limit, which gives rise to an infinite idealization
proving indispensable, exactly as our statement (iii) dictates. Other formulations
of the paradox have been put forward in the literature, which differ from the one
just presented only regarding how the allegedly conflicting conditions are stated,
but they all basically agree on the content of the problem. There is also a variety
of proposed solutions. Callender himself suggests that one should not take “too
seriously” the theory that prompts one to introduce a discontinuity in the de-
scription of phase transitions, namely thermodynamics: accordingly, condition
(3) can be dropped, which means that one does not need to appeal to the thermo-
dynamical limit and therefore denies statement (ii) of our formulation of the par-
adox. Shech (2013), on the other hand, suggests that one can resolve the paradox
by noticing that the terms related to the existence of infinite systems do not refer
to concrete physical systems but just to mathematical objects, which do not carry
any ontological significance. However, as he correctly recognizes, this cannot be
the attitude of scientific realists, who are interested in our abstract scientific ac-
counts getting something right about the real world. Alternatively, Liu (2019)
suggests that the alleged contradiction disappears if one adopts a form of con-
textual realism, whereby realist claims should be evaluated relative to anchoring
assumptions formulated within the background theory: in particular, the claim
that the number of molecules grows to infinity can be regarded as true in the
context of a microscopic theory holding that condensed matter is continuous.
However, the jury is still out as to whether these proposals effectively dissolve the
paradox of phase transitions.
Instead, our preferred strategy to elude the paradox and thus salvage sci-
entific realism goes along the lines of Butterfield’s (2011) own dissolution
of the mystery of singular limits, which is endorsed, at least in connection
with classical phase transitions, by Menon and Callender (2013) as well as by
Norton (2014) and Palacios (2019). It develops into two steps: first of all, one
ought to make a careful choice of the physical quantities to work with, so as
to avoid those quantities for which the infinite idealization appears essential
336 Contemporary Scientific Realism
in that the limit is singular; then, one employs the notion of approximation
to show that for the selected quantities empirically correct results obtain on
the way to the limit, without having to commit to the reality of the infinite
system. Thus, the underlying idea of the proposed strategy is that of focusing
on just the properties that are relevant for the behavior that we intend to
explain, instead of requiring that all properties of the finite target system
extend smoothly to the limit system. More to the point, Butterfield observes
that the alleged mystery of singular limits is simply a consequence of looking
at functions that do not give information regarding the behavior of real sys-
tems as N increases toward infinity. For instance, if one restricts one’s atten-
tion only to the fact that some function f be discontinuous, like in the case
of the derivative of the free energy phase, one would lose insight of what
happens for very large but finite N, since the expected discontinuity obtains
only at the limit “ N = ∞ .” To the contrary, one ought to turn one’s attention
to physical quantities represented by different functions, so that the corre-
sponding properties of the actual target system SN are approximated by the
0
properties of S∞. In other words, even though it appears to be an essential
idealization for some quantities, the limit system is an idealization yielding
approximations for other properties that one regards as physically salient for
the phenomenon to be explained. In this way the desired behavior is recov-
ered on the way to the limit, and hence one can dispense from the infinite
idealization, by demonstrating that this apparent “essential idealization” is a
case of idealizations yielding approximations.
In order to make this proposal more concrete, let us look at a specific ex-
ample of first-order phase transitions, that is a ferromagnet at sub-critical
temperature. The Ising model portrays a ferromagnet as a chain of N spins,
wherein a physical quantity called magnetization is represented as a func-
tion of the applied magnetic field. At very low temperatures, there are two
possible phases available: a state where all spins are oriented in the up-
direction, for which the magnetization takes on the value +1; and a state
where all spins are oriented in the down-direction, for which the magneti-
zation takes on the value −1. When the sub-critical temperature is reached,
even if the applied magnetic field is null, one observes an abrupt flip from
one phase to the other. That is formally captured by the magnetization func-
tion being discontinuous. Arguably, one can describe this phenomenon just
in case one takes the thermodynamical limit whereby the number N of spins
grows to infinity. Butterfield’s proposed strategy can then be illustrated by
means of a toy model. He defines a sequence of real-valued functions {g1, g2,
..., gN} with argument x belonging to the real numbers , which take on the
following form:
The Paradox of Infinite Limits 337
1
−1 iff x ≤ − N
1 1
g N ( x ) : = Nx iff − ≤ x ≤
N N
+1 iff x ≥ 1
N
All such functions are continuous, in that they remain equal to –1 until x = –1/N
and then they start to grow linearly with gradient N up to the value +1 for x = 1/N,
after which they constant again. However, when one takes the limit for N → ∞, the
resulting function is no more continuous: in fact, the limit is given by
−1 iff x < 0
g ∞ ( x ) = 0 iff x = 0
+1 iff x > 0
1 iff g N continuous
f N :=
0 iff g N discontinuous
However, the mystery can be explained away if one looks at the beha-
vior of the functions gNs instead of the functions fN’s. In fact, as N grows, the
functions gN become more and more similar to the step function g∞, even
though, contrary to the latter, they remain continuous. Specifically, when N0
is extremely large, for most points x one has g N 0 (x ) = g ∞ (x ) , which is true in
particular when the argument is x = 0; furthermore, whenever the values of
these functions are not strictly equal, namely around the singular point x = 0,
they will still be close to each other, so that g N 0 (x ) ≈ g ∞ (x ) . Therefore, the
empirically correct values of the magnetization function can be recovered
by the properties of the real system SN without having to resort to the infi-
0
nite system S∞. In this way, the values obtained when taking the limit N → ∞
just yield approximations of the relevant property of the real target system.
More importantly, as a theoretical analysis also shows for realistic values of
N, the gradient in the derivatives of the free energy is sufficiently steep that
the difference in the limit values of the thermodynamic quantities as N → ∞
and realistic systems with finite N0 becomes negligibly small (Schmelzer and
Ulbricht 1987; Fisher and Berker 1982).
A straightforward justification similar to the one obtained in cases
of idealizations yielding approximations can now be given also for infi-
nite limits that appear to give rise to essential idealizations. In fact, the use
of the latter is justified based on the two desirable properties of mathemat-
ical convenience and empirical adequacy. Here, though, we would like
to emphasize a point that Butterfield does not develop, namely the fact
that the choice of a certain topology over the other determines different
degrees of empirical adequacy. For this purpose, let us consider the func-
tion fn indexed by the parameter n that maps the independent variable x
onto the real numbers: accordingly, the notation fn(x) indicates the value
that the function takes on for each x in the domain, where in concrete phys-
ical cases the variable x would represent the possible states of the system
under investigation. A natural topology is induced by the following type of
convergence:
Translated into our framework, the fact that the sequence of functions { fn (x )}n
converges pointwise to the function f ∞ (x ) means that there is a real finite
system Sn for which the value of the relevant function is approximated by the
0
The Paradox of Infinite Limits 339
value of the limit function for a given state x, that is fn 0 (x ) ≈ f ∞ (x ) . But one may
as well adopt a different type of convergence, that is:
Since in this case n0 depends only on ε, and not even on the variable x like in
the case of pointwise convergence, uniform convergence proves stronger
than the latter (in fact, it is strictly stronger in that one can show by means of
counterexamples that pointwise convergence does not imply uniform conver-
gence). Indeed, here one can choose a natural number n0 for which the sought-
after approximation fn 0 (x ) ≈ f ∞ (x ) holds for all the possible states x. To the
contrary, pointwise convergence allows for exceptions, in the sense that once n0 is
fixed together with the margin of approximation determined by ε > 0 there may
be some state x for which | fn ( x ) − f ∞ ( x )|< ε does not hold for any n > n0 . The up-
shot of this analysis is that the standard hierarchy of convergence conditions for
functions representing physical quantities entails different degrees of accuracy
up to which empirical accuracy is satisfied. Hence, just as the precise notion of
approximation rests on what one means by the expression “sufficiently close,”
Butterfield’s straightforward justification of infinite limits ultimately depends on
the topology under which one takes the limits. In fact, in the explanation of first
order phase transitions the criterion of empirical adequacy is satisfied only in
its weak degrees of accuracy: for, it is just when one chooses the topology in-
duced by pointwise convergence, and not by the stronger uniform convergence,
that a sequence of continuous functions approaches a discontinuous function in
the limit.
This analysis shows that the property of empirical adequacy is sensitive
to the choice of topology. Thus, the extent to which one satisfies empirical
correctness, namely the condition that the value of a function representing a
given physical quantity is approximately equal to the empirical data, depends
on how the limit is taken. On this point it is worth making an important clari-
fication: the issue whether the use of the limit is justified and the issue whether
the use of the limit raises a threat to scientific realism, even though they have
a common root, they should be kept separate. The common root is that the
problem that, if the variable n represents a physical parameter in that, then
when n becomes infinite the limit system would not be real. The former issue
asks the question: idealizations arising in the limit in physical applications?
The second issue, instead, arises because it appears as if one violates statement
(I) of the Paradox of Infinite Limits, which is a basic desideratum for scientific
340 Contemporary Scientific Realism
Let us move on to address the question whether the purported strategy to cope
with the indispensability of infinite limits can be generalized to other cases where
there appear essential idealizations. Butterfield (2011) conjectures that the solu-
tion of the Paradox of Phase Transitions proposed in the classical context holds
generally. However, matters are less straightforward in other cases such as in the
12 For a complete investigation of this issue one should actually discuss in much greater details
than we can here the sense in which the limit system can be said to represent the target system. To
this extent, Shech (2014) put forward a distinction between epistemologically and ontologically
faithful representations. However, it goes beyond the scope of the present paper to survey the notion
of representation.
The Paradox of Infinite Limits 341
jury is still out as to whether or not phase transitions in quantum statistical me-
chanics constitute a case of essential idealizations where one can provide an ex-
planation of the phenomenon without committing to the reality of the infinite
system constructed in the thermodynamical limit.13
Be that as it may, there is a further point that is worth emphasizing re-
garding the general strategy to cope with other cases of essential idealizations,
besides classical phase transitions. As pointed out by Palacios (2018), it does
not suffice to demonstrate that approximately the same behavior that occurs
in the limit, also occurs “on the way to the limit,” but in addition we need
to demonstrate that it arises for realistic values of the parameter n that goes
to infinity. This latter condition is important because there are cases, such as
the ergodic approach to equilibrium, in which the expected behavior arises
for finite but unrealistic values of the parameter that goes to infinity. In more
detail, in the ergodic theory of equilibrium, one takes the infinite-time limit
t → ∞ in order to assure that phase-averages and time-averages coincide.
Even in simple examples like a small sample of diluted hydrogen, though, one
can estimate that the desired behavior can occur for times t0 that are finite but
unimaginably longer than the age of universe. In such cases, even if we can ac-
tually demonstrate that the behavior arises for finite values of the parameter,
we would not have succeeded in demonstrating the empirical adequacy of the
theory.
To conclude, the analysis we have developed in the present section indicates
that the challenge posed by essential idealizations to scientific realism ought to
be resolved on a case-by-case basis, and that ultimately calls for empirical consid-
erations. In general, the strategy to dispense with the infinite-limit system in the
explanation of a certain phenomenon involves a suitable selection of the phys-
ical quantities to focus on, so as to yield approximations of the relevant prop-
erties of the target system already “on the way to the limit,” where the degrees
of inexactness that one may accept depends on the particular situation at stake
and the choice of an adequate topology. Accordingly, as the example of classical
phase transitions shows, one can assuage worries concerning scientific realism.
In fact, if the idealization is not genuinely essential, one does not need to assume
statement (II) of the Paradox of Infinite Limits. Moreover, the limit values of the
relevant physical quantities can still give good, albeit inexact, descriptions of the
properties of realistic systems, which are supposed to be finite in accordance with
statement (I), and so in order to evade the paradox we do not even need to cast
doubts upon the validity of the enhanced indispensability argument contained
in statement (III). The Paradox of Infinite Limits can thus be avoided in spite of
13 Another example of apparent essential idealizations are continuous phase transitions. We will
For the sake of completeness, in this last section we address a different role played
by mathematical limits that has been much less discussed in the philosoph-
ical literature than approximations and idealizations, namely the use of limits
as abstractions. We mentioned earlier that the thermodynamical limit, in which
the number of particles N as well as the volume V go to infinity, can be used
to recover the quantities that successfully describe phase transitions in thermo-
dynamics. Another important role of the thermodynamic limit is to enable the
removal of irrelevant contributions such as “surface” and “edge” effects. More
to the point, any finite lattice system will include contributions to the partition
function coming from the edges and surfaces, which may be considerably dif-
ferent from those coming from the center of the sample. Taking the thermody-
namical limit allows one to treat the system as a bulk, leaving out surface effects
that are mostly irrelevant for the behavior. In this sense, the thermodynamical
limit enables one to abstract away details that do not make a difference for the
phenomenon under investigation (see also Mainwood 2006; Butterfield 2011;
Jones 2006).
A more interesting case of infinite limits used as abstractions is given in
the context of continuous phase transitions. In contrast to first-order phase
transitions that involve discontinuities in the derivatives of the free energy, in the
case of continuous phase transitions there are no discontinuities but rather there
are divergences in the response functions (e.g. specific heat, susceptibility for a
magnet, compressibility for a fluid). An example of a continuous phase transition
is the transition in magnetic materials from the phase featuring spontaneous
magnetization—the ferromagnetic phase—to the phase where the spontaneous
magnetization vanishes—the paramagnetic phase. Continuous phase transitions
are also characterized by the divergence of a quantity called the correlation
length ξ, which measures the distance over which the particles are correlated.
A typical way of dealing with these long correlations is by introducing renor-
malization group methods, which are mathematical and conceptual tools that
consist in defining a transformation that successively coarse-grains the effective
344 Contemporary Scientific Realism
degrees of freedom while keeping the partition function and the free energy (ap-
proximately) invariant. Palacios (2019) pointed out that such a process can be
interpreted as assuming an infinite limit, in which the number of iterations n of
the renormalization group transformation goes to infinity. An important aspect
of the infinite limit for n → ∞ is that, at each step of the sequence, irrelevant cou-
pling constants are abstracted away, so as to retain only factors that are relevant
for the behavior under description. That follows because, in this process, the par-
tition function and the free energy that determine the behavior at a phase tran-
sition remain (approximately) invariant. After an infinite number of iterations,
the sequence of systems may converge toward non-trivial fixed points, where the
values of the coupling constants no longer change by applying the transforma-
tion. Note that the limit for the number n of iterations going to infinite does not
represent any physical parameter of the system under description, but rather the
number of transformations that one needs to apply by means of the renormaliza-
tion group procedure.
The beauty of renormalization group methods is that linearizing around fixed
points allows one to calculate the critical exponents of the power laws that de-
termine the behavior close to the transition. Furthermore, it allows one to give
an account of universality, namely the remarkable fact that physical systems as
heterogeneous as fluids and magnets exhibit the same behavior near a phase
transition (details in Goldenfeld 1992). There is good reason to consider the in-
finite iteration limit n → ∞ as an abstraction. In fact, in Section 14.2, we de-
fined abstraction as a process that consists in leaving some factors out without
misrepresenting the properties of the original system. In fact, when applying a
renormalization group transformation to the description of a given system, one
does exactly this: that is, at each stage n, one leaves out irrelevant details while re-
taining those factors that make a difference for the phenomenon to be explained.
To be sure, in doing that, one might have to use approximations; yet, one does
not introduce any idealization. Moreover, it is also important to point out that
the renormalization group process has the same epistemic role attributed to
abstractions, namely that it enables us to give an explanation of the common
behavior generated by systems that are diverse from each other at a fine-grained
level, thereby accounting for universality.
In recent philosophical literature, the use of renormalization group methods
has often been considered as an example of essential idealizations, e.g. Batterman
(2011), Batterman (2017) and Morrison (2012). Indeed, Batterman says in a
footnote:
It seems to me that if one is going to hold that the use of the infinite limits is a
convenience, then one should be able to say how (even if inconveniently) one
might go about finding a fixed point of the RG transformation without infinite
The Paradox of Infinite Limits 345
iterations. I have not seen any sketch of how this is to be done. The point is that
the fixed point, as just noted, determines the behavior of the flow in its neigh-
borhood. If we want to explain the universal behavior of finite but large systems
using the RG, then we need to find a fixed point and, to my knowledge, this
requires an infinite system. (2017, p. 571)
number of iterations is not a physical parameter, hence in principle one may not
violate statement (I) even if one takes the limit n → ∞. However, since this limit
does not commute with the thermodynamical limit, taking the former implies
taking the latter limit too, and thus if n grows to infinity so does the number N
of particles in the gas, thereby violating statement (I). Instead, the resolution to
the paradox comes once again by rejecting statement (II), which says that the
explanation of the phenomenon P can only be given by means of claims about an
infinite system constructed in the limit n → ∞. In fact, the procedure outlined
earlier shows how the two limits N → ∞ and n → ∞ can be dispensed by means
of the notion of approximation: accordingly, one can satisfy the condition of em-
pirical correctness without committing to the limit system S∞. In light of this
analysis, we conclude that one has good reason to believe that renormalization
group methods, at least in the example we considered, do not pose by themselves
any challenge for a form of scientific realism allowing for approximate truth.
14.7 Conclusion
In this paper, we addressed the issue of whether or not the use of infinite limits
in physics raises a problem of compatibility with scientific realism. For this pur-
pose, we surveyed various physical examples where infinite limits are invoked
in order to describe real target systems, so as to offer a taxonomy of the dif-
ferent uses of infinite limits (Section 14.3). In particular, we distinguished be-
tween approximations that do not constitute idealizations, idealizations yielding
approximations, essential idealizations, and abstractions, which we then
discussed in greater details in the subsequent sections. We argued in Section 14.5
that a challenge for scientific realism arises just when infinite limits are intended
as essential idealizations, due to the fact that it appears as if empirically correct
results can be recovered only in the limit. However, if one commits to the limit
system in that it seems indispensable, one runs against the Paradox of Infinite
Limits since the limit system is infinite whereas real systems are necessarily fi-
nite. We then went on to suggest how the ensuing worries for scientific realism
can be assuaged in concrete examples, e.g. in the much debated case of classical
phase transitions. The strategy to do so is to show, without referring to the in-
finite system, that the limit values of some physical quantities of interest yield
approximations of the values obtained for realistic systems. That is of course
compatible with a form of scientific realism that allows for theories to be just suf-
ficiently close to the truth. In the same vein, as we explained in Section 14.4 and
Section 14.6 by means of physical examples such as that of thermodynamically
reversible processes and that of continuous phase transitions, understanding
The Paradox of Infinite Limits 347
Acknowledgments
The authors thank two anonymous referees for helpful and constructive
comments. Giovanni Valente acknowledges financial support from the Italian
Ministry of Education, Universities and Research (MIUR) through the grant
n. 201743F9YE (PRIN 2017 project “From models to decisions”).
References
Aronson, J. (1990). Verisimilitude and types of hierarchies. Philosophical Topics 18, 5–28.
Baker, A. (2005). Are there genuine mathematical explanations of physical phenomena?
Mind 114(454), 223–238.
Baker, A. (2009). Mathematical explanation in science. British Journal for the Philosophy
of Science 60(3), 611–633.
Baron, S. (2016). The explanatory dispensability of idealizations. Synthese 193(2),
365–386.
Baron, S. (2019). Infinite lies and explanatory ties: Idealization in phase transitions.
Synthese 196, 1939–1961.
Batterman, R. W. (2001). The devil in the details: Asymptotic reasoning in explanation, re-
duction, and emergence. New York: Oxford University Press.
Batterman, R. W. (2005). Critical phenomena and breaking drops: Infinite idealizations
in physics. Studies in History and Philosophy of Science Part B: Studies in History and
Modern Physics 36(2), 225–244.
Batterman, R. W. (2011). Emergence, singularities and symmetry breaking. Foundations
of Physics 41(6), 1031–1050.
Batterman, R. W. (2017). Philosophical implications of Kadanoff ’s work on the renormal-
ization group. Journal of Statistical Physics 167(3–4), 559–574.
Boyd, R. (1983). On the current status of the issue of scientific realism. Erkenntnis
19(1–3), 45–90.
Butterfield, J. (2011). Less is different: Emergence and reduction reconciled. Foundations
of Physics 41(6), 1065–1135.
Callender, C. (2001). Taking thermodynamics too seriously. Studies in History and
Philosophy of Science Part B: Studies in History and Philosophy of Modern Physics 32(4),
539–553.
Cartwright, N. (1994). Nature’s capacities and their measurement. New York: Oxford
University Press.
Chakravartty, A. (2010). Truth and Representation in Science: Two Inspirations from
Art. In R. Frigg and M. Hunter (Eds.), Beyond Mimesis and Convention: Representation
in Art and Science, pp. 33– 50, Boston Studies in the Philosophy of Science.
Dordrecht: Springer.
348 Contemporary Scientific Realism
Department of Philosophy
University of Miami
[email protected]
It is the sense in which Tycho and Kepler do not observe the same thing which
must be grasped if one is to understand disagreements within microphysics.
Fundamental physics is primarily a search for intelligibility—it is a philosophy
of matter. Only secondarily is it a search for objects and facts (though the two
endeavors are as hand and glove). Microphysicists seek new modes of concep-
tual organization. If that can be done the finding of new entities will follow.
Norwood Russell Hanson (1965/1958, pp. 18–19)
Scientific realism is commonly understood as the idea that our best scientific
theories, read literally as descriptions of a mind-independent world, afford
knowledge of their subject matters independently of the question of whether
they are detectable with the unaided senses or, in some cases, detectable at all.
It is a staple of the field of history and philosophy of science to wonder whether
any such prescription for interpreting theories (and models and other scientific
representations; I will take this as read henceforth) is plausible given the history
of theory change in specific domains of the sciences. A lot of ink has been spilled
on the question of whether, or under what circumstances, a realist interpreta-
tion of theories is reasonable. Antirealists of various kinds have argued that given
the lessons of changing descriptions of targets of scientific interest over time,
adhering to realism is something of a fool’s errand. Conversely, realists of var-
ious kinds—often referred to as selective realists—have sought to identify some
principled part of theories regarding which there has been continuity across
theory change in the past, thus fostering the reasonableness of expectations of
continuity in the future.
My focus in this essay is not the historical framing of these particular debates
about realism per se, but rather a key feature of them that amounts to a more
general problematic for realism. The shared strategy among selective realists for
dealing with descriptive discontinuity across historical theory change has been,
unsurprisingly, selectivity in what they take to be correct about a theory, pro-
posed in discussions of how certain claims have had or do have greater epistemic
warrant than others, such that continuity regarding these claims may then serve
as a bulwark for realism even while discontinuity rules more generally. Hence the
now familiar maneuver of associating realism with only certain parts of theories,
such as those involved in making successful novel predictions, or concerning ex-
perimental entities or mathematical structures. In each case we find concomitant
arguments about the typical preservation of the relevant parts of theories across
theory change, both as a reading of history and as a promise for the future.
Here emerges the key feature of debates surrounding the shared strategy of
selective realism on which I will focus. The hope of selective realism is that less
is more. By associating realist commitments with less, the hope is that it will be-
come easier to defend—indeed, that it will amount to a plausible epistemology of
science. It is by no means easy, however, to know how much is enough. Consider,
in connection with any given theory, an imagined spectrum of epistemic
commitments one might make regarding its content. At one end of the spectrum
one believes almost nothing; at the other end, one believes everything the theory
states or suggests. Arguably, if realism is purchased at the cost of believing al-
most nothing, it is largely empty; if instead realism is made more substantial by
licensing ever greater quantities of substantive belief, it runs an ever greater risk
of (for example) falling prey to concerns arising from theory change. The realist,
then, in any given case, must perform a kind of balancing act appropriate to that
case. Let me label this challenge the realist tightrope. On one side, there is the
temptation to affirm less and less, and on the other, the temptation to affirm more
and more. Giving in to either of these temptations may spell disaster, but it is no
easy feat to get the balance just right.
The potential benefit to realism of walking the tightrope is wide-ranging, in
that it is relevant to both historical and ahistorical defenses of the position. As
noted, if the realist were able to get the balance just right in some particular do-
main of science, she might then be in a position to furnish a narrative of con-
tinuity of warranted belief across theory change in that domain, past, present,
and future. But the tightrope is something that must be walked not only in con-
nection with historical lineages of theories, but also in connection with any
given theory, for it is often a challenge to work out how any one theory should be
352 Contemporary Scientific Realism
interpreted in a realist way. As Jones (1991) notes, it can be rather unclear how
best to articulate the subject matters of theories in physics (as per his examples),
where different interpretations can amount to different explanatory frameworks,
each suggesting a different ontology. This he presents as a challenge to realism,
which “envisions mature science as populating the world with a clearly defined
and described set of objects, properties, and processes, and progressing by steady
refinement of the descriptions and consequent clarification of the referential
taxonomy to a full blown correspondence with the natural order” (p. 186). In
this characterization of realism we catch a glimpse of how the tightrope must be
walked both synchronically and diachronically.
How shall we understand the spectrum of commitment, from thinner to more
substantial? I take it that an understanding of this is implicit in most discussions
of realism; indeed, it is implicit in Jones’s characterization. The thinnest possible
realist commitment is to the mere existence of something, which we capture
by speaking of successful reference. We say that the term “ribonucleic acid” or
“black hole” refers, which is (typically) shorthand for saying that it refers deter-
minately to something in the world. From here, commitment becomes increas-
ingly substantial as realists assert descriptions of the properties and relations of
these things. A very substantial commitment may involve asserting all of the
descriptions comprising or entailed by a theory, but it is often expressed other-
wise, not by believing everything—often not in the cards in any case, since the-
ories often contain known idealizations and approximations—but by asserting
increasingly detailed descriptions of whatever the realist does, in fact, en-
dorse. For example, an entity realist might describe the natures of the entities
she endorses in terms of certain causally efficacious properties. A structural re-
alist might describe the natures of these same entities in terms of certain struc-
tural relations. One can imagine yet further, finer-grained descriptions of the
natures of these properties and structures. On the thin end of the spectrum we
have bare reference, and on the other end, ever more comprehensive or detailed
descriptions of the natures of the referents.
With this understanding in hand, a clearer picture of the realist tightrope
emerges. Believing as little as possible and thereby asserting successful refer-
ence alone, which might seem a more defensible position than believing sig-
nificantly more, might also seem to run the risk of rendering realism empty of
much content.1 The more one believes, however, the wider one opens the door to
both the epistemic peril of believing things that may be weeded out as theories
develop and improve, and the metaphysical peril of defending finer-and finer-
grained descriptions of the subject matters of realist commitment, in virtue of
1 Stanford 2015 argues that it would make realism into something that is, if not empty, so weak
that antirealists need not dispute it. I will return to this contention in section 15.5.
Realist Representations of Particles 353
the inevitably and increasingly abstruse concepts and objections to which meta-
physical theorizing is prone. In some cases the epistemic and metaphysical perils
come together: in these cases, succumbing to the temptation to articulate the
natures of the referents of a theory in some fine-grained detail yields descriptions
that may become outmoded by subsequent developments in the relevant science.
An examination of precisely this sort of case, to which I will turn now, forms the
backbone of what follows.
The Standard Model of particle physics, one of the landmark achievements of
twentieth-century science, itemizes a taxonomy of subatomic particles along with
their properties and interactions. Beyond mere reference to these entities, how-
ever, their nature has been subject to realist wonderment and debate throughout
the history of theorizing in this domain. In ways that are well known and which
I will consider momentarily, the particles of the Standard Model are radically
unlike what could be imagined in classical physics—thus providing an example
of how descriptions of the physical and metaphysical natures of something con-
ceived in connection with earlier theorizing would have to be relinquished in
light of subsequent theorizing. Just what the natures of these things enumer-
ated by the Standard Model are, however, is still far from clear. Indeed, if re-
alist attempts to characterize them are any guide, there is no consensus at all.
Advocates of different forms of selective realism, for example, have characterized
them in very different ways. And all the while, realists and antirealists alike have
suggested that in the absence of some unique characterization, realism is unten-
able. After considering various possibilities for thinking about the identity and
individuality of particles, for instance, van Fraassen (1991, p. 480) concludes that
we should say “good-bye to metaphysics.” Ladyman (1998, p. 420) holds that tol-
erating metaphysical ambiguity regarding these issues would amount to a merely
“ersatz form of realism.”
In earlier work I have offered judgments that might be construed as echoing
these kinds of sentiments. “One cannot fully appreciate what it might mean to
be a realist until one has a clear picture of what one is being invited to be a re-
alist about” (Chakravartty 2007, p. 26). But having a clear picture is compatible,
I submit, with a number of different and defensible understandings of how best
to walk the realist tightrope in any given case. In the remainder of this essay I en-
deavor to explain why this is so, taking particles as a case study. In the next section
I briefly substantiate the contention that many questions regarding the natures
of particles are still very much up for grabs in contemporary physics and phi-
losophy of physics with a synopsis of a handful of the conceptual conundrums
surrounding them. In sections 15.3 and 15.4, I examine, respectively, what I de-
scribe as the two main approaches to thinking about the natures of particles and
their properties, which have in turn shaped varieties of selective realism—“top-
down” approaches, emphasizing formal, mathematical descriptions furnished
354 Contemporary Scientific Realism
The electron is the veritable poster child of scientific realist commitment. Most
proponents of both realism and antirealism (rightly or wrongly) take various facts
about observable phenomena as uncontroversial, but seriously contest the status of the
unobservable. Though theories at every “level” of description—from social and psy-
chological phenomena to biological and chemical phenomena through to the subject
matters of physics—all theorize about putatively unobservable objects, events, pro-
cesses, and properties, realists often cite subatomic particles as a shining example of
entities conducive to realism. On the one hand, given the mind-boggling success of the
uses to which we have put twentieth-century theories concerning them, not least in
a host of startlingly effective technologies from computing to telecommunications to
medical imaging, this may not seem surprising. On the other hand, perhaps it should
be a cause for concern after all, because even a cursory foray into our best attempts to
grasp the natures of particles and particle behavior are fraught with conceptual diffi-
culties, and our best scientific theories are far from transparent on this particular score.
Trouble rears its head at the start with the term “particle.” What is a particle
in this domain? There is, of course, a classical conception of what a particle is,
which is easily graspable and conceptually undemanding, relatively speaking,
but this conception is simply inapplicable at atomic and subatomic scales in light
of twentieth-century developments in theory and experiment. Classical particles
are solid entities that can be envisioned colliding with and recoiling from one
another in the ways that billiard balls appear to behave, phenomenologically. The
natures of particles described by the Standard Model are not in this way intelli-
gible. They appear to behave in the manner of discrete entities in some contexts
but like continuous, wave-like entities in others. It is unclear whether all of their
properties are well defined at all times, though we can ostensibly detect them and
measure their values under certain conditions. As intimated earlier, there is con-
troversy as to whether they can be regarded in any compelling way as individ-
uals; if they can, it is certainly not in the way we commonly think about identity
conditions and individuation in classical contexts, in terms of differential pro-
perty ascription and allowing for their re-identification over time.2 The natures
2 Arguably, one may overstate this last point. Cf. Saunders 2006, p. 61: “there are many classical
objects (shadows, droplets of water, patches of colour) that likewise may not be identifiable over time.”
Realist Representations of Particles 355
of particles conceived today may become enmeshed over arbitrarily large spatial
distances—as per quantum entanglement—in ways not previously conceived.
All of this suggests the strangeness of particles from a classical point of view,
but this should not be inimical to realism all by itself—if one adopts a thinner
conception of realism about particles framed in terms of reference. It is in the na-
ture of theoretical development that sometimes the features of target systems of
scientific interest that come to be viewed in a new light will appear strange from
the point of view of what came before. This by itself suggests only a common and
understandable propensity to regard the unfamiliar as strange. The challenge to
realism here is no mere strangeness en passant, but the fact that in this case in
particular, our attempts to interpret our theories in more substantial ways, so
as to make their content intelligible to ourselves, have resulted in a great deal of
unsettled debate and lasting conceptual puzzlement. Taking this as a basis for re-
alism regarding a more substantial conception of particles, an antirealist might
be forgiven for wondering whether an argument here against realism is in fact re-
quired, since it would appear that collectively, realists cannot themselves decide
what they should believe.
Consider, for example, the question of the basic ontological category to which
particles belong. There is a long history here of being flummoxed, even upon
careful consideration of the relevant physics, regarding what this might be. The
term particle is commonly associated with the notion of objecthood, but we have
already noted that particles cannot be objects if “objecthood” is allowed to carry
classical connotations. In quantum field theory, particles are often described as
modes of excitation of a quantum field. This does not sound very object-like in
any traditional sense, which leads many to claim that particles are not objects
after all. Yet even physicists who observe that with the advent of quantum field
theory the ontology of the quantum realm might be thought of in terms of
fields rather than particles are happy to talk about particles at will. This suggests
that they either regard particle-talk as merely elliptical for states of fields, or
that they do in fact regard particles as non-classical objects of some sort, per-
haps standing in some sort of dependence relation to fields. Generally, there is
nothing like a precise specification of a basic ontology to be found, thus leaving
the answer to the question of an appropriate assignment of category ambiguous.
Here one might think that philosophers would lend a hand, but a sampling of the
views of philosophers merely reveals the trading of ambiguity for transparent
disagreements.3
3 The sample to follow is by no means comprehensive, given the long history of differing interpret-
ations, but representative of some of the most recent literature. Not everyone cited is a realist, neces-
sarily, but all are attempting to clarify the relevant ontology.
356 Contemporary Scientific Realism
At the very least, in the absence of an intuitive grasp of the natures of prop-
erties of the sort just suggested, perhaps we could say something about them in
terms that philosophers find perspicuous. Are these properties monadic, dyadic,
or polyadic? Are they intrinsic or extrinsic or essentially relational? Mass is com-
monly cited as an exemplar of an intrinsic property, but Bauer (2011) thinks that
it is extrinsic, since it is “grounded” in and thus ontologically dependent on the
Higgs field. French and McKenzie (2012) contend that there are no fundamental
intrinsic properties by means of an argument appealing to gauge theory,4 which
is integral to the Standard Model, but Livanios (2012) does not find this argu-
ment compelling. Lyre (2012, p. 170) maintains that properties such as mass,
charge, and spin are “structurally derived intrinsic properties,” which suggests
something of a hybrid, intrinsic-extrinsic nature. And in some cases an exami-
nation of the natures of these properties brings us full circle, back to a consider-
ation of the ontological category of things that best corresponds to particle-talk,
as when Berghofer (2018) holds that the relevant properties are, in fact, in-
trinsic and non-relational, but features of fields, not particles, and when Muller
(2015, p. 201) contends that we would be better off with a new conception of
objects: particles, he argues, are “relationals”; “objects that can be discerned by
means of relations only and not by properties.”
The purpose of the preceding whirlwind tour has not been to suggest that
progress cannot be made on questions surrounding the ontological natures of
particles. No doubt some and perhaps many of the issues disputed in the pre-
ceding discussion may ultimately be resolved in ways that produce a measure of
consensus. The point here is a different one. If, in order that realism be a tenable
epistemic attitude to adopt in connection with the Standard Model, we were to
require a degree of communally sanctioned, fine-grained clarity regarding de-
scription that could only follow from having resolved all such debates, this would
suggest a prima facie challenge to the very possibility of realism here and now. As
the brief glimpse into a number of contemporary debates just presented makes
plain, any attempt to clarify the ontology of the Standard Model quickly and in-
evitably draws one into contentious metaphysical discussions. In the following
two sections I will examine the two overarching approaches to prosecuting these
debates that have formed the basis of selective realist pronouncements regarding
the natures of particles and their properties, with the eventual goal of arguing for
a rapprochement between them qua realism—one that clarifies how realism can
be a tenable, shared epistemic commitment even in cases where realists disagree
about details of description.
4 Offering a different argument, McKenzie 2016 takes this position with respect to various proper-
As it happens, the two approaches to thinking about how best to interpret the
Standard Model I have in mind reflect a longstanding division of labor within
the community of physicists. On the one hand there is theoretical physics,
which views particles through the lens of formal, mathematical descriptions fur-
nished by theory, and on the other hand there is experimental physics, which
views particles through the lens of the sorts of detections and manipulations
of them that are part and parcel of laboratory practice. These communities of
scientists are not, of course, strictly isolated from one another; they must often
work together. Nevertheless, their approaches to the subject matter are of ne-
cessity shaped by the kinds of work they do. Corresponding to this rough di-
vision, in the philosophy of science, there are what I will refer to as “top-down”
approaches to interpretation, which place primary emphasis on mathematical
descriptions of the properties and interactions of particles found in high theory
as a source of insight into the natures of particles in the world; on the flipside
there are what I will call “bottom-up” approaches to interpreting the natures of
particles, which place primary emphasis on their behaviors in the trenches of
concrete interventions characteristic of experimental investigation. As we will
see, both approaches offer insight into the natures of particles, and both leave
important questions open.
Let us begin with the top-down approach to thinking about particles. The
Standard Model provides a remarkably elegant description of fundamental
particles, their properties, and their (electromagnetic, weak, and strong)
interactions, neatly systematizing them by means of symmetry principles.
A symmetry is a transformation (or a group of transformations) of an en-
tity or a theory in which certain features of these things are unchanged. That
is, the relevant features are preserved or remain invariant under the trans-
formation. The kinds of transformations relevant to physical descriptions in-
clude translations in space or time or spacetime, reflections, rotations, boosts
of certain quantities such as velocity, and gauge transformations. When the
states of systems are related by symmetries, they have the same values of cer-
tain quantities or properties—including those used to classify particles, such
as mass, charge, and spin. To take an everyday example, if one rotates a square
by 90 degrees, one gets back a square. “Squareness” is invariant under this
transformation, which maps the square onto itself. With the notion of a sym-
metry in hand we may define a symmetry group as a mathematical structure
comprising the set of all transformations that leave an entity unchanged to-
gether with the operation of composition of transformations on this set (satis-
fying the conditions: associativity; having an identity element; every element
having an inverse).
Realist Representations of Particles 359
There are a variety of accounts of realism about particles that one might char-
acterize as top down. What they have in common is the (explicit or implicit)
operating principle that insight regarding the natures of particles should be in-
timately and exclusively connected to interpreting the mathematical formalism
I have just described. This is all we need to understand the natures of particles,
nothing more. There is nothing in my description of the top-down approach that
suggests that it should provide exclusive insight into the natures of particles, but
in practice, this is how realists who take this approach proceed. To take this extra
step from adopting a top-down approach to thinking, furthermore, that this is
our best or only legitimate source of insight into the natures of particles requires
some further motivation or argument. Let me now briefly consider a couple of
arguments of this sort, and for each suggest one of two things: either the top-
down description of particles provided does not preclude supplementation with
bottom-up description; or if it does, it is unclear why the top-down characteri-
zation should be judged superior qua realism. Obviously, this will not amount
to a comprehensive survey of all possible arguments for an exclusive commit-
ment to the top-down approach. Nonetheless, I take it to be suggestive of a plau-
sible general moral, that the necessity or irresistible appeal of this commitment
is unproven.
Motivating at least some realists who are exclusively committed to top-down
characterizations of particles are desiderata such as descriptive or ontological
intelligibility or simplicity. In section 15.4 we will see in some detail how the
bottom-up approach is typified by an ontologically robust understanding of the
causal or modal natures of the properties of particles, but for the time being it
will suffice to note that some realists who focus their attention on symmetries
hold that this focus alone is sufficient for understanding the natures of these
properties, thus precluding any “inflation” of our ontological commitments in
ways recommended by bottom-up realists about particles. If the Standard Model
is simply interpreted as describing properties such as mass, charge, and spin as
invariants of certain symmetry groups, we might rest content with this purely
mathematical, theoretical apparatus for describing the natures of properties. On
this view it is unnecessary to appeal to the causal roles of things in order to iden-
tify or understand them. Armed with symmetries, we might then understand
the natures of the relevant properties without appealing to the notion of causal
features or roles at all, thus articulating realism in terms of a simpler ontological
picture.
Let us then consider whether the content of a realism about particles can
be provided solely through an examination of symmetries and invariants. It is
difficult to see how it could, given that questions about what realists justifiably
believe cannot be separated from matters of how evidence furnishes justifica-
tion. Detection and measurement are intimately connected to determining what
360 Contemporary Scientific Realism
things there are, in fact, in the world. Are properties thus conceived, as things
one might identify as existing or being exemplified in the physical world—things
about which one might be a scientific realist—identified independently of their
causal roles? Perhaps there are cases in which this happens, but the present case
does not seem like one. While there is no doubt that symmetry groups comprise
a beautiful framework for codifying particles and their properties, all that exam-
ining them can achieve in isolation is to generate descriptions of candidate enti-
ties that may then be put to the test of experimental detection. It is one thing to
describe the natures of some target of realist commitment in terms of the formal
or mathematical aspects of a theory, but generally, in order for descriptions
to have the sort of content required to support realism, they must be taken to
refer to some thing or things in the world, and establishing successful reference
requires more than the examination of a formalism. Some supplementation
seems necessary.
A nice illustration of this is furnished by permutation symmetry, which arose
earlier in section 15.2. Recall that in quantum theory, state representations of
particles of the same type in which the particles are interchanged do not count
as representing different states of affairs. An examination of the permutation
group yields certain “irreducible representations” corresponding to all of the
particles, the fermions and bosons, populating the Standard Model. However, in
addition to these fermionic and bosonic representations, there are also so-called
“paraparticle representations,” and unlike fermions and bosons, paraparticles do
not appear to exist—at least, not in the actual world subject to scientific realism.5
Thus, merely examining the mathematical formalism of the theory is insufficient
for the identification of entities to which realists should commit. To avoid being
misled about the ontology of the world, there would seem to be no substitute for
getting one’s hands dirty with the causal roles of properties in the context of ex-
perimental work using detectors, and this suggests that there may be something
to the thought that properties of particles have some sort of causal efficacy after
all, in virtue of which they are amenable to detection, measurement, manipu-
lation, and so on. But now we have entered the territory of the bottom-up ap-
proach to understanding the natures of particles, to which we will return in the
following section.
Let us consider a second possible motivation for an exclusive reliance on
the top-down approach for the purpose of illuminating particles. Perhaps the
boldest motivation yet proposed stems from a version of selective realism that
was designed specifically (in the first instance) to serve as an account befitting
5 The closest we have come to generating empirical evidence in this sphere is the detection of
paraparticle-like states, though not paraparticles themselves, under very special conditions. For a
brief discussion, see Chakravartty 2019, p. 14.
Realist Representations of Particles 361
fundamental physics: ontic structural realism. There are many variants of the
view, but generically, the common thread is what one might call a reversal of the
ontological priority traditionally associated with objects and properties relative
to their relations. In much traditional metaphysics, objects and/or properties are
conceived as having forms of existence whereby their relations are in some way
derivative (and not vice versa). Some variants of ontic structural realism simply
boost the ontological “weight” of the relevant relations relative to their relata
such that they are all on a par, ontologically speaking. Others take the relations to
have greater ontological priority, and the most revisionary formulations do away
with objects and properties altogether, eliminating them in favor of structural re-
lations which are then viewed as ontologically subsistent in their own right, thus
constituting the concrete furniture of the world.6 If particles are conceived as
being entirely dependent on relations described by symmetries—or stronger yet,
as epiphenomena of these relations—it may well seem that a top-down approach
to describing their natures should be sufficient.
As is true regarding any proposal for realism there are several aspects of this
view that one might seek to clarify, but perhaps the most fundamental concern
that has been raised is whether ontic structural realism can render intelligible the
idea that things described in purely mathematical terms—such as symmetries
and invariants, which are standardly regarded as (at best) abstract entities—can
be understood to constitute the world of the concrete. Merely stipulating that
some mathematical structures are subsistent appears to achieve no more than to
substitute the term “concrete” with “subsistent.” Something more is needed, and
no doubt with this in mind, advocates of the position sometimes explicate the
sense of concreteness or subsistence at issue by saying that the relevant structures
are causal or modal.7 Esfeld (2009, p. 180), for example, is explicit that on his
variant of ontic structural realism, “fundamental physical structures are causal
structures.” French (2014, p. 231) is clear that on his, “we should take laws and
symmetries—and hence the structure of which these are features—as inherently,
or primitively, modal,” and take this de re modality as serving the explanatory
functions commonly associated with attributions of causality, such as helping us
to explain what it means for something to be concrete.
If one goes this route, however, the first of our potential rationales for
favoring a top-down approach to interpreting the natures of particles based
on the promise of a comparatively simple or streamlined ontology is ruined,
because it is difficult to see how a reification of symmetries and other mathe-
matical structures endowed with causal or modal efficacy should count as less
6 For a detailed and comprehensive exploration of the many variants, see Ladyman 2014/2007.
7 Cf. Ben-Menahem 2018, p. 14: “causal relations and constraints go beyond purely mathematical
constraints; they are (at least part of) what we add to mathematics to get physics.”
362 Contemporary Scientific Realism
From the point of view of detection, which is intimately linked to many of the
strongest cases that can be made for realism in specific instances, more abstract
descriptions of the properties of things are somewhat removed from the work
of physics. Where the focus of experimental work is the physical discernment
of interactions between particles and between particles and detectors, often
requiring extraordinarily precise adjustments and manipulations of both the ex-
perimental apparatus and the target entities under investigation, more precise
descriptions of concrete natures are necessary. It is here that the determinate
properties of particles, whose values are detected and manipulated in such work,
take center stage. This is not to say that group theoretic structures are irrelevant to
describing these properties, but simply that the descriptions afforded by symme-
tries and invariants are at a remove from the specificities of experimental work.
As Morganti (2013, p. 101) puts it, “[w]hen one focuses on invariants . . . one
moves at a high level of abstractness.” In contexts of experimentation and detec-
tion, it is necessary to move in the direction of more determinate description;
the specific values of mass, charge, etc. (pertaining to different particles) at issue
in these contexts are not given by descriptions of symmetries (cf. Wolff 2012,
p. 617).
In the realm of experiment it is what we can do that is our best guide to what
there is and what these things are like—that is, to the ontology of our targets
of investigation. What we can do in the arena of particle physics is entirely de-
pendent on the precise values of the properties of particles. Since all of this
doing involves designing and engineering instruments to interact with those
parts of the world we aim to explore, and generating certain kinds of effects, it
is natural to describe it in terms of causal interactions, relations, and processes.
Thus it is no surprise that selective realists who take a bottom-up approach to
Realist Representations of Particles 363
any given particle. Similarly, consider principles of so-called least action: for any
given system and a specification of some initial and final conditions, the evolu-
tion of the state of the system will minimize a quantity referred to as “action.” It is
difficult to see how the causal profiles associated with a particle and its properties
could somehow generate the minimization of action in systems more generally.
In order to explain constraints on behavior such as those expressed in prin-
ciples of conservation and least action in terms of the causal natures or profiles
of properties, it would seem we must think of these properties as belonging to
the systems to which these principles apply, and this is inevitably controver-
sial. Taking the dispositional variant of the bottom-up approach as an illus-
tration once again, Harré (1986, p. 295) maintains that some dispositions may
be grounded in “properties of the universe itself ”—a phenomenon he labels
“ultragrounding”—attributing the idea to Mach’s discussion of inertia in the
context of Newtonian thought experiments. Imagine two globes connected by
a spring balance and rotating, alone in the universe. Newtonians held that there
would be a force tending to separate the globes, registered in the spring balance,
but on a Machian reading there is no reason to believe that the globes would have
inertia in an otherwise empty universe; it is better to think of the disposition to
resist acceleration as grounded in the universe itself. Bigelow, Ellis, and Lierse
(1992, pp. 384–385; cf. Ellis 2005) go so far as to contend that the actual world
is a member of a natural kind whose essence includes various symmetry prin-
ciples, conservation laws, and so on. It is a short step from this to thinking that
the system-level behaviors associated with these principles are properties of the
world—a very large system indeed.
The standard objection to this family of speculations is that it is ad hoc.8
Granted, explaining constraints on the behaviors of systems of particles in terms
of properties of the entire world may seem, prima facie, rather convenient, es-
pecially in the absence of any independent motivation for the explanans. If one
takes a bottom-up approach to understanding the natures of target systems of
scientific interest, however, it is simply a mistake to suggest that there is no in-
dependent motivation. Just as particles are investigated empirically in carefully
designed and executed experiments, systems of particles are likewise investi-
gated. It is an empirical fact, not a convenient fact, that certain kinds of systems
exhibit behaviors that conform to various principles of conservation and least
action. Having adopted a methodology of associating causal profiles with certain
types of particles and their characteristic properties, it is hardly an unmotivated
extension of this methodology to do likewise in connection with certain types of
systems, such as closed systems.
8 For recent discussion on both sides of this fence, see Smart & Thébault 2015 and Livanios 2018.
366 Contemporary Scientific Realism
Now, as it happens, the world itself is a system of this type. From the perspec-
tive of the bottom up, the attribution of a causal profile to it on the basis of a con-
sideration of empirical investigations into members of the type cannot be said to
be based on merely wishful speculation. And neither should it seem peculiar in
the era of quantum theory, in which systems are routinely viewed as having prop-
erties, such as entanglement, that cannot be reduced to the properties of their
parts. What may appear superficially as ad hoc speculation inevitably sounds
more credible when the metaphysical terminology in terms of which it is some-
times expressed (“natural kinds,” “essential properties”) is given a plausible in-
terpretation in the language of scientific description. Consider, for example, the
possibility entertained earlier that particles are in fact best understood as field
quanta. In that case the properties of systems suggested above would be proper-
ties of fields, and fields are global in the sense that they permeate the whole of the
world. Anyone moved by this description would then be in a position to ask yet
further questions about the natures of particles and their properties, depending
on whether one is a substantivalist about fields, or interprets the values of field
quantities as properties of spacetime points, or . . . . But let us stop here.
Having refined and extended a bottom-up description of particles in such a
way as to answer a preeminent concern, let us once again inquire into how this
approach fares in comparison to its counterpart, top down. Here, once again,
it is very difficult to make a case one way or the other on the basis of some
imagined criteria of descriptive or ontological intelligibility or simplicity. On
a permissive enough conception of causation one may see it as an appropriate
descriptor of many different things. Ben-Menahem (2018), for instance, applies
the label “causal” to any general constraint on change, where constraints deter-
mine what may happen, or is likely to happen, or what cannot happen. On such
a conception, symmetries, conservations laws, and variational principles (such
as the principle of least action) all qualify as causal. But are they causal in the
sense advocated by someone looking top down, as a primitive feature of certain
mathematical structures, or are they causal in the sense of, say, a dispositionalist
looking bottom up, where the (potential for) behaviors associated with the prop-
erties of various kinds of entities and systems determine their identities? And
does anything hang on this choice, from the point of view of defending realism?
In closing, let me attempt to shed some light on the latter question.
of our very best ones) are entirely correct, not merely in connection with the
idealizations and approximations we know of, but also in other ways we have
yet to discover, everyone appreciates that they will evolve over time as scientific
inquiry proceeds. Hence the various strategies found among realists, especially
selective realists, for identifying those aspects of theories and models that have
sufficient warrant to command realist commitment, both as a guide to interpre-
tation in the present and to reasonable expectations about what will survive into
the future. From this I distilled a more specific challenge to anyone hoping to be
successfully selective, which I called the realist tightrope: believing too little or
too much of a theory that is not entirely correct may well appear to spell trouble.
Believing too little—in the limit, the bare reference of central terms, or claims re-
garding the existence of their referents—may seem tenuous, but the more com-
prehensive or detailed the descriptions of such things one endorses, the more
one runs the risk of falling foul of future developments in the relevant science,
and/or metaphysical objections to the increasingly fine-grained natures pro-
posed. Where does the proper balance lie?
In the case of the Standard Model, walking the tightrope seems especially
fraught. Understanding the natures of the particles described by the theory has
always been difficult, and serious proposals for illuminating these natures have
inevitably required increasingly speculative and technical theorizing. Given this
state of affairs, it is hardly surprising that there is so much disagreement among
realists about how best to describe what particles are, exactly. On the surface this
may appear a victory for antirealism, for if there is nothing determinate here to
be found under the heading of “realism” to which all realists subscribe, but in-
stead a thousand splintered commitments to conflicting and (what will appear
to some as) increasingly esoteric interpretations, the camp of realism may look
more like a ball of confusion than anything endorsing a shared epistemic com-
mitment. This dismal portrait of the cognitive landscape of realism in connec-
tion with the Standard Model is, however, though perhaps understandable on
the basis of what we have seen, entirely misleading. I submit that there is in fact
something substantial to which all or most realists subscribe in the context of the
Standard Model, even as they debate how this commitment is best elaborated.
Here in conclusion I will attempt to explain how this can be so.
In order to understand how different and conflicting descriptive commitments
among realists regarding particles are compatible with a shared commitment
qua realism, we could do worse than to begin by looking at how scientists ap-
proach this area of physics. Here, just as Jones (1991, p. 191) notes in connection
with the analogous case of multiple candidate interpretations of quantum me-
chanics, one may reasonably worry about “the failure of any interpretation to
provide an ‘explanatorily satisfactory’ link between the mathematical formalism
and the world of laboratory experience.” He continues:
368 Contemporary Scientific Realism
The general approach of one interpretation may suit a physicist more than the
general approach of others, and he or she may spend some time adapting it to
issues that he or she thinks particularly important and developing arguments
as to why its lacunae are not devastating for its coherence. But every physicist
will admit that such allegiance is to some degree a matter of taste. No physicist
is unaware of competing interpretations, and none expects decisive evidence or
arguments for one against the others.
9 Galison 1997, pp. 833–835, tells the story of how Sidney Drell and James Bjorken aspired to write
a book on quantum field theory in the early 1960s, but ended up producing two separate volumes, one
geared to experimentalists (concerned more with measurable quantities) and the other to theorists
(concerned more with formal properties of theories, such as symmetries and invariances). Some of
the differences between the volumes amounted to a “radical difference in the ontology” (p. 835).
370 Contemporary Scientific Realism
10 Cf. Magnus 2012, p. 122: “Retail arguments [i.e., arguments stemming from evidence specific
to the case at hand] for believing in particular things can give us good reasons to believe that those
things exist on the basis of their connections to other things, while leaving questions of things’ funda-
mental nature either unmentioned or unresolved.”
372 Contemporary Scientific Realism
Acknowledgments
References
Bain, J. 2011: “Quantum Field Theory in Classical Spacetimes and Particles,” Studies in
History and Philosophy of Modern Physics 42: 98–106.
Baker, D. J. 2009: “Against Field Interpretations of Quantum Field Theory,” British Journal
for the Philosophy of Science 60: 585–609.
Bauer, W. A. 2011: “An Argument for the Extrinsic Grounding of Mass,” Erkenntnis
74: 81–99.
Ben-Menahem, Y. 2018: Causation in Science. Princeton: Princeton University Press.
Berghofer, P. 2018: “Ontic Structural Realism and Quantum Field Theory: Are There
Intrinsic Properties at the Most Fundamental Level of Reality?” Studies in History and
Philosophy of Modern Physics 62: 176–188.
Bigaj, T. 2018: “Are Field Quanta Real Objects? Some Remarks on the Ontology
of Quantum Field Theory,” Studies in History and Philosophy of Modern Physics
62: 145–157.
Bigelow, J., B. Ellis, & C. Lierse 1992: “The World as One of a Kind: Natural Necessity and
Laws of Nature,” British Journal for the Philosophy of Science 43: 371–388.
Bird, A. 2007: Nature’s Metaphysics: Laws and Properties. Oxford: Clarendon.
Chakravartty, A. 2007: A Metaphysics for Scientific Realism: Knowing the Unobservable.
Cambridge: Cambridge University Press.
Chakravartty, A. 2017: Scientific Ontology: Integrating Naturalized Metaphysics and
Voluntarist Epistemology. New York: Oxford University Press.
Chakravartty, A. 2019: “Physics, Metaphysics, Dispositions, and Symmetries— à la
French,” Studies in History and Philosophy of Science 74: 10–15.
Einstein, A., B. Podolsky, & N. Rosen 1935: “Can Quantum-Mechanical Description of
Physical Reality Be Considered Complete?” Physical Review 47: 777–780.
Ellis, B. 2005: “Katzav on the Limitations of Essentialism,” Analysis 65: 90–92.
Esfeld, M. 2009: “The Modal Nature of Structures in Ontic Structural Realism,”
International Studies in the Philosophy of Science 23: 179–194.
Fraser, D. 2008: “The Fate of ‘Particles’ in Quantum Field Theories with Interactions,”
Studies in History and Philosophy of Modern Physics 39: 841–859.
Realist Representations of Particles 373
For the benefit of digital users, indexed terms that span two pages (e.g., 52–53) may, on occasion,
appear on only one of those pages.
Figures are indicated by an italic f and tables by t following the page number.
Fahrbach, Ludwig, 203–4. See also pessimistic General Scholium (Newton), 78, 82
meta-meta-induction General Theory of Relativity (General
False Basis Argument, 146–49 Relativity), 89–92, 197, 221–22, 245,
false-but-successful theories, 34–35, 37, 48 292, 313, 323
false may entail true objection, 208 Glymour, Clark, 159–60
false theories. See also specific types Godfrey-Smith, P., 58, 313–14, 317
successful predictions, 56 (see also freezing gravity
point of water, pressure lowering) Huygens on, 79–80, 81
Farr, William, 15–19, 18f, 19f, 20f, 22 Leibniz on, 80–81
objections, 22–29 Newton’s theory, 81–84, 85–86
Feyerabend, Paul, 121, 246–47 Einstein’s retentionism and General
Field, Robert, The Spectra and Dynamics of Theory of Relativity, 89–92
Diatomic Molecules, 175 Gregory, David, Astronomiae physicae et
Fine, Arthur, 231–32, 233–34, 235 geometricae elementa, 80
first-order phase transitions, 334–40
Fisher, Ronald, 224–25 Hall, Robert, 258, 262–63, 270
Flavin, M. A., 267 Hamlin, C., 14–15
Fraser, J., 341–42 Hanson, Norwood Russell, 350
freezing point of water, pressure Harper, William, 85, 94–95
lowering, 56–68 Harré, R., 365
Carnot’s theory, 58–59 Havranek, T., 269–70
methodological hesitation, 63–64 heat, caloric theory of. See caloric theory
Thomson’s revival, 59–60, 61–63 of heat
concluding review, 67–68 heat-engine, 58–60, 61–62, 67–68
James Thomson’s prediction heat-pump, 59–60
analysis, 61–63 Heilbron, John, 117–18
methodological hesitation, 63–64 Heisenberg, Werner, 128–29, 371–72
no miracle, 62–63 Hempel, Carl, 209
quantitative calculation, 64–66 Hendry, Robin, 47–48
rationale behind, 60 Herzberg, Gerhard, Molecular Spectra and
William Thomson’s figure for James’ Molecular Structure, 163–64, 166–68,
engine’s efficiency, 66–67 166t, 173–74
no-miracle argument, 56–57, 62–64, 67 Hesse, Mary, 1, 12–13
French, Steven, 240–41, 245–51, 357, 361 Hicks, W. M., 117
Fresnel, Jean, 199–200 historiography, of science, 253–54
equations, 245–46 Hooke, Robert, 253–54
ether, 244–45 Hoover, K., 265–66, 272–73
theory of light, Poisson white spot, 24–25, Hoyningen-Huene, P., 127
26–27, 44–45 Hricko, Jonathon. See boron discovery
Friedman, M., 260 Hutchison, Keith. See freezing point of water,
permanent income hypothesis, 258, 260–61, pressure lowering
262, 267–69, 272 Huygens, Christiaan, 79–82
Fuhrer, J. C., 269–70 Treatise on Light, 95
fundamentalism, 248 hypothetico-deductive model, 73, 82–83
successful prediction from, using abandoned van’t Hoff, Jacob Henricus, 132
theory, 57–58 verisimilitude approach, 317–18
thermometry, absolute, 59–60 Vickers, Peter, 44, 99–101, 201, 235–36. See also
Thomson, James, 132, 140 specific topics
engine efficiency, William Thomson’s derivation external vs. internal posits,
figure, 66–67 41, 147
prediction on Farr and miasma theory, 22, 24–25
analysis, 61–63 on novel predictive success, 44
methodological hesitation, 63–64 impressive vs. unimpressive, 44–45
no miracle, 62–63 on posits, 222
quantitative calculation, 64–66 on Psillos and essential deployment, 188, 192,
rationale behind, 60 201–2, 209
Thomson, Thomas, 44 on quantum theory and quantum
Thomson, William, 57–58, 105, 108–9 mechanics, 100
Carnot’s theory, revival, 59–60 on Sommerfeld’s atomic theory, 100–1
James Thomson’s engine’s efficiency, William’s vortex theory of planetary motions, Newton
figure, 66–67 on, 79–89
Thomson brothers, pressure lowers freezing explanation, Newton’s account, 84–86
point of water novel prediction, 56–68. Huygens vs. Leibniz, 79–82
See also freezing point of water, pressure natural philosophy, doing, 82–84
lowering problem with vortices, 79–82
top down approach, Standard Model, 358–62 quam proxime, 86–89
Treatise on Light (Huygens), 95 vortices, fate of, 77–78
triumphalism, 254 Votsis, I., 199, 203
truth-likeness, 243, 317–18
truth-transfer, theories of molecular structure, water, pressure lowering freezing point, 56–68.
selective realism and, 130–57. See also See also freezing point of water, pressure
molecular structure theories, selective lowering
realism and truth-transfer Weak Evidence Argument, 149–50
Tulodziecki, Dana. See miasma theory Werner, Alfred, 136–39, 137f, 140, 147–48
Tycho, 75–76, 350 Werner’s coordination theory, 130–31,
type hierarchy approach, 317–18 134–35, 138–39, 140, 142–43, 147–
49, 155
unconceived alternatives problem, 131, 141–46 Whewell, William, 287
Dicken on, 154 Whiggism, 110, 253–54
No Final Theory Objection to the Problem of Wilson, K., 345–46
Unconceived Alternatives, 144–45 Wilson, Mark, 219–21, 222, 223, 236
underdetermination, 1, 239–40, 241–42, 253, Wilson, William, 117
269, 300 Wing, Vincent, 172–73
uniformitarianism, 217–19, 222, 236 Woodward, J., 265–66
unobservables, 1, 12–13, 20–21, 33, 73–74, 86, Worboys, M., 14
141, 197–98, 212, 220, 233, 241–42, working posits, 20–21, 147, 148, 149, 152, 155–
243–44, 312, 354 56, 218, 222–23, 228
use-novel predictions, 12, 13, 15, 17, 19–21, 22, Worrall, John, 11, 13, 20–21, 22, 101–2, 239–
24–25, 33 40, 264–73
Wray, K. Brad, 126, 128, 287, 300–1
valence, 132–34, 135–36, 139, 142–43, 147–48, Wright, John, 184
242–43
Valente, Giovanni, 330–31. See also XT statements, 273–75
infinite limits
van den Broek, Antonius, 104 Yablo S., 188–89
van Fraassen, Bas, 123, 159–60, 161, 162–63, 170,
173–74, 176–78, 232–33, 241–42, 250 Zenneck, Jonathan, 89–90
Scientific Image, The, 1 Zewail, Ahmed, 175–76