HEAD PREVIOUS

Chapter 2
Reproducibility.
What is nature?

Michael Faraday is a fascinating character in the history of science. Born the son of a poor blacksmith, his formal education, such as it was, ended by age 13; but it ended by his becoming apprenticed to a bookbinder and bookseller; his informal self-education was beginning! This unconventional background left him knowing practically no mathematics. But he was an incredibly insightful experimentalist and he had an intuitive way of understanding the world in mental pictures. No less than five laws or phenomena of science are named after him. His experimental discoveries came to dominate the science of electricity, and of chemistry too, for the first two thirds of the nineteenth century. His conceptualization of the effects of electromagnetism in terms of lines of force laid the foundation for Maxwell's mathematical electromagnetic equations, and for the modern concept of a "field", in terms of which much fundamental physics is now expressed. 24
FaradayYoungCochrane.png
Figure 2.1: Michael Faraday as a young man
It was said of Faraday that whenever he heard of some new result or phenomenon, reported in a public meeting or a scientific journal, the first thing he would do was to attempt to reproduce the effect in his own laboratory. The reason he gave for this insistence was that his imagination had to be anchored in what he called the "facts". He understood in his bones that science is concerned with reproducible phenomena which can be studied anywhere under controlled conditions and give confirmatory results. "Without experiment I am nothing," he once said.
Faraday's attitude is a reflection of what is often taken for granted in talking about science, that science deals with matters that show reproducibility. For a phenomenon to be a question of science, it had to give reproducible results independent of who carried out the experiment, where, and when. What the Danish Professor Hans Oersted observed during a lecture demonstration to advanced students at the university in Copenhagen in the spring of 1820 ought to be observed just the same when Faraday repeated the experiment later at the Royal Institution in London. And it was. Here, by the way, I am alluding to the discovery that a compass needle is affected by a strong electric current nearby, demonstrating for the first time the mutual dependence of electricity and magnetism. According to the students present at his demonstration, this discovery was an accident during the heating of a fine wire to incandescence using an electrical current. But Oersted's own reports claim greater premeditation on his part25.

2.1  The meaning of experiment

Imagine a family trip to the Australian beach. The youngster of the family, three-year-old Andrew, is there for the first time. He is fascinated at all the new experiences. He idly, perhaps accidentally, kicks the gravel on the way down to the sand, and pauses to hear it rattle. When seated in the sun he grabs handfuls of sand, and throws them awkwardly over himself, and anyone else who strays too near. He is fearful and wondering at the unexpected waves, even the gentle ones that surge up the smooth sand towards him. Sarah, the eight-year-old is more deliberate. She is on a trek down the beach to find treasures: smooth pebbles of special shape or color, sand dollars, shells, seaweed, and maybe a blue crab. She returns with her bucket full, and she proceeds to sort her collection carefully into different kinds and categories. Cynthia watches both with motherly affection; her gaze shifts to the surf. She delights in the almost mesmerizing rhythm: rolling in and out. She wonders, at an almost subconscious level, what makes the waves adopt that particular tempo. Dan, the husband, chooses a spot well up from the water for their base, to avoid having to move as the tide comes in. He erects the sun-shade, trying a number of different rocks till he finds the ones that best keep it upright in the soft sand. He lies alongside his wife where the shadow will continue, even as the sun moves in the sky, to protect his fair skin from excessive ultra-violet radiation caused by the antarctic ozone hole.
Humans experiment from their earliest conscious moments. They are fascinated by regularities perfect and imperfect, and by similarities and distinctions. In children we call this play. In adults it is often trial and error devoted to a specific purpose, but sometimes it is simply a fascination that seeks no further end than understanding. We are creatures who want to know about the regularities of the world. And the way we find out about them is largely by experiment.
Induction is often touted as the defining philosophical method of natural science. It takes little thought and no detailed philosophical analysis to recognize that the deductive logic of the syllogism is inadequate for the task of discovering general facts about the natural world. All boggles are biggles, no baggles are biggles, therefore no baggles are boggles, is the stuff of IQ tests, not a way to understand the universe. By contrast, induction, the generation of universal laws or axioms from the observation of multiple specific instances, is both more fraught with logical difficulty and also vastly more powerful. But as a practical procedure it is hardly more than a formalization of the everyday processes of discovery illustrated by our Australian beach.
bacon1.png
Figure 2.2: Francis Bacon
Francis Bacon (1561-1626) is often credited with establishing the inductive method as primary in the sciences, and thereby laying the foundations of modern science. Here is what Thomas Macaulay, in his (1837) essay thought of that viewpoint.
The vulgar notion about Bacon we take to be this, that he invented a new method of arriving at truth, which method is called Induction, and that he detected some fallacy in the syllogistic reasoning which had been in vogue before his time. This notion is about as well founded as that of the people who, in the middle ages, imagined that Virgil was a great conjurer. Many who are far too well-informed to talk such extravagant nonsense entertain what we think incorrect notions as to what Bacon really effected in this matter.
The inductive method has been practiced ever since the beginning of the world by every human being. It is constantly practiced by the most ignorant clown, by the most thoughtless schoolboy, by the very child at the breast. That method leads the clown to the conclusion that if he sows barley he shall not reap wheat. By that method the schoolboy learns that a cloudy day is the best for catching trout. The very infant, we imagine, is led by induction to expect milk from his mother or nurse, and none from his father26.
Bacon thought and claimed that his analysis of Induction provided a formulation of how to obtain knowledge. That's why he named what is perhaps his crowning work, the "New Organon", meaning it was the replacement for the old "Organon", the collection of Aristotle's works on logic, which dominated the thinking of the schoolmen of Bacon's day. Bacon did not invent or even identify Induction. It had in fact already been identified by Aristotle himself, as Bacon well knew. Bacon thoroughly analyzed induction. He offered corrections to the way it was mispracticed, emphasizing the need for many examples, for caution against jumping to conclusions, and for considering counter-examples as importantly as confirmatory instances. He advocated gathering together tables of such contrasting instances, almost as if by a process of careful accounting one could implement a methodology of truth. These methodological admonitions are interesting and in some cases insightful, but they fall far short of Bacon's hopes for them. Scientists don't need Bacon to tell them how to think. And they didn't in 1600. What it seems philosophers did need to be told, or at any rate what is arguably Bacon's key contribution, is captured in his criticism of prior views about the ends, that is purposes, of knowledge. He says that philosophy was considered "... a couch whereupon to rest a searching and restless spirit; or a terrace for a wandering and variable mind to walk up and down with a fair prospect; or a tower of state for a proud mind to raise itself upon; or a fort or commanding ground for strife and contention; or a shop for profit or sale; and not a rich storehouse for the glory of the Creator and the relief of man's estate."27 For the schoolmen and generations of philosophers before them, all the way back to Aristotle, true learning was for the development of the mind, the moral fiber, and the upright citizen, not for practical everyday provisions. The Christianized version was, everyone agreed, for the glory of the Creator. Bacon's innovation was that science must also be for the relief of man's estate; that it must be practical. This insistence on the practical transformed speculative philosophy into natural science. Macaulay's summary is this.
What Bacon did for inductive philosophy may, we think, be fairly stated thus. The objects of preceding speculators were objects which could be attained without careful induction. Those speculators, therefore, did not perform the inductive process carefully. Bacon stirred up men to pursue an object which could be attained only by induction, and by induction carefully performed; and consequently induction was more carefully performed. We do not think that the importance of what Bacon did for inductive philosophy has ever been overrated. But we think that the nature of his services is often mistaken, and was not fully understood even by himself. It was not by furnishing philosophers with rules for performing the inductive process well, but by furnishing them with a motive for performing it well, that he conferred so vast a benefit on society.
What Bacon was advocating was knowledge that led to what we would call today technology. This emphasis has drawn the fire of a school of modern critics of science as a whole (part of the Science Studies movement) whose argument is that science is not so much about knowledge as it is about power. Despite Francis Bacon's many failings of legal and personal integrity, there seems little reason to question the sincerity of his avowedly humanitarian motivation towards practical knowledge. He lived in the court of monarchic power and rose to become the most powerful judge in England before his conviction for corruption and bribery. So he was no naive idealist, and is a natural target for the suspicions of the science critics. But he was at the same time a convinced Christian who would not have been insensitive to the appeal of a philosophy motivated by practical charity, especially since it supported the escape from scholasticism that he also yearned for. Whatever may have been the sincerity of Bacon's theological arguments in favor of practical knowledge, there can be little doubt that they furnished a persuasive rationale that helped to establish the course of modern science, and that persists today.
Technology demands reproducibility. Technology has to be based upon a reliable response in the systems that it puts into operation. Technology seeks to be able to manipulate the world in predictable ways. The knowledge that gives rise to useful technology has to be knowledge about the world in so far as it is reproducible and gives rise to tangible effects. These are precisely the characteristics of modern science.
When we talk about experiments, however, we normally conjure up visions of laboratories with complicated equipment and studious, bespectacled, possibly white-coated scientists; not a day at the beach. The practical experimentation at the beach which is our symbol for the acquisition of everyday knowledge does not draw strong distinctions between the levels of confidence with which we expect the world to follow our plans. At any time we hold to a vast array of beliefs with a wide spectrum of certainties from tentative hypothesis to unshakable conviction; and most often we draw little conscious distinction between them. In most cases, we have no opportunity to do so, since the press of events obliges us every moment to make decisions about our conduct based on imperfect and uncertain knowledge. Establishing confidence in reproducible knowledge, certain enough for practical application, and meeting our expectations for natural science, requires a more deliberate approach to experimentation. This (capitalized) Experiment is a formalization of the notion of reproducibility.
A formal Experiment is generally conducted in the context of some already-articulated theoretical expectation. It can be considered the opposite end of a spectrum of different degrees of deliberateness in experimentation. The idle play of the beach is the other end. And in between are random exploratory investigations, fact and specimen gathering, systematic documentation and measurement, the trial and error of technique development and instrumentation, and the elimination of spurious ideas and mistakes.
At its purest an Experiment is devised specifically to test a theoretical model or principle. Isaac Newton's famous demonstration that white light is in fact composed of a spectrum of light of different colors is often cited as an illustration of experimental investigations leading up to a "crucial experiment". In his letter to the Royal Society of February 1672 he relates in a personal story-telling style his initial experiments with the refraction of light through a prism, and his demonstration by careful measurement that the greater than two degree spread of the refracted colors could not be caused by the angular size of the sun's disc. He talks about various ideas he ruled out as possible explanations of the observations and then says: 28
Figure 2.3: Newton's drawing of his Experimentum Crucis, published much later in his Opticks (1704) .
NewtonPrism.png
The gradual removal of these suspitions, at length led me to the Experimentum Crucis, which was this: I took two boards, and placed one of them close behind the Prisme at the window, so that the light might pass through a small hole, made in it for the purpose, and fall on the other board, which I placed at about 12 feet distance, having first made a small hole in it also, for some of that Incident light to pass through. Then I placed another Prisme behind this second board, so that the light, trajected through both the boards, might pass through that also, and be again refracted before it arrived at the wall. This done, I took the first Prisme in my hand, and turned it to and fro slowly about its Axis, so much as to make the several parts of the Image, cast on the second board, successively pass through the hole in it, that I might observe to what places on the wall the second Prisme would refract them. And I saw by the variation of those places, that the light, tending to that end of the Image, towards which the refraction of the first Prisme was made, did in the second Prisme suffer a Refraction considerably greater then the light tending to the other end. And so the true cause of the length of that Image was detected to be no other, then that Light consists of Rays differently refrangible, which, without any respect to a difference in their incidence, were, according to their degrees of refrangibility, transmitted towards divers parts of the wall29.
Even this report and Newton's conclusions were not without controversy. Others seeking to reproduce his results observed different dispersions of the light, presumably because of using prisms with different angles. There followed a correspondence lasting some years, but in a remarkably short time the acceptance of this demonstration became practically universal because the key qualitative features, and by attention to the full details even the quantitative aspects, could be reproduced at will by experimenters with only a moderate degree of competence.

2.2  Is reproducibility really essential to science?

Observational Science

Most thoughtful people recognize the crucial role that repeatable experiments play in the development of science. Nevertheless, there arises, an important objection to the view that science is utterly dependent on reproducibility for its operation. The objection is this. What about a discipline like astronomy? The heavenly bodies are far outside our reach. We cannot do experiments on them, or at least we could not in the days prior to space travel and we still cannot for those at stellar distances. Yet who in their right mind would deny to astronomy the status of science?
Or consider the early stages of botany or zoology. For centuries, those disciplines consisted largely of systematic gathering of samples of species; cataloging and classifying them, not experimenting on them. Of course today we do have a more fundamental understanding of the cellular and molecular basis of living organisms, developed in large part from direct manipulative experiments. But surely it would be pure physicist's arrogance to say that botany or zoology were not, even in their classification stages, science.
In short, what about observational sciences? Surely it must be granted that they are science. If they are exceptions to the principle that science requires reproducibility then that principle rings hollow.
Some commentators find this critique so convincing, that they adopt a specialized expression to describe the type of science that is based on repeatable experiments. They call it "Baconian Science". The point of this expression is to suggest that there are other types of science than the Baconian model. What I suppose people who adopt this designation have in mind is observational sciences. They think that observational sciences, in which we can't perform experiments on the phenomena of interest at will, don't fit the model of reproducibility. There is some irony in using the expression with this meaning, since actually Bacon was at great pains to emphasize the systematic gathering of observations, without jumping to theoretical conclusions, so he certainly did not discount observational science, even though he did emphasize the motivation of practicality.
However, we need to think carefully whether observational sciences are really exceptions to reproducibility. Let us first consider astronomy. It is an appropriate first choice because in many ways astronomy was historically the first science. Humans gazed into the heavens and pondered on what they saw. The Greeks had extensive knowledge of the constellations and their cycles. And it was the consideration of the motions of the planets, more than anything else that led to the Newtonian synthesis of gravity and dynamics. But astronomy, considered in its proper historical context, is not an exception to the scientific dependence on reproducibility. Far from it. Astronomy was for the pre-industrial age the archetype of reproducibility. It was just because the heavens showed remarkable systematically repeated cycles that it commanded the attention of so many philosophers in attempts to explain the motions of the heavenly bodies. It was because the repeatability gave astronomers the ability to predict with amazing precision the phenomena of the heavens that astronomy appeared almost mystical in its status.
What is more, the independence of place and observer was satisfied by astronomy with superb accuracy. Better than almost all other phenomena, the sky looks the same from where ever you see it. As longer distance travel became more commonplace, the systematic changes of the appearance of the heavens with global position (latitude for example) were soon known and relied upon for navigation. And what could be more common to the whole of humanity than the sky?
Far from being an exception to the principle of reproducibility, astronomy's success depends upon that principle. Astronomy insists that all observers are going to see consistent pictures of the heavens when they observe. Those observations are open to all to experience (in principle). And those observations can be predicted ahead of time with great precision.
One way to highlight the importance of reproducibility in the context of astronomy is to contrast Astronomy (the scientific study of the observational universe) with Astrology (the attempt to predict or explain human events from the configurations of the heavenly bodies). Many people still follow assiduously their daily horoscope. Regarded as cultural tradition, that is probably no more harmful than wondering, on the feast of Candlemas, if the groundhog Puxatawney Phil saw his shadow; and recalling that if so, by tradition there will be six more weeks of winter. Astrology, for most people, is a relatively harmless cultural superstition. But surely no thinking person today would put forward astrology as a science. Its results are not reproducible. Its predictions appear to have no value beyond those of common sense. And its attempts to identify shared particular characteristics of people born in certain months simply don't give reliable results. Once upon a time there was little distinction between astronomy and astrology. Their practice in the pre-scientific age seems to be a confusing mix of the two. A major success of the scientific revolution was the disentangling of astronomy and astrology. The most important principle that separates the two activities is that astronomy is describing, systematizing and ultimately explaining the observations of the heavens in so far as they are reproducible and clear to all observers.
There are, of course, unique phenomena in astronomy. Supernovae, for example, each have unique features, and are first observed on a particular date. In that sense they are observations of natural history. On 4th July 1054, astronomers in China first observed a new star in the constellation of Taurus. Its brightness grew visibly day by day. During its three brightest weeks it was reported as visible in daylight, four times brighter than the evening star (Venus). It remained visible to the naked eye for about two years. It is thought that Anastasi Indian art in Arizonan pictographs also records the event. Surprisingly, there seem to be no European records of the event that have survived.30
If this were the only supernova ever observed, then we would probably be much more reticent to regard the event with credence. But there are approximately twenty different recorded supernovae (or possibly novae) in our galaxy during the 2000 years before 1700. And with modern telescopes, supernovae in other galaxies can also be observed fairly frequently.
The SN1054 supernova is probably the best known because it gave rise to the beautiful Crab Nebula discovered in 1731. That gas in the Nebula is expanding was established in the early 20th century by observing the line splitting caused by the Doppler effect. The nearer parts of the Nebula are moving towards us and the further parts away from us. In 1968 a new type of pulsing radio emission was discovered coming from the center of the Nebula. This Crab Pulsar is also observable in the visible spectrum. It is now known to be an extremely compact neutron star, rotating at an astonishing 30 times per second. 31
crabnasa.png
Figure 2.4: Photograph by the Hubble Telescope of the Crab Nebula.
One can get readily accessible reproducible evidence of the date of the SN1054 supernova. The expansion rate of the Crab Nebula can be established by comparing photographs separated in time. One can then extrapolate that expansion backward and discover when the now-expanding rim must have been all together in the local explosion. This process, applied for example by an undergraduate at Dartmouth College to photographs taken 17 years apart, gives a date in the middle of the 11th century. In 100% agreement with the historic record.32 [For technical precision we should note that, since the nebula is 6000 light years away, the explosion and the emission of the light we now see took place 6000 years earlier.]
Notice the following characteristics. First the Crab Nebula's supernova, though having its own unique features, was an event of a type represented by numerous other examples. Second, the event itself was observable to, and recorded by, multiple observers. Third, the supernova left long-lived evidence that for years could be seen by anyone who looked, and still gives rich investigation opportunities to experts from round the world, who simply have to point their telescope in the right direction. These are the characteristics of reproducibility in the observational sciences.
What about botany and zoology, in their collection and classification stages? Again, careful consideration convinces us that these do rely on reproducibility. If only a single specimen is available, it remains largely a curiosity. Who is to say that this not simply some peculiar mutant, or even a hoax? But when multiple similar specimens are found, then it is possible to detect what is common to all specimens and to discount as individual variation those characteristics that are not. Indeed, in the life sciences the best option is to have breeding specimens, which guarantee the ability to establish new examples for which the reproducible characteristics are those on which the scientific classification is based.
All right, what about geology then? Its specimens don't reproduce. But again the observation of many different examples of the same types of rock, or formation, or other phenomenon, is essential to its scientific progress. In its earliest stages, before geophysics had more direct physical descriptions of its processes, geology progressed as a science largely by the identification of multiple examples of the same processes at work, that is by reliance on repeatability. As the scientific framework for understanding the earth's formation was gradually built up, the systematic aspects of the rock formations began to become clear. Events could be correlated to produce an ordered series of ages. Then additional techniques such as radioactive dating enabled geologists to assign a quantitative date to the different geological ages, based on multiple assessments of the time that must have elapsed since the formation of rocks identified as belonging to each period. All of this process requires the ability to make multiple observations, observe repeatable patterns, and perform repeatable physical tests on the samples.
So observational science requires multiple repeatable examples of the phenomenon or specimen under consideration. It does not require that these can be produced at will in the way that a laboratory experiment can in principle be performed at any hour on any day. Observations may be constrained by the fact that the examples of interest occur only at certain times (for example eclipses) or in certain places (for example in specific habitat), over which we might have little or no control. But it does require that multiple examples exist and can be observed.

Randomness

A second important objection to the assertion that science requires reproducibility concerns the occurrence in science of phenomena that are random. By definition, such events are not predictable, or reproducible, at least as far as their timing is concerned.
If science is the study of the world in so far as it is reproducible, how come probability, the mathematical embodiment of randomness, the ultimate in non-reproducibility, plays such a prominent role in modern physics? Here, I think, it is helpful to take for a moment a historical perspective of science. Pierre Laplace is famous (amongst other things) for his encounter with the emperor Napoleon. Laplace explained to him his deterministic understanding of nature. Bonapart is reported to have asked, "But where in this scheme is the role for God", to which Laplace's response was "I have no need of that hypothesis" . Regardless of his theological position, Laplace's philosophical view of science was rather commonplace for his age. It was that science was in the process of showing that the world is governed by a set of deterministic equations. And if one knew the initial conditions of these equations for all the particles in the universe, one could, in principle at least, solve those equations and thereby predict, in principle to arbitrary accuracy, the future of everything. In other words, Laplace's view, and that of probably the majority of scientists of his age, was that science was in the business of explaining the world as if it were completely predictable, subject to no randomness.
This view persisted until the nineteen twenties, when the formulation of quantum mechanics shocked the world of science by demonstrating that the many highly complicated and specific details of atomic physics could be unified, explained, and predicted with high accuracy using a totally new understanding of reality. At the heart of this new approach was a concession that at the atomic level events are never deterministic; they are always predictable only to within a significant uncertainty. Heisenberg's uncertainty principle is the succinct formulation of that realization. Quantum mechanics possesses the mathematical descriptions to calculate accurately the probability of events but not to predict them individually in a reproducible way. Perfect reproducibility, it seems, exists only as an ideal, and that ideal is approached only at the macroscopic scale of billions of atoms, not at the microscopic scale of single atoms.
Where does that leave the view that science demands reproducibility? Did 20th century science in fact abandon that principle?
It seems clear to me that science has not at all abandoned the principle. The principle is as intact as ever that science describes the world in so far as it is reproducible. For hundreds of years, science pursued the task with spectacular success. Quantum mechanics, shocking though it seemed and still seems, did not halt or even alter the basic drive. What it did was to show that the process of describing the world in reproducible terms appears to have limits, fundamental limits, that are built into the fabric of the universe. The quantum picture accepts that there are some boundaries beyond which our reproducible knowledge fails in principle (not just for technical reasons). Even as innovative a scientist and supple a philosopher as Albert Einstein was repelled by the prospect, famously resisting the notion that the randomness of the world is in principle impenetrable with his comment "God does not play dice". Einstein, like scientists before and after him was committed to discovering the world in so far as it is reproducible, not arbitrary. The overwhelming opinion of physics today, though, is that Einstein was wrong in this epigram. God does play dice, in the sense that some things are simply irreproducible .
But that does not stop science from proceeding to explore what is measurable and predictable. Science does that first by pressing up to the limits of what is reproducible. If individual events are not predictable, it calculates the probabilities of events. Quantum theory provides this reproducible measure in so far as reproducibility exists. The interesting thing about Quantum mechanics is that it is governed by deterministic equations. These equations are named after the great physicists Erwin Schrödinger and Paul Dirac. A problem in Quantum mechanics can be solved by finding a solution of these equations. The equations take an initial state of the system and then predict the entire future evolution of that state from that time forward. This is such a deterministic process that some people argue that Quantum mechanics does not undermine determinism. But such a view misses the key point. The function that is solved for and the future system state that is predicted are no longer the definite position or velocity of a particle, or one of the many other quantities that we are familiar with in classical dynamics (or the everyday world). Instead it is essentially the probability of a particle being in a certain position or having a certain velocity. This is an example of science pressing up against the limits of reproducibility. The world is not completely predictable even in principle by mathematical equations, but science wants to describe it as completely as possible, to the extent that it is. So when up against the non-predictability, science invokes deterministic mathematics, but uses the mathematics to govern just the probability of the occurrence of events. Probability is, in a sense, the extent to which random events display reproducibility. Science describes the world in terms of reproducible events to the extent that it can be described that way.
And science proceeds, second, by pressing on into the other areas of scientific investigation that still lie open to reproducible description: describing and understanding them as far as they are indeed reproducible.

The history of nature

A third important challenge to the principle of reproducibility lies in the types of events that are inherently unique. How could we possibly have knowledge of more than one universe? Therefore how can reproducibility be a principle applied to the Big Bang origin of the universe? Or, perhaps less fundamentally, but probably just as practically, how can we apply principles of repeatability to the origin of life on earth, or to the details of how the earth's species got here?
I think these issues all boil down to the question "What about natural history?" Is the history of nature part of science?
It is helpful to think first about the ways that science can tell us about the past. Today, our ability to analyze human DNA has become extremely important in legal evidence. It can help prove or disprove the involvement of a particular person in a crime under investigation. The high profile cases tend, of course, to be capital cases, but increasingly this type of forensic evidence is decisive in a wider variety of situations. Obviously this is an example of a way in which scientific evidence is extremely powerful in telling us something about history - not as overwhelming as it is often portrayed in the popular TV series such as CSI, but still very powerful. Important as this evidence may be, it still depends on the rest of the context of the case which is what determines the significance of the DNA test results. Moreover, the laboratory results themselves are rarely totally unimpeachable. We have to be sure that the sample really came from the place the police say it did. We have to be sure it was not tampered with before it got to the lab or while it was there. We have to be sure that the results are accurately reported, and so on. So science can provide us with highly persuasive evidence, in part because of its clarity and the difficulty of faking it. But when we adduce scientific evidence for specific unique events of history (even recent history) our confidence is, in principle, less than if we had access to multiple examples of the same kind of phenomenon.
To illustrate that difference in levels of confidence, ask yourself how successful a defense lawyer would likely be if they tried to defend against DNA evidence by arguing that it is a scientific fallacy that DNA is unique to each individual. Trying to impeach the principles of science would be a ridiculously unconvincing way to try to discredit evidence in a criminal case. Those principles have been established by innumerable laboratory tests by independent investigators over years of experience and subjected to intense scrutiny by experts. General principles of science, such as the uniqueness of DNA, need more compelling investigation and evidence to establish them than we require for everyday events. But they get that investigation, and the support of the wider fabric of science into which they are woven. So, once established, we grant them a much higher level of confidence. What's more, if there is reason to doubt them, we can go back and get some more evidence to resolve the question.
But notice that our confidence in scientific principles is not the same as confidence in knowledge of the particular historical question. We are confident that insofar as the world is reproducible, and can thus be scientifically described, DNA is unique to the individual. But that does not automatically decide the legal case. If in fact there is other extremely compelling evidence that contradicts the DNA evidence, we are likely to conclude that in this specific instance there is something that invalidates the DNA evidence and we should discount it.
In summary, then, for specific unique events of history, evidence based on scientific analysis can be important, but is not uniquely convincing.
However, even though it contains some questions about unique events, much of natural history is not of this type. Much is about the broad sweep of development of the universe, the solar system, the planet, or the earth's creatures. In other words, questions of natural history are usually about generalities, not particularities, about issues giving rise to repeated observational examples, not single instances. For these generalities, science is extremely powerful, but continues to rely upon its principles of reproducibility and clarity.
We believe the universe is expanding from an initial event, starting some 13 billion years ago, because we can observe the Doppler shifts of characteristic radiation lines that are emitted from objects over an enormous range of distances from our solar system. We see that the farther away the object is, the more rapidly it is receding from us, as demonstrated by the frequency downshift of the light we see. Anyone, anywhere can observe these sorts of effects with telescopes and instruments that these days are relatively commonplace. The observed systematic trend is consistent with a more-or-less uniform expansion of the universe. If we project back the tracks of the expansion, by mentally running the universe in reverse, we find that our reversing tracks more-or-less bring all the objects together at the same time: the time that we identify as the Big Bang, or the age of the universe. This picture is confirmed by thousands of independent observations of different stars and astrophysical objects. Thus, the Big Bang theory of the origin of the universe is a generality: that the universe had a beginning roughly 13 billion years ago, when all of the objects we can see were far closer together. And it is a generality confirmed by many observational examples that show the same result.
Vostok-ice-core-petit.png
Figure 2.5: Temperature, atmospheric carbon-dioxide, and dust record from the ice cores from Vostok, Antarctica.
What about earth's history? The history of the earth's climate is a topic of great current interest. How do we know what has happened to the climate in past ages? This information comes in all sorts of ways from tree rings to paleontology. Perhaps the most convincing and detailed information comes from various types of "cores" sampled from successively deposited strata. Examples include ice cores, and ocean sediments. Ice cores as long as three kilometers, covering the last three quarters of a million years have been drilled from the built-up ice arising from annual snowfall in Antarctica. The resulting history of the climate is laid out in amazing detail. By analyzing the fraction of different isotopes of hydrogen and oxygen in the water, scientists can estimate the mean temperature. By analyzing trapped gases, the fraction of carbon dioxide can be determined, and by observations of included microscopic particles, the levels of dust in the atmosphere can be documented, in each case over the entire history represented by the core. Figure 2.5 illustrates the results from the core analyzed in 1999.33
These results are reproducible. If a core is drilled in the same place, the results one gets are the same. Actually, the longer (3km) core was drilled more recently (2004) in a different place in Antarctica than the one I have illustrated,34 but shows almost exactly the same results for the periods of history that overlap. Moreover, deep-sea-bed sediment cores from all over the globe show global ice mass levels deduced from oxygen isotopes that correlate extremely well with the ice-core data for their history, but stretch back to 5 million years ago. The rapid progress in these observations and analysis over the past decade has enabled us to build up a remarkable record of the climatic history of the earth. It enables us to construct well-informed theories for what factors influence the different aspects of climate, and to say how the climate varied through time. But these vital additions to our knowledge are still about broad generalities: the global or regional climate, not usually the highly specific questions that preoccupy historians. For example, the uncertainties in the exact age of the different ice-core samples can be as large as a thousand years or more. In the scheme of the general picture of what happened in the last million years, this is a negligible uncertainty; but on the timescale of human lives (for example) it is large.

2.3  Inherent limitations of reproducibility

We have seen some examples of the great power of science's reliance upon reproducibility to arrive at knowledge. These examples are not intended to emphasize science's power; such a demonstration would be superfluous for most modern minds. They are to show the importance of reproducibility. We are so attuned to the culture of science that we generally take this reproducibility for granted. But we must now pause to recollect both that this reproducibility is not obvious in nature, and that in many fields of human knowledge the degree of reproducibility we require in science is absent.
Of course, substantial regularity in the natural world, and indeed in human society, was as self-evident to our ancestors as it is to us. But the extent to which precise and measurable reproducibility could be discovered and codified was not. The very concept and expression `law of nature' dates back only to the start of the scientific revolution, to Boyle and Newton. And in its original usage, it intended as much the judicial meaning of a legislated edict of the Creator as the impersonal physical principle, or force of nature that now comes to mind. Indeed, a case can be made that it was in substantial degree the expectation that law governed the natural world, fostered by a theology of God as law-giver, that provided the fertile intellectual climate for the growth of science. As late as the nineteenth century, Faraday motivated his search for unifying principles, and explained his approach to scientific investigation, by statements like "God has been pleased to work in his material creation by laws". By referring to God's pleasure, Faraday was not in the least intending to be metaphorical, and by laws he meant something probably much less abstract than would be commonplace today.
In drawing attention now to disciplines in which the reproducibility expected in science is absent, I want to start by reiterating that this absence does not in my view undermine their ability to provide real knowledge. The whole point of my analysis is to assert that non-scientific knowledge is real and essential. So I beseech colleagues from the disciplines I am about to mention to restrain any understandable impulse to bristle at the charge that their disciplines are not science. I remind you that I am using the word science to mean natural science, and the techniques that it depends upon. If the semantics is troubling, simply insert the qualification "natural" in front of my usage. Let us also stipulate from the outset that there are parts of each of these disciplines that either benefit from scientific techniques or indeed possess sufficient reproducibility to be scientifically analyzed. I am not at all doubting such a possibility. I am simply commenting that the core subject areas of these disciplines are not most fruitfully studied in this way for fundamental reasons to do with their content.
In his (1997) graduate text Science Studies, introducing various philosophical and sociological analysis of science, David J Hess acknowledges without hesitation the difficulties in applying scientific analysis to other disciplines "Probably the greatest weakness in this position comes when the philosophy of science is generalized from the natural sciences to the human sciences". He says specifically "Many social phenomena are far too complicated to be predictable".35 In other words, in my terminology, these phenomena are not science. Yet a few pages later he says. "One of the reasons social scientists lose patience with philosophers of science is that we are constantly told that we are in some sense deficient scientists - we lack a paradigm, predictive ability, quantitative exactness, and so on - instead of being seen as divergent or different scientists".36 This is an argument about titles and semantics. Sociologists today acknowledge that sociology does not offer the kind of reproducibility that is characteristic of the natural sciences. They feel they must insist on the title of science, which I believe is because of the scientism of the age; without the imprimatur of the title they feel their discipline is in danger of being dismissed as non-knowledge. Yet they resent it when the essential epistemological differences between their field and science are pointed out. No wonder there are difficulties in this discussion. As a physical scientist, I need to keep out of this argument, but I will observe that if we disavow scientism, then the whole of this discussion becomes more tractable. It is no longer a problem for sociology to be recognized as a field of knowledge in which reproducibility is not available.
History is a field in which there is thankfully less resentment towards an affirmation that it is not science. Obviously history, more often than not, is concerned with unique events in the past that cannot be repeated. Here is a commentary by a historian on King James the Second's frequent remark to justify his intransigence: "My father made concessions and he was beheaded". Macaulay writes, "Even if it had been true that concession had been fatal to Charles the First, a man of sense would have remembered that a single experiment is not sufficient to establish a general rule even in sciences much less complicated than the science of government; that, since the beginning of the world, no two political experiments were ever made of which all the conditions were exactly alike ..."37 Macaulay's typical, but confusing, use of `science' has already been noted, but the point he makes very clearly is that there is no reproducibility in history. No more than a small fraction of its concerns benefit from analysis that bears the stamp of natural science. Yet no thoughtful person would deny that historical knowledge is true knowledge, that history at its best has high standards of scholarship and credibility, and that the study of history has high practical and theoretical value.
Similarly the study of the law, jurisprudence, is a field whose research and practice cannot be scientific because it is not concerned with the reproducible. The circumstances of particular events cannot be subjected to repeated tests or to multiple observations. Moreover, the courts do not have the luxury of being able indefinitely to defer judgement until sufficient data might become available. They have to arrive at a judgement that is binding on the protagonists even with insufficient data. Consequently, the legal system's approach to decision-making is very different from science's.
In Britain in the early 1980s the government of Prime Minister Margaret Thatcher introduced policies in line with Milton Friedman's economic theories, which the press was fond of referring to as the `monetarist experiment'. Here was what an economist must surely dream about: the chance to see an experimental verification of his theory. What was the result of this experiment? Was monetarism thereby confirmed or refuted? To judge by current economic opinion, it seems neither. Economists don't really know how to assess the outcome unambiguously, because this was a real economy with all sorts of extraneous influences; and what is most important, one can't keep trying repeatedly till one gets consistent results. It may have been an experiment, but it was not truly a scientific experiment. Economics is an interesting case here, because economists have large quantities of precise measured data and usefully employ highly sophisticated mathematics for many of their theories, a trait that they share with some of the hardest physical scientists. Economics is a field of high intellectual rigor. But the absence of an opportunity for truly reproducible tests or observations and the impossibility of isolating the different components of economic systems means that economics as a discipline is qualitatively different from science.
Politics is to many a physical scientist baffling and mysterious. Here is a field, if there ever was one, that is the complete contradiction of what scientists seek in nature. In place of consistency and predictability we find pragmatism and the winds of public opinion. In place of dispassionate analysis, we have the power of oratory. And once again, nothing in the least approximating the opportunity for reproducible tests or observations offers itself to political practitioners. It seems a great pity, and perhaps a sign of wistful optimism, not to mention the scientism that is our present subject, that the academic field of study is referred to these days almost universally as Political Science.
We will discuss more, equally important, examples of inherently non-scientific disciplines later. But these suffice to illustrate that not only is science not all the knowledge there is, it may not be even the most important knowledge. And however much we might hope for greater precision and confidence in the findings of the non-scientific disciplines, it is foolishness to think they will ever possess the kind of predictive power that we attribute to science. Their field of endeavor does not lend itself to the epistemological techniques that underlie science's reliable models and convincing proofs. They are about more indefinite, intractable, unique, and often more human problems. In short, they are not about nature.

HEAD NEXT