1 What makes a claim scientific?

The standard answer taught in most science classes is that science uses the scientific method: you observe, you form a hypothesis, you test it, you revise. This is not wrong, but it is not sufficient either. Astrology also makes claims about the world and invites empirical testing. Psychoanalysis also generates hypotheses and accumulates confirming evidence. What, precisely, distinguishes science from these?

Kitzmiller v. Dover Area School District (2005)

In October 2004 the Dover Area School Board in Pennsylvania ordered all ninth-grade biology teachers to read aloud, before any unit on evolution, a four-paragraph statement noting that “Darwin’s Theory” had “gaps” and recommending that students consider an alternative account of biological diversity called intelligent design, available in the school library in the textbook Of Pandas and People.1 Eleven Dover parents, led by Tammy Kitzmiller, sued the school board on the grounds that the policy established religion in violation of the First Amendment.2Tammy Kitzmiller, et al. v. Dover Area School District, et al. was tried before US District Judge John E. Jones III in Harrisburg, Pennsylvania, in late 2005. The trial — six weeks, bench-only, transcripts published in full — became the most thorough public proceeding ever conducted on the demarcation question. Plaintiffs called Kenneth Miller (Brown University) and Kevin Padian (Berkeley) to testify on what made evolutionary biology scientific; defendants called Michael Behe (Darwin’s Black Box, 1996) and Scott Minnich to defend intelligent design.3 Behe was forced to admit on cross-examination that a definition of science permissive enough to include intelligent design would also admit astrology. On 20 December 2005, Judge Jones issued a 139-page opinion ruling for the plaintiffs.4 Three findings constitute one of the most carefully reasoned legal applications of Popper’s demarcation criterion in any judicial record: (1) intelligent design is not science because it invokes supernatural causation, which by methodological commitment science excludes; (2) the irreducible complexity arguments — that certain biological structures (the bacterial flagellum, the blood-clotting cascade) cannot have evolved gradually — had been rebutted by working biologists with extensive cited literature; (3) the school board members had explicitly invoked their religious motivations in board meetings before retreating to “scientific” framings during litigation, which the court treated as evidence of the policy’s religious purpose under the Establishment Clause. The school board lost, paid roughly $1 million in attorneys’ fees and costs, and had been unseated by voters in the November 2005 elections (the verdict was issued after the new board took office). The case is the most institutionally serious test the demarcation question has had since Karl Popper, in Vienna in 1919, watched Alfred Adler explain any human behaviour by reference to the inferiority complex and concluded that psychoanalysis was unfalsifiable (treated below).

1.1 Popper, Adler, and the Vienna Origin of the Demarcation Criterion

In Vienna in 1919, the young Karl Popper worked briefly with Alfred Adler and watched psychoanalytic theory account for a case of child psychology.5 Popper found that Adler’s theory could explain any human behaviour whatsoever: aggression confirmed the inferiority complex; generosity confirmed overcompensation for the inferiority complex. There was no observation that could, even in principle, count against it. This was the experience that led Popper to formulate the falsifiability criterion: a genuine scientific theory must be capable of being refuted by observable evidence. Psychoanalysis, he argued, was not science — not because it was false, but because it was structured to be unfalsifiable. The psychoanalytic establishment did not accept this characterisation, and the debate about whether psychoanalytic claims can be empirically tested has continued for decades. Seymour Fisher and Roger Greenberg, in a 1977 meta-analysis, concluded that several specific Freudian claims had testable implications and had received partial empirical support.6 The demarcation problem — identifying a principled criterion that separates science from non-science — is harder than Popper’s clean distinction suggests: Thomas Kuhn observed that scientists regularly ignore apparently falsifying evidence during periods of normal science. The 2005 Kitzmiller trial (info box above) is the institutional test of Popper’s criterion against contemporary intelligent-design literature; Judge Jones’s opinion is in effect a 139-page philosophical-scientific application of the criterion to specific empirical claims.

1.2 Popper’s Criterion

Karl Popper’s answer, developed in The Logic of Scientific Discovery (German 1934, English 1959) and Conjectures and Refutations (1963), is that falsifiability distinguishes science from non-science.7 A claim is scientific if and only if there is some possible observation that could show it to be false. The mark of science is not certainty but vulnerability to refutation.

Popper arrived at this criterion by noticing the asymmetry between confirmation and falsification. No number of confirming observations can prove a general claim — you cannot verify that all swans are white by observing a million white swans. But a single observation of a black swan proves the claim false. Science progresses by trying to prove its claims false; surviving severe attempts at falsification is the best evidence a theory can have.

The black swan example is also the source of Nassim Nicholas Taleb’s The Black Swan (2007). When Dutch explorers reached Western Australia in 1697, they found black swans — refuting a belief held for centuries in Europe as a paradigm of certainty. Taleb uses this as a metaphor for high-impact, unexpected events that retrospective narratives make seem predictable.8

Apply the criterion: Is astrology scientific? Astrological claims are highly flexible — an astrologist can interpret almost any outcome as consistent with a horoscope. If a claim can accommodate any possible observation, it predicts nothing, and Popper would say it is not scientific. Is Freudian psychoanalysis scientific? Freud’s theory of the unconscious can explain any patient behaviour — aggression and submission, sexuality and its suppression — because it has concepts that can be applied in both directions. Popper saw this not as evidence of the theory’s power but as evidence of its unfalsifiability. Is string theory scientific? This is a live and serious debate: string theory has been enormously productive mathematically, but its predictions (extra dimensions at the Planck scale, supersymmetric particles) may be permanently beyond the reach of any conceivable experiment.

1.3 The Demarcation Problem

Popper’s criterion is elegant and influential, but philosophers of science have found it increasingly difficult to apply cleanly.

Auxiliary hypotheses. Scientific theories do not confront evidence alone; they confront it together with a package of auxiliary assumptions, instrument calibrations, and background theories. When a prediction fails, you don’t know whether the core theory is wrong or whether one of the auxiliaries is wrong. Uranus was discovered in 1781, and over the following decades its observed orbit drifted from the predictions of Newtonian mechanics. Did this falsify Newton? No — Urbain Le Verrier computed the position of an unseen perturbing planet, and on 23 September 1846 Johann Galle observed Neptune at the Berlin Observatory within a degree of the predicted location.9 The anomaly that might have falsified the theory instead supported an extension of it. This is Imre Lakatos’s point: scientists work within “research programmes” with a “hard core” they protect from falsification and a “protective belt” of auxiliary hypotheses they adjust.10

Kuhn’s Reply. Thomas Kuhn, in The Structure of Scientific Revolutions (1962), argues that Popper’s account bears no resemblance to how science actually works.11 Real scientists do not attempt to falsify their theories. When anomalies arise — observations that don’t fit the theory — scientists typically assume the problem lies with the experiment, the observer, or some peripheral assumption. They do not abandon the theory. This is not irrationality; it is the only way science can function. A theory that was abandoned at the first anomaly would never survive long enough to be tested properly.

1.4 What Distinguishes Science, Then?

If falsifiability is too simple a criterion, we need a more complex answer. Several features working together distinguish science from non-science, without any single one being individually sufficient:

  • Empirical content: scientific claims make specific, testable predictions about the observable world.
  • Internal coherence: scientific theories must be internally consistent.
  • Communal scrutiny: scientific claims are subjected to peer review, replication attempts, and criticism by a community of informed sceptics.
  • Self-correction: when evidence against a theory accumulates sufficiently, the theory is revised or abandoned. This happens slowly and imperfectly — but it happens.
  • Progressive extension: good scientific theories don’t just explain what they were designed to explain; they successfully extend to new domains (Newtonian mechanics applied to tides, to the Moon, to projectile trajectories simultaneously).

None of these is perfectly sharp, and the boundary between science and non-science is genuinely fuzzy in some cases. This is the demarcation problem. But fuzziness at the boundary doesn’t mean the distinction is meaningless: the centre of the category is clear even if the edges are not.

Forced Fork: Is String Theory Science?

Position A: Popper’s falsifiability criterion correctly identifies what distinguishes genuine scientific claims from non-science. A theory that cannot in principle be shown to be false by observation makes no real claim about the world. String theory, insofar as it makes no testable predictions, is not science — it is mathematics dressed in physical language. This criterion is clean, applicable, and explains why we rightly dismiss astrology while accepting general relativity.

Position B: Falsifiability is not the right criterion, or not sufficient by itself. Kuhn’s point stands: scientists do not actually try to falsify their theories, and a theory that was abandoned at the first anomaly would never have time to develop. What distinguishes science is something richer — a combination of empirical constraint, communal scrutiny, self-correction over time, and progressive extension to new domains. String theory may still be science even if currently untestable, because it satisfies the other criteria.

Choosing Position A commits you to the consequence that large parts of contemporary cosmology and theoretical physics fail to qualify as science, and to explaining why this is the right verdict rather than a reductio of the criterion. Choosing Position B commits you to specifying what combination of criteria does the work that falsifiability was meant to do, and to explaining why those criteria do not simply permit unfalsifiable but internally coherent systems (astrology, psychoanalysis) to count as science.

Note: “I find both positions unsatisfactory” is not an answer. Pick the less wrong position and defend it against the hardest objection.

1.5 Questions to Argue About

  • Is string theory science? It makes predictions, but they may be permanently untestable. Does this matter?
  • Popper says psychoanalysis is not scientific because it cannot be falsified. Does that mean Freud was wrong? Or can a non-scientific theory be true?
  • Kuhn argues that scientists don’t actually try to falsify their theories. Does this undermine Popper’s criterion, or does it just show that scientists don’t always practise what philosophers of science preach?
  • If there is no sharp boundary between science and non-science, what are the practical implications? Does it matter whether climate science, economics, or nutritional research count as “really scientific”?

2 What is the scientific method — and does science actually follow it?

Every school science curriculum teaches the scientific method: observe, hypothesize, experiment, conclude. This story is clean, linear, and rational. Every scientist learns it. Almost no scientist actually uses it in this sequence — and several philosophers of science have made distinguished careers from pointing this out.12 The story is, as a description of how science is actually done, substantially false — or at least radically incomplete.

Robert Millikan and the Oil Drop Experiment

Between 1909 and 1913, the American physicist Robert Millikan conducted his famous oil drop experiment, in which tiny electrically charged droplets of oil were suspended in an electric field to measure the charge of a single electron. Millikan won the Nobel Prize for the work in 1923. When historian of science Gerald Holton examined Millikan’s laboratory notebooks in the 1970s, he found that Millikan had recorded 175 measurements of which he published 58.13 The unpublished measurements were those that deviated significantly from the result Millikan expected; the published measurements clustered closely around the accepted value for the electron’s charge. Meanwhile, Millikan’s competitor Felix Ehrenhaft was publishing measurements suggesting the existence of “sub-electrons” with charges smaller than Millikan’s value — Millikan’s papers referred dismissively to Ehrenhaft’s data as artifacts of experimental error. In retrospect, Millikan’s published value was very close to the modern accepted value, and Ehrenhaft was wrong — but Millikan’s method of arriving at his result was not the neutral, systematic hypothesis-testing described in introductory science textbooks. Paul Feyerabend, in Against Method (1975), used exactly such cases to argue that science advances through a heterogeneous mixture of theoretical commitment, data selection, and rhetoric.

2.1 Bacon’s Inductivism

Francis Bacon (Novum Organum, 1620) proposed what became the standard picture: knowledge advances by the patient accumulation of observations, from which general laws are induced. Collect enough data, without prejudice, and nature’s laws will emerge. This is inductivism: from particular observations to general conclusions.14

The logical objection is Hume’s problem of induction: no number of particular observations can logically entail a general law. This is not merely a logical technicality; it means that scientific knowledge is always provisional, however massive the evidence base.

The practical objection is that pure inductivism doesn’t describe how science works. Scientists don’t observe without prejudice; they observe through the lens of existing theory, which tells them what to observe, how to measure it, and what counts as a relevant result. As Hanson put it, observation is “theory-laden.”15 Bacon’s ideal of the blank-mind observer is impossible and, if it were possible, would produce meaningless data.

2.2 The Hypothetico-Deductive Model

The more sophisticated account is hypothetico-deductive: the scientist starts with a hypothesis (often creatively generated, not induced from observations), deduces what observations the hypothesis would predict, and tests those predictions. This is closer to what Einstein and Darwin actually did — they had theoretical ideas first and looked for evidence to test them.

Einstein is reported by Ilse Rosenthal-Schneider to have answered, when asked what he would have done if Eddington’s 1919 eclipse observations had failed to confirm general relativity, that he would have felt sorry for the dear Lord — the theory was correct anyway.16 Einstein is not being arrogant on this telling; he is reporting a profound internal conviction about theoretical coherence. From a Popperian perspective, the conviction is alarming.

But this model still produces an idealised picture. The step from “hypothesis” to “test” is not a simple logical deduction. Experiments must be designed, which requires decisions about what variables to control, what instruments to use, and what counts as a significant result. These decisions embed assumptions. The famous example: Millikan’s oil-drop experiments (1909–1913) to measure the electron’s charge involved a process of selection — he discarded data that didn’t fit what he expected. This is neither fraud nor good practice unambiguously; it is the messy reality of how knowledge gets made.

2.3 Feyerabend and “Anything Goes”

Paul Feyerabend, in Against Method (1975), pushed the critique to its extreme. There is no single scientific method; the history of science shows that progress has been achieved by violating any methodological rule you care to name. Galileo’s Dialogue Concerning the Two Chief World Systems (1632) is Feyerabend’s central exhibit: Galileo defended Copernicanism in a literary dialogue between Salviati (the Copernican spokesman), Sagredo (the receptive layman), and Simplicio (whose name punningly marks him as the simpleton); his strongest physical argument for the Earth’s motion — that the tides are caused by the daily and annual motions sloshing the oceans — was wrong, while several apparently strong contemporary objections to Copernicanism (the absence of detectable stellar parallax; the lack of a felt wind from the Earth’s rotation) were on the available evidence reasonable. Galileo, on Feyerabend’s reading, used propaganda and psychology as well as observation; he was sometimes wrong about the data while being right about the theory. The rule “believe what the observations show” would have prevented the Copernican revolution.17

Feyerabend’s conclusion is deliberately provocative:

“the only principle that does not inhibit progress is: anything goes.” — Paul Feyerabend, Against Method, Introduction.18

He is not saying science is irrational. He is saying that the rationality of science cannot be captured in any neat set of rules.

2.4 How Kepler Actually Worked

Johannes Kepler’s discovery of elliptical planetary orbits (published in Astronomia Nova, 1609) is a case study in how science is actually done versus how it is described. Kepler spent eight years trying to make Tycho Brahe’s precise observational data fit circular orbits — the traditional Platonic assumption that the heavens required perfect circular motion. When the data consistently showed an 8-arc-minute discrepancy from circular predictions (a discrepancy he knew was too large to be observational error, because Brahe’s instruments were far more accurate than their predecessors), he was forced to abandon the circle and try the ellipse.19

Kepler’s path was neither pure induction (he had a prior commitment to circle-perfect heavens that he was reluctant to abandon) nor clean hypothetico-deduction (the “hypothesis” of elliptical orbits came from trying things that fit, not from a theoretical framework that predicted ellipses). It was stubbornness, creativity, data-respect, and eventually the willingness to overturn centuries of assumption — a process that looks nothing like the textbook method.

2.5 What Counts as an Explanation? Hempel’s Deductive-Nomological Model

Method gets you to a hypothesis that survives testing. But suppose it does survive — what kind of answer have you given? Ignaz Semmelweis, working in the Vienna General Hospital between 1844 and 1848, was confronted with a puzzle: the First Maternity Division, staffed by physicians and medical students, had a maternal mortality from childbed fever of around 18% in April 1847; the Second Division, staffed by midwives, ran at roughly 2%.20 Semmelweis’s eventual hypothesis was that the physicians, who arrived from morning autopsies, were carrying what he called “cadaverous particles” on their hands and introducing them into wounds during examination. He instituted handwashing in chlorinated lime in mid-May 1847; mortality in the First Division dropped to 1.2% by July. In what sense had he explained childbed fever?

Carl Hempel, writing in 1966, argued that the structure of every genuine scientific explanation is the same. An explanation is a deductive argument: its premises (the explanans) must include at least one general law together with statements of antecedent conditions, and its conclusion (the explanandum) must follow logically. Hempel’s own term for the pattern is canonical:

“Explanatory accounts of this kind will be called explanations by deductive subsumption under general laws, or deductive-nomological explanations. (The root of the term ‘nomological’ is the Greek word ‘nomos’, for law.) The laws invoked in a scientific explanation will also be called covering laws for the explanandum phenomenon.” — Carl G. Hempel, Philosophy of Natural Science (1966), Ch. 5.21

Hempel opens the book with Semmelweis precisely because the case fits the pattern so cleanly. Semmelweis’s eventual diagnosis — that childbed fever was caused by “decomposed animal matter” (his “cadaverous particles”) carried from the dissection room into wound surfaces — implicitly invokes the general law that introduction of such matter into the bloodstream produces blood poisoning, plus the particular fact that medical students went directly from cadavers to the maternity ward. Together they entail the high mortality in the First Division.22 The diagnosis was correct in its operative content (handwashing prevents transmission) but its proposed mechanism was pre-bacterial: Semmelweis had no germ theory and could not say what the “particles” actually were. The Vienna medical establishment rejected his findings, his contract was not renewed, and he died in a Viennese asylum in August 1865 — beaten by guards and infected through a wound on his hand — two decades before Pasteur and Lister vindicated the underlying claim by way of bacteriology.23

The DN model is elegant. It is also too liberal. Samir Okasha’s standard counter-example: a flagpole of known height casts a shadow of calculable length, and from that shadow length (with the laws of optics and the sun’s elevation) we can deduce the flagpole’s height. The deduction is valid; it satisfies Hempel’s schema; and it gets the explanatory direction backwards. The flagpole’s height explains the shadow’s length; the shadow does not explain the flagpole.

“Hempel’s covering law model does not respect this asymmetry. For just as we can deduce the length of the shadow from the height of the flagpole, given the laws and additional facts, so we can deduce the height of the flagpole from the length of the shadow … So Hempel’s model fails to capture fully what it is to be a scientific explanation.” — Samir Okasha, Philosophy of Science: A Very Short Introduction (2002), Ch. 3.24

The flagpole case suggests that scientific explanation involves something the DN model omits — most plausibly causal direction, which deductive entailment cannot capture. The alternative often advanced in its place is inference to the best explanation (IBE): we should believe the hypothesis that, if true, would best explain the evidence. IBE has its own difficulties — “best of which available alternatives?” — but it tracks the asymmetry the flagpole exposes.

The deeper point: even when method has done its work, a separate philosophical question remains about what an explanation is, and the cleanest available answer turns out to be insufficient.

2.6 Questions to Argue About

  • Feyerabend says “anything goes” in science. Is this a defence of scientific freedom or an abdication of rational standards?
  • If scientific observations are always theory-laden, does that mean we can never get an objective test of a theory? Or does “theory-laden” just mean “informed by background knowledge,” which seems unavoidable and not necessarily distorting?
  • Millikan discarded inconvenient data points. In hindsight, his final value for the electron’s charge was right. Does this vindicate his data selection, or does it just show that in this case he got lucky?
  • The textbook “scientific method” is acknowledged by philosophers of science to be an oversimplification. Should schools continue to teach it, and if so, why?

Forced Fork: Was Galileo Doing Science When He Used Rhetoric?

Position A (no-method pluralism is incoherent): Scientific progress requires defeasible but real norms — the requirement of independent confirmation, the publication of negative results, the asymmetric treatment of anomaly. Without them we lose any principled way to distinguish lucky successful pseudoscience (astrology occasionally getting a prediction right) from genuine science. Feyerabend’s case histories show only that any single rule has exceptions; that does not show no rules govern the practice. What did the work in Galileo’s case was not the rhetoric itself but the eventual convergence of independent evidence (Newtonian mechanics, then stellar parallax) onto the heliocentric model.

Position B (Feyerabend’s case is right): No methodological rule captures what scientific progress required at the moments that mattered. “Always follow the current best-established evidence” would have killed the Copernican revolution at the stage where heliocentrism predicted worse than Ptolemaic models with epicycles; “abandon a theory at the first anomaly” would have abandoned Newtonian mechanics at Uranus. What good science requires is contextual judgement — knowing when to push past current evidence on theoretical grounds and when to defer. The “rules” of method are post-hoc rationalisations of decisions whose actual basis was tacit.

Choose one. Position A must specify the minimum set of non-negotiable commitments and explain how they would have permitted Copernicus and Galileo to proceed against the available evidence of their day. Position B must specify what distinguishes “contextual judgement” from “post-hoc rationalisation of whatever the scientist already believed.”


3 How do scientific models relate to reality?

A map of London is not London. It does not contain the smell of the Underground, the particular grey of the Thames on a November afternoon, or the experience of being stuck behind a tourist who stops on the escalator. It is a representation — simplified, selective, oriented toward a purpose. A scientific theory is not a direct description of reality either. It is a model — a simplified, formalised representation that captures some aspects of what is being studied while necessarily abstracting away from others. The question is what we mean when we say it is true. The relationship between the model and the reality is one of the deepest questions in the philosophy of science.

Vera Rubin and the Mass That Wasn’t There

In the early 1970s, the astronomer Vera Rubin, working with Kent Ford at the Department of Terrestrial Magnetism in Washington and the Kitt Peak observatory in Arizona, used a sensitive new image-tube spectrograph to measure how spiral galaxies rotate. Galaxies are bright in the middle and dim at the edges; the obvious expectation, given the visible mass, is that the outer stars should orbit slowly and the inner stars quickly — Kepler applied to a galaxy. Rubin and Ford’s measurements showed almost the opposite. The outer stars in galaxy after galaxy moved at roughly the same speed as the inner stars; the rotation curves were flat. By Newtonian mechanics, this is only possible if there is far more mass present than the visible light reveals, distributed in an extended halo around the galaxy rather than concentrated where the stars are. Rubin’s 1970 paper on Andromeda with Ford, and her 1980 Astrophysical Journal survey of twenty-one spiral galaxies, are the empirical foundation of the modern dark-matter picture. Roughly five-sixths of the matter in the universe, by current models, is made of something no experiment has ever directly detected. (The headline figures for the total mass-energy budget are different: roughly 5% ordinary matter, 27% dark matter, and 68% dark energy.)25 The full story — and the question of whether “inserting it to make the equations work” is science or the evasion of science — is in §Dark Matter below.

3.1 Box’s Aphorism

The statistician George Box wrote in 1976:

“all models are wrong but some are useful.” — George E. P. Box, “Science and Statistics,” Journal of the American Statistical Association (1976).26

This aphorism has become a touchstone of scientific self-understanding, and its simplicity conceals real depth. “Wrong” in Box’s sentence is doing more work than the surface reading allows: Box was a statistician writing about idealised models of complex systems for forecasting, where “wrong” means strictly false as a description of the data-generating process, not “approximately incorrect.” Cartwright (How the Laws of Physics Lie, 1983) extends the same point to the fundamental laws of physics; van Fraassen (The Scientific Image, 1980) directs it at the unobservable entities those laws postulate.27

The operative question is therefore not “is the model wrong?” — they are all wrong in Box’s strict sense — but “useful for what?” Useful for prediction, for explanation, for control, and for pedagogy can come apart.

A weather model is not the weather; it is a mathematical system that represents certain features of atmospheric dynamics (pressure, temperature, humidity, wind) at a certain level of resolution. It is “wrong” in the sense that it omits an enormous amount — every butterfly, every microclimatic variation, every human exhaled breath. But it is enormously useful: it predicts tomorrow’s weather with an accuracy that was impossible before numerical weather prediction. The question is not whether it is true but whether it is accurate enough for the purpose at hand.

The same applies to Newton’s laws of motion: they describe the motion of everyday objects with extraordinary accuracy. They are also wrong — or rather, they are an approximation that breaks down near the speed of light (where special relativity takes over) and at very small scales (where quantum mechanics takes over). For most engineering purposes, however, Newton is not just “good enough”; it is the right level of description for the phenomena being modelled.

3.2 The Periodic Table as Model

The periodic table of elements, organised by Dmitri Mendeleev in 1869, is one of the most successful models in the history of science. It organises 118 elements by atomic number and chemical properties; it reveals periodic patterns in reactivity, electronegativity, and ionisation energy; it predicted the existence of then-undiscovered elements — Mendeleev’s eka-aluminium, eka-boron, and eka-silicon, later identified as gallium (1875), scandium (1879), and germanium (1886) — which were subsequently found with precisely the properties the table predicted.28

The philosopher Ian Hacking distinguishes between “representing” and “intervening” in science.29 We might debate whether scientific theories represent reality correctly; but the fact that we can intervene in the world using scientific knowledge — build transistors using quantum mechanics, make vaccines using virology — seems like strong evidence that we are latching onto something real.

What does the periodic table model? The atomic structure of matter — but the table itself contains no explicit statement about atomic structure. It is an organisation of empirical regularities that turned out to reflect underlying physical reality (the filling of electron shells). The model preceded the explanation of why it worked.

3.3 Lavoisier and the Overthrow of Phlogiston

The textbook example of one chemical model giving way to a successor is the late-eighteenth-century overthrow of the phlogiston theory by Antoine Lavoisier’s oxygen chemistry. Phlogiston, on the account systematised by Becher (1669) and Stahl (1703), was a substance released from combustible bodies during burning: wood released phlogiston into the air, leaving a calx (ash) behind; metals released phlogiston when heated, leaving the metal calx; the air absorbed phlogiston until saturated, at which point combustion stopped. The theory was productive: it organised chemistry for nearly a century, made successful predictions about which substances would burn, and unified phenomena that the ancient four-element scheme had treated as unconnected.30

Lavoisier’s Traité élémentaire de chimie (1789) overturned it. Lavoisier had shown, by careful weighing — a methodological move that pre-Lavoisierian chemistry had not made central — that combustion added mass to the burning substance rather than removing it. If phlogiston was released, the residue should weigh less; in fact, combusted metals weighed more. Lavoisier proposed that combustion was a chemical combination with a constituent of the air (which he named oxygène, “acid-forming”); that respiration was the same process; that water was a compound, not an element; and that mass was conserved across all chemical reactions. The 1789 Traité is the founding text of modern chemistry.31

The case is philosophically rich. It is the cleanest example of a Kuhnian paradigm shift in chemistry: the same experimental data (calxes weigh more than the metals they came from) had been available for fifty years before Lavoisier; what shifted was the theoretical framework that made the data significant. It is also the case Larry Laudan repeatedly cites against scientific realism in his “pessimistic meta-induction”: phlogiston was empirically successful, predictively useful, and eventually had to be abandoned as referring to nothing — and there is no general reason to think the theoretical posits of contemporary chemistry will fare any better in two centuries. Whether that is a fatal worry for chemical realism or a salutary humility about it is the open question.

3.4 Dark Matter: Model Postulated to Save Theory

Dark matter is one of the most striking examples of a theoretical entity postulated not because it was observed but because existing models fail without it. Observations of galaxy rotation rates in the 1970s — beginning with Vera Rubin and Kent Ford’s 1970 Andromeda paper, and consolidated in their 1978–1980 flat-rotation studies of other spirals — showed that galaxies rotate too fast: the outer stars move as if there is far more mass present than is visible.32 The standard response was to postulate an invisible form of matter — “dark matter” — that does not emit or absorb light but exerts gravitational effects.

This postulation has a good precedent: Neptune was similarly postulated to explain anomalies in Uranus’s orbit before it was observed. Dark matter has accumulated strong indirect evidence (gravitational lensing, cosmic microwave background data). But as of 2025, no particle detector has directly detected a dark matter particle. Candidate particles like WIMPs (Weakly Interacting Massive Particles) have been ruled out across most of their predicted parameter space; alternatives like axions are being tested.33

The question is which way the situation reads. Defenders of dark matter as a scientific hypothesis emphasise the convergent indirect evidence (rotation curves, lensing, the CMB peaks, the Bullet Cluster), the specificity of the predictions, and the fact that the candidate particles are being actively ruled in or out by experiment. Critics — those favouring modified-gravity alternatives like MOND, or those reading dark matter as a Lakatosian auxiliary protecting the standard cosmological model from falsification — emphasise that dark matter has been the auxiliary postulate for 90 years without direct detection and that the parameter space for the originally favoured candidates has been progressively eliminated. The Forced Fork below puts the question to the reader.

3.5 Scientific Realism vs. Instrumentalism

The deepest question about models is whether they are representations of reality or merely instruments for prediction:

Scientific realism: the entities postulated by successful scientific theories really exist. Electrons, quarks, dark matter, and the curvature of spacetime are real things, not just calculational conveniences. The predictive success of science is best explained by assuming that our theories are approximately true.

Instrumentalism: scientific theories are neither true nor false; they are tools. We should evaluate them solely by their predictive success. Whether electrons “really exist” is a meaningless question; what matters is whether the electron model correctly predicts the results of experiments.

Bas van Fraassen’s constructive empiricism (The Scientific Image, 1980) offers a middle position: we should believe that the observable consequences of successful theories are true, but remain agnostic about unobservable entities like electrons.34 This is epistemically cautious but creates strange consequences: it implies we should believe that atoms in a container exert pressure (observable) but remain agnostic about whether the atoms exist (unobservable).

Nancy Cartwright presses the anti-realist case from the other direction. Where van Fraassen targets the unobservable entities, Cartwright (How the Laws of Physics Lie, 1983) targets the equations themselves: “the fundamental laws of physics do not describe true facts about reality. Rendered as descriptions of facts, they are false; amended to be true, they lose their fundamental, explanatory force.”35 Newton’s law of universal gravitation, on Cartwright’s analysis, is true of objects in a model and not of any actual planet, which is also subject to electromagnetic forces, atmospheric drag, and other effects the law abstracts from. The phenomenological laws used by engineers are closer to reality but less general; the more universal a law looks, the less literally true it is in any concrete situation.

3.6 Questions to Argue About

  • If all scientific models are wrong (Box), does that mean science never achieves truth? Or does “wrong” in this context mean something different from “false”?
  • Dark matter is postulated to explain anomalies but has never been directly detected. Is this good scientific practice (like the Neptune case) or special pleading to protect a theory?
  • Scientific realists say electrons really exist because the electron model predicts observable results so well. But good predictive success is compatible with the model being false. Does predictive success give us reason to believe in the existence of unobservable entities?
  • Hacking argues that we know electrons exist because we can manipulate them to produce other effects. Does this “manipulation” argument for scientific realism work?

Forced Fork: Does Dark Matter Exist, or Is It an Equation-Saving Device?

Position A: Scientific realism is the best explanation of the dark matter evidence — we should believe dark matter exists because multiple independent lines of evidence (galaxy rotation curves, gravitational lensing of distant galaxies by foreground clusters, the angular power spectrum of the CMB, the matter distribution in the Bullet Cluster) converge on the same quantitative postulate. The predictive success of science — Maxwell predicting radio waves mathematically before Hertz detected them; Dirac predicting antimatter; the Standard Model entailing a Higgs-like scalar 48 years before ATLAS and CMS announced 5σ discovery on 4 July 201236 — is best explained by the no-miracles argument (Putnam): it would be a cosmic coincidence if our theories were systematically false but reliably predictive. Instrumentalists do not deny this convergence; they re-describe it as “empirical adequacy”. The realist’s claim is that the re-description is gratuitous.

Position B: Constructive empiricism (van Fraassen) is correct: dark matter is empirically adequate as a postulate, and that is all we are entitled to assert. The history of science is a history of empirically successful theories subsequently abandoned — caloric, phlogiston, the luminiferous ether all predicted well for a time and were later judged to refer to nothing. The “no-miracles” argument proves too much: it would have licensed nineteenth-century physicists to assert the existence of the ether on identical grounds. The right realist response is structural realism (Worrall) — what survives theory change is the mathematical structure, not the postulated entities. But structural realism is a retreat from full realism, not a vindication of it.

Choose one. Position A must address Laudan’s pessimistic meta-induction directly37 — why should we trust current theories when ether and caloric and phlogiston were equally backed by their generation’s “convergent evidence”? — and must say what your view is on whether those theories’ theorists were warranted in their realist commitment. Position B must say whether the retreat to structural realism is genuinely an instrumentalist position or a realist concession dressed in instrumentalist clothing — and must specify whether you would actually use the word “fictional” of electrons, quarks, and the curvature of spacetime in conversation, given that those are the entities your view declines to commit on.


4 Can science be objective?

The ideal of scientific objectivity is that the method — observation, experiment, analysis — produces results that are independent of the observer’s identity, values, and expectations. It is a powerful ideal, and the institutions built on it have produced reliable knowledge across many domains. The evidence from both philosophy and the history of science complicates it. The question is whether the complications require abandoning the ideal, weakening it, or refining what “objectivity” means in scientific practice — and reasonable people who agree on the evidence disagree about the answer.

The Replication Crisis in Social Priming

In 2011, the social psychologist Diederik Stapel was suspended from his position at Tilburg University after it was established that he had fabricated data across dozens of published studies.38 Stapel’s work had included influential studies on stereotype threat and environmental effects on behaviour — precisely the field of social priming. John Bargh’s famous 1996 “elderly priming” study — in which participants exposed to words associated with old age walked more slowly down a corridor — became a touchstone of popular social psychology, cited in Malcolm Gladwell’s Blink.39 When Stéphane Doyen attempted to replicate it in 2012 under double-blind conditions, he found no effect.40 Daniel Kahneman wrote an open letter in September 2012 to Bargh and other priming researchers urging them to conduct systematic replications before the field’s credibility collapsed entirely.41 Norwood Russell Hanson’s argument that all observation is theory-laden — that what scientists see is partly determined by what they expect — finds its darkest illustration in the priming literature, where the expectation of an effect may have been systematically producing the appearance of one across thousands of individual experimental sessions.

4.1 Theory-Ladenness of Observation

Norwood Russell Hanson, in Patterns of Discovery (1958), argues that what we see depends on what we already know.42 He asks: what does a naive observer see when they watch the sun rise? What does a Keplerian astronomer (who knows the Earth orbits the sun) see when they watch the same event?

The naive observer sees the sun rising. The Keplerian sees the horizon tilting away from a stationary sun. The same visual input generates different perceptions because the conceptual framework within which observation is interpreted differs. This is not a matter of what the observer believes while looking; it is a matter of what they see.

The implications for scientific objectivity are uncomfortable: if all observation is theory-laden, then no observation can serve as a neutral test of a theory, because the observation is already conditioned by theoretical assumptions. Scientists see what their paradigm trains them to see — and may systematically fail to see what their paradigm says cannot be there.

4.2 Heisenberg’s Uncertainty Principle

At the quantum level, certain pairs of properties cannot both be sharply defined at the same time. The Heisenberg uncertainty principle (1927) is usually written \(\Delta x \cdot \Delta p \geq \hbar/2\): the product of the standard deviations of position and momentum in any quantum state has an irreducible lower bound. This is not a statement that instruments disturb particles (the popular “observer-effect” gloss), and it is not a statement about the accuracy of individual measurements — it is a statement about the intrinsic non-commutativity of the quantum operators themselves.43 It is a fundamental feature of quantum mechanics rather than a limitation of current technology.

This has been taken by some philosophers to imply that the observer is necessarily implicated in what is observed at the quantum level — that there is no quantum world independent of measurement. This interpretation (associated with the Copenhagen interpretation of quantum mechanics, developed by Bohr and Heisenberg) is controversial; other interpretations (the Everett “many worlds” interpretation, for example) deny that measurement is special. But the question of what quantum mechanics implies about the observer-independent character of reality remains genuinely open.

4.3 The Replication Crisis

Starting around 2011, psychology began experiencing what is now called the replication crisis: a systematic failure to reproduce the results of landmark published studies. The Open Science Collaboration (2015) attempted to replicate 100 published psychology experiments; only 39% of the replications were judged by the replicating teams themselves to have reproduced the original result, and on a stricter measure — statistically significant effects in the same direction — the figure was 36%.44

The replication crisis is not limited to psychology. A 2016 survey of 1,576 scientists by Nature found that 70% had tried and failed to replicate another scientist’s experiments, and 50% had failed to replicate their own previous results.45 The problem is structural: publication bias rewards positive results, small samples are common, and statistical thresholds are often poorly understood.

Cross-reference: The Human Sciences unit’s lesson “What can experiments tell us about human nature?” covers the replication crisis as it specifically affects psychology. The two treatments together give a fuller picture: the crisis is real across both natural and human sciences, but its causes and remedies differ by discipline.

The implications for scientific objectivity: if a substantial proportion of published results cannot be reproduced, what does this say about the knowledge claims those studies generate? The problem is partly statistical (underpowered studies, p-hacking, flexible analysis), partly incentive-driven (journals publish novel positive results, not replications), and partly about the gap between the clean story told in published papers and the messy, iterative process that actually produced the data.

4.4 Feminist Critiques

Sandra Harding, in Whose Science? Whose Knowledge? (1991), argues for what she calls “strong objectivity” — a methodological demand that goes beyond conventional objectivity.46 Standard objectivity demands that researchers not let their values distort their research. Strong objectivity demands that researchers examine the values and social positions that shape which questions are asked, which methods are used, and which phenomena are considered worth studying.

Strong objectivity is not a demand that science be partisan; it is a demand that the social conditions of knowledge production be made part of the analysis. Historical examples: the biological sciences’ long history of providing “natural” explanations for social hierarchies (women’s “naturally” limited intellectual capacity, racial differences in intelligence) turned out to reflect the social assumptions of the scientists, not biological reality.47 The objectivity that would have exposed this was precisely the kind that examined the assumptions the science brought to its questions.

Harding’s argument is sometimes assimilated to a stronger constructionist claim — that scientific findings are socially constructed in the sense of being merely contingent products of their social setting. Ian Hacking’s The Social Construction of What? (1999) is sharp on this conflation: a “social construction of X” claim, on Hacking’s analysis, characteristically packs together three theses (X is not inevitable; X is bad as it is; we would be better off without X), and only the first is shared with Harding’s methodological demand.48

4.5 Inductive Risk and the Value-Free Ideal

A separate question is whether non-epistemic values — social, ethical, political — have any legitimate role inside scientific reasoning, not merely in selecting topics or interpreting results for policy. Heather Douglas, in Science, Policy, and the Value-Free Ideal (2009), argues that they do, and that the standard ideal of value-free science must be rejected:

“the value-free ideal must be rejected precisely because of the importance of science in policymaking. In place of the value-free ideal, I articulate a new ideal for science, one that accepts a pervasive role for social and ethical values in scientific reasoning, but one that still protects the integrity of science.” — Heather Douglas, Science, Policy, and the Value-Free Ideal, Ch. 1.49

Douglas’s central technical move revives the inductive risk argument of Richard Rudner (1953): because no scientific hypothesis is ever conclusively verified, the scientist who decides to accept or reject one must weigh how bad it would be to be wrong in each direction — a weighing that requires non-epistemic values whenever the claim has practical consequences. Setting the p-value threshold at 0.05 rather than 0.01, choosing how strong evidence of carcinogenicity must be before reporting a chemical as carcinogenic, deciding whether to model uncertainty conservatively or optimistically: each of these decisions has, on Douglas’s account, a value-laden dimension that the value-free ideal misdescribes when it treats the decision as purely epistemic. Whether her diagnosis applies to all cases or only to the regulatory-science cases she chiefly draws on is itself contested in the literature.50

4.6 Questions to Argue About

  • If observation is theory-laden, can any observation provide a genuinely independent test of a theory? Or is this a problem that can be partially solved by using theories from different domains to constrain each other?
  • Does Heisenberg’s uncertainty principle show that the observer is necessarily part of quantum reality — or is it just a limitation on measurement?
  • The replication crisis shows that a substantial proportion of published psychological findings cannot be reproduced. Does this mean psychology is not a science? Or does it just mean science has a quality control problem?
  • Harding argues for “strong objectivity” that examines the social assumptions of science. Is this a strengthening of objectivity, or is it a political intervention that threatens objectivity?

Forced Fork: Was Bargh’s Elderly-Priming Study Ever Knowledge?

Position A: Bargh’s 1996 elderly-priming literature was not knowledge, and the pre-registration reforms cannot retroactively make it so. The replication crisis reveals something deep about the epistemological limitations of much social and psychological science: the phenomena are too context-dependent, sample sizes too small, and researcher degrees of freedom too large for the published literature to constitute genuine knowledge. The crisis is not a bug in an otherwise sound system; it is the visible failure of methods that were never adequate for the complexity of their subject matter.

Position B: Had Bargh-style studies been pre-registered with the Doyen-era protocols, the replicable effects (if any) would have surfaced and the spurious ones would have been filtered. The replication crisis is an incentives problem with known solutions, not an epistemological failure of the science itself. Publication bias, p-hacking, and underpowered studies are all addressable through pre-registration of studies, open data requirements, and journals that publish null results. Psychology and social science can produce genuine knowledge; they need better institutional structures, not methodological abandonment.

Choose one. If you choose Position A, specify which areas of psychology and social science you regard as reliable — or, if none, what follows for policy-making that currently uses social-scientific findings as its evidence base. If you choose Position B, explain what the system of pre-registration and open data has already achieved — and what it has not yet fixed.


5 How does science change?

Science is often presented as accumulating knowledge steadily, building toward a more complete and accurate picture of the world. Thomas Kuhn showed that this picture, while not entirely wrong, conceals the more dramatic and disruptive process by which science actually changes.

The Higgs Boson at CERN, 4 July 2012

In a 1964 paper in Physical Review Letters, the British physicist Peter Higgs proposed a mechanism by which the otherwise massless gauge bosons of electroweak theory could acquire mass; François Englert and Robert Brout in Brussels had submitted a paper with substantially the same mechanism a few weeks earlier; Gerald Guralnik, Carl Hagen, and Tom Kibble in London published a third independent version in October 1964.51 The mechanism worked only if there existed a new scalar field pervading all space and a corresponding particle — what would be called the Higgs boson. The Standard Model of particle physics, completed by 1973, depended on this mechanism. The particle’s mass was not predicted by the theory; it had to be measured. By the late 1990s the Large Electron–Positron collider at CERN had narrowed the allowed mass window through electroweak precision fits, but the particle itself remained beyond direct detection. The Large Hadron Collider — at roughly $9 billion the most expensive scientific instrument ever built — was constructed largely to find it. On 4 July 2012, in a packed seminar room at CERN, the spokespersons for the ATLAS and CMS collaborations, Fabiola Gianotti and Joe Incandela, jointly announced a 5σ discovery of a new boson at approximately 125 GeV/c² with properties consistent with the Standard Model Higgs.52 Peter Higgs, then 83, was in the audience; cameras caught him removing his glasses and wiping his eyes. The 1964 papers had predicted what the 2012 instrument found, with no intervening experimental contact. The interval was 48 years. Englert and Higgs received the Nobel Prize in Physics on 10 December 2013; Brout had died in May 2011, missing the discovery by 13 months. Eugene Wigner had described this kind of event in his 1960 essay “The Unreasonable Effectiveness of Mathematics in the Natural Sciences” as almost miraculous — pure mathematical structures, developed without reference to physical application, turning out to describe physical reality with uncanny precision.53 James Clerk Maxwell’s 1865 unification of electricity, magnetism, and light (treated below) had been the canonical historical case of the same phenomenon. Higgs is the contemporary version, with named scientists in the audience to see the prediction confirmed.

5.1 Kuhn’s Paradigms

In The Structure of Scientific Revolutions (1962), Kuhn introduces the concept of a paradigm — a framework of shared assumptions, methods, and problems that defines what “normal science” is doing at any time.54 Normal science is puzzle-solving within the paradigm: it does not question fundamental assumptions but fills in details, resolves inconsistencies, and extends the theory’s scope.

In the 1969 Postscript Kuhn admitted that “paradigm” had been doing two distinct jobs in the original book and offered a sharper vocabulary: a disciplinary matrix (the whole bundle of values, methods, symbolic generalisations, and shared exemplars binding a scientific community) and a concrete exemplar (a worked-out problem-solution that trains scientists how to handle similar problems).55 Most lay use of “paradigm shift” elides this distinction.

Plate tectonics is a clean example of a disciplinary-matrix shift: an entire community’s organising framework changed in roughly a decade. The Copernican revolution is messier — a multi-stage process more easily described in terms of successive exemplars (Copernicus, Kepler, Galileo, Newton) than as a single matrix-replacement.

Anomalies — observations that cannot be accommodated within the paradigm — accumulate. At first, they are ignored or explained away. When they become impossible to ignore, a scientific revolution occurs: the old paradigm is replaced by a new one. The Copernican revolution replaced the Ptolemaic geocentric model. The germ theory of disease replaced miasma theory. The plate tectonics revolution replaced the “permanentist” view of stable continents.

Crucially, Kuhn argues that different paradigms are incommensurable: scientists working in different paradigms are not just disagreeing about facts, they are working in different conceptual worlds, asking different questions, seeing different phenomena. This is why paradigm shifts are so difficult and so slow: the scientists defending the old paradigm are not being irrational — they are operating within a framework that, for them, still works.

5.2 Continental Drift

Alfred Wegener, a German meteorologist, proposed in The Origin of Continents and Oceans (1915) that the continents had once been joined and had drifted apart.56 His evidence was substantial: the shapes of Africa and South America fit together like torn pieces of a map; identical fossil species appeared on both sides of the Atlantic; geological strata matched across coastlines thousands of miles apart. The theory was rejected, sometimes contemptuously, by the geological establishment for more than four decades. At a 1926 symposium of the American Association of Petroleum Geologists convened specifically to evaluate Wegener’s hypothesis, the geologist Rollin T. Chamberlin argued that if Wegener’s theory were accepted, geologists would have to “forget everything which has been learned in the last 70 years and start all over again.”57

The Wegener case illustrates a real epistemological problem: a correct theory can be rationally rejected when it lacks a mechanism. How do we distinguish “reasonable scepticism” from “paradigmatic conservatism”? Hindsight makes this easy; foresight does not.

The rejection was not entirely irrational. Wegener proposed no plausible mechanism for how continents could move through oceanic crust — his own tentative suggestions (centrifugal force from the Earth’s rotation; tidal drag from the Sun and Moon) were quantitatively far too weak to do the work, and his critics knew this. By the standards of physics available at the time, the idea seemed physically impossible. It was only in the 1950s and 1960s, when the discovery of mid-ocean ridges and paleomagnetism provided both evidence and a different mechanism (seafloor spreading driven by mantle convection — not Wegener’s continents-through-crust picture), that plate tectonics was accepted — and rapidly became the organising paradigm of the earth sciences.58

A more sceptical reading of this story is worth noting. Naomi Oreskes (The Rejection of Continental Drift, 1999) argues that the standard “Wegener was vindicated” narrative is partly a retrospective reconstruction: the modern theory of plate tectonics differs in important respects from what Wegener actually proposed (the mechanism, the timescale, several specific claims about geological correlations).59 On Oreskes’s reading the narrower lesson is that the geological community resisted a theoretical framework with substantial evidence in its favour for longer than the evidence alone warranted, and that the resistance was loosened only when an independent line of evidence (paleomagnetism) and a workable mechanism (seafloor spreading) became available together. On the older reading the narrative survives intact: Wegener was substantively right; the community took fifty years to admit it. Which reading is right matters for what the case shows about scientific resistance to outsider theories.

5.3 Maxwell, Higgs, and the 48-Year Gap

The Higgs case (info box above) is the contemporary version of a phenomenon Eugene Wigner identified in 1960 as “the unreasonable effectiveness of mathematics in the natural sciences.” Its canonical historical version is Maxwell. In 1865, the Scottish physicist James Clerk Maxwell published a unified mathematical theory of electricity and magnetism, in which he derived as a consequence of the equations themselves that electromagnetic disturbances should propagate through space at approximately the speed of light — already known to be roughly 300,000 km/s.60 Maxwell concluded that light was an electromagnetic wave, and his equations predicted the existence of electromagnetic waves at other frequencies. Twenty-two years later, in 1887, Heinrich Hertz confirmed the existence of what are now called radio waves, working entirely from Maxwell’s mathematical prediction.61 The radio, television, mobile telephone, and wireless internet are the practical descendants of a set of mathematical symbols manipulated by Maxwell at King’s College London during his professorship there from 1860 to 1865.

The Higgs interval (48 years) and the Maxwell interval (22 years) raise a question about what counts as knowledge during the gap. Did physicists know the Higgs boson existed before 2012, on the strength of the theoretical prediction? They had strong indirect grounds: electroweak precision fits at LEP and the Tevatron narrowed the allowed mass window, and the Standard Model without some such mechanism was internally inconsistent at high energies. This is not nothing. But it is also not the same epistemic state as a 5σ direct detection. The question of what to call the intermediate state — between a successful mathematical prediction and its experimental confirmation — is a permanent feature of physics on Wigner’s reading and a temporary inconvenience of measurement on the deflationary alternative. The philosophical question — whether the fit between mathematics and physics is a deep fact about the universe or a selection effect in which we remember the mathematics that worked and forget what did not — remains genuinely open.

5.4 Why Scientists Resist New Ideas

If paradigm shifts are how science advances, why do scientists resist them? Several mechanisms:

Professional investment. A career spent on one theoretical framework is threatened by its replacement. This is not cynical self-interest; it is a rational response to epistemic uncertainty. You believe your framework; you’ve built your understanding of the field within it.

The rationality of conservatism. New theories typically explain less, at first, than the established ones. The new paradigm often works well in the narrow domain that motivated it and badly everywhere else. It takes time for a new theory to prove itself adequate to the full range of phenomena.

Incommensurability. Scientists trained in one paradigm may genuinely not understand what the new paradigm is claiming, because the concepts are different. Communication between paradigms requires translation, which is always approximate.

5.5 Questions to Argue About

  • Kuhn’s incommensurability thesis implies that scientists in different paradigms can’t really communicate with each other. Is this too strong? Are there historical cases of genuine communication across paradigm shifts?
  • The Wegener case suggests that correct theories can be rationally rejected. Does this mean the sociology of scientific communities is as important as the logic of scientific evidence in determining what science accepts?
  • Scientists resisted the germ theory of disease and continental drift, both of which turned out to be correct. How should this history affect our attitude toward current scientific consensus? Does it give grounds for scepticism — or just caution?
  • The Higgs boson was predicted in 1964 and confirmed in 2012. During those 48 years, did physicists “know” it existed? What would it mean to say they did?

Forced Fork: When Is It Rational to Resist Scientific Consensus?

Position A: For a non-expert on a question within an expert domain, the rational default is to defer to the standing scientific consensus, and the evidential threshold for departing from this default is very high. The history of scientists resisting correct theories (Wegener, plate tectonics) is not a model for laypeople second-guessing climate science or vaccine safety — those scientists had domain expertise and access to the primary evidence. The layperson who rejects consensus on climate or evolution is typically not doing what Wegener did; they are typically relying on motivated reasoning, manufactured doubt (Oreskes and Conway document the financed campaigns), or misunderstanding of scientific method. Deference is not surrender of judgement; it is the recognition that the layperson lacks the means to evaluate primary evidence directly.

Position B: “Scientific consensus” is heterogeneous and the rational stance toward it is differentiated, not uniformly deferent. First, the consensus on the temperature record (very strong, multiple independent lines) is epistemically different from the consensus on specific policy responses (where non-epistemic values legitimately enter — see Douglas above). Second, scientific consensus has been substantively wrong on empirical questions within living memory: ulcers (Marshall and Warren); the dietary-fat hypothesis (substantially revised in the 2010s); the recommendation against early peanut introduction (reversed in 2015 after the LEAP trial). The rational layperson distinguishes the core empirical claim from policy applications, attends to the quality of underlying evidence, and notices when a “consensus” is being used as a conversation stopper rather than a conclusion.

Choose one. Whose reading of the Marshall–Warren case (Barry Marshall drank a broth culture of H. pylori in 1984 to demonstrate that ulcers were caused by bacterial infection, against the consensus that they were caused by stress and acid;62 he and Warren shared the 2005 Nobel) is more defensible — Position A’s “the case shows the system working, since Marshall was a gastroenterologist publishing in the Lancet, not a layperson,” or Position B’s “it shows that the consensus was substantively wrong for two decades and that the system corrected it slowly enough for many of the patients it killed in the meantime to count” — and what does each side concede?


6 What can science explain — and what can’t it?

Science is the most powerful method for producing knowledge about the natural world. That claim would have struck most educated people in the 17th century as hubristic. It strikes most educated people today as obvious. That change — from hubristic to obvious — is itself one of the most consequential intellectual events in human history. But even standing at the summit of that achievement, there is a serious question about what “the natural world” includes, and whether there are genuine domains of inquiry that scientific methods cannot reach.

Adrian Owen, the Vegetative-State Patient, and the Detection of Awareness

In 2006 the neuroscientist Adrian Owen, then at the MRC Cognition and Brain Sciences Unit in Cambridge, scanned a 23-year-old woman who five months earlier had been left in a vegetative state by a road-traffic accident.63 By every behavioural criterion she was unaware: no responses to verbal commands, no visual tracking, no purposive movement. Owen’s team placed her in an fMRI scanner and asked her, while her brain was monitored, to imagine playing tennis, and then to imagine walking from room to room through her own home. Her supplementary motor area lit up for the tennis task and her parahippocampal gyrus, premotor cortex, and posterior parietal cortex lit up for the navigation task — the same neural signatures observed in healthy controls performing the same imagery on instruction. Whatever was happening inside her, it was responsive, sustained, and following instructions. Published in Science as “Detecting Awareness in the Vegetative State,” the case was followed in 2010 by a New England Journal of Medicine paper from Owen and Steven Laureys’s group: 4 of 54 patients in vegetative or minimally-conscious states showed willful brain responses, and one patient — a 22-year-old man with severe traumatic brain injury — was able, by selecting tennis or house imagery as “yes” or “no,” to answer factual questions about his own pre-injury life that the medical team could independently verify.64 What Owen’s protocol can do is determine, against negative behavioural evidence, that there is someone in there. What it cannot do — what the protocol’s structural design makes it unable to do — is tell us what it is like to be that person, trapped inside a body that cannot speak. The neuroscientist’s ability to detect consciousness from outside, and the philosopher’s claim that there is something fundamental about consciousness no third-person account reaches, are demonstrated in the same patient on the same scan.

6.1 Reductionism and Its Limits

Reductionism is the view that complex phenomena can be fully explained by their component parts. Chemistry is explained by physics (atoms and their interactions); biology is explained by chemistry; psychology is explained by neuroscience; sociology is explained by psychology. In principle, if you knew enough physics, you could explain everything.

The reductionist programme has had extraordinary successes. The discovery that DNA encodes genetic information in sequences of four bases — and that this code can be read, edited, and engineered — is a triumph of biochemical reductionism that has transformed medicine.

Philip Anderson’s influential paper “More Is Different” (1972) argues that reductionism as a research strategy is correct (you should look at lower levels for mechanisms) but that it does not imply constructivism (you cannot reconstruct higher-level behaviour from lower-level knowledge). Each level of complexity has its own regularities.65

Emergence is the counter-claim: that complex systems have properties that are not predictable from, and cannot be reduced to, the properties of their components. The “wetness” of water is not a property of individual water molecules; it emerges from the interaction of vast numbers of them. The behaviour of an ant colony cannot be predicted from the chemistry of individual ants; it is a collective property of the interaction. Ernst Mayr argued that biology is irreducibly autonomous: biological concepts like fitness, adaptation, and selection operate at a level that cannot be reduced without loss of explanatory power.66

The reductionist response distinguishes explanatory irreducibility (we cannot in practice predict colony behaviour from ant chemistry) from ontological irreducibility (there is something present in the colony beyond the ants and their interactions). On this reading the first is uncontroversial — a reflection of computational limits and the usefulness of higher-level concepts — and the second would require evidence of physical effects unaccounted for by lower-level laws. Wetness, in this reading, is just the macroscopic behaviour of large numbers of water molecules under standard conditions; no ontologically extra “wetness” is doing causal work. Strong-emergence defenders (Jaegwon Kim’s later work on downward causation, certain biologists working on developmental systems, some philosophers of mind) reply that the reductionist begs the question by demanding lower-level evidence for irreducibility — and that the difficulty of providing such evidence is precisely what one would expect if the higher-level kinds were genuinely real. The dispute is genuinely live within philosophy of mind and biology, less so within physics; how widely either side of the dispute extends across the sciences depends on the science.

6.2 The Hard Problem of Consciousness

David Chalmers, in The Conscious Mind (1996), distinguishes the “easy problems” of consciousness (explaining how the brain processes information, controls behaviour, reports its states — “easy” because they are tractable by standard scientific methods, even if technically difficult) from the “hard problem”: why is there subjective experience at all?67

The hard problem asks: why, when information is processed in the brain, is there something it is like to be the brain processing that information? Why doesn’t neural processing occur “in the dark,” without any accompanying inner experience? This question resists scientific explanation not because we lack data but because the standard scientific vocabulary — causes, mechanisms, functions — seems inadequate to it. Even a complete functional account of how the brain produces colour perception would leave open the question: why does red look like this?

The most mathematically developed response is Giulio Tononi’s Integrated Information Theory (IIT). Tononi proposes that consciousness is identical to integrated information (Φ): a measure of how much a system’s cause-effect structure is irreducible to the cause-effect structure of its parts. The technical move: cut the system across its weakest link (the partition that destroys the least information) and see how much information is lost; a system with high Φ is one whose causal powers cannot be recovered from any decomposition.68 IIT is mathematically precise and makes testable predictions, but it implies that simple networks of logic gates would be conscious — which most working scientists find counterintuitive enough to count against the theory. Two other frameworks are worth knowing: Global Workspace Theory (Baars, Dehaene) proposes that consciousness arises when information is broadcast to a workspace accessible to multiple cognitive systems — it explains access but says nothing about why the broadcast is experienced.69Predictive processing (Clark, Friston) treats the brain as a prediction machine constantly minimising error against its model of the world.70 None has solved the hard problem.

Thomas Nagel’s famous formulation, in “What Is It Like to Be a Bat?” (Philosophical Review, October 1974), is the canonical statement of the limit. Bats navigate their environment through echolocation — emitting ultrasonic pulses and processing the returning echoes to construct a spatial map of striking precision. We know, in extraordinary scientific detail, how the system works: the frequency modulation of the calls, the neural processing, the cochlear mechanics. What Nagel argued is that this scientific knowledge leaves something entirely untouched: the subjective character of the bat’s experience.71 What is it like, from the inside, to perceive a moth through sound? There is, Nagel argued, something it is like to be a bat — there is a subjective experience — and no amount of neuroscientific information about its echolocation system will tell us what that experience is. The bat is the cleanest illustration of the limit because the bat is real and because its perceptual world is alien enough to make the gap between physical description and subjective experience intuitively clear. The Owen fMRI findings (treated in the info-box above) sharpen the point. Owen’s protocol can establish, against negative behavioural evidence, that a vegetative-state patient is conscious. It cannot tell us what is is like to be locked inside that body. The third-person science and the first-person experience separate cleanly.

6.3 What Science Cannot Explain

The scope question has several layers:

Value questions. Science can tell you that a particular pesticide causes cancer in laboratory rats at a certain dose. The standard view — going back to Hume’s distinction between “is” and “ought” — is that science cannot tell you whether the risk is acceptable given the agricultural benefits, because that is a question about values, not facts. This view is contested: empirical findings about welfare, suffering, and human flourishing arguably bear on normative questions even if they do not settle them, and Heather Douglas has argued that non-epistemic values already enter scientific reasoning at the point where evidence is weighed. The minimal claim — that science alone, without value premises, does not deliver normative conclusions — is the textbook position. The stronger claim, that science is wholly silent on what we should do, is what evolutionary ethicists, neo-Aristotelians on naturalised goodness, and Sam Harris’s moral landscape programme each, in different ways, deny.

Questions about existence. Science explains the structure of the universe and the processes by which complex structures emerged. It does not explain why there is a universe at all rather than nothing. This may not be a scientific question.

The problem of induction. Science assumes that the future will resemble the past — that the laws governing the universe are stable across time and space. This assumption cannot itself be justified scientifically (it would be circular). It is a presupposition of science, not a result of it.

Historical singularities. Science works best on repeatable phenomena. The origin of life, the origin of the universe, the particular sequence of evolutionary steps that led to consciousness — these are historical singularities that can be studied but not experimentally replicated.

6.4 Questions to Argue About

  • Is emergence a genuine scientific phenomenon, or does “emergent properties” just mean “properties we don’t yet know how to reduce”? What would evidence for genuine irreducibility look like?
  • Chalmers’ “hard problem” suggests that consciousness cannot be explained scientifically. Does this imply consciousness is supernatural, or just that science needs new concepts?
  • “Science can tell us what is but not what ought to be.” Is this a satisfying account of science’s limits? Or can scientific findings (about welfare, suffering, flourishing) bear on normative questions?
  • If the assumption that the laws of physics are stable across time cannot be scientifically justified, does that undermine science — or is it an acceptable foundational assumption?

Forced Fork: Will a Complete Neuroscience of Owen’s Patient Tell Us What It Is Like to Be Locked In?

The case is in the info-box above. Adrian Owen’s protocol can determine, against negative behavioural evidence, that a vegetative-state patient is conscious — by the brain’s response to instructions to imagine playing tennis. What it cannot do is tell us what it is like to be that patient, trapped in a body that cannot speak. Nagel’s 1974 bat (treated above in the body) is the canonical philosophical anchor; Owen’s protocol is the empirical case where the limit becomes life-and-death.

Position A: The hard problem of consciousness is a genuine limit of scientific explanation, and Owen’s case demonstrates the limit clinically. Chalmers is right that no account of information processing, neural correlates, or functional organisation will explain why any of this is experienced — why there is something it is like to be the patient who has just answered “yes” by imagining tennis. Owen’s protocol is impressive precisely because it shows the third-person limit: we now know, in principle, that she is conscious; we still do not know what it is like to be her. This is not because we lack data; it is because the explanatory vocabulary of science (causes, mechanisms, functions) is insufficient for the phenomenon Nagel and Chalmers identified. The bat is a stumbling block; the locked-in patient makes the stumbling block urgent.

Position B: The hard problem is a product of confused intuitions rather than a genuine explanatory gap. The sense that consciousness cannot be explained by physical processes is a cognitive phenomenon produced by the limited access we have to our own neural machinery. When we have a sufficiently rich functional and neurological account of how information is integrated, broadcast, and reflected upon — and Owen’s protocol is the kind of integration the better account will be built on — the “hard problem” will dissolve. Not because it has been answered, but because we will see it was never coherently posed. Owen’s vegetative-state patient is conscious in the way the patient’s brain is conscious; the seeming further question (what it is like) is a confusion the integrated account will retire.

Choose one. If you choose Position A, explain why the hard problem is different from historical “hard problems” that turned out to be solvable — like the apparent impossibility of explaining life without vitalism. If you choose Position B, explain what you say to Owen’s twenty-two-year-old patient who answered yes/no questions through fMRI: state specifically what, in a complete neuroscience of his case, would constitute an explanation of what it is like to be locked in.


7 What are the ethics of science?

“The scientist is not responsible for the laws of nature,” goes the defence, “but only for discovering them.” It is a clean argument with a long history in scientific self-understanding. It is also disputed: critics argue that scientific methods involve subjects whose consent and welfare matter, that scientific findings have foreseeable consequences whose risks must be weighed, and that scientific applications reshape the world in ways that require governance. The cases below — Tuskegee, He Jiankui, Oppenheimer, Haber — are the standard exhibits in that dispute. Some illustrate clear ethical failure (Tuskegee’s denial of effective treatment); others (Oppenheimer’s role in a wartime weapons programme) admit defensible arguments on more than one side. Whether the “science is neutral; responsibility lies with users” defence holds up across the spectrum, or breaks down at some specific point on the spectrum, is for the reader to argue.

The Many Labs Replication Project and the Psychology Textbook Problem

In 2015, the Open Science Collaboration published in Science the results of the Reproducibility Project: 270 researchers had attempted to replicate 100 published studies in psychology. Only 36 per cent of the replications produced a statistically significant effect in the same direction as the original — the stricter of the several figures reported in the paper.72 The failure rate was highest in social psychology, where modest sample sizes and researcher degrees of freedom — the many small, undisclosed choices researchers make about data collection and analysis — had produced a literature substantially inflated by false positives. Stanford epidemiologist John Ioannidis had predicted this outcome in a 2005 paper, “Why Most Published Research Findings Are False,” which used statistical modelling to show that under conditions of low prior probability, modest sample sizes, and publication bias toward positive results, the majority of published findings in certain fields would be expected to be false.73 The crisis had immediate practical consequences: pharmaceutical companies had been running clinical trials based on psychological findings; policy programmes had been built on social priming research; millions of textbook pages contained claims that had never been subjected to serious replication attempts. What the crisis revealed was not that scientists were dishonest, but that the incentive structures of academic science — publish or perish, journals that favour positive results — systematically produce errors that the self-correcting mechanism of science is slow to catch.

7.1 The Tuskegee Study

The US Public Health Service Syphilis Study at Tuskegee (1932–1972) recruited 399 Black American men with syphilis and 201 uninfected controls in rural Alabama.74 The men were told they were being treated for “bad blood”; they were not. When penicillin was established as an effective treatment for syphilis in the mid-1940s, the men were not offered treatment. The study continued for 40 years so that researchers could observe the “natural progression” of untreated syphilis. It ended only when a whistleblower, Peter Buxtun, went to the press in 1972.

The Tuskegee study was not a rogue operation; it was funded by the federal government, published in peer-reviewed journals, and known to public health officials for four decades. Its exposure led directly to the National Research Act (1974) and the Belmont Report (1979), which established the ethical principles — respect for persons, beneficence, justice — that govern human subjects research in the US.75

The Nuremberg Code (1947), produced in the wake of Nazi medical experiments, had already established the requirement for informed voluntary consent.76 Tuskegee showed that these principles had not been internalised or enforced.

7.2 CRISPR and He Jiankui

In November 2018, Chinese biophysicist He Jiankui announced the birth of the world’s first gene-edited human babies — twin girls, given the pseudonyms Lulu and Nana, whose embryos had been edited to disable the CCR5 gene in an attempt to mimic the naturally-occurring CCR5-Δ32 deletion that confers partial resistance to HIV infection. He timed the announcement to coincide with the Second International Summit on Human Genome Editing in Hong Kong (27–29 November 2018), where he was due to present; the work was not preceded by peer review or regulatory approval and had been conducted in secret. The Shenzhen court that later sentenced him confirmed the existence of a third edited child born to a second couple in 2019.77

CRISPR-Cas9 technology (developed by Jennifer Doudna and Emmanuelle Charpentier, Nobel Prize 2020) is genuinely transformative: it enables precise gene editing with applications in medicine, agriculture, and research.78 The question is not whether CRISPR should exist, but who should govern its use, under what constraints, and with what consent requirements.

The scientific community’s response was immediate condemnation. The reasons are multiple: He’s own sequencing data showed that the editing had not in fact reproduced the natural CCR5-Δ32 deletion — Lulu carried at least one off-target intergenic mutation, and both twins showed mosaicism (different cell lines carrying different edits in the same embryo); the editing was performed on healthy embryos (the babies’ father was HIV-positive, but standard sperm-washing IVF protocols already prevent paternal-to-embryo HIV transmission); the long-term safety for the children and their potential offspring is unknown; and the consent process was ethically compromised. He was sentenced to three years in prison by a Shenzhen court on 30 December 2019, and released in April 2022.79

7.3 Oppenheimer and Responsibility

J. Robert Oppenheimer directed the Manhattan Project at Los Alamos, which developed the atomic bombs dropped on Hiroshima and Nagasaki in August 1945. After watching the first test explosion at Trinity, New Mexico, on 16 July 1945, he later recalled:

“We knew the world would not be the same. A few people laughed, a few people cried. Most people were silent. I remembered the line from the Hindu scripture, the Bhagavad Gita; Vishnu is trying to persuade the Prince that he should do his duty, and to impress him, takes on his multi-armed form and says: ‘Now I am become death, the destroyer of worlds.’ I suppose we all thought that, one way or another.” — J. Robert Oppenheimer, in the NBC News documentary The Decision to Drop the Bomb (1965), directed by Fred Freed.80

The question of scientific responsibility is sharpest here. The scientists who built the bomb knew its purpose. Many had been motivated by the fear that Nazi Germany would build one first. When Germany surrendered in May 1945, the project continued toward its use against Japan.

The Szilard petition. Leo Szilard — the physicist who had originally conceived the nuclear chain reaction, and who in 1939 had drafted with Einstein the letter to President Roosevelt that initiated the American bomb programme81 — spent the summer of 1945 trying to stop the bomb’s use. In July 1945, Szilard drafted a petition urging President Truman not to use atomic weapons against Japan without a public demonstration and an explicit surrender ultimatum first. Approximately seventy Manhattan Project scientists signed it, primarily at the Metallurgical Laboratory in Chicago.82 General Leslie Groves, the military head of the project, blocked the petition from reaching the President before the bomb was dropped on Hiroshima on 6 August 1945; it was subsequently classified, and its signatories were, in several cases, punished in the security-clearance reviews of the following decade. The petition is the ethical fulcrum of the whole Manhattan Project history: the scientists closest to the work were not silent about what they feared it would be used for, and the channels through which scientific conscience could reach political decision-makers were closed by the very institutional arrangements that had made the work possible in the first place.

The ethical question, then, is not only about what scientists know when they produce a weapon. It is also about what they can do with that knowledge inside an institutional architecture designed to use their work while silencing their judgement. The claim that science is morally neutral — that knowledge is just knowledge, and responsibility lies with those who apply it — is a claim that the Szilard petition already refutes: the scientists attempted responsibility and were structurally denied it.

7.4 Questions to Argue About

  • The Tuskegee researchers argued (implicitly) that the scientific knowledge gained justified the harm to the participants. Can any scientific knowledge justify research that violates consent? Or is consent an absolute constraint?
  • He Jiankui conducted gene editing on human embryos without adequate oversight. But if the editing was safe and the children were not harmed, was there a genuine ethical violation? Or only a procedural one?
  • Oppenheimer helped build the atomic bomb, then spent the rest of his life with ambivalence about having done so. Was he morally responsible for Hiroshima and Nagasaki? What is the right account of scientists’ responsibility for applications of their work?
  • Is there scientific research that should simply not be done — regardless of whether it could be done safely — because the knowledge it would produce is too dangerous? What would determine this?

Forced Fork: Does a Scientist Bear Moral Responsibility for Applications of Their Work?

Position A: Scientists bear genuine moral responsibility for foreseeable applications of their work, scaled by the specificity and proximity of the foresight. Oppenheimer knew the specific intended use of the bomb (and Szilard’s petition shows the scientific community knew); He Jiankui knew the specific edited embryos he was implanting; Haber personally directed the chlorine release at Ypres. The “science is neutral; responsibility lies with users” defence works for diffuse, hard-to-foresee applications (number theory in cryptography decades later) but breaks down precisely where the application is specific, proximate, and known to the researcher. Where it breaks down, scientists carry responsibility, and the institutional architectures that prevent them from acting on it (Groves blocking the Szilard petition) carry institutional responsibility.

Position B: Holding scientists responsible for foreseeable applications is practically unworkable except in extreme cases. “Foreseeable” admits no clean boundary: almost any basic research has some foreseeable misuse if you pull the chain far enough. The chemist who studies molecular bonding is not responsible for explosives; the mathematician who develops number theory is not responsible for surveillance; the geneticist who sequences a pathogen is not responsible for its weaponisation. Moral responsibility attaches in the first instance to those who decide on use — funders, militaries, regulators, governments. Extending it backward to researchers as a default makes basic science morally intolerable as a profession and produces no observable improvement in actual decisions about use, which are made by people far removed from the laboratory. The He Jiankui case is correctly handled by Position B: he is responsible because he himself made and implanted the embryos, not because he discovered something dangerous.

Choose one. The hardest case for Position B is Fritz Haber, who actively directed chemical weapons deployment — and whose Haber-Bosch process now feeds half the world. Does the distinction between researching and directing deployment give Position B a principled escape (Haber bears responsibility for the second, not the first), or does it simply concede that the unit of moral analysis is the act rather than the person — at which point Position B has lost the ability to call any researcher categorically immune?


8 What is the relationship between science and truth?

Science is widely taken to be the most reliable method for producing knowledge about the natural world. That assessment is widely shared — and, interestingly, one that science itself cannot fully justify without circularity, since any argument for science’s reliability must ultimately use the methods of science. (Hume identified the structurally similar problem about induction; the parallel is exact.) Some philosophers have argued that this circularity is benign and unavoidable for any knowledge-producing practice; others (Feyerabend, some constructivists) have argued that it should make us more modest about science’s claim to special epistemic status compared with other practices. What “most reliable” means — reliable for what purposes, by what measure, in comparison with what alternatives — is itself part of the philosophical question, as is what the relationship between scientific knowledge and truth turns out to be.

Fritz Haber and the Nitrogen Problem

Fritz Haber was awarded the Nobel Prize in Chemistry in 1918 for developing, together with Carl Bosch, a process for synthesising ammonia from atmospheric nitrogen — the Haber-Bosch process, which made nitrogen-based fertiliser available at industrial scale for the first time. The process is now responsible for feeding approximately half the world’s population: without synthetic nitrogen fertiliser, the soil could not support current human numbers.83 Haber is also the chemist who developed and supervised the first large-scale use of chlorine gas as a weapon of war at Ypres on 22 April 1915, and who spent the First World War directing Germany’s chemical weapons programme.84 His wife Clara Immerwahr, herself a chemist and pacifist, shot herself in their garden on the night of 2 May 1915, hours after Haber had returned from supervising the Ypres attack.85 Haber’s biography is the most concentrated available illustration of the relationship between scientific knowledge and technological power. The same intellectual achievement that prevents mass starvation also enables mass killing; the same process of scientific development cannot be neatly separated into its beneficial and its destructive applications. Whether the scientist bears moral responsibility for the uses to which knowledge is put is a question Haber’s life makes impossible to answer comfortably in either direction.

8.1 Scientific Consensus

Climate science provides the most politically charged example. The Intergovernmental Panel on Climate Change (IPCC) represents a global consensus among climate scientists: Earth’s average temperature has increased approximately 1.1°C since pre-industrial times, the cause is primarily the emission of greenhouse gases from human activity, and continued emissions will cause further, potentially catastrophic changes.86

This consensus is about as solid as scientific consensus gets: thousands of independent researchers, multiple independent lines of evidence (temperature records, sea level rise, ice cores, ocean acidification, phenological changes), multiple independent methodologies.

The “97%” figure that sticks in public discourse comes from one specific study — Cook et al. (2013) — which classified the abstracts of 11,944 climate papers (using AI plus human raters) as endorsing, rejecting, or being neutral on anthropogenic climate change; of those that took a position, 97.1% endorsed it.87 Subsequent surveys with different methodologies — Powell (2017) examining 11,602 peer-reviewed articles, Lynas et al. (2021) sampling more recent literature — report consensus levels above 99%. “97%” sticks in public memory because the Cook paper landed there first; what the underlying literature actually offers is one methodologically explicit estimate among several converging estimates, not a single magic number.

Naomi Oreskes and Erik Conway, in Merchants of Doubt (2010), document how a small group of scientists with political commitments strategically exploited the media’s “balance” convention (presenting “both sides”) to create a misleading picture of scientific controversy about climate change, acid rain, and the ozone hole.88

Yet public scepticism persists — particularly in certain political communities. Why? The reasons are not primarily epistemological: they are political, economic, and motivated. The fossil fuel industry has funded deliberate campaigns to cast doubt on climate science, importing the tobacco industry’s strategy of manufacturing uncertainty where the scientific picture is clear. This is a case where the sociology of knowledge — who is producing doubt, for what purposes, with what funding — is directly relevant to epistemology.

8.2 Does Science Approach Truth?

Three positions:

Scientific realism: science converges toward truth. The success of science in predicting novel phenomena, manipulating the world, and extending to new domains is best explained by assuming our theories are approximately true. As theories improve, they become more accurate descriptions of reality.

Anti-realism (instrumentalism): science produces models that work, not descriptions that are true. “Truth” doesn’t apply to scientific theories; “empirical adequacy” is the appropriate criterion. Science doesn’t converge toward truth; it converges toward better prediction.

Structural realism (John Worrall): science correctly describes the structure of reality (the mathematical relationships), even if we can’t know the nature of the underlying entities. What persists through scientific revolutions is the mathematical structure (Fresnel’s equations for light, for example, were preserved in Maxwell’s electromagnetism even as the ontology changed from elastic ether to electromagnetic field).89

The historical record gives some support to anti-realism: the history of science is littered with theories that were empirically successful for a time and then abandoned — the caloric theory of heat, phlogiston, the luminiferous ether. If past empirically successful theories were false, why should we expect current ones to be true? This is Larry Laudan’s pessimistic meta-induction: the history of science is a history of theories that were empirically successful and subsequently judged false.90

8.3 COVID-19 and Scientific Updating in Real Time

The COVID-19 pandemic offered an unusually compressed view of scientific consensus forming, failing, and reforming in public.

Two specific cases. On 14 January 2020 the WHO publicly reported that “preliminary investigations conducted by the Chinese authorities have found no clear evidence of human-to-human transmission” of the novel coronavirus identified in Wuhan; within two weeks the same agency had reversed the claim and was preparing to declare a Public Health Emergency of International Concern (30 January 2020).91 More substantively, WHO’s early transmission guidance treated SARS-CoV-2 as a droplet-borne pathogen — hence the 1–2 metre / 6 ft distancing rule — and discounted airborne aerosol transmission. In July 2020 Lidia Morawska and Donald Milton, with 239 co-signing scientists, published an open letter in Clinical Infectious Diseases arguing that the existing evidence already established aerosol transmission and that ventilation, not just distance, was the operative variable.92 WHO partially acknowledged airborne transmission on 30 April 2021, more fully in December 2021, and in April 2024 redefined “airborne” in a way that absorbed the aerosol-scientists’ position.93

This is the question of how scientific claims are revised as evidence accumulates, played out at unusual speed and in public view. On a Kuhnian reading the COVID episode runs the familiar pattern — anomaly, contested expert dissent, gradual incorporation — at compressed speed; on a Bayesian-updating reading it is just rapid revision under a strong evidential signal, with no need for paradigm-talk. Whether what science was doing in those eighteen months counts as converging on the truth about how the virus actually spreads, or merely as replacing one useful model with a better one, is what Position A and Position B below disagree about.

8.4 AI and the Production of Scientific Knowledge

Large language models can now produce plausible-sounding scientific abstracts, generate hypotheses, and write up experimental results — in some cases passing peer review.94 This raises a version of the epistemological questions about shared knowledge: when an AI system produces a scientific claim, what is the epistemic status of that claim? Is it knowledge? Whose?

The scientific community’s current response — developing detection tools, requiring data and code sharing, tightening peer review — is essentially an extension of existing epistemic practices: provenance, transparency, independent replication. What AI-generated scientific content reveals is that these practices were always the epistemologically relevant ones; they matter more, not less, as the cost of generating plausible content falls.

The more interesting question is the upstream one: AI systems trained on the existing scientific literature may systematically amplify the biases, errors, and publication distortions in that literature. If the replication crisis means that the published literature contains many false positives, then AI systems trained on that literature will “know” things that are false — and will express that false knowledge with the confident, technically correct language of scientific writing.

8.5 Questions to Argue About

  • Laudan’s pessimistic meta-induction argues that since past scientific theories were false, current ones probably are too. Does this undermine confidence in science? Or is there an asymmetry between past failures and current success?
  • Scientific consensus is how science communicates its collective judgement about well-established findings. But scientific consensus has been wrong before (continental drift was rejected by consensus for decades). When, if ever, is it rational to dissent from scientific consensus?
  • The WHO’s transmission guidance for SARS-CoV-2 changed substantially between January 2020 and April 2024 — from “no clear evidence of human-to-human transmission” through droplet-only to a redefined notion of airborne. Does this look more like science approaching the truth about how the virus spreads, or more like science replacing one useful model with a better one without either being “true”?
  • If science doesn’t converge on truth but only on better models, does it matter? Should our practical attitude toward science depend on our answer to this philosophical question?

Forced Fork: From Ptolemy to the Higgs — Is This Convergence on Truth or Just Better Prediction?

Position A: Science converges on truth in a meaningful sense. The pessimistic meta-induction points to failed theories, but it ignores the asymmetry between past and present science: modern theories make orders-of-magnitude more precise predictions, survive far more severe tests, and extend successfully to far more domains. The Ptolemaic system predicted planetary positions; it did not predict gravitational lensing, the expansion of the universe, or the Higgs boson. Convergence is real, even if imperfect.

Position B: The history of science — caloric, phlogiston, the luminiferous ether, Laudan’s full catalogue — gives us no reason to believe in convergence toward truth. Every era has thought its theories were approximately true and its successor theories have shown them to be radically wrong. Structural realism (Worrall) is the best available defence of realism, but what it shows is that the mathematical structures survive theory change — not that our ontological commitments do. We are convergent on better prediction and better mathematical structure; whether this tracks truth in any deeper sense is not established.

Choose one. If you accept Position A, address Laudan’s catalogue directly: name the specific feature of modern theories (precision? unification? novel-prediction success?) that makes you confident the present case is unlike phlogiston, and explain why nineteenth-century physicists could not have made the same speech about ether. If you accept Position B, address the action-guidance question: an instrumentalist or constructive empiricist still has to decide whether to act on climate projections, build particle accelerators, prescribe a vaccine. What is the difference, in practice, between deciding to act on a model that “predicts well” versus a model that is “approximately true” — and if there is no practical difference, has Position B forfeited its philosophical bite?


9 Media

  • Richard Feynman, The Pleasure of Finding Things Out (1981 lecture, published 1999) — Feynman on what science feels like from the inside: the pleasure of uncertainty, the discipline of not knowing, the joy of discovery. One of the most honest accounts of scientific practice by a great scientist.
  • Morten Tyldum, The Imitation Game (2014) — Dramatisation of Alan Turing’s work at Bletchley Park; more relevant to mathematics and logic than to the philosophy of science, but raises the ethics of secrecy, sexuality, and the treatment of scientific genius by the state.
  • Carl Sagan, Contact (1985, novel; 1997 film) — A meditation on the relationship between scientific evidence and belief; unusually rigorous about epistemological questions for popular science fiction. The film’s ending handles the tension between evidence and faith more carefully than most accounts of the science-religion conflict.
  • Primo Levi, The Periodic Table (1975) — Memoirs structured around the elements of the periodic table; a chemist’s account of his own experience, including Auschwitz. Not primarily about the philosophy of science, but about what science can mean to a person and what it cannot protect against.
  • Christopher Nolan, Oppenheimer (2023) — Three hours on the Manhattan Project, the Trinity test, and Oppenheimer’s security hearing. The film is primarily about political power and conscience, not physics; but it forces the question of scientific responsibility in a form that is difficult to evade.
  • Stanley Kramer, Inherit the Wind (1960) — A dramatisation of the 1925 Scopes “Monkey Trial,” in which a Tennessee schoolteacher was prosecuted for teaching evolution. The film is more useful than the trial transcript for thinking about the rhetorical structure of science-vs.-religion confrontations and the conditions under which scientific evidence is or is not accepted as politically relevant.
  • James Watson, The Double Helix (1968) — Watson’s account of the discovery of the structure of DNA; read against the documented record of Rosalind Franklin’s contribution, which Watson minimised. The book is an invaluable case study in the social dimensions of scientific discovery, credit, and gender bias in science.

10 Bibliography

Anderson, Philip W. “More Is Different.” Science 177 (1972): 393–396.

Baars, Bernard J. A Cognitive Theory of Consciousness. Cambridge: Cambridge University Press, 1988.

Bacon, Francis. Novum Organum. 1620. Trans. Peter Urbach and John Gibson. Chicago: Open Court, 1994.

Baker, Monya. “1,500 Scientists Lift the Lid on Reproducibility.” Nature 533 (2016): 452–454.

Bargh, John A., Mark Chen and Lara Burrows. “Automaticity of Social Behavior: Direct Effects of Trait Construct and Stereotype Activation on Action.” Journal of Personality and Social Psychology 71.2 (1996): 230–244.

Box, George E. P. “Science and Statistics.” Journal of the American Statistical Association 71.356 (1976): 791–799.

Cartwright, Nancy. How the Laws of Physics Lie. Oxford: Clarendon Press, 1983.

Carter, K. Codell and Barbara R. Carter. Childbed Fever: A Scientific Biography of Ignaz Semmelweis. Westport, CT: Greenwood, 1994.

Charles, Daniel. Master Mind: The Rise and Fall of Fritz Haber, the Nobel Laureate Who Launched the Age of Chemical Warfare. New York: Ecco, 2005.

Chalmers, David J. “Facing Up to the Problem of Consciousness.” Journal of Consciousness Studies 2.3 (1995): 200–219.

Chalmers, David J. The Conscious Mind: In Search of a Fundamental Theory. Oxford: Oxford University Press, 1996.

Chamberlin, Rollin T. Contribution to the Symposium on Continental Drift, in Theory of Continental Drift: A Symposium on the Origin and Movement of Land Masses (American Association of Petroleum Geologists, 1928).

Clark, Andy. “Whatever Next? Predictive Brains, Situated Agents, and the Future of Cognitive Science.” Behavioral and Brain Sciences 36.3 (2013): 181–204.

Cook, John, et al. “Quantifying the Consensus on Anthropogenic Global Warming in the Scientific Literature.” Environmental Research Letters 8 (2013): 024024.

Douglas, Heather E. Science, Policy, and the Value-Free Ideal. Pittsburgh: University of Pittsburgh Press, 2009.

Doyen, Stéphane, Olivier Klein, Cora-Lise Pichon and Axel Cleeremans. “Behavioral Priming: It’s All in the Mind, But Whose Mind?” PLoS ONE 7.1 (2012): e29081.

Else, Holly. “Abstracts Written by ChatGPT Fool Scientists.” Nature 613 (12 January 2023): 423.

Englert, François and Robert Brout. “Broken Symmetry and the Mass of Gauge Vector Mesons.” Physical Review Letters 13 (1964): 321–323.

Feyerabend, Paul. Against Method: Outline of an Anarchistic Theory of Knowledge. London: New Left Books, 1975.

Fisher, Seymour and Roger P. Greenberg. The Scientific Credibility of Freud’s Theories and Therapy. New York: Basic Books, 1977.

Gould, Stephen Jay. The Mismeasure of Man. Revised and expanded ed. New York: W. W. Norton, 1996.

Greely, Henry T. “CRISPR’d Babies: Human Germline Genome Editing in the ‘He Jiankui Affair’.” Journal of Law and the Biosciences 6.1 (2019): 111–183.

Guralnik, Gerald S., C. R. Hagen and Tom Kibble. “Global Conservation Laws and Massless Particles.” Physical Review Letters 13 (1964): 585–587.

Hacking, Ian. Representing and Intervening: Introductory Topics in the Philosophy of Natural Science. Cambridge: Cambridge University Press, 1983.

Hacking, Ian. The Social Construction of What? Cambridge, MA: Harvard University Press, 1999.

Hanson, Norwood Russell. Patterns of Discovery: An Inquiry into the Conceptual Foundations of Science. Cambridge: Cambridge University Press, 1958.

Harding, Sandra. Whose Science? Whose Knowledge? Thinking from Women’s Lives. Ithaca: Cornell University Press, 1991.

Hempel, Carl G. Philosophy of Natural Science. Englewood Cliffs, NJ: Prentice-Hall, 1966.

Hertz, Heinrich. Electric Waves: Being Researches on the Propagation of Electric Action with Finite Velocity through Space. Trans. D. E. Jones. London: Macmillan, 1893.

Hess, Harry H. “History of Ocean Basins.” In A. E. J. Engel, H. L. James and B. F. Leonard (eds.), Petrologic Studies: A Volume in Honor of A. F. Buddington. Boulder, CO: Geological Society of America, 1962, pp. 599–620.

Higgs, Peter. “Broken Symmetries and the Masses of Gauge Bosons.” Physical Review Letters 13 (1964): 508–509.

Holton, Gerald. “Subelectrons, Presuppositions, and the Millikan-Ehrenhaft Dispute.” Historical Studies in the Physical Sciences 9 (1978): 161–224.

Intergovernmental Panel on Climate Change. Climate Change 2021: The Physical Science Basis. Contribution of Working Group I to the Sixth Assessment Report of the Intergovernmental Panel on Climate Change. Cambridge: Cambridge University Press, 2021.

Ioannidis, John P. A. “Why Most Published Research Findings Are False.” PLoS Medicine 2.8 (2005): e124.

Jinek, Martin, Krzysztof Chylinski, Ines Fonfara, Michael Hauer, Jennifer A. Doudna and Emmanuelle Charpentier. “A Programmable Dual-RNA–Guided DNA Endonuclease in Adaptive Bacterial Immunity.” Science 337.6096 (2012): 816–821.

Jones, James H. Bad Blood: The Tuskegee Syphilis Experiment. New York: Free Press, 1981.

Kepler, Johannes. Astronomia Nova. 1609. Trans. William H. Donahue. Cambridge: Cambridge University Press, 1992.

Kuhn, Thomas S. The Structure of Scientific Revolutions. Chicago: University of Chicago Press, 1962.

Lakatos, Imre. The Methodology of Scientific Research Programmes. Ed. John Worrall and Gregory Currie. Cambridge: Cambridge University Press, 1978.

Laudan, Larry. “A Confutation of Convergent Realism.” Philosophy of Science 48.1 (1981): 19–49.

Lewis, Dyani. “Why the WHO Took Two Years to Say COVID Is Airborne.” Nature 604 (2022): 26–31.

Marshall, Barry J. and J. Robin Warren. “Unidentified Curved Bacilli in the Stomach of Patients with Gastritis and Peptic Ulceration.” Lancet 323.8390 (1984): 1311–1315.

Maxwell, James Clerk. “A Dynamical Theory of the Electromagnetic Field.” Philosophical Transactions of the Royal Society of London 155 (1865): 459–512.

Mayr, Ernst. The Growth of Biological Thought: Diversity, Evolution, and Inheritance. Cambridge, MA: Belknap/Harvard, 1982.

Medawar, Peter. “Is the Scientific Paper a Fraud?” The Listener 70 (12 September 1963): 377–378.

Morawska, Lidia and Donald K. Milton. “It Is Time to Address Airborne Transmission of Coronavirus Disease 2019 (COVID-19).” Clinical Infectious Diseases 71.9 (2020): 2311–2313.

Nagel, Thomas. “What Is It Like to Be a Bat?” Philosophical Review 83.4 (1974): 435–450.

National Commission for the Protection of Human Subjects of Biomedical and Behavioral Research. The Belmont Report: Ethical Principles and Guidelines for the Protection of Human Subjects of Research. DHEW Publication No. (OS) 78-0012. Washington, DC: U.S. Government Printing Office, 1979.

Nuland, Sherwin B. The Doctors’ Plague: Germs, Childbed Fever, and the Strange Story of Ignác Semmelweis. New York: W. W. Norton, 2003.

Okasha, Samir. Philosophy of Science: A Very Short Introduction. Oxford: Oxford University Press, 2002.

Open Science Collaboration. “Estimating the Reproducibility of Psychological Science.” Science 349.6251 (2015): aac4716.

Oreskes, Naomi. The Rejection of Continental Drift: Theory and Method in American Earth Science. New York: Oxford University Press, 1999.

Oreskes, Naomi and Erik M. Conway. Merchants of Doubt. New York: Bloomsbury Press, 2010.

Ozawa, Masanao. “Universally Valid Reformulation of the Heisenberg Uncertainty Principle on Noise and Disturbance in Measurement.” Physical Review A 67 (2003): 042105.

Planck Collaboration. “Planck 2018 Results. VI. Cosmological Parameters.” Astronomy & Astrophysics 641 (2020): A6.

Popper, Karl. The Logic of Scientific Discovery. German 1934, English trans. London: Hutchinson, 1959.

Popper, Karl. Conjectures and Refutations: The Growth of Scientific Knowledge. London: Routledge and Kegan Paul, 1963.

Rhodes, Richard. The Making of the Atomic Bomb. New York: Simon & Schuster, 1986.

Rosenthal-Schneider, Ilse. Reality and Scientific Truth: Discussions with Einstein, von Laue, and Planck. Detroit: Wayne State University Press, 1980.

Rubin, Vera C. and W. Kent Ford, Jr. “Rotation of the Andromeda Nebula from a Spectroscopic Survey of Emission Regions.” Astrophysical Journal 159 (1970): 379–403.

Semmelweis, Ignaz. Die Aetiologie, der Begriff und die Prophylaxis des Kindbettfiebers. Pest, Vienna and Leipzig: C. A. Hartleben, 1861.

Smil, Vaclav. Enriching the Earth: Fritz Haber, Carl Bosch, and the Transformation of World Food Production. Cambridge, MA: MIT Press, 2001.

Stoltzenberg, Dietrich. Fritz Haber: Chemist, Nobel Laureate, German, Jew. Philadelphia: Chemical Heritage Foundation, 2004.

Stroebe, Wolfgang and Fritz Strack (Levelt Committee et al.). Flawed Science: The Fraudulent Research Practices of Social Psychologist Diederik Stapel (Levelt Committee Final Report). Tilburg, 2012.

Taleb, Nassim Nicholas. The Black Swan: The Impact of the Highly Improbable. New York: Random House, 2007.

Tononi, Giulio. “An Information Integration Theory of Consciousness.” BMC Neuroscience 5 (2004): 42.

van Fraassen, Bas C. The Scientific Image. Oxford: Clarendon Press, 1980.

Vine, F. J. and D. H. Matthews. “Magnetic Anomalies over Oceanic Ridges.” Nature 199 (1963): 947–949.

Wegener, Alfred. The Origin of Continents and Oceans. 1915. Trans. John Biram. New York: Dover, 1966.

Wigner, Eugene. “The Unreasonable Effectiveness of Mathematics in the Natural Sciences.” Communications on Pure and Applied Mathematics 13.1 (1960): 1–14.

World Health Organization. Global Technical Consultation Report on Proposed Terminology for Pathogens that Transmit through the Air. Geneva: WHO, 18 April 2024.

Worrall, John. “Structural Realism: The Best of Both Worlds?” Dialectica 43.1–2 (1989): 99–124.


  1. The Dover Area School Board’s policy was adopted on 18 October 2004; the four-paragraph statement read aloud to ninth-grade biology students included: “Because Darwin’s Theory is a theory, it is still being tested as new evidence is discovered. The Theory is not a fact. Gaps in the Theory exist for which there is no evidence. … Intelligent Design is an explanation of the origin of life that differs from Darwin’s view.” Full text in the trial record. [VERIFY]↩︎

  2. Tammy Kitzmiller, et al. v. Dover Area School District, et al., 400 F. Supp. 2d 707 (M.D. Pa. 2005). Plaintiffs were represented by the American Civil Liberties Union of Pennsylvania, Americans United for Separation of Church and State, and the law firm Pepper Hamilton LLP. The defendants were represented by the Thomas More Law Center. [VERIFY]↩︎

  3. For the trial witnesses see the trial transcripts (available on the National Center for Science Education website) and Edward Humes, Monkey Girl: Evolution, Education, Religion, and the Battle for America’s Soul (New York: Ecco, 2007). Plaintiffs’ expert witnesses: Kenneth R. Miller, Robert T. Pennock, Brian Alters, Kevin Padian, Barbara Forrest, John F. Haught. Defendants’ expert witnesses: Michael Behe, Scott Minnich, Steve Fuller. [VERIFY]↩︎

  4. John E. Jones III, Memorandum Opinion in Kitzmiller v. Dover Area School District, 400 F. Supp. 2d 707, 20 December 2005 (139 pages). The opinion’s central legal holding rests on both the Lemon test and the endorsement test under the Establishment Clause; its philosophical-scientific findings (intelligent design as a non-scientific because supernaturalist; irreducible-complexity arguments rebutted; religious purpose evidenced in board minutes) appear at pages 64–89. The board did not appeal. [VERIFY]↩︎

  5. Popper, Conjectures and Refutations (1963), Ch. 1 (“Science: Conjectures and Refutations”). Popper describes the 1919 Vienna encounter with Adler; Adler died in 1937.↩︎

  6. Fisher and Greenberg, The Scientific Credibility of Freud’s Theories and Therapy (1977).↩︎

  7. Popper, The Logic of Scientific Discovery (German 1934, English trans. 1959), Ch. 1 §6 (“Falsifiability as a Criterion of Demarcation”); elaborated in Conjectures and Refutations (1963), Ch. 1 §§I–V.↩︎

  8. Taleb, The Black Swan: The Impact of the Highly Improbable (2007), Prologue (“Before and After”).↩︎

  9. Le Verrier communicated the predicted position to Johann Galle at the Berlin Observatory, who recorded Neptune on the night of 23 September 1846.↩︎

  10. Imre Lakatos, The Methodology of Scientific Research Programmes (Cambridge University Press, 1978; the centrepiece essay was first published in 1970 in Lakatos and Musgrave, eds., Criticism and the Growth of Knowledge). Lakatos distinguishes a programme’s “hard core” — protected from refutation by the negative heuristic — from its “protective belt” of auxiliary hypotheses, which absorbs the modus tollens of failed predictions. He grades programmes as progressive (the theory predicts novel facts) or degenerating (the theory is patched to accommodate facts already known).↩︎

  11. Kuhn, The Structure of Scientific Revolutions (1962), Ch. 7 (“Anomaly and the Emergence of Scientific Discoveries”) and Ch. 8 (“The Response to Crisis”).↩︎

  12. Medawar, “Is the Scientific Paper a Fraud?” The Listener (1963).↩︎

  13. Holton, “Subelectrons, Presuppositions, and the Millikan-Ehrenhaft Dispute,” Historical Studies in the Physical Sciences (1978).↩︎

  14. Bacon, Novum Organum (1620), Book I, Aphorisms 19–26 (on the method of progressive induction from particulars), and Book II, Aphorisms 11–20 (the tables of presence, absence and degree).↩︎

  15. Hanson, Patterns of Discovery (1958), Ch. 1 (“Observation”). The “theory-laden” formulation is Hanson’s.↩︎

  16. Reported in Ilse Rosenthal-Schneider, Reality and Scientific Truth: Discussions with Einstein, von Laue, and Planck (Detroit: Wayne State University Press, 1980), in the “Conversations with Einstein” section. The widely-circulated form of the remark — “Then I would have been sorry for the dear Lord. The theory is correct anyway” — is given here as a paraphrase. [VERIFY: Rosenthal-Schneider not held; verbatim uncheckable]↩︎

  17. Feyerabend, Against Method (1975), Chs. 6–11 (“Galileo”).↩︎

  18. Feyerabend, Against Method (1975), Introduction (chapter abstract) and end of Ch. 1: “It is clear, then, that the idea of a fixed method, or of a fixed theory of rationality, rests on too naive a view of man and his social surroundings. To those who look at the rich material provided by history, and who are not intent on impoverishing it in order to please their lower instincts, their craving for intellectual security in the form of clarity, precision, ‘objectivity’, ‘truth’, it will become clear that there is only one principle that can be defended under all circumstances and in all stages of human development. It is the principle: anything goes.” Feyerabend later glossed the slogan as “the terrified exclamation of a rationalist who takes a closer look at history” rather than a positive methodology.↩︎

  19. Kepler, Astronomia Nova (1609), Ch. 19. The “eight arc-minute” discrepancy Kepler describes as a gift “from which a reformation of all of astronomy can be constructed.”↩︎

  20. K. Codell Carter and Barbara R. Carter, Childbed Fever: A Scientific Biography of Ignaz Semmelweis (Greenwood, 1994), Ch. 3, narrate the chlorine intervention from late May 1847. The cited monthly percentages — April 1847 First Division mortality at 18.3%, falling to 2.2% in June and 1.2% in July — come from Semmelweis’s own Etiology (1861), trans. Carter (Wisconsin, 1983).↩︎

  21. Hempel, Philosophy of Natural Science (1966), Ch. 5 (“Laws and Their Role in Scientific Explanation”), §5.2 (“Deductive-nomological explanation”). The DN model receives its earlier full statement in Hempel and Oppenheim, “Studies in the Logic of Explanation,” Philosophy of Science 15 (1948): 135–175.↩︎

  22. Hempel, Philosophy of Natural Science (1966), Ch. 2 (“Scientific Inquiry: Invention and Test”), §2.1, opens the book with Semmelweis’s investigation of childbed fever in the Vienna General Hospital (1844–48). Hempel returns to the case in Ch. 5, §5.3, to show how Semmelweis’s eventual hypothesis fits the deductive-nomological schema once the implicit covering law (that introduction of decomposing animal matter into the bloodstream causes blood poisoning) is made explicit.↩︎

  23. Semmelweis was not reappointed in 1849; he returned to Pest and eventually published Die Aetiologie in 1861. On 29 July 1865 he was committed to a Viennese asylum on Lazarettgasse, beaten by guards, and died there on 13 August 1865 from a gangrenous wound on his hand — almost certainly the same blood poisoning he had spent his career trying to prevent. See Carter and Carter, Childbed Fever (1994), Ch. 4, and Sherwin B. Nuland, The Doctors’ Plague (Norton, 2003).↩︎

  24. Okasha, Philosophy of Science: A Very Short Introduction (2002), Ch. 3 (“Explanation in science”). The flagpole counter-example is due to Sylvain Bromberger in unpublished work of the early 1960s; Okasha presents it as the standard objection to the symmetry of explanation and prediction in Hempel’s DN model. The contrast with inference to the best explanation is developed in Okasha’s Ch. 2 (“Scientific reasoning”) and goes back to Gilbert Harman, “The Inference to the Best Explanation,” Philosophical Review 74 (1965): 88–95.↩︎

  25. Planck Collaboration, “Planck 2018 Results. VI. Cosmological Parameters,” Astronomy & Astrophysics 641 (2020): A6. The base \(\Lambda\)CDM fit to the Planck CMB temperature, polarisation and lensing data gives a present-day matter density \(\Omega_m \approx 0.315\) (of which baryons account for \(\Omega_b \approx 0.049\) and cold dark matter for \(\Omega_c \approx 0.265\)) and a dark-energy density \(\Omega_\Lambda \approx 0.685\).↩︎

  26. Box, “Science and Statistics,” Journal of the American Statistical Association (1976). The canonical aphorism “all models are wrong but some are useful” appears without the conventional comma added by later citation.↩︎

  27. Cartwright, How the Laws of Physics Lie (1983), Essay 3, attacks the literal truth of fundamental physical laws while defending phenomenological laws. Van Fraassen, The Scientific Image (1980), Ch. 2, restricts truth-claims to the observable consequences of theories while allowing the unobservable apparatus to be retained for its usefulness in inference. Both treat Box’s “useful for what?” as the live question; they disagree about which parts of a theory are “wrong” in Box’s strict sense.↩︎

  28. Gallium was isolated by Lecoq de Boisbaudran (1875), scandium by Lars Fredrik Nilson (1879), and germanium by Clemens Winkler (1886); all three matched Mendeleev’s 1869/1871 predictions for eka-aluminium, eka-boron, and eka-silicon closely.↩︎

  29. Hacking, Representing and Intervening (1983), Ch. 16 (“Experimentation and Scientific Realism”). Hacking’s manipulation argument: “if you can spray them, they are real.”↩︎

  30. Georg Ernst Stahl, Specimen Becherianum (1703); for the modern reconstruction see Hasok Chang, Inventing Temperature (Oxford University Press, 2004), Chapter 4, and James R. Partington, A History of Chemistry, Vol. 2 (Macmillan, 1961), Ch. XVIII. The standard scholarly verdict is that phlogiston was a productive theory whose empirical content survives in modern chemistry’s account of reduction-oxidation as the reverse of oxidation-combination; what was abandoned was the substantial-thing reading of phlogiston, not the relational pattern.↩︎

  31. Antoine-Laurent Lavoisier, Traité élémentaire de chimie (Paris: Cuchet, 1789); trans. Robert Kerr as Elements of Chemistry (Edinburgh, 1790). For Lavoisier’s mass-balance methodology see Frederic Lawrence Holmes, Antoine Lavoisier — The Next Crucial Year (Princeton University Press, 1998).↩︎

  32. Rubin and Ford, “Rotation of the Andromeda Nebula from a Spectroscopic Survey of Emission Regions,” Astrophysical Journal (1970). The definitive flat-rotation-curve evidence for dark matter in spiral galaxies appeared in Rubin, Ford and Thonnard’s 1978 and 1980 follow-up papers.↩︎

  33. For the experimental status of WIMP searches see Aprile et al. (XENON Collaboration), “First Dark Matter Search with Nuclear Recoils from the XENONnT Experiment,” Physical Review Letters 131 (2023): 041003; J. Aalbers et al. (LZ Collaboration), “Dark Matter Search Results from 4.2 Tonne-Years of Exposure of the LUX-ZEPLIN (LZ) Experiment,” arXiv:2410.17036 (2024); and Z. Bo et al. (PandaX Collaboration), “Dark Matter Search Results from 1.54 Tonne-Year Exposure of PandaX-4T,” Physical Review Letters 134 (2025). The leading 2024 LZ limit on the spin-independent WIMP–nucleon cross-section is ≈ 2.2 × 10⁻⁴⁸ cm² at a WIMP mass of ≈ 40 GeV/c² — roughly five orders of magnitude below the canonical 1990s supersymmetric expectations of ~10⁻⁴² to 10⁻⁴³ cm². For axion searches, see C. Bartram et al. (ADMX Collaboration), “Search for Invisible Axion Dark Matter in the 3.3–4.2 μeV Mass Range,” Physical Review Letters 127 (2021): 261803, and the MADMAX Collaboration, “A New Experimental Approach to Probe QCD Axion Dark Matter,” European Physical Journal C 79 (2019): 186. No confirmed direct detection has been reported as of 2025.↩︎

  34. van Fraassen, The Scientific Image (1980), Ch. 2 (“To Save the Phenomena”), §3 (“Constructive Empiricism”).↩︎

  35. Cartwright, How the Laws of Physics Lie (1983), Essay 3 (“Do the Laws of Physics State the Facts?”), opening section on “the facticity view of laws.” The book’s Introduction states the trade-off thesis in compact form: the more fundamental and explanatory a law looks, the less literally true it is in any concrete situation; phenomenological laws are closer to reality but less general.↩︎

  36. ATLAS Collaboration, “Observation of a New Particle in the Search for the Standard Model Higgs Boson with the ATLAS Detector at the LHC,” Physics Letters B 716 (2012): 1–29; CMS Collaboration, “Observation of a New Boson at a Mass of 125 GeV with the CMS Experiment at the LHC,” Physics Letters B 716 (2012): 30–61. Both papers report a 5σ excess; the joint announcement was made at CERN on 4 July 2012. Subsequent measurements of spin-parity (J^P = 0⁺) and coupling strengths confirmed the new particle as consistent with the Standard Model Higgs boson.↩︎

  37. Laudan, “A Confutation of Convergent Realism,” Philosophy of Science (1981). Often misattributed to Hilary Putnam, who in fact defended the “no-miracles” argument for realism in Mathematics, Matter and Method (1975).↩︎

  38. Levelt Committee, Flawed Science: The Fraudulent Research Practices of Social Psychologist Diederik Stapel (Final Report, Tilburg, 28 November 2012). Stapel was suspended from Tilburg University on 7 September 2011.↩︎

  39. Bargh, Chen and Burrows, “Automaticity of Social Behavior,” Journal of Personality and Social Psychology (1996).↩︎

  40. Doyen, Klein, Pichon and Cleeremans, “Behavioral Priming: It’s All in the Mind, But Whose Mind?” PLoS ONE (2012).↩︎

  41. Daniel Kahneman, open letter to social-priming researchers, 26 September 2012.↩︎

  42. Hanson, Patterns of Discovery (1958), Ch. 1. The Kepler–Tycho Brahe sunrise thought experiment is Hanson’s; strictly, the contrast is between a Keplerian/Copernican observer and the Ptolemaic/Tychonic astronomer Tycho Brahe, not between a “naive” observer and a Keplerian.↩︎

  43. The popular “measurement-disturbance” gloss derives from Heisenberg’s 1927 \(\gamma\)-ray microscope thought experiment; the rigorous \(\Delta x \cdot \Delta p \geq \hbar/2\) inequality (Kennard 1927; Robertson 1929) is instead a statement about state standard deviations. Masanao Ozawa, Physical Review A (2003), disentangles the two.↩︎

  44. Open Science Collaboration, “Estimating the Reproducibility of Psychological Science,” Science (2015). The paper reports several reproducibility measures for the same 100 studies: 39% of replications were judged by the replicating teams to have reproduced the original, while only 36% achieved a statistically significant result in the same direction. The two figures describe the same dataset under different criteria.↩︎

  45. Baker, “1,500 Scientists Lift the Lid on Reproducibility,” Nature (2016). Sample of 1,576 respondents.↩︎

  46. Harding, Whose Science? Whose Knowledge? (1991), Ch. 6 (“‘Strong Objectivity’ and Socially Situated Knowledge”).↩︎

  47. Gould, The Mismeasure of Man (1981; revised and expanded 1996). Gould reconstructs the nineteenth- and twentieth-century history of craniometric, IQ, and hereditarian arguments for fixed racial and sex hierarchies — Morton, Broca, Goddard, Yerkes, Burt — and shows in each case how the published numbers were shaped by the investigators’ prior commitments. The 1996 revision answers the resurgent hereditarianism of Herrnstein and Murray’s The Bell Curve (1994).↩︎

  48. Hacking, The Social Construction of What? (1999), Ch. 1 (“Why Ask What?”), §“Against Inevitability.” Hacking distinguishes the contingency claim (1) from the evaluative claims (2) “X is quite bad as it is” and (3) “We would be much better off if X were done away with, or at least radically transformed.” Most “social construction” disputes in the science wars, on Hacking’s diagnosis, conflate these.↩︎

  49. Douglas, Science, Policy, and the Value-Free Ideal (2009), Ch. 1 (“Introduction”), opening paragraph. The book’s structure: Chs. 2–3 on the historical roots of the value-free ideal (Reichenbach, Hempel, Rudner), Ch. 4 on the autonomy of science, Ch. 5 on the role of values, Ch. 6 on objectivity, Ch. 7 on science in policy.↩︎

  50. Douglas, Science, Policy, and the Value-Free Ideal (2009), Ch. 5 (“The Structure of Values in Science”), reconstructing Richard Rudner, “The Scientist Qua Scientist Makes Value Judgments,” Philosophy of Science 20 (1953): 1–6. Douglas’s earlier statement of the argument is “Inductive Risk and Values in Science,” Philosophy of Science 67 (2000): 559–579.↩︎

  51. Englert and Brout, “Broken Symmetry and the Mass of Gauge Vector Mesons,” Physical Review Letters (1964); Higgs, “Broken Symmetries and the Masses of Gauge Bosons,” Physical Review Letters (1964); Guralnik, Hagen and Kibble, “Global Conservation Laws and Massless Particles,” Physical Review Letters (1964). The particle’s mass is not predicted by the theory and had to be determined experimentally.↩︎

  52. ATLAS Collaboration, “Observation of a New Particle in the Search for the Standard Model Higgs Boson with the ATLAS Detector at the LHC,” Physics Letters B 716 (2012): 1–29; CMS Collaboration, “Observation of a New Boson at a Mass of 125 GeV with the CMS Experiment at the LHC,” Physics Letters B 716 (2012): 30–61. Both papers report a 5σ excess; the joint announcement was made at CERN on 4 July 2012. Subsequent measurements of spin-parity (J^P = 0⁺) and coupling strengths confirmed the new particle as consistent with the Standard Model Higgs boson.↩︎

  53. Wigner, “The Unreasonable Effectiveness of Mathematics in the Natural Sciences,” Communications on Pure and Applied Mathematics (1960).↩︎

  54. Kuhn, The Structure of Scientific Revolutions (1962), Ch. 2 (“The Route to Normal Science”) and Ch. 5 (“The Priority of Paradigms”). Kuhn’s later Postscript (1969) distinguishes the “disciplinary matrix” sense of paradigm from the “exemplar” sense.↩︎

  55. Kuhn, Structure, Postscript (1969), §§2–3. The “disciplinary matrix” gathers what a community shares (symbolic generalisations, models, values, exemplars); the “exemplar” is one element within it — the concrete problem-solution that students learn to work and from which they acquire tacit knowledge of what counts as a well-formed scientific question. Kuhn elaborated the distinction further in “Second Thoughts on Paradigms” (1977, in The Essential Tension).↩︎

  56. Wegener, The Origin of Continents and Oceans (1915; Eng. trans. of 4th ed. 1966), Ch. 1 (“The First Notions of Displacement”) and Ch. 3 (“The Geodetic Arguments”).↩︎

  57. Rollin T. Chamberlin, contribution to the 1926 American Association of Petroleum Geologists symposium on continental drift, published in Theory of Continental Drift: A Symposium (AAPG, 1928). The oft-quoted “utter rot” characterisation that appears in secondary sources is not directly attested in the proceedings; Chamberlin’s surviving written remark complains that Wegener’s theory would require geologists to “forget everything which has been learned in the last 70 years and start all over again.”↩︎

  58. Harry H. Hess, “History of Ocean Basins,” in A. E. J. Engel, H. L. James, and B. F. Leonard (eds.), Petrologic Studies: A Volume in Honor of A. F. Buddington (Geological Society of America, 1962), pp. 599–620 (the seafloor-spreading hypothesis); F. J. Vine and D. H. Matthews, “Magnetic Anomalies over Oceanic Ridges,” Nature 199 (1963): 947–949 (the symmetric magnetic-stripe pattern that confirmed it). The synthesis into modern plate tectonics is traced in Naomi Oreskes, The Rejection of Continental Drift: Theory and Method in American Earth Science (Oxford UP, 1999), Chs. 9–10.↩︎

  59. Oreskes, The Rejection of Continental Drift (1999), esp. Ch. 11 (“The Triumph of Plate Tectonics”). Oreskes documents the substantial differences between Wegener’s 1915 proposal and the 1960s plate-tectonics synthesis, and resists the “lone genius vindicated by history” reading.↩︎

  60. Maxwell, “A Dynamical Theory of the Electromagnetic Field,” Philosophical Transactions of the Royal Society (1865), §§VI–VIII (electromagnetic theory of light).↩︎

  61. Hertz first generated and detected electromagnetic waves in 1886–1888; the key papers are collected in Electric Waves (Eng. trans. 1893). The often-cited “1887” refers to Hertz’s paper “Über sehr schnelle electrische Schwingungen” (Annalen der Physik, 1887).↩︎

  62. Marshall and Warren, “Unidentified Curved Bacilli in the Stomach of Patients with Gastritis and Peptic Ulceration,” Lancet (1984). Nobel Prize in Physiology or Medicine (2005).↩︎

  63. Adrian M. Owen, Martin R. Coleman, Melanie Boly, Matthew H. Davis, Steven Laureys, and John D. Pickard, “Detecting Awareness in the Vegetative State,” Science 313.5792 (8 September 2006): 1402. The case is also recounted at length in Adrian Owen, Into the Gray Zone: A Neuroscientist Explores the Border Between Life and Death (New York: Scribner, 2017), Chapter 1. [VERIFY]↩︎

  64. Martin M. Monti, Audrey Vanhaudenhuyse, Martin R. Coleman, Mélanie Boly, John D. Pickard, Luaba Tshibanda, Adrian M. Owen, and Steven Laureys, “Willful Modulation of Brain Activity in Disorders of Consciousness,” New England Journal of Medicine 362.7 (18 February 2010): 579–589. Of 54 patients tested, 5 (4 vegetative-state, 1 minimally conscious) demonstrated willful modulation; one was able to communicate yes/no answers via the protocol. [VERIFY]↩︎

  65. Anderson, “More Is Different,” Science (1972).↩︎

  66. Mayr, The Growth of Biological Thought (1982), Ch. 2 (“The Place of Biology in the Sciences and Its Conceptual Structure”). Mayr’s argument for biological autonomy is developed further in This Is Biology (1997).↩︎

  67. Chalmers, “Facing Up to the Problem of Consciousness,” Journal of Consciousness Studies (1995); elaborated in The Conscious Mind (1996), Chs. 1–3. The “easy problems / hard problem” distinction is introduced in the 1995 article.↩︎

  68. Giulio Tononi, “An Information Integration Theory of Consciousness,” BMC Neuroscience (2004), gives the foundational version; the mathematical refinement is Tononi, Boly, Massimini and Koch, “Integrated Information Theory: From Consciousness to Its Physical Substrate,” Nature Reviews Neuroscience 17 (2016): 450–461. The popular monograph Phi (Pantheon, 2012) presents the cause-effect / irreducibility intuition in narrative form.↩︎

  69. Baars, A Cognitive Theory of Consciousness (1988). Dehaene’s neuronal-workspace extension is developed in Dehaene and Naccache, “Towards a Cognitive Neuroscience of Consciousness,” Cognition (2001).↩︎

  70. Clark, “Whatever Next? Predictive Brains, Situated Agents, and the Future of Cognitive Science,” Behavioral and Brain Sciences (2013); Friston, “The Free-Energy Principle: A Unified Brain Theory?” Nature Reviews Neuroscience (2010).↩︎

  71. Nagel, “What Is It Like to Be a Bat?” Philosophical Review (1974); reprinted in Mortal Questions (1979), Ch. 12.↩︎

  72. Open Science Collaboration, “Estimating the Reproducibility of Psychological Science,” Science (2015). The paper reports several reproducibility measures for the same 100 studies: 39% of replications were judged by the replicating teams to have reproduced the original, while only 36% achieved a statistically significant result in the same direction. The two figures describe the same dataset under different criteria.↩︎

  73. Ioannidis, “Why Most Published Research Findings Are False,” PLoS Medicine (2005).↩︎

  74. Jones, Bad Blood: The Tuskegee Syphilis Experiment (1981). Peter Buxtun supplied documentation to the Associated Press in 1972, ending the study.↩︎

  75. National Commission for the Protection of Human Subjects of Biomedical and Behavioral Research, The Belmont Report: Ethical Principles and Guidelines for the Protection of Human Subjects of Research (DHEW Publication No. (OS) 78-0012, 1979). The Commission was established by the National Research Act of 1974 (Pub. L. 93-348) following the 1972 exposure of the Tuskegee study. The three principles are stated in Part B (“Basic Ethical Principles”).↩︎

  76. The Nuremberg Code was formulated in 1947 as part of the judgment in United States v. Karl Brandt et al. (the Doctors’ Trial).↩︎

  77. The 30 December 2019 verdict from the Shenzhen Nanshan District People’s Court (reported by Xinhua, 30 December 2019) refers to “three genetically-edited babies”, confirming a second pregnancy beyond the Lulu and Nana twins; the third child was reportedly born in 2019 to a different couple in the same trial. The identities of all three children remain protected.↩︎

  78. Jinek, Chylinski, Fonfara, Hauer, Doudna and Charpentier, “A Programmable Dual-RNA–Guided DNA Endonuclease in Adaptive Bacterial Immunity,” Science (2012). Doudna and Charpentier received the 2020 Nobel Prize in Chemistry for the development of CRISPR-Cas9 genome editing.↩︎

  79. He Jiankui announced the births on 25 November 2018 in five YouTube videos timed to coincide with the Second International Summit on Human Genome Editing (Hong Kong, 27–29 November 2018); the Shenzhen Nanshan District People’s Court sentenced him on 30 December 2019 to three years’ imprisonment and a fine of three million yuan. He was released in April 2022. The off-target intergenic mutation in Lulu and the mosaicism in both twins are documented in He’s own slides shown at the Hong Kong summit and discussed in Henry T. Greely, “CRISPR’d Babies: Human Germline Genome Editing in the ‘He Jiankui Affair’,” Journal of Law and the Biosciences 6.1 (2019): 111–183, esp. §III.↩︎

  80. NBC News, The Decision to Drop the Bomb (1965), directed by Fred Freed. In Oppenheimer’s paraphrase of the Bhagavad Gita the speaker of the verse is Krishna rather than Vishnu, though Krishna in the Gita reveals himself as an avatar of Vishnu.↩︎

  81. Rhodes, The Making of the Atomic Bomb (1986), Ch. 9 (“An Extensive Burst”). The letter, drafted by Szilard in consultation with Eugene Wigner and Edward Teller and signed by Einstein on 2 August 1939 at Peconic, Long Island, was delivered to Roosevelt by the economist Alexander Sachs on 11 October 1939; it led directly to the formation of the Advisory Committee on Uranium.↩︎

  82. Rhodes, The Making of the Atomic Bomb (1986), Ch. 18 (“Trinity”). The 17 July 1945 petition collected roughly seventy signatures, primarily at the Metallurgical Laboratory in Chicago.↩︎

  83. Smil, Enriching the Earth: Fritz Haber, Carl Bosch, and the Transformation of World Food Production (Cambridge, MA: MIT Press, 2001), Ch. 1 (“Agricultures and Populations”) and the closing summary, which states that without synthetic ammonia “today’s global population would not stand at six billion people” and quantifies the share of dietary nitrogen now traceable to Haber-Bosch synthesis at roughly 40% of humanity’s protein intake — the basis for the rough “feeds half the world” formulation used in the lesson. The dependence is projected to rise.↩︎

  84. The first large-scale chlorine gas attack was launched at the Second Battle of Ypres on 22 April 1915, under Haber’s supervision.↩︎

  85. Daniel Charles, Master Mind: The Rise and Fall of Fritz Haber, the Nobel Laureate Who Launched the Age of Chemical Warfare (New York: Ecco, 2005), Ch. 14 (“A Suicide”), pp. 183–185, dates Clara Immerwahr’s suicide to the night of 1–2 May 1915, hours after Haber returned from supervising the chlorine release at Ypres on 22 April 1915. The standard scholarly biography is Dietrich Stoltzenberg, Fritz Haber: Chemist, Nobel Laureate, German, Jew (Philadelphia: Chemical Heritage Foundation, 2004). [VERIFY: Stoltzenberg not held; substituted via Charles Master Mind (id 1466)]↩︎

  86. IPCC, Climate Change 2021: The Physical Science Basis. Contribution of Working Group I to the Sixth Assessment Report (Cambridge UP, 2021), “Summary for Policymakers,” §A.1.2: “The likely range of total human-caused global surface temperature increase from 1850–1900 to 2010–2019 is 0.8°C to 1.3°C, with a best estimate of 1.07°C.” §A.1: “It is unequivocal that human influence has warmed the atmosphere, ocean and land.”↩︎

  87. Cook et al., “Quantifying the Consensus on Anthropogenic Global Warming in the Scientific Literature,” Environmental Research Letters (2013). The specific Cook methodology and its 97.1% figure have been contested (see responses from Tol and others), but several independent analyses (Doran & Zimmermann 2009; Anderegg et al. 2010; Powell 2017) converge on similarly high consensus estimates.↩︎

  88. Oreskes and Conway, Merchants of Doubt (2010), esp. Ch. 6 (“The Denial of Global Warming”) and Ch. 7 (“Denial Rides Again”).↩︎

  89. Worrall, “Structural Realism: The Best of Both Worlds?” Dialectica (1989).↩︎

  90. Laudan, “A Confutation of Convergent Realism,” Philosophy of Science (1981). Often misattributed to Hilary Putnam, who in fact defended the “no-miracles” argument for realism in Mathematics, Matter and Method (1975).↩︎

  91. World Health Organization, official Twitter/X statement, 14 January 2020 (status 1217043229427761152): “Preliminary investigations conducted by the Chinese authorities have found no clear evidence of human-to-human transmission of the novel #coronavirus (2019-nCoV) identified in #Wuhan, #China.” On the same day WHO’s technical lead Maria Van Kerkhove cautioned in a Geneva press briefing that limited human-to-human transmission could not be ruled out. Sustained human-to-human transmission was officially confirmed on 22 January 2020; the Public Health Emergency of International Concern was declared on 30 January 2020.↩︎

  92. Lidia Morawska and Donald K. Milton, “It Is Time to Address Airborne Transmission of Coronavirus Disease 2019 (COVID-19),” Clinical Infectious Diseases 71.9 (2020): 2311–2313. Co-signed by 239 scientists from 32 countries.↩︎

  93. WHO Scientific Brief, “Coronavirus disease (COVID-19): How is it transmitted?”, updated 30 April 2021, was the first WHO document to acknowledge airborne aerosol transmission alongside droplet transmission; the December 2021 update placed aerosol transmission on equal footing. WHO and partners’ global technical consultation report, Global Technical Consultation Report on Proposed Terminology for Pathogens that Transmit through the Air (Geneva: WHO, 18 April 2024), formally redefined “through the air” transmission so as to absorb the aerosol-scientists’ position. See also Dyani Lewis, “Why the WHO Took Two Years to Say COVID Is Airborne,” Nature 604 (2022): 26–31.↩︎

  94. Holly Else, “Abstracts Written by ChatGPT Fool Scientists,” Nature 613 (12 January 2023): 423. Reports the preprint by Catherine A. Gao et al., “Comparing Scientific Abstracts Generated by ChatGPT to Original Abstracts Using an Artificial Intelligence Output Detector, Plagiarism Detector, and Blinded Human Reviewers,” bioRxiv 2022.12.23.521610, in which blinded reviewers misidentified 32% of ChatGPT-generated abstracts as genuine.↩︎