1 What is an argument?

In everyday speech, “argument” means a quarrel. In logic, it means something precise and technical: a set of propositions in which some (the premises) are offered as reasons to believe another (the conclusion). The confusion between these two senses is itself philosophically interesting — and tells us something about what distinguishes the logical enterprise from mere persuasion.

Surgisphere, The Lancet, and the Disappearance of an Argument

In April 2020, in the early months of the COVID-19 pandemic, the Chicago-based health-data company Surgisphere supplied datasets to two major studies of cardiovascular and antimalarial drugs in COVID-19 patients. On 22 May 2020 The Lancet published “Hydroxychloroquine or chloroquine with or without a macrolide for treatment of COVID-19: a multinational registry analysis,” authored by Mandeep Mehra and colleagues, drawing on Surgisphere data covering 96,032 patients across 671 hospitals on six continents.1 The argument’s structure was clean: from a very large multinational observational dataset, controlling for known confounders, the authors inferred that hydroxychloroquine was associated with increased in-hospital mortality. The World Health Organization paused the hydroxychloroquine arm of its global Solidarity trial within days. The New England Journal of Medicine had published a related Surgisphere-based paper on 1 May. Within two weeks, an open letter from James Watson and over 200 scientists in 24 countries raised specific concerns: the data showed Australian COVID-19 deaths exceeding the country’s actual COVID-19 deaths at the time of analysis; African hospitals reported electronic-health-record completeness incompatible with the local clinical infrastructure; the company refused independent audit.2 On 4 June 2020 the lead non-Surgisphere authors of both papers issued retraction requests citing their inability to verify the underlying data; The Lancet retracted its paper the same day, NEJM the next.3 The argument’s logical structure had been impeccable. Every empirical premise it rested on was, on examination, independently unverifiable. The classic illustration of an argument whose premises can be independently verified is John Snow’s cholera-map analysis (treated below); Surgisphere is the contemporary illustration of what happens when the same logical structure rests on premises that the institutions of peer review cannot, in fact, check.

1.1 John Snow and the Broad Street Pump

In August 1854, cholera killed more than five hundred people within ten days in the Soho district of London. The prevailing medical theory — miasma, the idea that disease spread through foul-smelling air — had the support of the General Board of Health and most of the medical establishment. The physician John Snow was convinced the disease was waterborne but had no microscope and no germ theory. What he had was an argument. He mapped every death against its street address and found they clustered around a single water pump on Broad Street. He then examined the exceptions systematically: workers at a local brewery who drank only beer had not died; a widow in Hampstead who received Broad Street water by cart had died. Each exception, rather than refuting his hypothesis, reinforced it. Snow presented his evidence to the local Board of Guardians, who removed the pump handle on 8 September 1854 — famously regarded as one of the first acts of evidence-based public health.4 Snow’s case is the textbook illustration of the difference between a rhetorical performance and a genuine argument: his premises were independently verifiable, his conclusion followed from them, and the argument was structured so that a single clear counterexample could have destroyed it. The Surgisphere retraction shows what happens when an argument appears to have this structure but does not.

1.2 The Logical Skeleton of a Claim

Every genuine argument has a structure. Strip away the rhetoric, the examples, the tone, and what remains is a skeleton of premises and a conclusion. Logic is the study of what makes that skeleton strong or weak.

The simplest structure: modus ponens (Latin: “the way that affirms”):

  1. If P, then Q.
  2. Therefore, Q.

Example:

  1. If the bacteria is sensitive to penicillin, the patient will recover.
  2. The bacteria is sensitive to penicillin.
  3. Therefore, the patient will recover.

This is valid. The conclusion must follow from the premises — if the premises are true, the conclusion cannot be false. The logical form guarantees the inference.

A small caveat for what follows: “valid” here means classically valid — valid in the system formalised by Frege, Russell, and Whitehead. Modus ponens is its central inference rule, but it is not the only candidate. Relevance logics weaken classical inference to block the so-called “paradoxes of material implication” (a contradiction does not entail every proposition). Intuitionistic logic, descended from Brouwer, denies the law of excluded middle and refuses double-negation elimination, with the result that some classical proofs cease to be proofs at all. Paraconsistent and dialetheic logics permit some contradictions without explosion.

These are not eccentric private alternatives — they are working systems used in mathematics, computer science, and the philosophy of vagueness. Graham Priest’s Logic: A Very Short Introduction is the most economical entry point.5

This structure appears everywhere, often hidden. “We should tax carbon because it damages the atmosphere and damaging the atmosphere is wrong” conceals:

  1. (Implied) We should stop things that are wrong.
  2. Damaging the atmosphere is wrong.
  3. Carbon emissions damage the atmosphere.
  4. Therefore, we should stop carbon emissions [i.e., tax them].

Making the structure explicit is the first philosophical move — and often reveals gaps, hidden assumptions, or questionable premises that were invisible in the original prose.

1.3 Deductive vs. Inductive Arguments

There are two fundamentally different types of inference:

Deductive: The conclusion is guaranteed by the premises. If the premises are true and the argument is valid, the conclusion cannot be false. Mathematics and formal logic are deductive. The price: the conclusion can only contain what the premises already implicitly contained. Deduction cannot generate genuinely new information — only make explicit what was latent.

Inductive: The premises support the conclusion, but do not guarantee it. All observed ravens have been black; therefore (probably) all ravens are black. The conclusion goes beyond the premises. This is how science works — but it is also where Hume’s problem of induction bites.

The difference is important for TOK. When someone says “studies show that X leads to Y, therefore policy Z is correct,” this involves both an inductive step (from studies to generalisation) and often a deductive step that conceals a value premise.

1.4 Distinction from Persuasion

Aristotle distinguished three modes of appeal in rhetoric (Rhetoric, c. 322 BCE):6

  • Logos: appeal to reason and logic
  • Ethos: appeal to the character and credibility of the speaker
  • Pathos: appeal to the emotions of the audience

Logos is what logic studies. Ethos and Pathos are effective — often more effective than Logos — but they are not arguments in the logical sense. They can produce belief without justification.

A conclusion reached through skilful pathos can be exactly as well-supported as a conclusion reached through valid deduction — or much less so — and we often cannot tell the difference while it’s happening to us.

1.5 The Socratic Method as Argument

Socrates (as depicted in Plato’s early dialogues) did not deliver lectures. He asked questions. The elenchus (Socratic refutation) is a method of argument by cross-examination: take the interlocutor’s claim, show that it implies something else they also believe, show that the two beliefs contradict each other. The interlocutor must revise.

Example (Euthyphro, c. 380 BCE): Euthyphro claims that piety is what the gods love. Socrates asks: Do the gods love something because it is pious, or is it pious because the gods love it? (The Euthyphro Dilemma.)7 Either answer creates problems for Euthyphro’s definition. He cannot hold all his beliefs at once.

The elenchus is a tool of negative dialectic: it doesn’t tell you what the right answer is, but it eliminates bad answers by showing their internal contradictions. Philosophy progresses largely through elenchus.

The popular tag often attached to this method — “I know that I know nothing” — is not a sentence Socrates ever utters in the dialogues. What Plato actually has him say at Apology 21d is more careful: he is wiser than the man he has just questioned only in that he does not think he knows what he does not know. The crisper Latin formula (scio me nihil scire) is a much later distillation.8

1.6 The Regress Problem: Carroll’s Tortoise

Lewis Carroll (“What the Tortoise Said to Achilles,” Mind, 1895) posed a problem about the justification of logical inference that has never been fully resolved.9

To see the paradox, hold one distinction firmly in mind: a premise is a claim inside the argument (it can be true or false, and it is what the argument rests on); an inference rule is the pattern that licenses getting from the premises to the conclusion (it is what the argument moves by). “All humans are mortal” is a premise; “from ‘all A are B’ and ‘c is A’ conclude ‘c is B’” is an inference rule. Carroll’s Tortoise accepts every premise Achilles offers — and still refuses to move.

Achilles and the Tortoise discuss the following argument:

    1. Things that are equal to the same are equal to each other.
    1. The two sides of this Triangle are things that are equal to the same.
    1. Therefore the two sides of this Triangle are equal to each other.

The Tortoise accepts (A) and (B) but refuses to accept (Z). Achilles says: “But if you accept A and B, you must accept Z.” The Tortoise agrees — and asks Achilles to add that as a new premise: call it (C). Now the Tortoise accepts A, B, and C — but still refuses to accept Z without a further premise explaining why A, B, and C together entail Z. The regress continues without end.

Carroll’s point: the rule of inference (modus ponens — if the premises are true, the conclusion follows) cannot itself be justified by adding it as a premise. Any justification of an inferential rule already uses inferential rules. The rules of logic cannot be grounded from within logic.

This is not a puzzle to be solved. It is a structural feature of formal systems. Inference rules are not conclusions of arguments — they are the machinery that produces conclusions. They must be accepted as basic, not derived. What justifies that acceptance is an epistemological question, not a logical one.

This question — what does ground inference rules, if not further inferences — is left deliberately open here. It is taken up later, alongside the broader question of what grounds the choice of logical system (Quine’s web-of-belief entrenchment vs. BonJour’s a priori rational insight); the two questions are distinct but share a structure.

1.7 Questions to Argue About

  • Can you be persuaded of something true by an invalid argument? And if so, is that a problem?
  • “You can’t get from is to ought” (Hume). If an argument contains only factual premises, can its conclusion be a moral claim? What has to be smuggled in?
  • The Socratic method is negative — it destroys positions rather than building them. Is destruction a valid form of philosophical progress? Or does philosophy need to construct as well as demolish?
  • Every argument has premises. Premises need to be justified. Their justification requires further premises. Does this lead to an infinite regress — or are some premises just foundational? (Compare Carroll’s Tortoise: does the same regress apply to inference rules?)

Forced Fork: Could Lancet Readers in May 2020 Have Accepted Mehra et al.’s Conclusion Without a Premise Surgisphere Refused to Supply?

The case is in the info-box above. On 22 May 2020 The Lancet published Mehra et al.’s hydroxychloroquine study; within ten days James Watson and 200+ co-signatories accepted the form of the argument but refused the conclusion, demanding that Surgisphere release the underlying dataset for independent audit. Surgisphere refused. The Lancet withdrew the paper on 4 June. The argument’s logical structure had been intact; the methodological community refused to certify the premises without verification the authors would not supply. The earlier case of John Snow’s Broad Street cholera analysis (treated above in the body) is the contrasting historical anchor — premises that were independently verifiable, and an inference that survived hostile scrutiny.

Position A: Watson et al.’s open letter is the contemporary version of Carroll’s Tortoise. Acceptance of a published argument’s conclusion always requires more than the argument’s stated premises and inference rules — it requires further commitments (here, that the dataset is what it claims to be) that the argument cannot produce from inside itself. Carroll’s regress is unstoppable: data verification, methodological audit, peer-review norms cannot be justified by adding them as premises. Snow’s case was settled by the methodological community of public-health practitioners; Lancet’s retraction was settled by the methodological community of clinical statisticians. Logic is not self-grounding; this is a structural feature of how science actually concludes anything, not a puzzle to be solved.

Position B: The Watson signatories are not extending Carroll’s regress; they are doing something different. Carroll’s Tortoise is about inference rules — the demand that modus ponens itself be justified by adding it as a premise. The Surgisphere objection is about premises — the demand that the empirical content of the argument be verifiable. The two are not the same. Inference rules are the machinery of thought, not objects of thought; premises are objects of thought, examinable individually. The Watson letter is sound methodological practice; Carroll’s miasma theorist who keeps demanding further inferential premises is making a confused logical demand. Conflating them obscures both.

Choose one. If you choose Position A, explain how science ever concludes anything if every demand for further premises is on the same continuum. If you choose Position B, explain what distinguishes the methodologically legitimate Watson demand from the logically confused Carrollian one — and what would happen if a peer reviewer demanded that modus ponens itself be defended in the methods section.


2 What makes an argument valid — and what makes it sound?

Validity and soundness are not synonyms. They are technical terms, precisely defined, and the distinction between them is one of the most important tools in all of philosophy. Confusing them is a mark of philosophical inexperience; understanding them clearly unlocks enormous analytical power.

State v. Loomis and the Logic of an Algorithmic Risk Score

In February 2013, Eric Loomis was charged in Wisconsin with five offences arising from a drive-by shooting in La Crosse; he pleaded no contest to two of them. At sentencing, the trial court relied on a presentence investigation report that included a COMPAS risk assessment — a proprietary algorithmic score, produced by the company Northpointe (now Equivant), classifying defendants by their statistical likelihood of recidivism on the basis of 137 questions and a national database.10 The court sentenced Loomis to six years in prison and five years of extended supervision, citing COMPAS among other factors. Loomis appealed on the ground that he had been sentenced on the basis of a population-level statistical instrument whose internal workings were a trade secret and which he had no way to challenge: he had been sentenced as a member of a class, not as an individual. The Wisconsin Supreme Court ruled against him on 13 July 2016 (5–2), holding that COMPAS scores could be used at sentencing provided they were not determinative and that judges were warned of certain methodological limitations.11 In May 2016, the journalists Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner had published in ProPublica a major investigation of COMPAS data showing that, across more than 7,000 Florida defendants, Black defendants were almost twice as likely as white defendants to be falsely classified as “high risk,” while white defendants were more likely to be falsely classified as “low risk.”12 Northpointe responded that the analysis used the wrong fairness criterion. Computer-science researchers Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan demonstrated later in 2016 that the relevant fairness criteria — calibration across groups, equal false-positive rates, equal false-negative rates — are jointly mathematically incompatible whenever the base rates of recidivism actually differ between groups.13 The case puts the validity/soundness distinction into a system whose logical form is impeccable but whose premises (the algorithm’s outputs) cannot, by commercial design, be examined for individual cases. The earlier McCleskey v. Kemp (1987) — Warren McCleskey’s challenge to Georgia death sentences on the basis of the Baldus study showing a 4.3× racial disparity in capital sentencing14 — has exactly the same logical structure: a population-level argument that rises or falls on whether population-level evidence is admissible for individual constitutional claims. The Wisconsin court’s answer to Loomis is Justice Powell’s answer to McCleskey, thirty years on.

2.1 Validity

An argument is valid if the conclusion cannot be false when all the premises are true. Validity is about the logical relationship between premises and conclusion, not about whether the premises are actually true.

This argument is valid:

  1. All fish can fly.
  2. Salmon are fish.
  3. Therefore, salmon can fly.

Premise 1 is false. The conclusion is false. But the argument is valid — because if premise 1 were true, and if premise 2 were true, the conclusion would have to be true. The logical structure is watertight.

This argument is invalid:

  1. All cats are mammals.
  2. Dogs are mammals.
  3. Therefore, dogs are cats.

This is the fallacy of the undistributed middle: from “All A are B” and “All C are B,” it does not follow that “C is A,” because the middle term (“mammals”) is not distributed — it does not refer to all mammals in either premise, so the two premises do not pin the cats and the dogs to the same subset of mammals. (The related but distinct fallacy of affirming the consequent has propositional, not categorical, form: “If P then Q; Q; therefore P” — for example, “If it rained, the street is wet; the street is wet; therefore it rained.” Both fallacies are common; conflating them is itself a small case of imprecision worth catching.) The premises here are true; the conclusion is false; therefore the argument is invalid.

Key insight: a valid argument guarantees the truth of the conclusion given the truth of the premises. An invalid argument does not — the premises might all be true and the conclusion still false.

2.2 Soundness

An argument is sound if it is both (1) valid and (2) has all true premises. Soundness is what we actually want in practice: a sound argument guarantees a true conclusion.

ValidInvalid
True premisesSound (conclusion must be true)Not sound (conclusion may or may not be true)
False premisesNot sound (conclusion may be false)Not sound

In philosophy, we often work with valid but unsound arguments — thought experiments and hypotheticals where premises are stipulated rather than established. The point is to test the logical consequences of certain assumptions, not to assert the assumptions are true.

2.3 Modus Ponens and Modus Tollens

The two most important valid argument forms:

Modus ponens (“affirming the antecedent”):

  • If P, then Q.
  • Therefore, Q.

Modus tollens (“denying the consequent”):

  • If P, then Q.
  • Not Q.
  • Therefore, not P.

Modus tollens is how falsification works in science. The general theory of relativity predicts that light bends around massive objects (P → Q). Eddington’s 1919 observation found that light does bend (Q — prediction confirmed). But if the light had not bent (¬Q), that would have falsified the theory (¬P), by modus tollens. This is the logical structure of Popper’s falsificationism.

A wrinkle worth knowing. Modus ponens and modus tollens are not on equal footing in every logic. The standard derivation of MT from MP relies on classical contraposition (P → Q ⊢ ¬Q → ¬P) and on double-negation elimination — moves the intuitionist refuses.

In constructive mathematics, you cannot in general infer ¬P from a refutation of Q together with P → Q without a separate constructive reason. MP is unproblematic; MT is not the immediate twin it appears to be. The two rules look symmetric only against the silent backdrop of classical logic.15

2.4 The “True Premise, False Conclusion” Test

A powerful diagnostic: if you can construct a scenario where all the premises are true and the conclusion is false, the argument is invalid. This is the method of counterexample. It is one of the core techniques of philosophical analysis.

“All politicians are untrustworthy; the local librarian is untrustworthy; therefore, the local librarian is a politician.” Is this valid? Construct a counterexample of the same form: “All cats are mammals; all dogs are mammals; therefore all dogs are cats.” Premises true, conclusion false. Same logical form (the undistributed middle, as in §Validity). The form is invalid.

2.5 Questions to Argue About

  • Why do people often find invalid arguments compelling? (Connect to base-rate neglect and the prosecutor’s fallacy.)
  • A valid argument with a false conclusion must have a false premise. So: if you find yourself rejecting a conclusion, you are committed to rejecting at least one premise. What are the consequences of this for political and ethical debate?
  • Can an argument be good — worth believing — without being sound? (Think of arguments from analogy, or arguments to the best explanation. Are these valid in the formal sense?)
  • Mathematics consists entirely of valid deductive arguments from axioms. Does this mean mathematical conclusions are certain? What would it mean for a mathematical theorem to be uncertain?

Forced Fork: Was Loomis Sentenced on a Sound Argument?

The case is in the info-box above. The state’s argument, in compressed form: COMPAS reliably predicts recidivism risk on a population basis; risk-of-recidivism is a relevant sentencing consideration; therefore Loomis’s COMPAS score was a legitimate input to his sentence. Loomis’s reply: the form is fine but the application is unsound — the premise that COMPAS reliably predicts his recidivism risk had not been established and could not be examined.

Position A (with the Wisconsin Supreme Court, 5–2): Population-level statistical instruments are valid sentencing inputs as long as they are not determinative and the judge is alerted to their limitations. Demanding individual-case validation of every premise would make sentencing impossible — courts would have to reverify general criminological knowledge for every defendant. Validity of the argument structure is what the system can verify; soundness is satisfied at the level the legal system can actually assess.

Position B (with Loomis, and with Brennan’s dissent in the earlier McCleskey v. Kemp (1987), where Warren McCleskey had challenged Georgia death sentences on the basis of Baldus’s 4.3× racial-disparity study): When the population-level instrument is opaque, racially disparate in its error rates, and protected from examination by trade-secret law, the argument’s factual premises cannot be examined at all. An argument whose premises are inscrutable to the person it is being used against is not merely unsound — it is not an argument the person has been given. Powell’s McCleskey majority repeated thirty-three years on becomes a standard of unverifiability that should never have survived its first application.

Choose one. Position A must say what the limit is: at what point does an algorithmic premise become too opaque to count as a legitimate input to a sentence? Position B must say what should replace algorithmic risk-scoring at sentencing — a return to unaided judicial discretion, which has its own well-documented disparities, or some other regime?


3 How should I update what I believe when new evidence comes in?

Most reasoning in real life is not deductive. The premises are not certain, the conclusion is not entailed, and what we want to know is not “does this follow?” but “does this evidence make this hypothesis more likely?” The mathematics of that question — how a prior belief should be revised in the light of new data — is Bayes’s theorem, and the most common way to misuse it is to ignore the base rate: the prior probability of the hypothesis before the evidence arrives. Smart people, including expert witnesses, get this catastrophically wrong.

The Prosecution of Sally Clark

In November 1999, the solicitor Sally Clark was convicted of murdering her two infant sons, who had died in 1996 and 1998. The Crown’s expert witness, the paediatrician Sir Roy Meadow, told the jury that the probability of two cot deaths occurring in a single affluent non-smoking family was 1 in 73 million — a figure he obtained by squaring the estimated single-cot-death rate.16 The figure was wrong twice over. First, it assumed the two deaths were independent, which they are not: cot deaths cluster within families because of shared genetics and environment. Second, and more dangerously, what the jury heard as “the chance Mrs Clark is innocent” was in fact “the chance of two cot deaths given she is innocent” — a different quantity. To convert one into the other, the jury would have needed to weigh that figure against the probability of two infant murders within the same family, which is also vanishingly small. The Royal Statistical Society wrote publicly to the Lord Chancellor in 2002 to say so.17 The conviction was quashed in January 2003. Sally Clark, having served three years in prison, never recovered, and died of acute alcohol poisoning in March 2007. The Lucia de Berk case in the Netherlands followed an almost identical statistical pattern with similar consequences.18

3.1 Bayes’s Theorem (the toy version)

The theorem, due to the eighteenth-century English clergyman Thomas Bayes, describes how a belief should change in response to evidence.19 The intuition is one line:

posterior is proportional to prior × likelihood.

In words: the probability of a hypothesis given some evidence is proportional to the probability of the hypothesis before the evidence (the prior) multiplied by the probability of seeing this evidence if the hypothesis were true (the likelihood). Graham Priest puts the failure mode bluntly: many bad arguments are “seductive because people often confuse probabilities with their inverses, and so slide over a crucial part of the argument.”20 The probability of evidence given a hypothesis (\(P(E \mid H)\)) is not the probability of the hypothesis given the evidence (\(P(H \mid E)\)). The prior is what closes the gap.

3.2 Base-rate Neglect: A Worked Example

A disease affects 1 person in 1,000 (the base rate). A test for the disease is 99% accurate — 99% of sufferers test positive (sensitivity), and 99% of non-sufferers test negative (specificity). You test positive. What is the probability you have the disease?

The intuitive answer, and the one most doctors give when polled, is “about 99%.” The Bayesian answer is roughly 9%. Imagine 1,000 people. One has the disease and almost certainly tests positive. Of the 999 who do not, about 1% (roughly 10) test positive anyway. So out of about 11 positive results, only 1 is a true positive: \(1/11 \approx 9\%\). The test is fine; the base rate is decisive. Ignore the prior and you misjudge the posterior by an order of magnitude.

3.3 Representativeness and the Base-rate Failure

Daniel Kahneman and Amos Tversky’s chapter “Tom W’s Specialty” in Thinking, Fast and Slow shows the same error in plain prose. Asked to guess whether “Tom W” — described as quiet, neat, and orderly — is studying computer science or humanities, both lay subjects and trained graduate students in psychology pick computer science, even when explicitly told that the humanities cohort is many times larger.21 We judge by representativeness — how well the description matches a stereotype — and ignore the prior probability of the category. Kahneman’s chapter title for the cab-problem analysis is exact: Causes Trump Statistics. We reach for the causal story (the witness saw a blue cab, so the cab was blue) and discard the statistical fact (most cabs in the city are green). The cab problem is the textbook companion to the medical-test problem above, with one crucial difference: the witness is the prosecutor’s case.

3.4 The Prosecutor’s Fallacy, Named

A “1 in 10 million” DNA match in a database of 30 million people produces, on average, three matches by chance alone. The figure does not, by itself, identify a guilty party.

The Sally Clark case has a name in the statistical literature: the prosecutor’s fallacy. It consists in asserting \(P(\text{evidence} \mid \text{innocent})\) — the probability of seeing this evidence if the defendant were innocent — and inviting the jury to hear it as \(P(\text{innocent} \mid \text{evidence})\). The two quantities can differ by orders of magnitude. The Royal Statistical Society’s letter, the appeal court’s judgment, and three decades of statistical commentary have not stopped the fallacy from recurring; it appeared again in the conviction of the Dutch nurse Lucia de Berk (2003, quashed 2010), in DNA-match cases (“the chance of a random match is 1 in 10 million”) presented without reference to the size of the database searched, and in epidemiological courtroom testimony generally.

3.5 Bayesianism as a Theory of Rational Belief

Bayes’s theorem is a calculation. Bayesianism is the broader epistemological claim that rational belief is graded — that beliefs come in degrees, that those degrees obey the probability axioms, and that the rational response to evidence is to update one’s degrees of belief in line with the theorem. This is a substantive claim, not an obvious one. Classical logic treats belief as binary: a proposition is accepted or it is not. The Bayesian thinks this is a coarsening of what rationality actually requires. The classicist replies that probabilistic belief leaves no place for the classical notions of validity and proof — that a proposition either is or is not entailed, and degrees of belief are about psychology, not logic. The argument is alive.

3.6 Questions to Argue About

  • The medical-test calculation gives 9%. The doctor’s intuition gives 99%. Whose answer should determine policy on screening healthy populations for rare diseases? What goes wrong when the doctor’s intuition wins?
  • Sally Clark’s jury was told a number — 1 in 73 million — that was both technically wrong and rhetorically devastating. Is the deeper failure innumeracy (jurors and judges who could not interrogate the figure) or epistemic theatre (an expert witness deploying a figure designed to overpower interrogation)?
  • A defendant’s DNA matches the crime scene with a “1 in a million” random-match probability. The police database contains five million profiles. How should this evidence be presented? Whose job is it to present the base rate?
  • If rational belief is a matter of degree (Bayesianism), what becomes of the classical logical notion of acceptance — the binary yes/no judgement that a proof produces? Are these two pictures compatible, or does one have to give way?

Forced Fork: Should Statistical Expert Testimony Be Admissible at Trial?

The Sally Clark case poses an institutional question, not just a statistical one: Sir Roy Meadow’s “1 in 73 million” was technically wrong (it assumed independence between correlated events) and inferentially wrong (it presented \(P(\text{evidence} \mid \text{innocent})\) as if it were \(P(\text{innocent} \mid \text{evidence})\) — the prosecutor’s fallacy). Two opposing reform proposals follow.

Position A (training, not exclusion): The remedy is mandatory statistical instruction for judges, juries, and expert witnesses. Statistical evidence is the only available grip on whole classes of case (DNA-database matches, epidemiological causation, forensic reliability disputes); excluding it punishes future defendants whose cases would benefit from Bayesian re-framing. Keep the evidence, fix the participants. The UK Royal Statistical Society’s Statistics and the Law protocols (since 2010) are the right model.

Position B (exclude, with narrow exception): A “1 in 73 million” figure cannot be heard by a jury as a contribution to a Bayesian calculation; it can only be heard as an overwhelming claim about guilt. The adversarial system selects for experts whose figures dominate deliberation. The remedy is structural: presumptive exclusion of single-figure statistical claims, with admission only via court-appointed Bayesian expert presenting the calculation in posterior-probability form. Training has been tried for thirty years; the institutional incentive is the cause.

Choose one. Position A must say what specific training intervention would break a thirty-year pattern of recurrence. Position B must say how a court-appointed expert would itself be insulated from the same pressures, and which evidence the proposed exclusion would lose at the cost of justice.


4 Can we be certain through induction?

Almost everything we believe about the world — that aspirin cures headaches, that the sun will rise tomorrow, that jumping from heights causes injury — is based on induction. We generalise from what we have observed to what we have not yet observed. And yet induction, as David Hume showed with devastating clarity in the 18th century, cannot be rationally justified without circularity.

The Royal Bank of Scotland and the Black Swan of 2008

In the years before the 2008 financial crisis, the risk models used by the Royal Bank of Scotland, Lehman Brothers, and most major banks were built on Value at Risk (VaR) calculations derived from approximately twenty years of market data. VaR purports to give a number of the form “we will not lose more than £X on 99% of trading days”; the figure is calculated from the historical distribution of price moves. The models had performed well across thousands of trading days. Their developers had every inductive reason to trust them: the sample was large, the pattern was consistent, the predictions had been repeatedly confirmed. What the models had never encountered — because it lay outside their data range — was a correlated collapse of multiple asset classes simultaneously, combined with a shutdown of the short-term wholesale funding markets on which RBS had become dependent. The same inductive overconfidence shaped the bank’s strategic decisions: in October 2007, on the basis of recent market conditions that had been benign for as long as the data went back, RBS led a consortium that acquired the Dutch bank ABN AMRO for €71 billion on what the FSA later called “two lever arch folders and a CD” of due-diligence material — a deal whose thinness was not an oversight but a confidence, of exactly the kind the VaR figures licensed, that the recent past was a sound guide to the near future. The acquisition consumed the capital cushion that might have absorbed the 2008 shock; the UK government then injected £45.5 billion of equity capital across October 2008 and December 2009 in exchange for an eventual 84% stake.22 Nassim Nicholas Taleb had published The Black Swan in 2007, arguing that the very reliability of inductive models in finance made them dangerous — because the events that matter most are precisely those that fall outside the distribution of past observations.23 Hume’s problem of induction, which had seemed a philosophical puzzle with limited practical bite, turned out to describe the exact mechanism by which an institution employing thousands of highly numerate professionals destroyed itself: the future resembled the past — in the models, in the trading book, and in the boardroom — until, with tremendous violence, it did not.

4.1 Hume’s Problem

David Hume (A Treatise of Human Nature, 1739; An Enquiry Concerning Human Understanding, 1748) posed the problem of induction with a precision that has never been answered satisfactorily:24

The observation: past regularities give us no logical guarantee about future regularities. We have observed the sun rising every day. But: does the past rising of the sun provide any logical reason to expect it to rise tomorrow?

The would-be justification: of course past regularities are a guide to the future — nature is uniform, things that have always happened will continue to happen.

Hume’s response: How do you know nature is uniform? By observing that it has been uniform in the past. But this is circular — you are using inductive reasoning to justify inductive reasoning.

“There can be no demonstrative arguments to prove that those instances of which we have had no experience resemble those of which we have had experience.” — David Hume, A Treatise of Human Nature, Book I, Part III, Section VI25

This is Hume’s Fork: all propositions are either relations of ideas (like mathematics — necessarily true, but empty of empirical content) or matters of fact (empirically known through experience, but not necessarily true). Inductive reasoning tries to make matters of fact feel as certain as relations of ideas — and that is not legitimate.

4.2 Russell’s Inductivist Chicken

Bertrand Russell (The Problems of Philosophy, 1912) gave a now-famous illustration. A chicken is fed every morning by the farmer. Every morning, without exception, the farmer comes and feeds it. The chicken, by induction, comes to expect this. On Christmas Eve, the farmer comes — and wrings its neck.26

The chicken’s inductive reasoning was impeccable. Its conclusion was disastrously wrong. The past regularity provided no logical guarantee of the future. Russell’s moral, in Problems of Philosophy Chapter VI, is more careful than the chicken story alone suggests: he holds that induction must be backed by a separate Principle of Induction, which itself cannot be proved from experience and must be accepted as a piece of a priori knowledge if induction is to have any rational warrant. Russell is a defender of induction, not a sceptic about it; his point is that defending induction requires reaching outside what experience itself can supply.

Russell’s chicken is itself a repurposing of Hume’s earlier example, in Enquiry §IV, of bread that has nourished us in the past and is expected to nourish us again. Hume’s point was that “there is no known connexion between the sensible qualities and the secret powers” — having seen one loaf nourish gives no rational ground to expect the next to do so, even though we cannot help expecting it.27

Taleb’s Black Swan (2007) makes a sharper point than that induction is fallible. He argues that event-rarity scales with consequence: that long-run outcomes are dominated by rare, high-impact events drawn from “fat-tailed” distributions, and that mainstream statistical practice systematically underestimates them. The RBS info-box above is the live case. Where to go from here is the contemporary live debate: Bayesian inductivism (Howson and Urbach), Vapnik’s statistical-learning theory, John Norton’s “material” theory of induction, and Peter Lipton’s inference-to-the-best-explanation are four salient options among others, and which (if any) addresses Hume’s underlying problem rather than relocating it is itself disputed.28

4.3 The Black Swan

Before Europeans reached Australia, every swan ever observed was white. “All swans are white” was, for Europeans, a very well-confirmed inductive generalisation. Then black swans were found in Australia.

Karl Popper (The Logic of Scientific Discovery, 1934) used this to argue that no number of confirming instances can prove a universal generalisation — but a single disconfirming instance can disprove it. Hence his criterion: scientific claims should be falsifiable. We should try to refute our theories, not accumulate confirmations.29 (Taleb’s modern extension of this point is the RBS case in the info-box above.)

4.4 Falsifiability vs. Verificationism

The Logical Positivists (Vienna Circle, 1920s–1930s) — Moritz Schlick, Rudolf Carnap, A.J. Ayer — proposed verificationism: a statement is meaningful if and only if it is in principle verifiable through observation. Metaphysical claims (“God exists”) and ethical claims (“murder is wrong”) are not verifiable and hence meaningless — not false, but empty.30

Popper rejected verificationism and proposed falsificationism instead: what marks science is not that its claims can be verified but that they can be falsified — there is some observation that would count against them.

Popper also used this to critique Marxism and Freudianism: he felt these theories could accommodate any possible observation, which meant they made no genuine predictions and were therefore unfalsifiable — and unscientific.

The asymmetry: No number of white swans proves “all swans are white.” But one black swan disproves it. Verification is logically weak; falsification is logically strong. This asymmetry is why Popper preferred falsification as the hallmark of scientific knowledge.

4.5 Responses to Hume

No one has solved Hume’s problem. But responses range:

  • Pragmatic vindication (Hans Reichenbach): induction is the best strategy we have, even if it can’t be justified a priori. If nature is regular, induction will find it; if not, no method will.31
  • Bayesian inference: we update our probability estimates in response to evidence according to Bayes’ Theorem. This is not a solution to the problem (Bayes’ Theorem requires prior probabilities, which are themselves inductively based), but it formalises the reasoning.
  • Naturalism (Quine): the question “is induction justified?” is itself answerable only by empirical investigation — there is no standpoint outside our epistemic practices from which to evaluate them.32

4.6 Questions to Argue About

  • If Hume is right that induction cannot be rationally justified, does that make science irrational? Or just differently rational than we thought?
  • Russell’s chicken generalised from excellent evidence and reached a fatal conclusion. Does this show that the problem of induction is practically serious — or just a philosophical puzzle without real-world implications?
  • Popper said science advances through falsification, not confirmation. But in practice, scientists often protect their theories from falsification by adjusting auxiliary hypotheses. Is this bad science or rational practice?
  • What is the relationship between Hume’s problem of induction and the reliability of memory? If past experience doesn’t logically justify expectations about the future, does past experience of your own identity justify beliefs about who you are?

Forced Fork: Were the RBS Risk Modellers Rational?

Return to the Royal Bank of Scotland case in the info-box above. The Value-at-Risk models were built on twenty years of market data, had been repeatedly confirmed by performance, were mathematically sophisticated, and were produced by thousands of numerate professionals who knew their craft. They failed catastrophically in 2008 and destroyed the bank. The question is not whether they failed — everyone agrees — but whether the modellers were rational to trust them up to that moment.

Position A (the modellers were not rational, in the philosophically serious sense): Hume’s problem of induction describes exactly what went wrong. The modellers’ confidence came from past observation alone, and past observation can never rule out futures outside the observation window. A process that cannot distinguish good evidence from the absence of counter-evidence is habitual, not rational. Calling the practice rational because it usually works is calling the chicken’s expectation of breakfast rational.

Position B (the modellers were rational, given what rationality can actually be): Rationality is not the impossible standard of inference immune to sceptical challenge. It is the best-available-methods standard: Bayesian updating, diversification of models, sensitivity analysis, stress-testing against historical extremes. By that standard, the RBS modellers were mostly rational; their specific failure was a local failure of that method (insufficient stress-testing of correlated-collapse scenarios) rather than a vindication of Hume. The correct response to 2008 is not philosophical humility but better stress-testing — which the post-crisis regulatory regime has partly implemented.

Choose one. If you pick A, explain whether there is any method of forecasting in a domain with tail risks that you would be willing to call rational — and if not, what epistemic status you assign to anything that resembles scientific prediction. If you pick B, specify the technical failure at RBS precisely enough that your prescription (“better stress-testing”) would have produced a meaningfully different answer in 2007.


5 What are the limits of formal systems?

In 1931, a 25-year-old mathematician named Kurt Gödel published a paper with a title so technical it could clear a room: “On Formally Undecidable Propositions of Principia Mathematica and Related Systems.” Inside it were two theorems that shook the foundations of mathematics and, by extension, of formal knowledge.33

Boeing 737 MAX MCAS — A Formal System Outside Its Specification

On 29 October 2018 Lion Air Flight 610, a Boeing 737 MAX 8, crashed into the Java Sea thirteen minutes after takeoff from Jakarta, killing all 189 people aboard.34 Less than five months later, on 10 March 2019, Ethiopian Airlines Flight 302 — also a 737 MAX 8 — crashed near Bishoftu six minutes after takeoff from Addis Ababa, killing all 157 people aboard. Both crashes were attributed to the Manoeuvring Characteristics Augmentation System (MCAS) — a flight-control law Boeing had added to the MAX to compensate for the larger, more forward-mounted engines, which gave the aircraft a tendency to pitch up during certain manoeuvres. MCAS read a single angle-of-attack sensor; if it detected the nose was too high, it commanded the horizontal stabiliser to trim down, lowering the nose. The system had been formally specified, certified, and tested. Inside its design envelope it worked correctly. The two crashes occurred because in each case a single angle-of-attack sensor had been damaged or misaligned: in the Lion Air aircraft, by improper installation of a replacement vane in Denpasar two days earlier; in the Ethiopian aircraft, by what investigators believe was a bird strike on the sensor probe.35 Outside the conditions MCAS had been specified for, the system did not fail silently. It activated repeatedly — twenty-one times on the Lion Air flight before the final dive — issuing nose-down commands against the pilots’ increasingly desperate efforts to pull up. Pilots had not been told MCAS existed; the differences-training course Boeing supplied to airlines did not mention it. The Joint Authorities Technical Review report (2019) concluded that the FAA certification process had treated MCAS as a non-safety-critical refinement of an existing system rather than a new flight-control law that pilots needed to know about and could override.36 346 people died because a formal system that was correct inside its specification had no way of recognising that it had moved outside it. Gödel’s lesson — that no sufficiently strong consistent formal system can prove its own consistency, and that any such system contains truths it cannot reach — is here reproduced in metal and software: a formal system cannot tell you when its own inputs have left the conditions under which its derivation rules were proved valid.

The theorems tell us that any sufficiently powerful formal system is either incomplete (there are truths it cannot prove) or inconsistent (it can prove contradictions). You cannot have both completeness and consistency. This is not a technical result that concerns only specialists. It says something deep about the limits of formalisation, proof, and certainty.

5.1 The Dream of Formalism

By the early 20th century, mathematicians — led by David Hilbert — had an ambitious programme: to place all of mathematics on an absolutely secure axiomatic foundation. Propose a complete set of axioms; show the axioms are consistent (they don’t lead to contradictions); show they are complete (every mathematical truth can be derived from them). Mathematics would then be a perfect, closed system — a realm of certain knowledge.37

The crack in that ambition was already visible thirty years before Gödel. In June 1902 Bertrand Russell wrote a polite and devastating letter to the German mathematician Gottlob Frege, who had spent two decades constructing the Grundgesetze der Arithmetik — an attempt to derive all of arithmetic from pure logical axioms.38 Russell had found a contradiction at the heart of the system. Frege’s Basic Law V permitted sets to be defined by any property whatsoever, including the property of containing all sets that do not contain themselves. When Russell asked whether this set contained itself, the answer was: if it does, it doesn’t; if it doesn’t, it does. The paradox could not be patched within the existing axioms without rebuilding the foundation. Frege wrote back within days, acknowledging that Russell had shaken the ground on which he had meant to build arithmetic. The second volume of the Grundgesetze was already at the printer; Frege added a despairing appendix acknowledging the flaw. Russell’s letter is one of the most consequential pieces of correspondence in the history of mathematics — the moment at which the dream of grounding all mathematics in a complete, consistent formal system began to look not just difficult but perhaps impossible. Russell and Alfred North Whitehead spent the next decade attempting the repair: their Principia Mathematica (1910–1913) is a nearly 2,000-page formal derivation of mathematics from logical axioms, with a “theory of types” devised to block the Russell paradox. (It takes 379 pages to prove that 1 + 1 = 2.)39

5.2 Gödel’s First Incompleteness Theorem

Kurt Gödel, in “On Formally Undecidable Propositions of Principia Mathematica and Related Systems” (1931), proved:

First Incompleteness Theorem: Any consistent formal system capable of expressing basic arithmetic contains true statements that cannot be proved within the system.

The proof uses a piece of self-reference. Gödel devised a method of encoding statements about the formal system within the system itself — Gödel numbering, which assigns a unique whole number to every formula, allowing statements about formulas to be translated back into statements about arithmetic. Using this coding, he constructed a statement G whose informal content is:

G — This statement cannot be proved in this system.

Before the formal argument, a self-descriptive warm-up. “This sentence has five words” is true not because someone proved it but because the sentence correctly describes itself. Look at the sentence; count the words; the content of the claim and the state of the world coincide. G is built to do the same thing, except what it describes is its own unprovability rather than its word count.

The two horns. Suppose the system is consistent — it never proves a false statement. Then:

  • If G were provable, the system would have proved a statement that asserts its own unprovability. That is a false claim being proved. The system would be inconsistent.
  • If G is not provable, then G’s claim about itself — that it cannot be proved — is accurate.

Why unprovability entails truth. In a consistent system, the first horn is ruled out. So G is not provable. But G asserts precisely that it is not provable — and that assertion, by the argument just given, is correct. A statement whose content is correct is true. So G is true and unprovable at once. Truths exist in mathematics that formal proof, within any given consistent system, cannot reach.

A precision worth registering. Gödel’s 1931 proof in fact required a slightly stronger condition than mere consistency — a property called ω-consistency — to rule out the system proving the negation of G as well. J. B. Rosser strengthened the result in 1936 by constructing a different sentence for which simple consistency suffices. The teaching exposition above silently relies on Rosser’s strengthening; the historical Gödel argument is a hair more delicate. Nothing in the philosophical conclusion changes.40

5.3 Gödel’s Second Incompleteness Theorem

Even more devastating for Hilbert’s programme:

Second Incompleteness Theorem: A consistent formal system cannot prove its own consistency.

Douglas Hofstadter’s Gödel, Escher, Bach (1979) is the most accessible long treatment of Gödel’s theorems. It is also one of the strangest and most rewarding books about minds and formal systems ever written.41

In other words: you cannot use mathematics to prove that mathematics doesn’t contain a hidden contradiction. Any proof of consistency would itself have to be produced within a formal system — and that system’s consistency would remain unverified.

5.4 What This Means — and What It Doesn’t

What it means:

  • No formal system is everything. There are always truths that outrun provability within a given system.
  • Mathematical truth is not identical to mathematical provability. There is a gap.
  • Hilbert’s programme of securing all mathematics on a finite set of consistent, complete axioms is impossible.

What it doesn’t mean:

  • It does not mean mathematics is unreliable or useless. The vast majority of mathematics is provable within standard systems. Gödel sentences are constructed specifically to be unprovable; they don’t arise in ordinary mathematical practice.
  • It does not directly imply that the human mind is not a formal system — though some philosophers (J.R. Lucas, Roger Penrose) have argued this. The argument is controversial.42
  • It does not mean “anything goes” or that all logical systems are equally valid.

5.5 Consistency and Completeness

The tradeoff:

  • A consistent system never proves a contradiction (not both P and ¬P).
  • A complete system can prove or refute every statement in its language.

Gödel showed these cannot both be achieved in any system powerful enough to encode arithmetic. We must choose. Mathematicians choose consistency — a system that proves contradictions is useless, since from a contradiction, anything at all can be derived (the principle of explosion: ex contradictione quodlibet).

5.6 Questions to Argue About

  • Gödel showed there are mathematical truths that cannot be proved. Does this make mathematics less certain — or just different from what we thought?
  • If a formal system cannot prove its own consistency, how do mathematicians know their work is reliable? What is the basis of our confidence in mathematics?
  • Some philosophers argue that Gödel’s theorems show human minds transcend formal systems (because we can “see” that G is true, even though the system cannot prove it). Does this argument work?
  • Are there analogues to Gödel’s incompleteness outside mathematics — domains of knowledge where there are truths that the domain’s own methods cannot reach?

Forced Fork: Can a Human Mathematician “See” That Gödel’s G Is True in a Way No Formal System Can?

A competent mathematician reading Gödel’s 1931 proof concludes, on reflection, that G is true. The lesson has been working toward the J. R. Lucas / Roger Penrose argument (Lucas, “Minds, Machines and Gödel,” Philosophy 1961; Penrose, The Emperor’s New Mind 1989, Shadows of the Mind 1994) — which says that act of seeing is what no formal system can do.43 You now decide whether that argument works.

Position A (with Lucas and Penrose): For any formal system F proposed as capturing human mathematical reasoning, F has a Gödel sentence G(F) that F cannot prove. A competent mathematician, given F and the assumption that F is consistent, can see G(F) is true. Penrose argues: in actual practice mathematicians do see the consistency of PA and ZFC with rational confidence — not as proof within those systems but as rational insight. If that confidence is added to F as an axiom, the argument iterates; the regress favours the conclusion that no fixed F captures the open-ended human capacity.

Position B (Putnam, Feferman): The mathematician’s “seeing” that G is true is conditional on the system’s consistency — which humans cannot reliably check for arbitrary strong systems. Gödel’s Second Incompleteness Theorem says no sufficiently strong consistent system can prove its own consistency; humans are not exempt. So the mathematician’s seeing is itself conditional; formalise the conditional, and you get a formal system that proves the same thing. The argument equivocates between confident assertion and provable theorem.

Choose one. If you pick A, respond directly to Putnam’s charge of equivocation: explain why the mathematician’s “seeing” that G is true is not itself smuggling in an assumption a formal system could have been given. If you pick B, explain why Penrose’s microtubule proposal for a quantum-mechanical basis of non-computable reasoning is worse than a philosophical shrug — i.e. what is missing from the standard reply that Penrose is trying (however speculatively) to supply.44


6 What is a paradox — and what can we learn from one?

The Ship of Theseus and the Athens–Piraeus Replica

Athens maintains, in the harbour at Piraeus, a full-scale reconstruction of an ancient trireme — a three-banked Greek warship. The vessel Olympias, launched in 1987, was built to test whether ancient descriptions of trireme warfare were accurate.45 But the project immediately encountered a Sorites question of the kind the city of Athens had grappled with in antiquity. The mythological ship of Theseus, preserved as a relic, had its timbers replaced one by one over centuries as they rotted, until no original plank remained. The Sorites structure is exact: each replacement is, by itself, plainly insufficient to change what the ship is, just as adding a single grain of sand is plainly insufficient to make a heap; iterate the move enough times and you reach a conclusion the premises seem to forbid. Whether the predicate “the original ship” admits a non-arbitrary boundary is itself disputed (see the responses listed in §The Sorites Paradox); what is not disputed is that museums, insurance contracts, legal inheritances, and national identities require some answer, defensible or stipulative, when the question reaches them. The seventeenth-century Swedish warship Vasa — which sank on its maiden voyage on 10 August 1628 about 1,300 metres into Stockholm harbour, its top-heavy hull capsized by a light gust because the ballast was insufficient for the weight of the upper gun deck and decorative carvings — is the live case. Salvaged in 1961 after 333 years on the seabed, its timbers were continuously sprayed with polyethylene glycol (PEG) from 1962 to 1979 to displace the water that had soaked into the wood; the PEG itself has since slowly been chemically transformed by sulphuric acid forming inside the wood from sea-derived iron and sulphur, and conservators have spent the 2000s and 2010s neutralising and replacing the failing material.46 The museum claims roughly 98% original wood, but the Sorites bites at a finer grain than the plank: every long fibre of timber has been chemically infiltrated and partly substituted, and there is no first molecule whose replacement turns “original” into “reconstructed.” Where precisely the heap of sand becomes not-a-heap is, it turns out, a practical question with multi-million-pound consequences for museums, insurers, and the public who buy a ticket expecting to see the ship that sank in 1628.

A paradox is not a mere contradiction. A contradiction is straightforwardly wrong. A paradox is different: it is an argument that proceeds from apparently sound premises by apparently valid steps to a conclusion that is either absurd or contradictory. Paradoxes are not failures of thought; they are diagnostic instruments. They reveal that our concepts — motion, truth, set membership, heaps, identity — are not as clear as we assumed.

6.1 Zeno’s Paradoxes

Around 450 BCE, Zeno of Elea argued (according to Aristotle’s later account) that motion is impossible. His most famous argument: the race between Achilles and a tortoise.47

The tortoise has a head start. Achilles runs faster. To catch the tortoise, Achilles must first reach the point where the tortoise was. But by then, the tortoise has moved forward. Achilles must now reach that point. But again, the tortoise has moved. This sequence has infinitely many steps. Can an infinite number of steps be completed?

Zeno argued: no. Therefore, Achilles never catches the tortoise. Therefore, motion as we perceive it is an illusion.

This sounds absurd — of course Achilles catches the tortoise. But the argument seems valid. Where is the flaw?

The mathematical resolution (Cauchy, early 19th century): an infinite series can have a finite sum. The series \(1/2 + 1/4 + 1/8 + \ldots = 1\). So infinitely many steps can be completed in finite time, if the steps get correspondingly shorter. Modern analysis dissolves the paradox mathematically.

But the philosophical question remains: does the mathematical resolution tell us what actually happens in nature, or does it just tell us how to calculate? Is space and time continuous (as calculus assumes) or discrete (as quantum mechanics might suggest)?

6.2 The Liar Paradox

“This sentence is false.”

If the sentence is true, then what it says is the case — so it is false. If it is false, then what it says is not the case — so it is true. The sentence seems to be both true and false simultaneously, which violates the law of non-contradiction.

This was known in antiquity. Epimenides the Cretan reportedly said: “All Cretans are liars.” Does Epimenides speak truly? If yes, he (a Cretan) is a liar, so he speaks falsely. If no, there exists at least one truthful Cretan — possibly Epimenides himself. The latter version avoids strict paradox; the pure Liar (“This sentence is false”) does not.48

The significance: The Liar paradox motivated Tarski’s work on truth and Gödel’s incompleteness theorems. Gödel’s construction is, at its heart, a formalised version of the Liar: a sentence that says “I am not provable.” Self-reference generates paradoxes in truth and provability alike.

6.3 Russell’s Set Paradox

In 1901, Bertrand Russell discovered a contradiction at the heart of naive set theory. Consider the set of all sets that do not contain themselves.49

  • Does this set contain itself? If yes: it is a set that contains itself — so it should not be in our set (contradiction). If no: it is a set that does not contain itself — so it should be in our set (contradiction).

Russell wrote to Frege, who had just published his Basic Laws of Arithmetic, which relied on naive set theory. Frege’s response is one of the most poignant moments in intellectual history:

“Your discovery of the contradiction has surprised me beyond words and, I should almost like to say, left me thunderstruck, because it has rocked the ground on which I meant to build arithmetic.” — Gottlob Frege, letter to Bertrand Russell, 190250

The resolution requires restricting set formation — the axioms of Zermelo-Fraenkel set theory prohibit the kind of self-reference that generates Russell’s paradox. But the resolution is technical and leaves the intuitive concept of “collection” as less transparent than we thought.

6.4 The Sorites Paradox (The Heap Problem)

While Russell’s paradox attacks the foundations of mathematics, the Sorites paradox attacks something more fundamental still: the vague predicates of ordinary language. Sorites comes from the Greek soros: heap.

  1. One grain of sand is not a heap.
  2. Adding one grain of sand to a non-heap does not make a heap.
  3. Therefore, one million grains of sand is not a heap.

The argument is valid. The premises seem true. The conclusion is false. Where is the error?

The Sorites paradox arises for any vague predicate: tall, bald, rich, young, red. There is no sharp boundary between tall and not-tall. Remove one hair at a time from a full head: when exactly does the person become bald? There is no fact of the matter.

The philosophical responses:

  • Epistemicism (Timothy Williamson): there is a precise boundary; we just can’t know where it is.51
  • Supervaluationism: “John is bald” is neither true nor false in borderline cases — a truth value gap.
  • Fuzzy logic: truth comes in degrees. “John is bald” is 0.7 true for borderline cases.
  • Contextualism: “heap” is context-sensitive; the boundary shifts with the conversational context.

The paradox is practically important: law, medicine, and policy deal constantly with vague predicates — when does a fetus become a person? at what point does a company have “significant” market power? The paradox is not academic.

6.5 Questions to Argue About

  • Zeno’s paradoxes were “solved” by calculus. But is a mathematical model of motion the same as an account of what actually happens? Does mathematics dissolve philosophical paradoxes, or only silence them?
  • The Liar paradox suggests that unrestricted self-reference leads to contradiction. Are there other areas of thought — apart from formal logic — where self-reference is dangerous?
  • Russell’s paradox destroyed a perfectly intuitive concept (the set of all sets not containing themselves). What does this tell us about the reliability of intuition in mathematics?
  • The Sorites paradox arises because our concepts are vague. Is vagueness a defect of language that could in principle be eliminated? Or is it essential to how language works?

Forced Fork: Is the Conserved Vasa the Same Object That Sank in 1628?

The case is in the info-box above: the seventeenth-century Swedish warship Vasa, salvaged in 1961, has had its timbers continuously chemically infiltrated — first with PEG (1962–1979), then with successive treatments to neutralise the iron-sulphate acid attack discovered in the early 2000s. The museum’s “98% original wood” claim is true at the level of which fibres are where and false at the level of what those fibres are made of. Suppose you are advising the Swedish state on whether to designate the conserved object as continuous with the 1628 vessel for cultural-heritage and insurance purposes. Reasonable people disagree. You must decide.

Position A (there is a fact of the matter; the Vasa either is or is not the same object): There is a sharp boundary between “the same ship” and “a reconstruction”; what the Sorites shows is only that we cannot always know where it lies (this is Williamson’s epistemicism — see §The Sorites Paradox). The court does not need to create the boundary; it needs to apply the best available criterion (proportion of original keel? continuous registration as the same vessel? institutional intent?) and accept that some borderline cases will be decided incorrectly. Indeterminacy in our judgements is not the same as indeterminacy in the world. Legal decisions can — and do — proceed on the assumption that there are facts about identity even when those facts are hard to access, just as criminal verdicts proceed on the assumption that there are facts about what happened on the night of the crime.

Position B (the Sorites reveals a genuine metaphysical indeterminacy; the court is constructing, not discovering): There is no fact of the matter — full stop — about whether a 78% original-timber ship is “the same” object as its pre-sinking predecessor. Identity of an object persisting through gradual material replacement is not a natural kind; it is a stipulation we impose for practical purposes. The court is not discovering the correct answer; it is making a new one, and any criterion it picks (51%? The keel? The registered name?) will be stipulative. The right honesty is to acknowledge that law, insurance, and heritage designation all produce identities rather than recognising pre-existing ones.

Choose one. If you pick A, state the criterion you would apply to the Vasa and explain why it is not arbitrary — specifically, why that criterion rather than another survives the Sorites challenge applied to it. If you pick B, explain how courts can make principled rulings across cases if identity itself is stipulative — what stops Vasa jurisprudence from collapsing into arbitrary consistency with past rulings?


7 Can logic tell us what ought to be?

The Tuskegee Syphilis Study and the Is-Ought Gap

Between 1932 and 1972, the United States Public Health Service conducted a study in Macon County, Alabama, in which 399 Black men with syphilis were deliberately left untreated — even after penicillin became the established standard of care in the late 1940s — so that researchers could observe the disease’s natural progression.52 The study’s architects did not lack facts: they knew exactly what syphilis did to the human body, they knew penicillin cured it, they knew their subjects were suffering. What their scientific framework could not generate, and what was simply never articulated within the study’s logic, was any bridge from those facts to the conclusion that the study should therefore stop. The doctors involved were not monsters by their own self-understanding; they were scientists committed to the acquisition of knowledge. The is-ought gap — Hume’s observation that no statement of fact logically entails a statement about what ought to be done — is nowhere more visible than in institutional science, where the commitment to knowledge-gathering can become so entrenched that it crowds out the question of whether this particular knowledge ought to be gathered in this particular way. When the study was finally exposed by journalist Jean Heller in 1972, the public response was immediate moral outrage — suggesting that the ought was not mysterious, merely suppressed.

Logic is exquisitely powerful at telling us what follows from given premises. If you accept certain things, you must accept certain other things. But here is a question logic cannot directly answer: what should you accept in the first place? And, more acutely, can any set of purely factual premises logically entail a moral conclusion? Hume noticed, in a footnote that changed the history of philosophy, that it apparently cannot — and moral philosophers have been arguing with that footnote ever since. This is one of the most important questions at the intersection of logic and ethics.

7.1 Hume’s Is-Ought Gap

David Hume noticed something in moral philosophy (A Treatise of Human Nature, Book III, Part I, Section I, 1739):53

“In every system of morality, which I have hitherto met with, I have always remarked, that the author proceeds for some time in the ordinary way of reasoning, and establishes the being of a God, or makes observations concerning human affairs; when of a sudden I am surpriz’d to find, that instead of the usual copulations of propositions, is and is not, I meet with no proposition that is not connected with an ought, or an ought not.”

Hume observed that moral philosophers of his time would describe facts about human nature, God, or society — and then, without any logical bridge that he could find, slide into moral conclusions. “People desire happiness. Therefore, we ought to promote happiness.” The word “therefore” conceals a gap. (Natural-law theorists, the targets of Hume’s critique, deny that the gap exists at all: their factual premises about human nature already include teleological content from which obligations follow without any further normative addition. Whether this dissolves the gap or merely smuggles the ought into the description of human nature is exactly the question Hume opened — and which answer is right is what the rest of this section is about.)

The gap: no set of purely descriptive premises can logically entail a normative conclusion without an additional normative premise. From “people suffer when tortured” you cannot logically derive “you ought not torture people” without some further premise like “you ought not cause suffering.” That further premise is not itself derivable from facts alone.

This is sometimes called Hume’s Guillotine — the is-ought gap separates factual and normative discourse as cleanly as a blade.

7.2 Moore’s Naturalistic Fallacy

G.E. Moore (Principia Ethica, 1903) made a related point. He argued that the property of being “good” is a simple, non-natural property — it cannot be defined in terms of any natural property (pleasure, survival, what God commands, etc.). Any attempt to define good in natural terms commits the naturalistic fallacy.54

His test: the Open Question Argument. Suppose you define “good” as “what produces pleasure.” Then “Is pleasure good?” should be a closed question — trivially true by definition. But it isn’t. It remains a genuinely open question whether pleasure is good. Therefore “good” and “productive of pleasure” cannot mean the same thing. Repeat for any natural property: the question remains open.55

Moore thought the naturalistic fallacy explained why all previous ethical theories failed. They tried to define “good” and couldn’t.

7.3 The Structure of Ethical Arguments

Every ethical argument must contain at least one ethical premise. You cannot derive an “ought” from a pile of “is” statements. This means ethical arguments always have a normative foundation — and that foundation must itself be defended ethically, not empirically.

Example:

  1. The death penalty does not deter crime. [Empirical claim]
  2. Punishments that don’t deter crime should not be used. [Normative premise]
  3. Therefore, the death penalty should not be used. [Normative conclusion]

Premise 1 can be contested empirically. Premise 2 requires moral argument. Someone might accept 1 and reject 2 on the grounds that punishment has other purposes (retribution, justice). The empirical evidence, by itself, settles nothing — because it is the normative premise that does the moral work.

When knowledge claims in ethics are at issue, distinguishing the empirical and normative components of an argument is a crucial TOK skill.

7.4 Can Science Legislate Values?

A recurring question in modern thought: can neuroscience, evolutionary biology, or psychology tell us what is good? Sam Harris (The Moral Landscape, 2010) argues yes: if morality is about wellbeing, and science can study wellbeing, then science can study morality.56 E.O. Wilson (Sociobiology, 1975) suggested that human moral intuitions are adaptations selected for their fitness benefits.57

Hume’s guillotine, on the standard reading, cuts these arguments: even if science can tell us what maximises wellbeing, it cannot — without a further normative step — tell us that we ought to maximise wellbeing. That step requires a moral commitment that science cannot supply.

The standard reading is not unanimous. John Searle (“How to Derive ‘Ought’ from ‘Is’”, 1964) argued that constitutive facts about institutional practices — promising, marrying, signing a contract — generate ought-conclusions from is-premises, because the institutional facts already encode normative content.58 Philippa Foot (Natural Goodness, 2001) revived an Aristotelian line on which “good” for a living thing is grounded in the natural form of life it has: a sound oak is one that grows tall and bears acorns; a good human is one whose practical reasoning is in order — and these are not normative additions to the natural facts but their proper description.59 Whether these responses dissolve Hume’s gap or merely relocate the normative premise inside “constitutive rule” or “natural form of life” is itself contested. The unit takes no position on the resolution; it flags the dispute.

7.5 Questions to Argue About

  • “Natural selection has produced in us a tendency to care for our children. Therefore, it is natural — and right — to care for our children.” Does this argument commit the naturalistic fallacy? What are the missing steps?
  • Is Hume’s is-ought gap a logical point or a metaphysical one? Could there be a world in which “is” and “ought” were more directly connected?
  • If all ethical arguments require at least one normative premise, where do those foundational premises come from? Are they intuitions? Conventions? Revealed by God? Constructed?
  • Moore’s Open Question Argument says “Is X good?” is always an open question, for any natural X. Does this argument work? Can you find a case where it breaks down?

Forced Fork: How Should the Tuskegee Doctors Have Stopped?

Return to Tuskegee (info-box above). The doctors had every fact they needed: syphilis kills untreated; penicillin cures it; their subjects were dying. They did not stop, for forty years, until a journalist exposed the study in 1972. Suppose you are one of the PHS doctors in 1955, seven years after penicillin became standard care. You are asked to explain why you will continue, or refuse to continue, the withholding of treatment. Which of the two positions actually has a chance of getting you to refuse?

Position A (Hume’s gap is fatal): No collection of medical facts entails “we ought to stop.” To refuse, you must bring a normative premise — the dignity of the subjects matters more than the knowledge gained, or consent is necessary for human research. That premise is not provable from facts; it is a commitment. Ethical refusal is possible, but only by adding something to the empirical record. Pretending facts alone do the work is how the ethics failed to land for forty years.

Position B (Cornell realism / Foot): Hume’s gap presupposes a sharp factual/evaluative split closer inspection cannot sustain. On the Cornell-realist programme (Boyd, Brink, Sturgeon), moral properties are natural properties — wrongness reducible to features like unnecessary infliction of suffering. Philippa Foot’s Natural Goodness (2001) runs the parallel Aristotelian case. On either view, the Tuskegee facts entail the obligation to stop, without further normative premise.

Choose one. If you pick A, specify the minimal normative premise you would bring to a 1955 PHS review board to shut the study down — one that another doctor could reject, and say what you would say to them. If you pick B, explain what it is about “considered judgement in reflective equilibrium” that the Tuskegee doctors lacked, given that they presumably judged themselves to be reasonable people in the context of their institutions.


8 Does logic have to be classical?

Japanese Water Tribunal and the Logic of Degrees

Methylmercury discharged from the Chisso Corporation’s acetaldehyde plant into Minamata Bay, Kumamoto Prefecture, Japan, between 1932 and 1968 caused severe neurological injury — ataxia, peripheral numbness, tunnel vision, deafness, dysarthria, congenital cerebral damage in babies whose mothers had eaten contaminated fish during pregnancy, and death — to thousands of residents. The disease was officially identified in 1956. After Chisso was forced into the courts, the Kumamoto District Court ruled against the company in March 1973, ordering ¥937 million in damages and accepting that Chisso had had the foreseeability and the means to prevent the discharge.60 The 1973 judgment settled the binary causation question for the plaintiffs in front of it: Chisso did it. What it could not settle, and what dominated the next fifty years of Minamata litigation, was the boundary: who counts as a “Minamata-disease patient” entitled to compensation, given that mercury exposure produced a continuous spectrum of symptoms from severe Hunter–Russell syndrome at one end to subtle somatosensory disturbance at the other? The 1977 official certification criteria required a combination of multiple severe symptoms, leaving thousands of partially affected residents uncertified; successive lawsuits, the 1995 political settlement, the 2004 Supreme Court ruling against the state, and the 2009 Minamata Disease Special Measures Law all repeatedly redrew the line. Classical tort law asks “caused or not?”; Minamata jurisprudence has spent half a century fighting over a threshold the underlying biology does not respect. Lotfi Zadeh’s 1965 paper introducing fuzzy logic assigns truth values between 0 and 1.61 The Minamata tribunals were not formally fuzzy; they were classical tribunals trying to apply a binary predicate (“Minamata-disease patient”) to a continuum, and the cost of doing so has been counted in lives.

Classical logic — the logic Aristotle articulated and that Frege and Russell formalised — rests on certain principles that feel self-evident. Among them: the principle of non-contradiction (nothing can be both true and false), the law of excluded middle (every proposition is either true or false), and the principle of explosion (from a contradiction, anything follows). These feel like bedrock — like the very frame within which thought is possible. But that they feel like bedrock is not by itself an argument that they are bedrock. There are well-developed logics that reject each of them, used today in working mathematics, computer science, and the analysis of vagueness. The classicist will reply that these systems are local engineering tools that presuppose classical reasoning at the meta-level; the non-classicist replies that “presupposes classical reasoning at the meta-level” is precisely the question being begged.

8.1 The Principle of Non-Contradiction

Aristotle called it the most certain of all principles:

“It is impossible for the same thing to belong and not to belong at the same time to the same thing and in the same respect.” — Aristotle, Metaphysics, Book IV, Chapter 362

Non-contradiction is the logical backbone of rational argument. If a statement can be both true and false, any argument becomes valid — because from a contradiction, the principle of explosion allows you to derive any conclusion whatsoever.

But: is the principle a discovered truth about reality, or a constructed rule for our logical practices? Could reality be, in some domains, genuinely contradictory?

8.2 Fuzzy Logic

The lesson on the Sorites paradox above showed how badly classical logic handles vague predicates: one grain is not a heap; adding a single grain to a non-heap produces a non-heap; therefore (by induction) no number of grains is a heap. The argument is valid; the conclusion is absurd; one of the premises must go. Lotfi Zadeh’s 1965 proposal — fuzzy logic — pays for the absurdity by abandoning bivalence: truth values are not just 0 (false) or 1 (true) but any real number in the interval [0, 1]. “John is tall” can be 0.7 true. “These 100 grains are a heap” can be 0.3 true.63

This is not just a philosophical move. Fuzzy logic is the working logic of camera autofocus, washing-machine water-level sensors, anti-lock braking systems, and a generation of medical diagnostic decision-support tools. Each is a domain where the world supplies a continuous signal — a focus distance, a soil-moisture reading, a wheel-slip angle, a haemoglobin saturation — and the older binary engineering practice (sharp threshold, with hysteresis around it) misclassified at the boundary often enough to matter. A fuzzy system instead computes a degree of membership in each of several overlapping categories (“dry / damp / wet”), evaluates rules at degrees, and outputs a defuzzified action.

The philosophical move underneath the engineering: classical logic models a world of sharp distinctions, even where the underlying signal is continuous. Whether that produces errors at the boundary or merely registers a useful idealisation depends on what the system is supposed to do. For an anti-lock braking system the binary classical predicate “wheel locked” misses what matters; the engineering moved to fuzzy logic and the system improved. The Minamata case in the info-box puts the same question to the legal system: is bivalent classical reasoning a load-bearing feature of how rights are adjudicated, or a habit of inheritance whose cost is paid at the boundary by people who do not write the rules?

The cleanest objection is that fuzzy logic just relocates the problem: if “John is tall” is 0.7 true, is that statement (the assignment of degree 0.7) precisely true, or itself only approximately so? The fuzzy logician’s reply — that the assignment is itself a fuzzy proposition, with its own degree — invites a regress that has been a standing topic in the literature; see Williamson, Vagueness (1994), for the case that vague predicates are best handled with classical logic plus epistemicism (there is a sharp boundary; we just cannot know where).

8.3 Paraconsistent Logic

Paraconsistent logics reject the principle of explosion: they allow contradictions to coexist within a system without licensing arbitrary conclusions from them. This sounds alarming, but it is closer to working practice than the principle of explosion is:

  • A legal system may contain inconsistent laws (old law says X; new law says not-X; the contradiction hasn’t been resolved by repeal). Lawyers still reason within it.
  • A database may contain contradictory entries (data-entry errors, source conflicts, time-of-update mismatches). Queries still run.
  • Quantum mechanics and general relativity are locally inconsistent at certain scales. Working physicists continue to use both, picking which apparatus to invoke per problem.

Graham Priest (In Contradiction, 1987) pushes the claim further into dialetheism — the view that some contradictions are not merely tolerable but true. The Liar sentence (“this sentence is false”) is the canonical case: it appears to be both true and false. Dialetheism is not mainstream, but it is philosophically serious; the alternative (some hierarchy or restriction that blocks the Liar) has its own costs.64

The dialetheist position has a deeper non-Western precedent in the Buddhist catuṣkoṭi (Skt. “four corners,” sometimes “tetralemma”), systematically deployed by the 2nd-century Madhyamaka philosopher Nāgārjuna. Mūlamadhyamakakārikā I introduces the four-fold scheme: for any proposition P, the catuṣkoṭi asks whether P, not-P, both P and not-P, or neither P nor not-P. Nāgārjuna’s central use of the device is to argue that the four corners exhaust the available positions on, e.g., dependent origination — and that ultimate reality (paramārtha) cannot be fixed by any of them, including the second and the fourth.65 Whether the catuṣkoṭi is best read as an early dialetheist logic (Priest’s reading: at the limits of conceptualisation, contradictions are accepted), as a paraconsistent move (Garfield: the third corner is genuinely entertained without explosion), or as a strictly non-logical device for breaking the reader’s attachment to any conceptual position (the standard Madhyamaka self-understanding), is itself contested in modern Buddhist philosophy. The point for present purposes: classical logic’s three principles were not the only logic that working philosophers tried to organise reasoning by — and the contemporary non-classical logics did not arrive on a blank field.

A third major non-classical system, intuitionistic logic (L.E.J. Brouwer, early 20th c.), rejects the law of excluded middle — the rule that every proposition is either true or false. Intuitionists insist that to assert “P or not-P” you must already be in a position to assert one of them; for mathematical propositions that are neither proved nor disproved, neither half is yet available. The practical consequence: classical proofs by contradiction (assume not-P, derive absurdity, conclude P) are not generally valid in intuitionistic mathematics, only the constructive ones. This is not eccentric — it is the working logic of large parts of constructive mathematics and of theorem-proving systems like Coq and Lean.66

8.4 Logic as Discovery or Construction?

The deepest question: is logic something we discover or something we construct?

Quantum mechanics has features that resist classical logic. Quantum superposition seems to require that a particle is in two states simultaneously — a violation of the law of non-contradiction in some interpretations. Whether this really requires revising logic, or revising our interpretation of quantum mechanics, is disputed.

  • Platonism about logic: logical laws are objective, necessary truths about abstract structures. We discover them.
  • Formalism: logic is a formal game with rules we choose. Different games (classical, fuzzy, intuitionistic, paraconsistent) are tools for different purposes.
  • Naturalism: logic is continuous with science. Our logical beliefs are revisable in light of experience, just as empirical beliefs are. Quine argued the general revisability thesis in “Two Dogmas” — no statement, including a logical one, is in principle immune from revision in the face of recalcitrant experience.67 Hilary Putnam pushed the specific application to quantum mechanics in “Is Logic Empirical?” (1968), arguing that the strange behaviour of quantum systems gives us empirical reason to adopt a non-distributive quantum logic. Both moves are contested; the Quine reading especially has been resisted on the grounds that he never quite committed to revising any particular logical law on empirical grounds.68

8.5 An Open Question

This lesson has established that there are multiple logical systems — classical, fuzzy, paraconsistent, intuitionistic — and that each is internally consistent while differing from the others on fundamental principles. This raises a question that logic itself cannot answer:

What grounds the choice of logical system?

This is not a logical question. You cannot use logical inference to justify the choice of inferential rules — as Carroll’s Tortoise argument shows — because whichever rules you use in the justification are already the ones you are assuming. To ask why we accept classical logic rather than intuitionistic logic, or why we accept modus ponens at all, is to ask an epistemological question about the status of logical truths.

Two positions bracket the debate:

  • Quine (Two Dogmas of Empiricism, 1951, §§5–6): logical laws are simply the most entrenched nodes in our web of belief. They are not known a priori in any special sense — they are just the last things we would revise, because revising them would require revising almost everything else. Logical truths are not qualitatively different from empirical truths; they are merely more central.69
  • BonJour (In Defense of Pure Reason, 1998, Ch. 1 §§1.1–1.3 and Ch. 4 §§4.2–4.5): some knowledge is genuinely a priori — rational insight into necessary truths that no experience could undermine. Logical laws are known a priori because their negation cannot be coherently entertained by any rational mind. This is not merely a matter of entrenchment but of necessity.70

This question is carried into the Core unit. The two positions are incompatible, and the dispute is unresolved. Hold it open.

Which of these is right changes what it means to claim that logic is a form of knowledge.

8.6 Questions to Argue About

  • Aristotle thought the principle of non-contradiction was the most certain of all principles, beyond any possible doubt. But is this because it is true, or because any attempt to doubt it uses it? Is the principle self-validating?
  • Fuzzy logic admits truth values between 0 and 1. But doesn’t this just push the problem back — is “0.7 true” itself precisely true or just approximately true?
  • Paraconsistent logic allows for some contradictions without collapse. But if you accept any contradiction, how do you decide which contradictions are “tolerable” and which are disqualifying?
  • Quine said that logic is revisable in principle. If that’s right, what is the relationship between logic and knowledge? Can you even reason about the revision of logic using logic?

Forced Fork: Should the Minamata Courts Have Gone Fuzzy?

The Minamata case is in the info-box above. The 1973 Kumamoto District Court answered the binary question — Chisso caused the disease — and ordered damages. The fifty years since have been spent fighting over a different binary: who counts as a certified Minamata-disease patient, given that the underlying mercury exposure produced a continuous spectrum of harm. The 1977 certification criteria, the 1995 settlement, the 2004 Supreme Court ruling, and the 2009 Special Measures Law each redrew the line; tens of thousands of people have spent careers contesting which side of it they fall on. The question for you: should the courts have abandoned the binary “patient or not?” predicate altogether and adopted a degree-valued one — the move Zadeh’s fuzzy logic would have made formally available — or would that have changed the subject from law to engineering?

Position A (keep the binary predicate, even at the cost of moving the line): Classical logic is right — non-contradiction, excluded middle, bivalence are not engineering preferences. Law adjudicates rights; rights are binary (you have a claim against Chisso, or you do not); a fuzzy-valued “0.62 of a Minamata-disease patient” is not a status the legal system can confer. The fifty-year boundary fight is the system working as designed: criteria are revisable but they have to be criteria.

Position B (logical pluralism — Haack, Beall, Restall): There is no single “correct” logic, only logics better or worse fitted to their domain.71 Mercury exposure produces a dose-response curve, not a binary signal. The classical predicate forces the board to treat someone with 98% of the syndrome and someone with 49% as members of fundamentally different classes; the moral cost is visible in the lifetimes spent fighting the boundary. Graduated patient-status with graduated entitlements would track the underlying biology.

Choose one. If you pick A, explain what should happen to the plaintiff whose symptoms fall just below the certification threshold: is the answer “nothing” (and is that just?), and does your answer change if their neighbour with marginally worse symptoms is certified? If you pick B, specify what else has to be degree-valued once you let patient-status go fuzzy: does guilt in criminal law become fuzzy? Does legal personhood? Where does the graduated instrument stop being the right instrument?


9 Media

Novels, films, and artworks that illuminate the questions above:

  • Lewis Carroll, Alice’s Adventures in Wonderland (1865) and Alice Through the Looking-Glass (1871) — Carroll was a logician (Charles Dodgson, lecturer in mathematics at Oxford). The books are saturated with logical puzzles, non-sequiturs, and violations of inference. The Mad Hatter’s riddle (“Why is a raven like a writing desk?”) has no answer — a joke about the expectation that arguments have conclusions.
  • Douglas Hofstadter, Gödel, Escher, Bach: An Eternal Golden Braid (1979) — Pulitzer Prize-winning exploration of self-reference, formal systems, consciousness, and Gödel’s theorems, structured around parallels between Bach’s counterpoint and Escher’s impossible drawings. Difficult and brilliant.
  • Jorge Luis Borges, “The Library of Babel” (in Ficciones, 1944) — A universe consisting of an infinite library containing all possible books. A meditation on completeness, infinity, and the limits of formal systems.
  • Tom Stoppard, Rosencrantz and Guildenstern Are Dead (1967) — A play in which two minor Shakespeare characters discover they are living through a script they cannot control. Questions of logical necessity, free will, and the impossibility of reasoning outside one’s own system.
  • Christopher Nolan, Inception (2010) — Nested realities that cast doubt on the reliability of any single level of “truth.” Connects to questions of consistency, self-reference, and what it means for a system to be grounded.
  • Edwin Abbott Abbott, Flatland: A Romance of Many Dimensions (1884) — Creatures living in a two-dimensional world encounter beings from three dimensions; the logic of their world is consistent but radically limited. An allegory of the limits of formal systems.

10 Bibliography

Aristotle. Metaphysics. Trans. W. D. Ross. Oxford: Clarendon Press, c. 350 BCE.

Aristotle. Rhetoric. Trans. W. Rhys Roberts. Oxford: Clarendon Press, c. 322 BCE.

Ayer, A. J. Language, Truth and Logic. London: Victor Gollancz, 1936.

Bayes, Thomas. “An Essay Towards Solving a Problem in the Doctrine of Chances.” Philosophical Transactions of the Royal Society 53 (1763): 370–418.

Bonjour, Laurence. In Defense of Pure Reason: A Rationalist Account of A Priori Justification. Cambridge: Cambridge University Press, 1998.

Borges, Jorge Luis. Ficciones. 1944. Trans. Anthony Kerrigan. New York: Grove Press, 1962.

Brouwer, L. E. J. “Intuitionism and Formalism.” Bulletin of the American Mathematical Society 20 (1913): 81–96.

Carroll, Lewis. “What the Tortoise Said to Achilles.” Mind 4.14 (1895): 278–280.

Foot, Philippa. Natural Goodness. Oxford: Clarendon Press, 2001.

Frege, Gottlob. Grundgesetze der Arithmetik. 2 vols. Jena: Hermann Pohle, 1893–1903.

Frege, Gottlob. The Frege Reader. Ed. Michael Beaney. Oxford: Blackwell, 1997.

Gödel, Kurt. “On Formally Undecidable Propositions of Principia Mathematica and Related Systems.” 1931. Trans. B. Meltzer. Edinburgh: Oliver and Boyd, 1962.

Harris, Sam. The Moral Landscape: How Science Can Determine Human Values. New York: Free Press, 2010.

Hilbert, David. “Die Grundlagen der Mathematik.” Abhandlungen aus dem Mathematischen Seminar der Hamburgischen Universität 6 (1928): 65–85.

Hofstadter, Douglas. Gödel, Escher, Bach: An Eternal Golden Braid. New York: Basic Books, 1979.

Hume, David. A Treatise of Human Nature. 1739. Ed. L. A. Selby-Bigge, rev. P. H. Nidditch. Oxford: Clarendon Press, 2nd ed. 1978.

Hume, David. An Enquiry Concerning Human Understanding. 1748. In Hume: Dialogues Concerning Natural Religion and Other Writings, ed. Dorothy Coleman. Cambridge: Cambridge University Press, 2007.

Kahneman, Daniel. Thinking, Fast and Slow. New York: Farrar, Straus and Giroux, 2011.

Lipton, Peter. Inference to the Best Explanation. 2nd ed. London: Routledge, 2004.

Franzén, Torkel. Gödel’s Theorem: An Incomplete Guide to Its Use and Abuse. Wellesley, MA: A. K. Peters, 2005.

Lucas, J. R. “Minds, Machines and Gödel.” Philosophy 36.137 (1961): 112–127.

Moore, G. E. Principia Ethica. Cambridge: Cambridge University Press, 1903.

Penrose, Roger. The Emperor’s New Mind: Concerning Computers, Minds, and the Laws of Physics. Oxford: Oxford University Press, 1989.

Penrose, Roger. Shadows of the Mind: A Search for the Missing Science of Consciousness. Oxford: Oxford University Press, 1994.

Plato. Euthyphro. Trans. G. M. A. Grube. In Five Dialogues. Indianapolis: Hackett, c. 399 BCE.

Popper, Karl. The Logic of Scientific Discovery. 1934. London: Hutchinson, 1959.

Priest, Graham. In Contradiction: A Study of the Transconsistent. Dordrecht: Martinus Nijhoff, 1987.

Priest, Graham. Logic: A Very Short Introduction. 2nd ed. Oxford: Oxford University Press, 2017.

Quine, Willard Van Orman. “Two Dogmas of Empiricism.” Philosophical Review 60.1 (1951): 20–43.

Reichenbach, Hans. Experience and Prediction. Chicago: University of Chicago Press, 1938.

Russell, Bertrand. The Problems of Philosophy. London: Williams and Norgate, 1912.

Russell, Bertrand and Alfred North Whitehead. Principia Mathematica. 3 vols. Cambridge: Cambridge University Press, 1910–1913. Abridged ed.: Principia Mathematica to 56*. 2nd ed. Cambridge: Cambridge University Press, 1962.

Searle, John R. “How to Derive ‘Ought’ from ‘Is’.” Philosophical Review 73.1 (1964): 43–58.

Taleb, Nassim Nicholas. The Black Swan: The Impact of the Highly Improbable. New York: Random House, 2007.

van Heijenoort, Jean, ed. From Frege to Gödel: A Source Book in Mathematical Logic, 1879–1931. Cambridge, MA: Harvard University Press, 1967.

Williamson, Timothy. Vagueness. London: Routledge, 1994.

Wilson, E. O. Sociobiology: The New Synthesis. Cambridge, MA: Harvard University Press, 1975.

Zadeh, Lotfi A. “Fuzzy Sets.” Information and Control 8.3 (1965): 338–353.

11 Notes


  1. Mandeep R. Mehra, Sapan S. Desai, Frank Ruschitzka, and Amit N. Patel, “Hydroxychloroquine or chloroquine with or without a macrolide for treatment of COVID-19: a multinational registry analysis,” The Lancet, published online 22 May 2020 (later retracted). The paper drew on data attributed to the firm Surgisphere, which Desai owned. [VERIFY]↩︎

  2. James Watson et al., “Open letter to MR Mehra, SS Desai, F Ruschitzka, and AN Patel, authors of ‘Hydroxychloroquine or chloroquine with or without a macrolide for treatment of COVID-19: a multinational registry analysis’,” posted online 28 May 2020 with over 200 signatories from 24 countries. The letter detailed specific anomalies in the dataset (including Australian death counts exceeding the official total at the relevant date) and called for independent third-party audit. [VERIFY]↩︎

  3. Mandeep R. Mehra, Frank Ruschitzka, Amit N. Patel, “Retraction—Hydroxychloroquine or chloroquine with or without a macrolide for treatment of COVID-19: a multinational registry analysis,” The Lancet, 4 June 2020. The companion paper in NEJM (Mehra et al., “Cardiovascular Disease, Drug Therapy, and Mortality in Covid-19,” 1 May 2020) was retracted on 4 June 2020. The non-Surgisphere co-authors stated they could not verify the underlying data because Surgisphere refused to release it for independent audit. [VERIFY]↩︎

  4. For Snow’s own account of the Broad Street investigation and his map, see John Snow, On the Mode of Communication of Cholera, 2nd ed. (London: John Churchill, 1855). Steven Johnson, The Ghost Map (2006), gives a narrative reconstruction.↩︎

  5. Graham Priest, Logic: A Very Short Introduction (Oxford: Oxford University Press, 2000; 2nd ed. 2017) gives compact chapters on intuitionism, paraconsistency, relevance logic, and conditionals. For the longer technical statement, Priest, An Introduction to Non-Classical Logic, 2nd ed. (Cambridge: Cambridge University Press, 2008).↩︎

  6. Aristotle, Rhetoric, Book I, Chapter 2, distinguishes three pisteis (modes of proof) — those that depend on the character of the speaker (ethos), the emotion of the hearer (pathos), and the argument itself (logos).↩︎

  7. Plato, Euthyphro, 10a: the dilemma is put “Is what is holy holy because the gods approve of it, or do they approve of it because it is holy?”↩︎

  8. The popular sentence “I know that I know nothing” does not appear in any Platonic dialogue. The genuine source is Plato, Apology 21d, where Socrates concludes after questioning the politician that “neither of us appears to know anything great and good; but he fancies he knows something, although he knows nothing; whereas I, as I do not know anything, so I do not fancy I do” — and is therefore wiser only “in this trifling particular.” The crisper Latin formula scio me nihil scire is a much later distillation with no single textual original.↩︎

  9. Lewis Carroll, “What the Tortoise Said to Achilles,” Mind 4.14 (1895): 278–280. For discussion, see the SEP entry on Carroll’s Tortoise and the extensive literature on the rule-following regress.↩︎

  10. State v. Loomis, 881 N.W.2d 749 (Wis. 2016). Eric Loomis was charged in February 2013 with five offences related to a drive-by shooting in La Crosse, Wisconsin, and pleaded no contest to two of them on 17 August 2013. The trial court’s reliance on the COMPAS risk assessment is detailed in the Wisconsin Supreme Court’s opinion at ¶¶12–15. [VERIFY]↩︎

  11. State v. Loomis, 881 N.W.2d 749 (Wis. 13 July 2016). Majority opinion by Bradley, J.; concurrence by Roggensack, C.J., and Abrahamson, J. The US Supreme Court denied certiorari on 26 June 2017 (Loomis v. Wisconsin, 137 S.Ct. 2290). For analysis, see Harvard Law Review, “Criminal Law — Sentencing Guidelines — Wisconsin Supreme Court Requires Warning Before Use of Algorithmic Risk Assessments in Sentencing — State v. Loomis,” 130 Harv. L. Rev. 1530 (March 2017). [VERIFY]↩︎

  12. Julia Angwin, Jeff Larson, Surya Mattu, and Lauren Kirchner, “Machine Bias,” ProPublica, 23 May 2016. The investigation analysed COMPAS scores and two-year recidivism records for 7,214 defendants in Broward County, Florida. Northpointe (the firm that produces COMPAS, since renamed Equivant) responded with William Dieterich, Christina Mendoza, and Tim Brennan, “COMPAS Risk Scales: Demonstrating Accuracy Equity and Predictive Parity” (Northpointe Inc., 8 July 2016). [VERIFY]↩︎

  13. Jon Kleinberg, Sendhil Mullainathan, and Manish Raghavan, “Inherent Trade-Offs in the Fair Determination of Risk Scores,” Innovations in Theoretical Computer Science (ITCS) 2017; arXiv:1609.05807, posted September 2016. The result, also derived independently by Alexandra Chouldechova (“Fair Prediction with Disparate Impact: A Study of Bias in Recidivism Prediction Instruments,” Big Data 5.2 (2017): 153–163), shows that calibration across groups, balanced false-positive rates, and balanced false-negative rates cannot all be jointly satisfied when base rates differ between groups. [VERIFY]↩︎

  14. McCleskey v. Kemp, 481 U.S. 279 (1987). For the Baldus study, see David C. Baldus, George Woodworth, and Charles A. Pulaski Jr., Equal Justice and the Death Penalty: A Legal and Empirical Analysis (Boston: Northeastern University Press, 1990).↩︎

  15. For the constructive treatment of conditionals and the asymmetry between MP and MT, see Arend Heyting, Intuitionism: An Introduction (Amsterdam: North-Holland, 1956; 3rd rev. ed. 1971), Chapter 7; for the contemporary picture, Michael Dummett, Elements of Intuitionism, 2nd ed. (Oxford: Clarendon Press, 2000), §§1.3–1.4. Contraposition (P → Q ⊢ ¬Q → ¬P) is intuitionistically valid in only one direction; the converse direction requires double-negation elimination, which intuitionists reject.↩︎

  16. Roy Meadow’s “1 in 73 million” testimony was given at trial in November 1999 (Chester Crown Court). The figure is derived by squaring the cot-death incidence rate of approximately 1 in 8,500 for an affluent non-smoking household, on the (false) assumption of independence between the two deaths. Meadow’s evidence — and his use of similar statistical reasoning in other prosecutions — is treated in detail at R v Clark [2003] EWCA Crim 1020, paras. 96–180.↩︎

  17. R v Clark [2003] EWCA Crim 1020 (the Second Appeal judgment, 11 April 2003, quashing the conviction). For the statisticians’ intervention see Royal Statistical Society, “Royal Statistical Society Concerned by Issues Raised in the Sally Clark Case,” press release, 23 October 2001, and the subsequent letter from Peter Green (RSS President) to the Lord Chancellor, 23 January 2002.↩︎

  18. Lucia de Berk, a Dutch paediatric nurse, was convicted in 2003 of seven murders and three attempted murders on the basis of expert testimony that the probability of so many deaths occurring on her shifts by chance was 1 in 342 million. The figure was reanalysed by the statistician Richard Gill and others and shown to be the product of multiple statistical errors, including selection bias in the choice of comparison shifts. The Dutch Supreme Court ordered a retrial in 2008 and de Berk was acquitted in April 2010.↩︎

  19. Thomas Bayes, “An Essay Towards Solving a Problem in the Doctrine of Chances,” communicated by Richard Price after Bayes’s death and published in Philosophical Transactions of the Royal Society 53 (1763): 370–418. For a modern accessible presentation see Graham Priest, Logic: A Very Short Introduction, 2nd ed. (Oxford: Oxford University Press, 2017), Chapter 12 (“Inverse Probability”).↩︎

  20. Graham Priest, Logic: A Very Short Introduction, 2nd ed. (Oxford: Oxford University Press, 2017), Chapter 12 (“Inverse Probability: You Can’t Be Indifferent About It!”), discussing the Argument to Design. The full surrounding sentence reads: “It is seductive because people often confuse probabilities with their inverses, and so slide over a crucial part of the argument.”↩︎

  21. Daniel Kahneman, Thinking, Fast and Slow (New York: Farrar, Straus and Giroux, 2011), Chapter 14 (“Tom W’s Specialty”) and Chapter 16 (“Causes Trump Statistics”). The cab problem is from Chapter 16; the Tom W experiment from Chapter 14. The original journal source is Daniel Kahneman and Amos Tversky, “On the Psychology of Prediction,” Psychological Review 80 (1973): 237–251.↩︎

  22. The UK government’s equity-capital injection into RBS reached £45.5 billion across October 2008 and December 2009; total state exposure including the Asset Protection Scheme was substantially larger. The definitive post-mortem is Financial Services Authority, The Failure of the Royal Bank of Scotland, FSA Board Report (December 2011), which identifies the ABN AMRO acquisition (October 2007), over-reliance on short-term wholesale funding, and a thin capital position as the proximate causes; the “two lever arch folders and a CD” characterisation of due diligence appears at p. 158. See also National Audit Office, HM Treasury: The Asset Protection Scheme, HC 567 (21 December 2010). On the inductive logic of VaR specifically, see Nassim Nicholas Taleb, The Black Swan, 2nd ed. (2010), Chapter 16 (“The Aesthetics of Randomness”), and the technical critique in Jon Danielsson, “The Emperor Has No Clothes: Limits to Risk Modelling,” Journal of Banking and Finance 26 (2002): 1273–1296.↩︎

  23. Nassim Nicholas Taleb, The Black Swan: The Impact of the Highly Improbable (New York: Random House, 2007), especially Chapter 1 (“The Apprenticeship of an Empirical Sceptic”) and Chapter 4 (“One Thousand and One Days, or How Not to Be a Sucker”).↩︎

  24. David Hume, A Treatise of Human Nature (1739), Book I, Part III, Sections VI and XII; and An Enquiry Concerning Human Understanding (1748), Section IV (“Sceptical Doubts Concerning the Operations of the Understanding”) and Section V (“Sceptical Solution of These Doubts”). The Treatise passages are verified against the Selby-Bigge / Nidditch revised edition (Oxford: Clarendon Press, 2nd ed. 1978), Book I, Part III, §§VI and XII; the Enquiry passages against the Hackett Hume: Dialogues Concerning Natural Religion and Other Writings edition (ed. Coleman, Cambridge: Cambridge University Press, 2007), §IV.↩︎

  25. David Hume, A Treatise of Human Nature (1739), Book I, Part III, Section VI (“Of the Inference from the Impression to the Idea”). Quotation verified against the Selby-Bigge / Nidditch revised edition (Oxford: Clarendon Press, 2nd ed. 1978), p. 89 (= p. 115 of the held reprint), opening sentences of the section.↩︎

  26. Bertrand Russell, The Problems of Philosophy (1912), Chapter VI (“On Induction”).↩︎

  27. David Hume, An Enquiry Concerning Human Understanding (1748), Section IV (“Sceptical Doubts Concerning the Operations of the Understanding”), Part II, ¶16. The “no known connexion between the sensible qualities and the secret powers” sentence is in the second paragraph of §IV Part II; cited from the Hackett Hume: Dialogues Concerning Natural Religion and Other Writings, ed. Dorothy Coleman (Cambridge: Cambridge University Press, 2007), p. 91. The bread-nourishment example is the running illustration through Section IV; Hume returns to it explicitly in ¶16 (the “secret powers” passage) and ¶21.↩︎

  28. For the Bayesian inductivist response, see Colin Howson and Peter Urbach, Scientific Reasoning: The Bayesian Approach, 3rd ed. (Chicago: Open Court, 2006); for the statistical-learning-theory line, Vladimir N. Vapnik, The Nature of Statistical Learning Theory, 2nd ed. (New York: Springer, 2000); for inference to the best explanation, Peter Lipton, Inference to the Best Explanation, 2nd ed. (London: Routledge, 2004); for the material theory, John D. Norton, The Material Theory of Induction (Calgary: University of Calgary Press, 2021). Taleb’s specific claim about fat-tailed distributions is developed in Nassim Nicholas Taleb, The Black Swan: The Impact of the Highly Improbable, 2nd ed. (New York: Random House, 2010), Part III.↩︎

  29. Karl Popper, The Logic of Scientific Discovery (Logik der Forschung, 1934; English trans. 1959), especially Chapter I (“A Survey of Some Fundamental Problems”) and Chapter IV (“Falsifiability”).↩︎

  30. For the classic English-language statement of the verification principle, see A. J. Ayer, Language, Truth and Logic (London: Victor Gollancz, 1936), Chapter I. The Vienna Circle’s founding manifesto is Hans Hahn, Otto Neurath, and Rudolf Carnap, Wissenschaftliche Weltauffassung: Der Wiener Kreis (1929).↩︎

  31. Hans Reichenbach, Experience and Prediction (Chicago: University of Chicago Press, 1938), §§38–42. The “pragmatic vindication” is the argument that if any method can succeed at extrapolation, induction will.↩︎

  32. W. V. O. Quine, “Epistemology Naturalized,” in Ontological Relativity and Other Essays (New York: Columbia University Press, 1969).↩︎

  33. Kurt Gödel, “Über formal unentscheidbare Sätze der Principia Mathematica und verwandter Systeme I,” Monatshefte für Mathematik und Physik 38 (1931): 173–198. English translation as “On Formally Undecidable Propositions of Principia Mathematica and Related Systems,” trans. B. Meltzer (Edinburgh: Oliver and Boyd, 1962), and in Jean van Heijenoort (ed.), From Frege to Gödel (1967), 596–616.↩︎

  34. Komite Nasional Keselamatan Transportasi (Indonesia, KNKT), Aircraft Accident Investigation Report: PT. Lion Mentari Airlines Boeing 737-8 (MAX); PK-LQP, Tanjung Karawang, West Java, Republic of Indonesia, 29 October 2018, Final Report KNKT.18.10.35.04 (Jakarta: KNKT, 25 October 2019). The 21 MCAS activations on Lion Air JT610 are documented in Section 1.11 (flight recorder data). [VERIFY]↩︎

  35. KNKT, Final Report (2019), Sections 2 (analysis) and 3 (conclusions); Ethiopian Civil Aviation Authority, Aircraft Accident Investigation Bureau, Final Report on the Accident to Ethiopian Airlines Group B737-8 (MAX) Registered ET-AVJ on 10 March 2019 (Addis Ababa: ECAA, 23 December 2022); United States National Transportation Safety Board, Safety Recommendation Report ASR-19-01 (Washington: NTSB, 26 September 2019), Section 2.3 (the role of the AOA-DISAGREE alert that had been sold as an option rather than included as standard). [VERIFY]↩︎

  36. Joint Authorities Technical Review (FAA), Boeing 737 MAX Flight Control System: Observations, Findings, and Recommendations (Washington: FAA, 11 October 2019). The JATR was chaired by Christopher A. Hart, former chairman of the NTSB. The report’s central finding (Observation O3.2) is that MCAS had been classified during certification as a function whose failure mode was non-catastrophic, on the basis of an analysis that did not adequately account for repeated activation in conjunction with stick-shaker activation, master-caution alerts, and air-data disagreement — i.e. for the actual cognitive load on the flight crew. [VERIFY]↩︎

  37. David Hilbert, “Die Grundlagen der Mathematik” (1928), reprinted in Jean van Heijenoort (ed.), From Frege to Gödel (1967), 464–479. For Hilbert’s 1900 Paris address and the tenth problem, see “Mathematische Probleme,” Archiv der Mathematik und Physik 1 (1901): 44–63, 213–237.↩︎

  38. Russell’s letter is dated 16 June 1902; Frege’s reply 22 June 1902. Both are reproduced in Jean van Heijenoort (ed.), From Frege to Gödel (1967), 124–128. Russell’s opening paragraph and Frege’s full reply are also given in The Frege Reader, ed. Michael Beaney (Oxford: Blackwell, 1997), 271–273.↩︎

  39. Bertrand Russell and Alfred North Whitehead, Principia Mathematica, 3 vols. (Cambridge: Cambridge University Press, 1910–1913); 2nd ed. 1925–1927. The proof of “1 + 1 = 2” appears as proposition 54·43; Russell’s 1958 introduction to the abridged Principia Mathematica to 56 (Cambridge: Cambridge University Press, 2nd ed. 1962, p. 407) places the demonstration in Volume I, with the often-quoted authorial note (from the first edition’s prose at *54·43) that “the above proposition is occasionally useful.” The aphorism that “it takes 379 pages to prove that 1 + 1 = 2” refers to the page number in the 1910 first edition of Volume I. [VERIFY: 1910 vol I not held; “occasionally useful” remark accepted on secondary]↩︎

  40. Gödel’s original 1931 proof required ω-consistency (a stronger property than ordinary consistency: roughly, that the system does not prove ∃x P(x) while also proving ¬P(0), ¬P(1), ¬P(2), …). J. Barkley Rosser, “Extensions of Some Theorems of Gödel and Church,” Journal of Symbolic Logic 1.3 (1936): 87–91, constructed a different undecidable sentence for which simple consistency is sufficient. Standard textbook expositions usually present the Rosser-strengthened form without flagging the historical refinement. See Torkel Franzén, Gödel’s Theorem: An Incomplete Guide to Its Use and Abuse (Wellesley, MA: A. K. Peters, 2005), Chapter 1 (esp. p. 13) and Chapter 2 (pp. 26, 30), for a careful exposition.↩︎

  41. Douglas Hofstadter, Gödel, Escher, Bach: An Eternal Golden Braid (New York: Basic Books, 1979). The direct treatment of Gödel’s theorems is in Chapters IX (“Mumon and Gödel”) and XIV (“On Formally Undecidable Propositions of TNT and Related Systems”).↩︎

  42. J. R. Lucas, “Minds, Machines and Gödel,” Philosophy 36.137 (1961): 112–127; Roger Penrose, The Emperor’s New Mind (Oxford: Oxford University Press, 1989) and Shadows of the Mind (Oxford: Oxford University Press, 1994). Penrose’s own statement of the argument’s first strand: it “endeavours to show, by appealing to results of Gödel (and Turing) that mathematical thinking (and hence conscious thinking generally) is something that cannot be encapsulated within any purely computational model of thought” (The Emperor’s New Mind, Preface to the new edition). For the Putnam/Feferman/Chalmers replies, see Hilary Putnam, “Review of Shadows of the Mind,” Bulletin of the American Mathematical Society 32 (1995): 370–373, and Solomon Feferman, “Penrose’s Gödelian Argument,” Psyche 2 (1996): 21–32.↩︎

  43. J. R. Lucas, “Minds, Machines and Gödel,” Philosophy 36.137 (1961): 112–127; Roger Penrose, The Emperor’s New Mind (Oxford: Oxford University Press, 1989) and Shadows of the Mind (Oxford: Oxford University Press, 1994). Penrose’s own statement of the argument’s first strand: it “endeavours to show, by appealing to results of Gödel (and Turing) that mathematical thinking (and hence conscious thinking generally) is something that cannot be encapsulated within any purely computational model of thought” (The Emperor’s New Mind, Preface to the new edition). For the Putnam/Feferman/Chalmers replies, see Hilary Putnam, “Review of Shadows of the Mind,” Bulletin of the American Mathematical Society 32 (1995): 370–373, and Solomon Feferman, “Penrose’s Gödelian Argument,” Psyche 2 (1996): 21–32.↩︎

  44. Roger Penrose, Shadows of the Mind (1994), Part II, Chapters 7–8, develops the proposal that non-computable cognition is realised by quantum-state reductions in cytoskeletal microtubules. The biological mechanism (Orchestrated Objective Reduction) is set out jointly with Stuart Hameroff: Stuart Hameroff and Roger Penrose, “Orchestrated Reduction of Quantum Coherence in Brain Microtubules: A Model for Consciousness,” in Toward a Science of Consciousness, ed. Hameroff, Kaszniak, and Scott (Cambridge, MA: MIT Press, 1996), 507–540; and “Consciousness in the Universe: A Review of the ‘Orch OR’ Theory,” Physics of Life Reviews 11.1 (2014): 39–78.↩︎

  45. The Olympias was constructed at Piraeus 1985–87 to a design by the naval architect John Coates, working from textual and iconographic evidence assembled by the historian John Sinclair Morrison. See J. S. Morrison, J. F. Coates, and N. B. Rankov, The Athenian Trireme: The History and Reconstruction of an Ancient Greek Warship, 2nd ed. (Cambridge: Cambridge University Press, 2000).↩︎

  46. For the sinking, salvage, and conservation history, see Fred Hocker, Vasa I: The Archaeology of a Swedish Warship of 1628 (Stockholm: National Maritime Museums of Sweden, 2011), and the Vasa Museum’s published timeline at vasamuseet.se. The PEG-impregnation programme (1962–1979) and the post-2000 acid-degradation problem are documented in Magdalena von Bonsdorff Lindmark and Yvonne Fors, “The Vasa Experience with Polyethylene Glycol: A Conservator’s Perspective,” Journal of Cultural Heritage 13.3 (2012, supplement): S175–S182, and in Yvonne Fors and Magnus Sandström, “Sulfur and Iron in Shipwrecks Cause Conservation Concerns,” Chemical Society Reviews 35 (2006): 399–415. The “98% original wood” figure is the museum’s standard public claim; the philosophical question raised in the Forced Fork is what work the word “original” can do once the cellulose has been stabilised by long-term chemical infiltration.↩︎

  47. Zeno’s arguments survive only indirectly, most fully through Aristotle’s Physics, Book VI, Chapters 2 and 9 (239b–240a). For Cauchy’s resolution via convergent infinite series, see Augustin-Louis Cauchy, Cours d’Analyse (Paris: Debure, 1821).↩︎

  48. The Liar is traditionally attributed to Eubulides of Miletus (4th century BCE); the Cretan version to Epimenides. For the formal treatment, see Alfred Tarski, “The Concept of Truth in Formalized Languages” (1933), and the SEP entry on the Liar paradox.↩︎

  49. Bertrand Russell discovered the paradox in May–June 1901; it is published in The Principles of Mathematics (Cambridge: Cambridge University Press, 1903), Chapter X (§§100–106). For Russell’s first communication of the paradox to Frege and Frege’s response, see The Frege Reader, ed. Michael Beaney (Oxford: Blackwell, 1997), 271–273; for Zermelo’s 1908 axiomatic treatment, see Jean van Heijenoort (ed.), From Frege to Gödel (1967), 199–215.↩︎

  50. Gottlob Frege to Bertrand Russell, 22 June 1902, quoted from The Frege Reader, ed. Michael Beaney (Oxford: Blackwell, 1997), p. 272; also in Jean van Heijenoort (ed.), From Frege to Gödel (1967), 127–128. Frege added a despairing appendix (Nachwort) to the second volume of Grundgesetze der Arithmetik (1903) acknowledging the contradiction.↩︎

  51. Timothy Williamson, Vagueness (London: Routledge, 1994), especially Chapters 7 and 8, for the epistemicist view that vague predicates have precise but unknowable boundaries. (Cited from secondary literature; primary not yet held.) [VERIFY: primary not held]↩︎

  52. For the historical record, see James H. Jones, Bad Blood: The Tuskegee Syphilis Experiment, new and expanded ed. (New York: Free Press, 1993); and U.S. Department of Health, Education, and Welfare, Final Report of the Tuskegee Syphilis Study Ad Hoc Advisory Panel (Washington, D.C.: 1973). Jean Heller’s original Associated Press story ran on 25 July 1972.↩︎

  53. David Hume, A Treatise of Human Nature (1739), Book III (“Of Morals”), Part I, Section I (“Moral Distinctions Not Derived from Reason”), final paragraph. Quotation verified against the Selby-Bigge / Nidditch revised edition (Oxford: Clarendon Press, 2nd ed. 1978), p. 469 (= p. 499 of the held reprint).↩︎

  54. G. E. Moore, Principia Ethica (Cambridge: Cambridge University Press, 1903), §§10–14 (Chapter I).↩︎

  55. G. E. Moore, Principia Ethica (1903), §13 (Chapter I, “The Subject-Matter of Ethics”).↩︎

  56. Sam Harris, The Moral Landscape: How Science Can Determine Human Values (New York: Free Press, 2010), especially Chapter 1 (“Moral Truth”).↩︎

  57. E. O. Wilson, Sociobiology: The New Synthesis (Cambridge, MA: Harvard University Press, 1975), Chapter 27 (“Man: From Sociobiology to Sociology”).↩︎

  58. John R. Searle, “How to Derive ‘Ought’ from ‘Is’,” Philosophical Review 73.1 (1964): 43–58. Searle’s example: from “Jones uttered the words ‘I hereby promise to pay you, Smith, five dollars’” together with the constitutive rule of promising, one can derive “Jones ought to pay Smith five dollars.” The standard objection is that the constitutive rule itself smuggles in the normative premise; the standard reply is that constitutive rules are not normative additions but descriptions of what promising is.↩︎

  59. Philippa Foot, Natural Goodness (Oxford: Clarendon Press, 2001), especially Chapters 2 and 3. Foot’s neo-Aristotelian view is that evaluations of living things are grounded in the natural-historical pattern of the species, and that “good” for an organism is a matter of fact about its form of life rather than a non-natural property in Moore’s sense.↩︎

  60. For the medical history and the methylmercury aetiology, see Masazumi Harada, Minamata Disease, trans. T. Sakamoto and T. George (Kumamoto: Iwanami Shoten / Kumamoto Nichinichi Shimbun, 2004), and the Lancet editorial “Japan Remembers Minamata,” The Lancet 367.9505 (2006): 99. The 1973 Kumamoto District Court judgment (Watanabe et al. v. Chisso Corporation, 20 March 1973) is summarised in Frank Upham, Law and Social Change in Postwar Japan (Cambridge, MA: Harvard University Press, 1987), Chapter 2 (“Litigation as Social Protest: The Big Four Pollution Suits”). The 1977 official certification criteria, the 1995 political settlement, the Kansai Minamata Supreme Court ruling of 15 October 2004 (which held the state and Kumamoto Prefecture liable for failing to regulate Chisso after 1959), and the 2009 Minamata Disease Victims’ Relief Special Measures Law together constitute the half-century boundary fight. The Japanese Ministry of the Environment’s official record is Minamata Disease: The History and Measures (Tokyo: MoE, 2002), available in English on the ministry’s website.↩︎

  61. Lotfi A. Zadeh, “Fuzzy Sets,” Information and Control 8.3 (1965): 338–353.↩︎

  62. Aristotle, Metaphysics, Book IV (Gamma), Chapter 3, 1005b19–20 (Ross translation). Aristotle’s defence of the principle extends through Chapters 3–6 of Book IV.↩︎

  63. Lotfi A. Zadeh, “Fuzzy Sets,” Information and Control 8.3 (1965): 338–353.↩︎

  64. Graham Priest, In Contradiction: A Study of the Transconsistent (Dordrecht: Martinus Nijhoff, 1987; 2nd ed., Oxford: Oxford University Press, 2006).↩︎

  65. Nāgārjuna, Mūlamadhyamakakārikā (c. 150 CE), Chapter I (on dependent origination); standard English translation Jay Garfield, The Fundamental Wisdom of the Middle Way (Oxford University Press, 1995). Garfield and Priest, “Nāgārjuna and the Limits of Thought,” Philosophy East and West 53 (2003): 1–21, argues for a paraconsistent reading; Mark Siderits, “The Madhyamaka Critique of Epistemology,” Journal of Indian Philosophy 8 (1980): 307–335, defends a non-logical interpretation. The dispute over how to read the catuṣkoṭi is the central methodological question in contemporary Madhyamaka scholarship.↩︎

  66. L. E. J. Brouwer, “Intuitionism and Formalism” (inaugural address, 1912), Bulletin of the American Mathematical Society 20 (1913): 81–96. For a later systematic statement, see Arend Heyting, Intuitionism: An Introduction (Amsterdam: North-Holland, 1956).↩︎

  67. W. V. O. Quine, “Two Dogmas of Empiricism,” Philosophical Review 60.1 (1951): 20–43; reprinted in From a Logical Point of View (Cambridge, MA: Harvard University Press, 1953), Chapter 2. The “web of belief” metaphor and the claim that no statement is immune to revision occur in §6.↩︎

  68. Hilary Putnam, “Is Logic Empirical?,” in Boston Studies in the Philosophy of Science V, ed. R. S. Cohen and M. W. Wartofsky (Dordrecht: Reidel, 1968), 216–241; reprinted as “The Logic of Quantum Mechanics” in Putnam, Mathematics, Matter and Method (Cambridge: Cambridge University Press, 1975). Putnam later moderated his position; for the dialectic, see Maria Luisa Dalla Chiara, Roberto Giuntini, and Richard Greechie, Reasoning in Quantum Theory: Sharp and Unsharp Quantum Logics (Dordrecht: Kluwer, 2004), Chapter 1.↩︎

  69. W. V. O. Quine, “Two Dogmas of Empiricism,” Philosophical Review 60.1 (1951): 20–43; reprinted in From a Logical Point of View (Cambridge, MA: Harvard University Press, 1953), Chapter 2. The “web of belief” metaphor and the claim that no statement is immune to revision occur in §6.↩︎

  70. Laurence BonJour, In Defense of Pure Reason: A Rationalist Account of A Priori Justification (Cambridge: Cambridge University Press, 1998). For the case that genuine, non-tautological a priori justification is indispensable, see Ch. 1, §§1.1–1.3 (esp. p. 11 on the “need for an account of genuine and non-tautological a priori justification”). For the positive account of rational insight into necessity — including its fallibility and corrigibility — see Ch. 4, §§4.2–4.5 (pp. 100–130).↩︎

  71. Susan Haack, Deviant Logic, Fuzzy Logic: Beyond the Formalism (Chicago: University of Chicago Press, 1996); JC Beall and Greg Restall, Logical Pluralism (Oxford: Clarendon Press, 2006), Chapter 2 (the “Generalised Tarski Thesis”: logical consequence is necessary truth-preservation in virtue of logical form, where “case” admits classical, relevant, and constructive precisifications) and Chapter 7 (defence against the monist objection). Beall and Restall defend a “one schema, many admissible specifications” view; Haack defends a more straightforwardly tool-pluralist view. The opposing monist view is most forcefully argued in Graham Priest, Doubt Truth to Be a Liar (Oxford: Clarendon Press, 2006), Chapters 12–13.↩︎