How Likely is That? The growth of mathematics in the 20th and early 21st centuries has been explosive. More new mathematics has been discovered in the last 100 years than in the whole of previous human history. Even to sketch these discoveries would take thousands of pages, so we are forced to look at a few samples from the enormous amount of material that is available. An especially novel branch of mathematics is the theory of probability, which studies the chances associated with random events. It is the mathematics of uncertainty. Earlier ages scratched the surface, with combinatorial calculations of odds in gambling games and methods for improving the accuracy of astronomical observations despite observational errors, but only in the early 20th century did probability theory emerge as a subject in its own right. Probability and statistics Nowadays, probability theory is a major area of mathematics, and its apphed wing, statistics, has a significant effect on our everyday lives - possibly more significant than any other single major branch of mathematics. Statistics is one of the main analytic techniques of IMIW I IK I IV 11 IIIAI / the medical profession. No drug comes to market, and no treat ment is permitted in a hospital, until clinical trials have ensured that it is sufficiently safe, and that it is effective. Safety here is a relative concept: treatments may be used on patients suffering from an otherwise fatal disease when the chance of success would be far too small to use them in less serious cases. Probability theory may also be the most widely misunderstood, and abused, area of mathematics. But used properly and intelligendy, it is a major contributor to human welfare. A few probabilistic questions go back to antiquity, In the Middle Ages, we find discussions of the chances of throwing various numbers with two dice. To see how this works, let's start with one die. Assuming that the die is fair — which turns out to be a difficult concept to pin down - each of the six numbers 1, 2, 3, 4, 5 and 6 should be thrown equally often, in the long run. In the short run, equality is impossible: the first throw must result in just one of those numbers, for instance. In fact, after six throws it is actually rather unlikely to have thrown each number exacdy once. But in a long series of throws, or trials, we expect each number to turn up roughly one time in six; that is, with probability V6. If this didn't happen, the die would in all likelihood be unfair or biased. An event of probability 1 is certain, and one of probability 0 is impossible. All probabilities lie between 0 and 1, and the probability of an event represents the proportion of trials in which the event concerned happens. Back to that medieval question. Suppose we throw two dice simultaneously (as in numerous games from craps to Monopoly™). What is the probability that their sum is 5? The upshot of numerous arguments, and some experiments, is that the answer is %. Why? Suppose we distinguish the two dice, colouring one blue and the other red. Each die can independently yield six distinct numbers, 332 333 TAMING THE INFINITE making a total of 36 possible pairs of numbers, all equally likely.The combinations (blue + red) that yield 5 are 1 + 4, 2 + 3, 3 + 2, 4 + 1; these are distinct cases because the blue die produces distinct results in each case, and so does the red one. So in the long run we expect to find a sum of 5 on four occasions out of 36, a probability of */„ = %. Another ancient problem, with a clear practical application, is how to divide the stakes in a game of chance if the' game is interrupted for some reason. The Renaissance algebraists Pacioli, Cardano, andTartaglia all wrote on this question. Later, the Chevalier de Mere asked Pascal the same question, and Pascal and Fermat wrote each other several letters on the topic. From this early work emerged an implicit understanding of what probabilities are and how to calculate them. But it was all very ill-defined and fuzzy. A working definition of the probability of some event is the proportion of occasions on which it will happen. If we roll a die, and the six faces are equally likely, then the probability of any particular face showing is '/<,. Much early work on probability relied on calculating how many ways some event would occur, and dividing that by the total number of possibilities. A basic problem here is that of combinations. Given, say, a pack of six cards, how many different subsets of four cards are there? One method is to list these subsets: if the cards are 1-6, then they arc 1234 1235 1236 1245 1246 1256 1345 1346 1356 1456 2345 2346 2356 2456 3456 - so there are 15 of them. But this method is too cumbersome for larger numbers of cards, and something more systematic is needed. HOW LIKELY IS THAT? Imagine choosing the four members of the subset, one at a time. We can choose the first in six ways, the second in only five (since one has been eliminated), the third in four ways, the fourth in three ways. The total number of choices, in this order, is 6 x 5 x 4 x 3 = 360. However, every subset is counted 24 times - as well as 1234 we find 1243, 2134, and so on, and there are 24 (4 x 3 x 2) ways to rearrange four objects. So the correct answer is 360/24, which equals 15. This argument shows that the number of ways to choose m objects from a total of n objects is /n\ n(n-l)...(n-m+1) \mj 1 x 2 X 3 X • • ■ X m These expressions are called binomial coefficients, because they also arise in the algebraic expansion of the binomial (a+b)fl. If we arrange them in a table, so that the nth row contains the binomial coefficients then die nth row looks like this: (;)(;)(;)•(:) In the sixth row we see the numbers 1, 6, 15, 20, 15, 6, 1. Compare with the formula (x + iy - x6 + 6xs 4 15x4 + 20x3 + 15x2 4 6x'4 1 and we see the same numbers arising as the coefficients. This is not a coincidence. The triangle of numbers (see next page) is called Pascal's triangle because it was discussed by Pascal in 165 5. However, it was known much earlier; it goes back to about 950 in a commentary on an ancient Indian book called the Chondos Shastra. It was also known to the Persian mathematicians Al-Karaji and Omar Khayyam, and is known as the Khayyam triangle in modern Iran. 334 335 TAMING THE INFINITE IIIIW LIKELY IS THAT? Pascal's triangle 1 7 15 21 1 5 10 10 5 1 20 35 J5 35 21 □ Binomial coefficients were used to good effect in the first book on probability: the Ars Conjectandi (Art of Conjecturing) written by Jacob Bernoulli in 1713.The curious title is explained in the book: 'We define the art of conjecture, or stochastic art, as the art of evaluating as exacdy as possible the probabilities of things, so that in our judgments and actions we can always base ourselves on what has been found to be the best, the most appropriate, the most certain, the best advised; this is the only object of the wisdom of the philosopher and the prudence of the statesman.' So a more informative translation might be The Art of Guesswork. Bernoulli took it for granted that more and more trials lead to better and better estimates of probability. 'Suppose that without your knowledge there are concealed in an urn 3000 white pebbles and 2000 black pebbles, and in trying to determine the numbers of these pebbles you take out one pebble after another (each time replacing the pebble...) and that you observe how often a white and how often a black pebble is drawn... Cm you do 1111•. si > often that it becomes ten times, one hundred times, one tin uisand times, etc, more probable... that the numbers of whiles and blacks chosen are in the same 3:2 ratio as the pebblei in the urn, rather than a different ratio?' Here Bernoulli asked a fundamental question, and also invented a standard illustrative example, balls in urns. He clearly helievei I ihai a ratio of 3:2 was the sensible outcome, but he also recognized ih.il actual experiments would only approximate this ratio. Bui he believed that with enough trials, this approximation shot1l