taming THE INFINITE Physics goes mathematical Newton's Principia was impressive, with its revelation of dei'i mathematical, laws underlying natural phenomena. But wJu happened next was even more impressive. Mathematicians tacklts the entire panoply of physics - sound, light, heat, fluid flov gravitation, electricity, magnetism. In every case, they came ul with differential equations that described the physics, often very accurately. The long-term implications have been remarkable. Many of thl most important technological advances, such as radio, television and commercial jet aircraft depend, in numerous ways, on the mathematics of differential equations. The topic is still the subject of intense research activity, with new applications emerging almost daily. It is fair to say that Newton's invention of differential equations, fleshed out by his successors in the 18th and 19th centuries, is in many ways responsible for the society in which wc now live. This only goes to show what is lurking just behind the scenes, if you care to look. mpossible Quantities [;m mm Mathematicians distinguish several different kinds of number, with different properties. What really matters is not the individual numbers, but the system to which they belong - the company they keep. Pour of these number systems are familiar: the natural numbers, 1, 2, 3, the integers, which also include zero and negative whole numbers; the rational numbers, composed of fractions p/q where p and q are integers and q is not zero; and the real numbers, generally introduced as decimals that can go on forever -whatever that means - and represent both the rational numbers, as repeating decimals, and irrational numbers like Vl, e and jt whose decimal expansions do not ever repeat the same block of digits. Integers The name integer just means whole; the other names give the impression that the systems concerned are sensible, reasonable tilings - natural, rational and of course real. These names reflect, and 180 181 ta min g THE INFINITE IMPOSSIBLE QUANTITIES encourage, a long-prevailing view that numbers are features of the world around us. Many people think that the only way you can do mathematical research is to invent new numbers.This view is almost always wrong; a lot of mathematics is not about numbers at all, and in any case the usual aim is to invent new theorems, not new numbers. Occasionally, however, 'new numbers' do arise. And one such invention, a so-called 'impossible' or 'imaginary' number, completely changed the face of mathematics, adding immeasurably to its power.That number was the square root of minus one. To early mathematicians, such a description seemed ridiculous, because die square of any number is always positive. So, negative numbers cannot have square roots. But just suppose they did. What would happen? It took mathematicians a long time to appreciate that numbers are artificial inventions made by human beings; very effective inventions for capturing many aspects of nature, to be sure, but no more a part of nature than one of Euclid's triangles or a formula in calculus. Historically, we first see mathematicians starting to struggle with this philosophical question when they began to realize that imaginary numbers were inevitable, useful and somehow on a par with the more familiar real ones. Problems with cubics Revolutionary mathematical ideas are seldom discovered in the simplest and (with hindsight) most obvious context. They almost always emerge from something far more complicated. So it was with die square root of minus one. Nowadays, we normally introduce this number in terms of the quadratic equation x2 + 1 = 0, the solution of which is the square root of minus one - whatever that means. Among the first mathematicians to wonder whether it had a sensible meaning were the Renaissance algebraists, who ran into square roots of negative numbers in a surprisingly indirect way: the solution of cubic equations. Recall that del Ferro andTartaglia discovered algebraic solutions to cubic equations, later written up by Cardano in his Ars Magna. In modern symbols, the solution of a cubic equation x3 + ax = b is 'Jl _b^ + /— + 2 V27 4 + 3 27 4 The Renaissance mathematicians expressed this solution in words and abbreviations, but the procedure was the same. Sometimes this formula worked beautifully, but sometimes it ran into trouble. Cardano noticed that when the formula is applied to the equation x3 = 15x + 4, with the obvious solution x = 4, the result is expressed as x =fi +J-121 + tfl-f-Tu This expression seemed to have no sensible meaning, however, because -121 has no square root. A puzzled Cardano wrote to Tartaglia, asking for clarification, butTartaglia missed the point and his reply was distinctly unhelpful. An answer of sorts was provided by Rafael Bombelli in his three-volume book L'Algebra, printed in Venice in 1572 and Bologna in 1579. Bombelli was worried that Cardano's Ars Magna was rather obscure, and he set out to write somediing clearer. He operated on the troublesome square root as if it were an ordinary number, noticing that (2 +f-\y = 2+fm and deducing the curious formula ^2 +7-121 = 2+/-1 Similarly, Bombelli obtained the formula 3Jl +7-121 = 2 v'-l 182 183 taming THE INFINITE IMPOSSIBLE QUANTITIES Now he could rewrite the sum of the two cube roots as (2+vFl) + (2-/=I)=4 So this strange method yielded the right answer, a perfecdy norma integer, but it got there by manipulating 'impossible' quantities. This was ah very interesting, but why did it work? Imaginary numbers To answer this question, mathematicians had to develop good ways to think about square roots of negative quantities, and do calculations with them. Early writers, among them Descartes and Newton, interpreted these imaginary numbers as a sign that a problem has no solutions. If you wanted to find a number whose square was minus one, the formal solution, square root of minus one, was imaginary, so no solution existed. But Bombelli's calculation implied that there was more to imaginaries than that. They could be used to find solutions; they could occur when solutions did exist. In 1673 John Walks invented a simple way to represent imaginary numbers as points in a plane. He started from the familiar representation of real numbers as a line, with the positive numbers on the right and the negative ones on the left. Then he introduced another line, at right angles to the first, and along this new line he placed the imaginaries. This is like Descartes's algebraic approach to plane geometry, using coordinate axes. Real numbers form one axis in the picture, imaginaries another. Wallis did not state the idea in quite this form - his version was closer to Fermat's approach to coordinates than Descartes's. But the underlying point is the same. The remainder of the plane corresponds to complex numbers, which consist of two parts: one real, one imaginary. In Cartesian coordinates, we measure the real part along the real line and measure the imaginary part parallel to the imaginary line. So 3 + 2i lies 3 units to the right of the origin and 2 units up. n/2 Ji \ » \ 6 -2-10 1 2 The real number line \ \ 1 -2i •-3i Two copies of the real number line, placed at right angles WaUis's idea solved the problem of giving meaning to imaginary numbers, but no one took the slightest notice. However, his idea slowly gained ground subconsciously. Most mathematicians stopped worrying that the square root of minus one could not occupy any position on the real line, and realized that it could live somewhere in the wider world of the complex plane. Some failed to appreciate the idea: in 1758 Francois Daviet de Foncenex, in a paper about • 3+2/ 1 1 \ -3 -2 -1 1 2 3 -Si i-3/ The complex plane according to Wessel 184 185 taming THE INFINITE IMPOSSIBLE QUANTITIES imaginary numbers, stated that it was pointless to think of imaginaries as forming a line at right angles to the real line. Bui others took it to heart and understood its importance. The idea that a complex plane could extend the comfortable re jI line and give imaginaries a home was implicit in Wallis's work, but slightly obscured by the way he presented it. It was made explicit by the Norwegian Caspar Wessel in 1797. Wessel was a surveyor, and his main interest was to represent the geometry of the plane in terms of numbers. Working backwards, his ideas could be viewed as a method for representing complex numbers in terms of planar geometry. But he published in Danish, and his work went unnoticed until a century later, when it was translated into French. The French mathematician Jean-Robert Argand independently published the same representation of complex numbers in 1806, and Gauss discovered it independently of them both in 1811. Complex analysis If complex numbers had been good only for algebra, they might have remained an intellectual curiosity, of little interest outside pure madiematics. But as interest in calculus grew, and it took on a more rigorous form as analysis, people began to notice that a really interesting fusion of real analysis with complex numbers - complex analysis - was not only possible, but desirable. Indeed, for many problems, essential. This discovery stemmed from early attempts to think about complex functions.The simplest functions, such as the square or the cube, depended only on algebraic manipulations, so it was easy to define these functions for complex numbers. To square a complex number, you just multiply it by itself, the same process that you would apply to a real number. Square roots of complex numbers are marginally trickier, but there is a pleasant reward for making the effort: every complex number has a square root. Indeed, every nonzero complex number has precisely two square roots, one equal to minus the odier. So not only did augmenting the real numbers with ,i new number, i, provide -1 with a square root, it provided square roots for everything in the enlarged system of complex numbers. What about sines, cosines, the exponential function and the logarithm? At this stage, things started to get very interesting, but .ilso very puzzling, especially when it came to logarithms. like i itself, logarithms of complex numbers turned up in purely real problems. In 1702 Johann Bernoulli was investigating the process of integration, applied to reciprocals of quadratics. He knew ,i clever technique to carry out this task whenever the quadratic equation concerned has two real solutions, r and s. Then we can rewrite the expression to be integrated in terms of 'partial fractions' l! ' X — S which leads to the integral A log (x - r) + B log (x - s) But what if the quadratic has no real roots? How can you integrate the reciprocal of x1 + 1, for instance? Bernoulli realized that once you have defined complex algebra, the partial fraction trick still works, but now r and s are complex numbers. So, for example, + 1 i/2 x + i +- i/2 x -1 and the integral of this function takes the form % log (x + i) + log (x - i) This final step was not fully satisfactory, because it demanded a definition of the logarithm of a complex number. Was it possible to make sense of such a statement? Bernoulli thought it was, and proceeded to use his new idea to excellent effect. Leibniz also exploited this kind of thinking. But 186 187 ........,......him taming THE INFINITE the mathematical details were not straightforward. By 1712 tli. two of them were arguing about a very basic feature of tli! approach. Forget complex numbers - what was the logarithm off negative real number? Bernoulli thought that the logarithm of , negative real number should be real; Leibniz insisted that it vM complex. Bernoulli had a kind of proof for his claim: assuming ihr usual formalism of calculus, the equation —x dx can be integrated to yield log(-x) = log(x) However, Leibniz was unconvinced, and believed that tin-integration was correct only for positive real x. This particular controversy was sorted out by Euler in 1749, and Leibniz was right. Bernoulli, said Euler, had forgotten that any integration involves an arbitrary constant. What Bernoulli should have deduced was that log(-x) =log(x) + c for some constant c. What was this constant? If logarithms of negative (and complex) numbers are to behave like logarithms of positive real numbers, which is the point of the whole game, then it should be true that log(-x) = log(-l x x) = log (-1) + log x so that c = log (-1). Euler then embarked on a series of beautiful calculations that produced a more explicit form for c. First, he found a way to manipulate various formulas involving complex numbers, assuming they behaved much like real numbers, and deduced a relation between trigonometric functions and the exponential: 188 IMPOSSIBLE QUANTITIES e'o = cos 6 + i sin 6 a formula that had been anticipated in 1714 by Roger Cotes, kitting 6 = it, Euler obtained the delightful result that em - _] relating the two fundamental mathematical constants e and n. It is remarkable that any such relation should exist, and even more remarkable that it is so simple. This formula regularly tops league tables for the 'most beautiful formula of all time'. Taking the logarithm, we immediately deduce that log (-1) = 131 revealing the secret of that enigmatic constant c above: it is in. As Mich, it is imaginary, so Leibniz was right and Bernoulli was wrong. However, there is more, and it opens Pandora's box. If we put B — 2ji, then e2w - i So log (1) = 2ijt.Then the equation x = x x 1 implies that log x = log x + 2iJt from which we conclude that if n is any integer whatever, log x = log x + 2nijt At first sight, this makes no sense - it seems to imply that 2niir = 0 for all n. But there is a way to interpret it that does make sense. Over the complex numbers, the logarithmic function is many-valued. Indeed, unless the complex number z is zero, the function log z can take infinitely many distinct values. (When z = 0, the value log 0 is not defined.) Mathematicians were used to functions that could take several distinct values, the square root being the most obvious example: here, even a real number possessed two distinct square roots, one 189 taming THE INFINITE What complex numbers did for them IMPOSSIBLE QUANTITIES The real and imaginary parts of a complex function satisfy the Cauchy-Riemann equations, which are closely related to the PDEs for gravitation, electricity, magnetism and some types of fluid flow in the plane. This connection made it possible to solve many equations of mathematical physics - but only for two-dimensional systems. positive and the other negative. But infinitely many values? This was very strange. Cauchy's theorem What really put the cat among the pigeons was the discovery that you could do calculus - analysis - with complex functions, and that the resulting theory was elegant and useful. So useful, in fact, that the logical basis of the idea ceased to be an important issue. When something works, and you feel that you need it, you generally stop asking whether it makes sense. The introduction of complex analysis seems to have been a conscious decision by the mathematical community — a generalization so obvious and compelling that any mathematician with any kind of sensitivity would want to see what transpired. In 1811 Gauss wrote a letter to a friend, the astronomer Friedrich Bessel, revealing his representation of complex numbers as points in a plane; he also mentioned some deeper results. Among them is a basic theorem upon which the whole of complex analysis hangs. Today we call it Cauchy's Theorem, because it was published by Cauchy but Gauss had die idea much earlier in his unpublished writings. This theorem concerns definite integrals of complex functions: that is, expressions Iwo distinct paths P and O from -1 to 1 in the complex plane I Az)dz where a and b are complex numbers. In real analysis this expression can be evaluated by finding an antiderivative F(z) of f(z), that is, a unction F(z) such that its derivative dF(z)/dz = f(z). Then the definite integral is equal to F(b) - F(a). In particular, its value depends only on the end points a and b, not on how you move from one to the other. Complex analysis, said Gauss, is different. Now the value of the integral may depend on the path that the variable z takes as it moves from a to b. Because the complex numbers form a plane, their geometry is richer than that of the real line, and this is where thai extra richness matters. For example, suppose you integrate f(z) = 1 /z from a = — 1 to b = 1. If the path concerned is a semicircle P lying above the real axis, then the integral turns out to be -Jti. But if the path is a semicircle Q lying below the real axis, then the integral turns out to be Jti.The two values are different, and the difference is 2jti. This difference, said Gauss, occurs because the function 1/z is badly behaved. It becomes infinite inside the region enclosed by the 190 191 taming THE INFINITE Augustin-Louis Cauchy 1789-1857 Augustin-Louis Cauchy was born in Paris during a time of political turmoil. Laplace and Lagrange were family friends, so Cauchy was exposed to higher mathematics at an early age. He went to the Ecole Polytechnique, graduating in 1807. In 1810 he carried out engineering work in Cherbourg, preparing for Napoleon's planned invasion of England, but he continued thinking about mathematics, reading Laplace's Mecanique Celeste {Celestial Mechanics) and Lagrange's Theorie des Fonctions (Theory of Functions), He continually sought academic positions, without success, but kept working on mathematics. His famous paper on complex integrals, which effectively founded complex analysis, appeared in 1814, and he finally achieved his goal of an academic post, becoming assistant professor of analysis at the Ecole Polytechnique a year later. His mathematics now flourished, and a paper on waves won him the 1816 prize of the Academy of Sciences. He continued to develop complex analysis, and in his 1829 Legons surle Calcul Differential he gave the first explicit definition of a complex function. After the revolution of 1830 Cauchy briefly went to Switzerland, and in 1831 he became professor of theoretical physics in Turin. His courses were reported as being highiy disorganized. By 1833 he was in Prague tutoring the grandson of Charles X, but the prince disliked both mathematics and physics, and Cauchy often lost his temper. He returned to Paris in 1838, regaining his post at the Academy but did not regain his teaching positions until Louis Philippe was deposed in 1848. Altogether, he published an astonishing 789 research articles on mathematics. two paths. Namely, at z = 0, which here is the centre of the circle formed by the two paths. 'But if this does not happen ... I affirm,' IMPOSSIBLE QUANTITIES (buss wrote to Bessel, 'that the integral has only one value even if i,i ken over different paths provided [the function] does not become infinite in the space enclosed by the two paths. This is a very beautiful theorem, whose proof I shall give on a convenient occasion'. But he never did. Instead, the theorem was rediscovered, and published by \ngustin-Louis Cauchy, the true founder of complex analysis. Gauss may have had the ideas, but ideas are useless if no one gets to see tliem. Cauchy published his work. In fact, Cauchy seldom stopped publishing. It is said that the rule, still in force today, that the n lurnal Comptes Rendus de I'Aademie Franchise accepts papers no more dian lour pages long, was introduced explicidy to stop Cauchy filling it with his huge output. But when the rule was introduced, Cauchy Hist wrote lots of short papers. From his prolific pen the main outlines of complex analysis quickly emerged. And it is a simpler, more elegant and in many ways more complete theory than real analysis, where the whole idea started. For instance, in real analysis a function may be differentiable, but us derivative may not be. It may be differentiable 23 times, but not 24. It may be differentiable as many times as you wish, but not I lossess a power series representation. None of these nasty things can happen in complex analysis. If a function is differentiable, then it can lie differentiated as many times as you wish; moreover, it has a power series representation.The reason - closely related to Cauchy's Theorem and probably a fact used by Gauss in his unknown proof-is that in order to be differentiable, a complex function must satisfy some very stringent conditions, known as die Cauchy-Riemann equations. These equations lead directly to Gauss's result that the integral between two points may depend on the path chosen. Equivalently, as Cauchy noted, the integral round a closed path need not be zero. It is zero provided the function concerned is differentiable (so in particular is not infinite) at all points inside the path. There was even a theorem - the residue theorem - that told you taming THE INFINITE IMPOSSIBLE QUANTITIES )l the value of an integral round a closed path, and it depended onl on die locations of the points at which the function became infinite, and its behaviour near diose points. In short, the entire structure of a complex function is determined by its singularities — the points at which it is badly behaved. And the most important singularities are its poles, the places where it becomes infinite. The square root of minus one puzzled mathematicians for centuries. Although there seemed to be no such number, it kept turning up in calculations. And there were hints that the concept must make some kind of sense, because it could be used to obtain perfectly valid results which did not themselves involve taking the square root of a negative number. As the successful uses of this impossible quantity continued to grow, mathematicians began to accept it as a useful device. lis status remained uncertain until it was realized that there is i logically consistent extension of the traditional system of real I numbers, in which the square root of minus one is a new kind of quantity — but one that obeys all of the standard laws of arithmetic. Geometrically, the real numbers form a line and the complex numbers form a plane; the real line is one of the two axes of this plane. Algebraically, complex numbers are just pairs of real numbers with particular formulas for adding the pairs or multiplying them. Now accepted as sensible quantities, complex numbers quickly spread throughout mathematics because they simplified calculations by avoiding the need to consider positive and negative numbers separately. In diis respect they can be considered as analogous to the earlier invention of negative numbers, which avoided the need to consider addition and subtraction separately. Today, complex numbers, and the calculus of complex functions, are routinely used as an indispensable technique in virtually all branches of science, engineering and mathematics. What complex numbers do for us Today, complex numbers are widely used in physics and engineering. A simple example occurs in the study of oscillations: motions that repeat periodically. Examples include the shaking of a building in an earthquake, vibrations in cars and the transmission of alternating electrical current. The simplest and most fundamental type of oscillation takes the form a cos art, where t is time, a is the amplitude of the oscillation and (o is its frequency. It turns out to be convenient to rewrite this formula as the real part of the complex function ef. The use of complex numbers simplifies calculations because the exponential function is simpler than the cosine. So engineers studying oscillations prefer to work with complex exponentials, and revert to the real part only at the end of the calculation. Complex numbers also determine the stabilities of steady states of dynamical systems, and are widely used in control theory. This subject is about methods for stabilizing systems that would otherwise be unstable. An example is the use of computer-controlled moving control surfaces to stabilize the space shuttle in flight. Without this application of complex analysis, the space shuttle would fly like a brick. 194 195