by John Gribbin
According to this view, space itself defined the framework against which things should be measured—absolute space, through which the Earth, the Sun, light, and everything else moved. This absolute space was also sometimes called the “aether” (or “ether”) and was conceived of as a substance through which electromagnetic waves moved, like water waves moving over the sea. The snag was, when experimenters tried to measure changes in the velocity of light caused by the motion of the Earth through absolute space (or “relative to the aether”), none could be found.
Because the Earth moves around the Sun in a roughly circular orbit, it should be moving at different speeds relative to absolute space at different times of the year. It’s like swimming in a circle in a fast-flowing river. Sometimes the Earth will be “swimming with the aether,” sometimes across the aether, and sometimes against the flow. If light always travels at the same speed relative to absolute space, common sense tells us this ought to show up in the form of seasonal changes in the speed of light measured from the Earth. It does not.
Einstein resolved the dilemma with his special theory. This says that all frames of reference are equally valid and that there is no absolute reference frame. Anybody who moves at a constant velocity through space is entitled to regard himself or herself as stationary. They will find that moving objects in their frame of reference obey Newton’s laws, while electromagnetic radiation obeys Maxwell’s equations and the speed of light is always measured to be the value that comes out of those equations, denoted by the letter c (for “constant” or the Latin celeritas, meaning “swiftness, celerity”). Furthermore, anybody who is moving at a constant speed relative to the first person (the first observer, in physicists’ jargon) will also be entitled to say that they are at rest and will find that objects in their laboratory obey Newton’s laws, while measurements always give the speed of light as c. Even if one observer is moving toward the other observer at half the speed of light and sends a torch beam out ahead, the second observer will not measure the speed of the light from the torch as 1.5c: it will still be c!
Starting out from the observed fact that the speed of light is a constant, the same whichever way the Earth is moving through space, Einstein found a mathematical package to describe the behavior of material objects in reference frames that move with constant velocities relative to one another—so-called “inertial” frames of reference. Provided the velocities are small compared with the speed of light, these equations give exactly the same “answers” as Newtonian mechanics. But when the velocities begin to become an appreciable fraction of the speed of light, strange things happen.
Two velocities, for example, can never add up to give a relative velocity greater than c. An observer may see two other observers approaching each other on a head-on collision course, each traveling at a speed of 0.9c in the first observer’s reference frame, but measurements carried out by either of those two fast-moving observers will always show that the other one is traveling at a speed less than c but bigger (in this case) than 0.9c.
The reason why velocities add up in this strange way has to do with the way both space and time are warped at high velocities. In order to account for the constancy of the speed of light, Einstein had to accept that moving clocks run more slowly than stationary clocks and that moving objects shrink in the direction of their motion. The equations also tell us that moving objects increase in mass the faster they go.
Strange and wonderful though all these things are, they are only peripheral to the story of modern cosmology and to the search for links between quantum physics and gravity. We stress, however, that they are not wild ideas in the sense that we sometimes dismiss crazy notions as “just a theory” in everyday language. To scientists, a theory is an idea that has been tried and tested by experiments and has passed every test. The special theory of relativity is no exception to this rule. All the strange notions implicit in the theory—the constancy of the speed of light, the stretching of time and shrinking of length for moving objects, the increase in mass of a moving object—have been measured and confirmed to great precision in very many experiments. Particle accelerators—“atom-smashing” machines like those at CERN, the European Center for Nuclear Research in Geneva—simply would not work if the theory were not a good one, since they have been designed and built around Einstein’s equations. The special theory of relativity as a description of the high-speed world is as securely founded in solid experimental facts as is Newtonian mechanics as a description of the everyday world; the only reason it conflicts with our common sense is that in everyday life we are not used to the kind of high-speed travel required for the effects to show up. After all, the speed of light, c, is ≈186,000 miles a second (≈300,000 kilometers a second), and the relativistic effects can be safely ignored for any speeds less than about 10 percent of this—that is, for speeds less than a mere 30,000 kilometers a second.
In essence, the special theory is the result of a marriage of Newton’s equations of motion with Maxwell’s equations describing radiation. It was very much a child of its time, and if Einstein hadn’t come up with the theory in 1905, one of his contemporaries would surely have done so within the next few years. Without Einstein’s special genius, though, it might have been a generation or more before anyone realized the importance of a far deeper insight buried within the special theory.
This key ingredient, to which we have already alluded, was the fruit of another marriage—the union of space and time. In everyday life, space and time seem to be quite different things. Space extends around us in three dimensions (up and down, left and right, forward and backward). We can see where things are located in space, and travel through it, more or less at will. Time, although we all know what it is, is almost impossible to describe. In a sense, it does have a direction (from past to future), but we can look neither into the future nor into the past, and we certainly cannot move through time at will. Yet the great universal constant, c, is a speed, and speed is a measure that relates space and time. Speeds are always in the form of miles per hour, or centimeters per second, or any other unit of length per unit of time. You cannot have one without the other when you are talking about speed. So the fact that the fundamental constant is a velocity must be telling us something significant about the Universe. But what?
If you multiply a speed by a time, you get a length. And if you do this in the right way (by multiplying intervals of time by the speed of light, c), you can combine measures of length (space) with measures of time in the same set of equations. The set of equations that combine space and time in this way consists of the equations of the special theory of relativity that describe time dilation and length contraction† and lead to the prediction that a mass m is equivalent to an energy E as described by the formula E = mc2. Instead of thinking about space and time as two separate entities, as long ago as 1905 Einstein was telling physicists that they should be thinking about them as different aspects of a single, unified whole—spacetime. But this spacetime, the special theory also said, was not fixed and permanent like the absolute space or absolute time of Newtonian physics—it could be stretched or squeezed. And therein lay the clue to the next great step forward.
Einstein used to say that the inspiration for his general theory of relativity (which is, above all, a theory of gravity) came from the realization that a person inside a falling elevator whose cable had snapped would not feel gravity at all. We can picture exactly what he meant because we have seen film of astronauts orbiting the Earth in spacecraft. Such an orbiting spacecraft is not “outside” the influence of the Earth’s gravity; indeed, it is held in orbit by gravity. But the spacecraft and everything in it is falling around the Earth with the same acceleration, so the astronauts have no weight and float within their capsule. For them, it is as if gravity does not exist, a phenomenon known as free fall. But Einstein had never seen any of this and had to picture the situation in a freely falling elevator in his imagination. It is as if the acceleration of the falling elevator, spe
eding up with every second that passes, precisely cancels out the influence of gravity. For that to be possible, gravity and acceleration must be exactly equivalent to one another.
The way this led Einstein to develop a theory of gravity was through considering the implications for a beam of light, the universal measuring tool of special relativity. Imagine shining a flashlight horizontally across the elevator from one side to the other. In the freely falling elevator, objects obey Newton’s laws: they move in straight lines, from the point of view of an observer in the elevator, bounce off each other with action and reaction equal and opposite, and so on. And, crucially, from the point of view of the observer in the elevator, light travels in straight lines.
But how do things look to an observer standing on the ground watching the elevator fall? The light would appear to follow a track that always stays exactly the same distance below the roof of the elevator. But in the time it takes the light to cross the elevator, the elevator has accelerated downward, and the light in the beam must have done the same. In order for the light to stay the same distance below the roof all the way across, the light pulse must follow a curved path as seen from outside the elevator. In other words, a light beam must be bent by the effect of gravity.
Einstein explained this in terms of bent spacetime. He suggested that the presence of matter in space distorts the spacetime around it, so that objects moving through the distorted spacetime are deflected, just as if they were being tugged in ordinary “flat” space by a force inversely proportional to the square of the distance. Having thought up the idea, Einstein then developed a set of equations to describe all this. The task took him ten years. When he had finished, Newton’s famous inverse-square law reemerged from Einstein’s new theory of gravity; but general relativity went far beyond Newton’s theory, because it also offered an all-embracing theory of the whole Universe. The general theory describes all of spacetime and therefore all of space and all of time. (There is a neat way to remember how it works. Matter tells spacetime how to bend; bends in spacetime tell matter how to move. But, the equations also insisted, spacetime itself can also move, in its own fashion.)
The general theory was completed in 1915 and published in 1916. Among other things, it predicted that beams of light from distant stars, passing close by the Sun, would be bent as they moved through spacetime distorted by the Sun’s mass. This would shift the apparent positions of those stars in the sky—and the shift might actually be seen, and photographed, during a total eclipse, when the Sun’s blinding light is blotted out. Just such an eclipse took place in 1919; the photographs were taken and showed exactly the effect Einstein had predicted. Bent spacetime was real: the general theory of relativity was correct.
But the equations developed by Einstein to describe the distortion of spacetime by the presence of matter, the very equations that were so triumphantly vindicated by the eclipse observations, contained a baffling feature that even Einstein could not comprehend. The equations insisted that the spacetime in which the material Universe is embedded could not be static. It must be either expanding or contracting.
Exasperated, Einstein added another term to his equations, for the sole purpose of holding spacetime still. Even at the beginning of the 1920s, he still shared (along with all his contemporaries) the Newtonian idea of a static Universe. But within ten years, observations made by Edwin Hubble with a new and powerful telescope on a mountaintop in California had shown that the Universe is expanding.
The stars in the sky are not moving farther apart from one another. The individual stars we can see from Earth all belong to a huge system, the Milky Way Galaxy, which contains about a hundred billion stars and is like an island in space. In the 1920s, astronomers discovered with the aid of new telescopes that there are many other galaxies beyond the Milky Way, many of them containing hundreds of billions of stars like our Sun. And it is the galaxies, not individual stars, that are receding from one another, being carried farther apart as the space in which they are embedded expands.
If anything, this was an even more extraordinary and impressive prediction of the general theory than the bending of light detectable during an eclipse. The equations had predicted something that even Einstein at first refused to believe but which observations later showed to be correct. The impact on scientists’ perception of the world was shattering. The Universe was not static, after all, but evolving; Einstein later described his attempt to fiddle the equations to hold the Universe still as “the greatest blunder of my life.” Even at the end of the 1920s, the observations and the theory agreed that the Universe is expanding. And if galaxies are getting farther apart, that means that long ago they must have been closer together. How close could they ever have been? What happened in the time when galaxies must have been touching one another and before then?
The idea that the Universe was born in a super-dense, super-hot fireball known as the Big Bang is now a cornerstone of science, but it took time—over fifty years—for the theory to become developed. Just at the time astronomers were finding evidence for the universal expansion, transforming the scientific image of the Universe at large, their physicist colleagues were developing the quantum theory, transforming our understanding of the very small. Attention focused chiefly on the development of the quantum theory over the next few decades, with relativity and cosmology (the study of the Universe at large) becoming an exotic branch of science investigated by only a few specialist mathematicians. The union of large and small still lay far in the future, even at the end of the 1920s.
As the nineteenth century gave way to the twentieth, physicists were forced to revise their notions about the nature of light. This initially modest readjustment of their worldview grew, like an avalanche triggered by a snowball rolling down a hill, to become a revolution that engulfed the whole of physics—the quantum revolution.
The first step was the realization that electromagnetic energy cannot always be treated simply as a wave passing through space. In some circumstances, a beam of light, for example, will behave more like a stream of tiny particles (now called photons). One of the people instrumental in establishing this “wave-particle duality” of light was Einstein, who in 1905 showed how the way in which electrons are knocked out of the atoms in a metal surface by electromagnetic radiation (the photoelectric effect) can be explained neatly in terms of photons, not in terms of a pure wave of electromagnetic energy. (It was for this work, not his two theories of relativity, that Einstein received his Nobel Prize.)
This wave-particle duality changes our whole view of the nature of light. We are used to thinking of momentum as a property to do with the mass of a particle and its speed (or, more correctly, its velocity). If two objects are moving at the same speed, the heavier one carries more momentum and will be harder to stop. A photon does not have mass, and at first sight you might think this means it has no momentum either. But, remember, Einstein discovered that mass and energy are equivalent to one another, and light certainly does carry energy—indeed, a beam of light is a beam of pure energy. So photons do have momentum, related to their energy, even though they have no mass and cannot change their speed. A change in the momentum of a photon means that it has changed the amount of energy it carries, not its velocity; and a change in the energy of a photon means a change in its wavelength.
When Einstein put all of this together, it implied that the momentum of a photon multiplied by the wavelength of the associated wave always gives the same number, now known as Planck’s constant in honor of Max Planck, another of the quantum pioneers. Planck’s constant (usually denoted by the letter h) soon turned out to be one of the most fundamental numbers in physics, ranking alongside the speed of light, c. It cropped up, for example, in the equations developed in the early decades of the twentieth century to describe how electrons are held in orbit around atoms. But although the strange duality of light niggled, the cat was only really set among the pigeons in the 1920s when a French scientist, Louis de Broglie, suggested using the wav
e-particle equation in reverse. Instead of taking a wavelength (for light) and using this to calculate the momentum of an associated particle (the photon), why not take the momentum of a particle (such as an electron) and use it to calculate the length of an associated wave?
Fired by this suggestion, experimenters soon carried out tests that showed that, under the right circumstances, electrons do indeed behave like waves. In the quantum world (the world of the very small, on the scale of atoms and below), particles and waves are simply twin facets of all entities. Waves can behave like particles; particles can behave like waves. A term was even coined to describe these quantum entities—“wavicles.” The dual description of particles as waves and waves as particles turned out to be the key to unlocking the secrets of the quantum world, leading to the development of a satisfactory theory to account for the behavior of atoms, particles, and light. But at the core of that theory lay a deep mystery.
Because all quantum entities have a wave aspect, they cannot be pinned down precisely to a definite location in space. By their very nature, waves are spread-out things. So we cannot be certain where, precisely, an electron is—and uncertainty, it turns out, is an integral feature of the quantum world. The German physicist Werner Heisenberg established in the 1920s that all observable quantities are subject, on the quantum scale, to random variations in their size, with the magnitude of these variations determined by Planck’s constant. This is Heisenberg’s famous “uncertainty principle.” It means that we can never make a precise determination of all the properties of an object like an electron: all we can do is assign probabilities, determined in a very accurate way from the equations of quantum mechanics, to the likelihood that, for example, the electron is in a certain place at a certain time.