1 A Misleading Problem: The Measurement One

The measurement problem is well known: at the end of an experiment where one measures the property of a particle that can take two values, the wave function (or quantum state) of the measuring device, or of the cat if we couple the device to the cat through a poison capsule, is (leaving aside normalization factors):

$$\begin{aligned} \Psi _{\text{ cat } \text{ alive }}+\Psi _{\text{ cat } \text{ dead }}. \nonumber \end{aligned}$$

And that cannot be a complete description of the cat, which is obviously either alive or dead but not both!

The way out of this problem from the point of view of ordinary quantum mechanics is to introduce the collapse postulate: when one looks at the cat, one sees whether she is alive or dead and, depending on what one sees, one reduces the wave function of the cat (and of the particle that was measured and is thus coupled to the state of the cat) to either \(\Psi _{\text{ cat } \text{ alive }}\) or \(\Psi _{\text{ cat } \text{ dead }}\).

Since this is a deus ex machina from the point of view of the linear Schrödinger evolution, justifying it is often viewed as the main problem in foundations of quantum mechanics.

However there is a deeper problem: neither \(\Psi _{\text{ cat } \text{ alive }}\) nor \(\Psi _{\text{ cat } \text{ dead }}\) are cats: they are functions defined on a high dimensional space \(\mathbb R^{3N}\) while cats are located in \(\mathbb R^3\). And it is not clear what it means to say that \(\Psi _{\text{ cat } \text{ alive }}\) or \(\Psi _{\text{ cat } \text{ dead }}\) are descriptions of cats, let alone “complete descriptions” of them.Footnote 1

What most people do is to mentally identify cats and wave functions of cats, which is illegitimate. So, the real problem is the meaning or the ontology one: what does the wave function mean outside of laboratories, and what does it say about what the world is made of?

2 An Intuitive Solution That Does Not Work

A simple and a priori attractive solution to this problem is the naïve statistical interpretation of quantum mechanics (and that is probably what is in the back of the minds of most of the “no worry about quantum mechanics” physicists): particles do have properties such as position, velocity, spin etc., but we cannot know or control them-we have only access to their wave function.

That object gives the statistical distribution of the values of those quantities (through the Born rule) over sets of particles having the same wave function. And, when we perform a measurement of a property of a given particle, we learn what that value is for that particle.

In that interpretation, the reduction of the wave function is no problem; we simply adjust our probabilities when we learn something new about the system.

And, if that worked, there would indeed be no reason to worry about the meaning of the wave function and we would have a decent meaning of that function outside of measurements. However, it cannot work, because of well-established theorems due to Bell [5] and to Kochen and Specker [25], but that unfortunately are not widely known among physicists.

Those theorems show that, if we assume that there exists a map v that assigns a value to each observable A corresponding to various properties that are simultaneously measurable according to ordinary quantum mechanics and that agrees with minimal quantum mechanical predictions concerning such observables, then one can deduce a contradiction.Footnote 2

These theorems are called the “no hidden variables theorems”.

Obviously, there cannot be a statistical distribution of maps that do not exist. In other words, what we call the statistical interpretation (individual quantum systems do have definite properties but we are only able to know their statistical distributions) does not work.

Note in passing that the ideas behind these no hidden variables theorems can be used to rule out the “decoherent histories” approach of Gell-Mann, Hartle, Griffiths and Omnès [17, 24, 31].

Indeed, as noticed by Goldstein [21] and Bassi and Ghirardi [4], this approach amounts to assigning simultaneous values to pairs of commuting observables; but if a series of such pairs is suitably chosen, a contradiction follows.

So, let us look for non obvious solutions.

3 Spontaneous Collapse Theories

These theories are modified versions of quantum mechanics, the first of which was introduced by Ghirardi, Rimini, and Weber [18] in which wave functions spontaneously collapse.Footnote 3

To be precise, in that model, the wave function evolves according to the Schrödinger equation most of the time, but there is a set of spacetime points \(({\textbf{y}}_i, t_i)\) chosen at random, such that the wave function \(\Psi ({\textbf{x}}_1, \dots , {\textbf{x}}_N, t) \) for a system of N particles is multiplied at the chosen times \(t_i\) by a Gaussian function in the variable \({\textbf{x}}_k\) (k chosen uniformly among \(1, \dots , N\)), centered in space at the chosen space points \({\textbf{y}}_i\).

The probability distribution of these random points is determined by the wave function of the system under consideration at the times when they occur, and is given by the familiar \(|\Psi |^2\) distribution. This ensures that the predictions of the GRW theory will (almost) coincide with the usual ones.

The above-mentioned multiplication factors localize the wave function in space, and, for a system of many particles in a superposed state, effectively collapse the wave function onto one of the terms. Now, the trick is to choose the parameters of the theory so that spontaneous collapses are rare enough for a single or for a few particles in order to ensure that they do not lead to detectable deviations from the quantum predictions, but are frequent enough to ensure that a system composed of a large number of particles, say \(N=10^{23}\), will not stay in a superposed wave function for more than a split second.

Spontaneous collapse theories are not the same as ordinary quantum mechanics, since they lead to predictions that differ from the usual ones, even for systems made of a small number of particles. But the parameters of the theory are simply adjusted so as to avoid being refuted by present experiments, which is not exactly an appealing move.

Moreover there is the problem of making sense of a pure wave function ontology (even when the latter collapses, since the collapsed wave function is still just a function defined on a high-dimensional space).

Two solutions have been proposed to give a meaning to the GRW theory beyond the pure wave function ontology: the matter density ontology often denoted GRWm [2, 19], and the flash ontology denoted GRWf [6].

The matter density ontology associates a continuous matter density to the wave function of a system of N particles. For each \({\textbf{x}}\in \mathbb R^3\), and \(t\in \mathbb R\), one defines:

$$\begin{aligned} m({\textbf{x}}, t)= \sum _{i=1}^N m_i \int _{\mathbb R^{3N}} \delta ({\textbf{x}}-{\textbf{x}}_i) |\Psi ({\textbf{x}}_1, \dots , {\textbf{x}}_N, t)|^2 d {\textbf{x}}_1 \dots d{\textbf{x}}_N\;, \end{aligned}$$
(1)

where \(\Psi ({\textbf{x}}_1, \dots , {\textbf{x}}_N, t)\) is the usual wave function of the system at time t. This equation makes a connection between the wave function defined on the high-dimensional configuration space and an object, the matter density, existing in our familiar space \( \mathbb R^3\).

In our three-dimensional world, there is just a continuous density of mass: no structure, no atoms, no molecules, etc., just an amount of “stuff”, with high density in some places and low density elsewhere.

In the flash ontology, one has a world made only of spacetime points at the center of the Gaussian multipliers of the wave function that collapse it. No particles, no fields, nothing at all, except a “galaxy” of spacetime points, called “flashes”.

Let us note that, if the God of the physicists was trying not to be malicious and if either the matter density or the flash ontologies are true, then He failed badly: indeed, it means that we were wrong all along when we “discovered” atoms, nuclei, electrons, etc., and that we are lying to schoolchildren when we tell them that matter is mostly void with a few pieces of matter (the atoms) here and there. Indeed, in the matter density ontology, matter is continuous after all, with higher and lower density in some places, and we have simply been fooled by this modified version of quantum mechanics into thinking that it is not.

On the other hand, if the flash ontology is true, then we have been fooled into thinking that there exists something most of the time (like atoms): if we take the visible universe since the Big Bang, it has contained only finitely many flashes. Since the flashes are all there is in that ontology, this means that, most of the time, the universe is just empty.

Of course, even if those ontologies are weird, we might be forced to accept one of them (after all the existence of atoms in empty space is also counterintuitive), if there were independent reasons for doing so, like a greater explanatory power or greater empirical adequacy.

But, and this is the most important point, empirical adequacy of any spontaneous collapse theory would mean that ordinary quantum mechanics is empirically wrong, since the predictions of both theories differ, at least in principle. So, if one found that a prediction of a spontaneous collapse theory is right, when it differs from ordinary quantum mechanics, it would be a major revolution in physics and we might be forced to worry about those weird ontologies. But this hasn’t happened yet and I would suggest not to worry about what to do after the revolution before that revolution has occurred.

To illustrate how odd the continuous matter and the flash ontologies are, compared to the particle ontology of the de Broglie–Bohm theory (to be discussed in Sect. 5), consider what happens, in those ontologies, with the thought experiment of Einstein’s boxes: imagine a box containing just one particle that is cut in two parts; one part is sent to New York, the other to Tokyo [29], [26, Chap. 10].

Fig. 1
A schematic rectangular, and 2 square-shaped Einstein's boxes, which are labeled B, B 1, and B 2.

Reproduced with permission from Norsen [29]. Copyright 2005 American Association of Physics Teachers

Einstein’s boxes.

The wave function of the particle is a superposition of a wave function located in box \(B_1\) \(+\) a wave function located in box \(B_2\). In ordinary quantum mechanics, when Alice in New York opens her half-box and sees the particle, the wave function collapses on the part of the wave function located in New York and, if she doesn’t see it, the wave function collapses on the part located in Tokyo (Fig. 1).

In the de Broglie–Bohm theory, nothing surprising happens: the particle is in one of the half-boxes all along and is found where it is.

But in the GRWm theory, there is one-half of the matter density of a single particle in each half-box. When Alice couples her half-box with a detector of particles, the evolution of the wave function of the particle is coupled with a macroscopic object and many collapses occur quite rapidly, so that the matter density suddenly jumps from being one-half of the matter density of a single particle in each half-box to being the full matter density of a particle in that half-box and nothing in the other.

There is a nonlocal transfer of matter in the GRWm theory, while there is no such thing in the de Broglie–Bohm theory, and not even anything nonlocal when one deals with only one particle.

In the GRWf theory, there is simply nothing in either half-box, just a wave function traveling so to speak with the half-boxes. When Alice couples her half-box with a detector of particles, the wave function of the particle becomes coupled with a macroscopic object, and there is suddenly a “galaxy of flashes” appearing (randomly) in that detector, either detecting the particle or not; we interpret the first possibility as meaning that the particle is in that half-box and the second one as meaning that the particle is in the other half-box.

Putting aside the weirdness of the ontologies, the spontaneous collapse theories are more nonlocal than the EPR-Bell result implies, since the latter concerns nonlocality for systems with at least two particles.

Here we have nonlocal effects or actions at a distance (Alice affects the physical situation in Tokyo by acting in New York) even for one particle!

In summary:

  • Spontaneous collapse theories have unnatural ontologies.

  • They are very ad hoc: parameters are chosen so as to avoid refutation and not on the basis of any evidence.Footnote 4

  • They can only be true if quantum mechanics itself is false.

  • They are more nonlocal than they have to be.

4 The Many-Worlds Interpretation

This interpretation postulates that, when the proverbial cat (or any other macroscopic device) finds itself in a superposed state, then, instead of undergoing a collapse by fiat as in ordinary quantum mechanics, both terms simply continue to exist. But how can that be possible? We always see the cat alive or dead but not both! The short answer is that they both exist, but in different “worlds”.

Hence, whenever an experiment leads to a macroscopic superposition, the world splits into two or more worlds, depending on the number of distinct macroscopic states produced by that experiment, one for each possible result.

Why do I always perceive only one of the results? It is simple: I, meaning my body, my brain (and thus also my consciousness) becomes entangled with the states of the cat, so there are two or more copies of me also, one seeing the dead cat in one world, another seeing the live cat in another world. And that, of course, is also true for everything else: every molecule in the entire world becomes copied twice (maybe not instantaneously, but that is another question).

In his original paper [16], Everett stressed that “all elements of a superposition are ‘actual’, none any more ‘real’ than the rest.” Everett felt obliged to write this because “some correspondents” had written to him saying that, since we experience only one element of a superposition, we have only to assume the existence of that unique element. This shows that some early readers of Everett were already baffled by the radical nature of the “many-worlds” proposal.

Putting aside the weirdness of this multiplication of “worlds”, one should ask whether the many worlds scheme is coherent. Consider the Born rule. Suppose that the probabilities of having the cat alive or dead, as a result of an experiment, are \((\frac{1}{2}, \frac{1}{2})\). And suppose that I decide to repeat the same experiment successively many times, with different particles (and cats) but all having the same initial wave function.

After one experiment, there are two worlds, one with a dead cat and a copy of me seeing a dead cat and one with a live cat and a copy of me seeing a live cat. Since both copies of me are in the same state of mind as I was before the first experiment (after all, both copies are just copies of me!), each of them repeats that experiment.

Then, we have four worlds, one with two consecutive dead cats, one with two consecutive live cats and two with one dead cat and one live cat. “I” (by that I mean each copy of me in each of those four worlds) repeat the experiment again: we have now eight worlds, with one “history” of worlds with three dead cats, one history of worlds with three live cats, three histories of worlds with one live cats and two dead ones, three histories of worlds with one dead cat and two live ones.

Now, continue repeating that experiment: for every possible sequence of outcomes, there will be some of my “descendants” (i.e. copies of me, that exist in all the future worlds) that will see it. There will be a sequence of worlds in which the cats are always alive and another sequence where they are always dead. There are also many sequences of worlds where the cats are alive one quarter of the time and dead three quarters of the time, and that is true for any other statistics different from \((\frac{1}{2}, \frac{1}{2})\). So that we can be certain that many of our descendants will not observe Born’s rule in their worlds.

But one could argue, on the basis of the law of large numbers, that, at least in the vast majority of worlds, the Born rule will be obeyed, since, in the vast majority of worlds, the frequencies of dead and live cats will be close to \((\frac{1}{2}, \frac{1}{2})\).

But what happens if, instead of being \((\frac{1}{2}, \frac{1}{2})\), the probabilities predicted by quantum mechanics are, say, \((\frac{3}{4}, \frac{1}{4})\)? We will still have two worlds coming out of each experiment, because these experiments have two possible outcomes. So, the structure of the multiplication of worlds is exactly the same as when the predicted probabilities were \((\frac{1}{2}, \frac{1}{2})\).

But now, if one applies the law of large numbers as above, one arrives at the conclusion that, in the vast majority of worlds, the quantum predictions will not be observed, since our descendants will still see the cats alive in approximately \(\frac{1}{2}\) of the worlds, and the cats dead also in approximately \(\frac{1}{2}\) of the worlds, instead of the \((\frac{3}{4}, \frac{1}{4})\) frequencies predicted by the Born rule.

This is a serious problem for the many-worlds interpretation. There have been many proposals to solve this problem and it would be too long and too technical to discuss all of them here.

Some authors have argued that one should count the worlds differently, by weighting them with the coefficients that appear in the Born rule [13, 16].

However, this does not answer the objection above.

Another “solution” is to give to low probability worlds (according to Born’s rule) a lower degree of existence or of reality, but it is unclear what it means to live in a low reality world since we cannot compare that life with one in a world with a high degree of existence: indeed, different worlds don’t interact with each other.

And this “solution” is quite contrary to Everett’s original idea that “all elements of a superposition are ‘actual’, none any more ‘real’ than the rest”.

Moreover, one also has the problem of ontology: what proliferates are wave functions, but the latter are mathematical objects not “worlds” in space-time.

One solution is to associate to the wave function the continuous matter density (1), as in the spontaneous collapse theories, see [3].

However, this does not solve the problem of the probabilities discussed above. Coming back to our example with two possible outcomes, one having probability \(\frac{3}{4}\) and the other \(\frac{1}{4}\), the density of matter will be different in the world where one sees the outcome having probability \(\frac{3}{4}\) from the one where one sees the outcome having probability \(\frac{1}{4}\).

But what difference does it make? In which way does having a smaller or larger matter density affect my states of mind ? And if it does not, we are back to the problem that, if one repeats many times the experiment whose outcomes have probabilities \((\frac{3}{4}, \frac{1}{4})\), most of my descendants (some of course having a small matter density) will see massive violations of the Born rule.

5 The de Broglie–Bohm Theory

Nature and Nature’s Laws lay hid in Copenhagen: God said, “Let de Broglie–Bohm be!” and all was light.Footnote 5

In the de Broglie–Bohm theory, the complete state of a system with N variables at time t is specified by \(\big (\Psi (t), \textbf{X}(t)\big )\), where \(\Psi (t)\) is the usual wave function, \(\Psi (t)=\Psi ({\textbf{x}}_1, \dots , {\textbf{x}}_N, t)\) and \({\textbf{X}(t)}= \big ({\textbf{X}}_1(t),\ldots , {\textbf{X}}_N (t)\big )\in \mathbb R^{3N}\) are the actual positions of the particles.Footnote 6

The theory assumes that the particles have positions at all times, and therefore trajectories, independently of whether one measures them or not.

The evolution of the state (\( \Psi , \textbf{X}\)) is given by two laws:

  1. 1.

    \(\Psi \) obeys the usual Schrödinger equation at all times. The wave function of an isolated system never collapses.

  2. 2.

    The evolution of the positions of the particles is guided by the wave function at time t.

Moreover, since the system is deterministic, one has to make statistical assumptions on the initial conditions of the system (as one does in chaos theory) in order to obtain statistical predictions.

These are the fairly natural “quantum equilibrium” ones: \(\rho =|\psi |^2\).

Let me list the main qualities of this theory:

  • Under the quantum equilibrium assumptions on initial conditions, one recovers the usual predictions of quantum mechanics.

  • The ontology of the theory is the same as in classical physics and is thus unproblematic, unlike the ontologies of the spontaneous collapse and many worlds theories.Footnote 7 It can be summarized by Detlef’s quote mentioned at the beginning of this article: “particles move”. And while in classical physics, that motion is guided by gravitational or electromagnetic fields, here it is guided by the wave function, a more abstract notion but a perfectly well-defined one.

  • The de Broglie–Bohm theory gives a clear physical meaning to the wave function, which is no longer simply a “probability wave” (whatever that means exactly) but a physical quantity determining the motion of particles, similar in some ways to classical Hamiltonians.

  • This theory explains the “contextuality” of measurements.

    Consider an idealized spin measurement: if the wave function has a symmetry along the z axis and if the particle starts above the line of symmetry \(z=0\), it will be deflected upwards, meaning that its spin is “up”, see Fig. 2. But if we reverse the orientation of the gradient of the magnetic field and do the same “measurement” with the same initial wave function and the same initial position of the particle, the particle will still be deflected upwards, but that means now that its spin is “down”, see Fig. 3.

    So, “measurements” don’t measure intrinsic properties of the system (except for position measurements).

    Measurements are interactions between an apparatus and a system.

    This was stated intuitively by Bohr, but here it follows from the equations of the theory.

  • The de Broglie–Bohm theory is a statistical theory, but, unlike the naïve one mentioned in the Sect. 2, it is consistent and is not refuted by the no hidden variables theorems of Bell and Kochen and Specker, because it does not introduce the “hidden variables” that are forbidden by these theorems, like spin values preexisting to their “measurement”.

    A subtle but crucial point: the de Broglie–Bohm theory is a hidden variable theory that is not refuted by the no hidden variables theorems.

    Often missed (to put it mildly)!

  • One can use this contextuality of measurements to illustrate how nonlocality works in the de Broglie–Bohm theory.

    If the result of a “measurement” on one side of an EPR-Bell experiment with an entangled pair of particles depends on how the orientation of the gradient of the field is oriented on that side, then changing that orientation will affect the behavior of the particle on the other side: since the spins of both particles have to be anti-correlated, if changing that orientation on one side changes the “value” of the spin on that side from up to down, then the spin on the other side must go from down to up. But that means that the trajectory followed by the particle on the other side must also change.

    This makes nonlocality explicit in the de Broglie–Bohm theory, which is a (big) quality rather than a defect since Bell has shown that nonlocality is a property of the world and not only of the quantum theory.

Fig. 2
Spin measurement of Z over X. The spin starts at 0 to the initial position of the electron is Psi of Z of modulus of 1 up arrow greater than + 1 down arrow greater than, the direction of the gradient of the field Psi of z t 1 up arrow greater than, and the direction opposite to the gradient of the field Psi of Z + t 1 down arrow greater than.

An idealized spin measurement

Fig. 3
Spin measurement of Z over X. It begins with the initial position of the electron Psi of Z of modulus of 1 up arrow greater than + 1 down arrow greater than, direction opposite to the field Psi of Z t 1 down arrow greater than, and direction of the field Psi of Z + t 1 up arrow greater than.

An idealized spin measurement with the orientation of the gradient of the magnetic field reversed with respect to the one in Fig. 2

6 Conclusions

While we have left out certain questions such as quantum field theory and relativity (that can be dealt with within de Broglie–Bohm theory but it would be too long to discuss that), we want to stress certain aspects of that theory:

The de Broglie–Bohm theory is not a different theory from ordinary quantum mechanics. It is the rational completion of ordinary quantum mechanics. The latter is just the algorithm allowing us to predict “results of measurements” and that algorithm can be derived from the de Broglie–Bohm theory.

Most physicists either don’t care about the meaning of their most fundamental theory (“shut up and calculate”) or adhere (in the back of their mind) to the naïve statistical interpretation.

Most physicists won’t be persuaded by spontaneous collapse theories (too ad hoc) unless some future experiments contradict ordinary quantum predictions.

If that happens, they are likely to look for an entirely different theory (nonlinear?)

An important minority of physicists “like” the Many-Worlds interpretation.

But I believe that this is because they haven’t thought it through.

Apart from its fantastic nature, one has to provide it with an ontology and solve the problem of the statistical predictions (unsolved since 1957).

So, that leaves us with the de Broglie–Bohm theory as the only option. I believe it is gaining popularity, due in part to the work of Detlef Dürr, although it is still very marginal.

For it to become more popular, one needs:

  • That physicists start to worry about the meaning of their most fundamental theory.

  • That they be better aware of the nature of the problem: not the measurement one, but the meaning or ontology. And that they also become aware of the no hidden variable theorems that refute their naïve statistical interpretation.

But that is a long way to go!