1 Introduction and statement of the results

Consider the Schrödinger-Newton equation

$$\begin{aligned} i \hbar \frac{\partial \varPsi (t,x)}{\partial t} = \left( -\frac{\hbar ^2}{2m} \varDelta - Gm^2 \int _{\mathbb {R}^3} \frac{|\varPsi (t,y)|^2}{|x-y|} \,dy \right) \varPsi (t,x), \end{aligned}$$
(1)

where \(\hbar \) is Planck’s constant; G is Newton’s constant; m is a real-valued parameter representing the mass; \(i^2 = -1\); \(\varDelta \) is the Laplacian in \(\mathbb {R}^3\); and \(\varPsi \) is the wave-function, a complex-valued function in \(\mathbb {R}^4\) of unit norm in \(L^2(\mathbb {R}^3)\). Here, \(\mathbb {R}^4\) is physically thought of as time and space, \(\mathbb {R}^4 = \mathbb {R}\times \mathbb {R}^3\), with coordinates denoted by (tx), where t is time and x the spatial position. For background on Eq. (1), see Sect. 1.1 below. Here, let us simply point out that Eq. (1) is based on some rather intriguing hypotheses about the coupling of gravity with matter, leading to a description of certain non-relativistic quantum-mechanical systems that differs from that provided by the ordinary Schrödinger equation. Thus, assessing the correctness of Eq. (1) can potentially provide a phenomenological window into the elusive subject of the behavior of gravity in the presence of quantum effects.

From a quick heuristic analysis of Eq. (1), it is seen that when \(m \gg 1\), the dynamics will be dominated by gravity, i.e., by the term proportional to G, and by the diffusion, i.e., the term involving \(\varDelta \), when \(m \ll 1\). Thus, it is natural to expect that there exists a value of m, henceforth called the critical mass and denoted \(m_c\), for which the gravitational and the diffusive terms are “balanced”. Previous works refer to \(m_c\) as a mass value above which the package “collapses” [54], or a value below which the dispersion is “gravitationally inhibited” [26]. Experimentally observing such a collapse would be tantamount to verifying a prediction based on the Schrödinger-Newton equation, hence suggesting the validity of the underlying assumptions that lead to Eq. (1) (some of such assumptions are discussed in Sect. 1.1 below). It becomes, therefore, important to estimate \(m_c\) quantitatively, particularly if such an estimate yields values of mass accessible to current technology.

Naturally, an initial condition \(\varPsi (0,x)\) has to be provided in order to solve Eq. (1), and one expects that \(m_c\) will depend upon \(\varPsi (0,x)\). When the initial condition is a (radially symmetric) Gaussian of width \(\sigma \), the value of \(m_c\) has been investigated initially by Carlip and Salzman [50], and subsequently in more detail by Giulini and Großardt [26] and Meter [54]. They found that \(m_c\) is of the order \(\left( \frac{\hbar ^2}{G \sigma } \right) ^\frac{1}{3} \equiv m_r\). With this estimate at hand, we can in principle test Eq. (1) experimentally by evolving a quantum-mechanical system of mass \(m \gtrsim m_c\) whose wave function at \(t=0\) is as just described. If, under these conditions, the wave-function of the system remains confined in a region of space rather than spreading out as predicted by standard quantum mechanics (or if the spreading is gnificantly slower than that predicted by the Schrödinger equation for a free particle), then we would have found evidence favoring a description based on the Schrödinger-Newton equation. Actually implementing this procedure, however, is technically challenging in that one has to construct, in a lab, a physical system that is described in terms of a wave-function of width \(\sigma \), having a mass of the order \(m_c\), and for which, at the same time, quantum-mechanical effects can be precisely measured. To understand where the difficulty lies, one simply has to point out that systems whose quantum mechanical nature is manifest in lab experiments tend to have very tiny masses and wave-packages of small width, a situation that precludes \( m \gtrsim \left( \frac{\hbar ^2}{G \sigma } \right) ^\frac{1}{3}\). This is the case even when one considers the so-called “heavy” systems for which quantum effects can still be experimentally observed. For instance, current day molecular interferometry technology allows experimentalists to build systems corresponding to wave-packages of size \(\sigma \sim 0.5 \mu m\) and masses as large as \(m \sim 10^3 u\) [3, 31]. This is still, however, several orders of magnitude below the corresponding \(m_c\) for \(\sigma \sim 0.5 \mu m\), namely, \(m_c \sim 10^9 u\).

In light of the above, an obvious question is whether it is possible to find initial conditions that produce smaller values for \(m_c\). In other words, dependence of \(m_c\) on \(\varPsi (0,x)\) raises the possibility that there exist initial conditions that lead to values \(m_c\) closer to what can be achieved in actual experiments of quantum mechanical systems. This is the problem addressed in this manuscript.

We take as \(\varPsi (0,x)\) a superposition of a Gaussian peaked at the origin and a Gaussian-like function peaked away from the origin. By Gaussian-like, we mean a radially symmetric function that is in fact a Gaussian centered at some \(R>0\) with respect to the radial variable r, i.e., a function of the form \(\sim e^{-\frac{(r-R)^2}{2\sigma ^2}}\). Obviously, such a function is not a Gaussian in \(\mathbb {R}^3\) (unless \(R=0\)), but it retains the Gaussian features of being radially symmetric and peaked at R. While a superposition of two actual Gaussian would also sound natural, this would break radial symmetry when one of them is not centered at the origin, making the analytic and numerical estimates more difficult. Figure 1 illustrates our initial conditions. We notice that simply taking \(e^{-\frac{(r-R)^2}{2\sigma ^2}}\) rather than a superposition will not have the desired effect of lowering \(m_c\) (see Sect. 1).

Fig. 1
figure 1

Illustration of a typical initial condition

Main result. Putting \(\psi _R = e^{-\frac{(r-R)^2}{2\sigma ^2}}\), we write explicitly \(\varPsi (0,x) = A (\psi _0 + \psi _R)\), where A is a normalization constant and \(\psi _0 = e^{-\frac{r^2}{2\sigma ^2}}\). We investigate how \(m_c\) varies with the separation R between the two peaks. The results are summarized in Fig. 5, where \(m_c\) is plotted, in units of \(m_r\), as a function of R. The two curves in Fig. 5 represent the analytic and numerical estimates. Both agree qualitatively, with the quantitative discrepancy probably stemming from some of the rough approximations we have employed in the analytic calculations as discussed below. Our results show that \(m_c\) decreases in comparison to \(m_r\) for certain choices of R. While this decrease is barely by an order of magnitude, thus still far from what is needed to reach values of masses achievable in lab experiments in the foreseeable future, the main message here is that it is possible to obtain values of \(m_c\) less than \( m_r \equiv \left( \frac{\hbar ^2}{G \sigma } \right) ^\frac{1}{3}\) upon suitable choices of initial conditions. This raises the interesting possibility that more complex initial conditions could in fact produce a significant decrease in the value of \(m_c\).

Figure 5 also shows that the critical mass of our system approaches \(m_r\) when \(R \gg 1\). This behavior is discussed in detail in the “Appendix 4”. In a nutshell, when R becomes very large, the system “decouples” into two subsystems associated with \(\psi _0\) and \(\psi _R\). For \(\psi _R\), a very large value of m is necessary to prevent the system from spreading, thus, \(\psi _0\) is solely responsible for the confining behavior when \(m=m_c\). Therefore, noticing that, by construction, \(m_c = m_r\) when \(R=0\), it can be concluded that the observed lowering of the critical mass in relation to \(m_r\) is a consequence of the interaction of the parts of \(\varPsi (t,x)\) associated with \(\psi _0\) and \(\psi _R\). In physical terms, this can be understood as follows. For \(R \gg 1\), the gravitational interaction between the packages concentrated near zero and near R is negligible, and thus confinement can only happen due to the self-interaction of each of these packages. However, when the two packages are near each other, not only does their self-gravitational interaction work against the diffusive term, but their mutual interaction also contributes to the confinement. This raises the obvious question of what happens when several (i.e., more than two) packages are employed, but this is not addressed in this work.

Notice that the question of whether wave functions whose associated probability density resembles some given intricate initial condition (or even our \(\varPsi (0,x)\)) can be realized in laboratories is left untouched here. If it turns that such constructions seem out of reach, this would definitely diminish the appeal for the investigations we have carried out. On the other hand, we should not underestimate the ability of experimentalists in producing actual far-reaching quantum states of matter, as many recent advances have demonstrated [3, 8].

A second question addressed in this paper is that of the robustness of earlier numerical results. Although the estimate \(m_c \sim \left( \frac{\hbar ^2}{G \sigma } \right) ^\frac{1}{3}\) for a single wave-packet as initial condition has been independently verified by Giulini and Großardt [26] and Meter [54], and their results agree with standard analytic estimates, leaving little doubt about the correctness of their results, we believe that a third and more thorough verification is beneficial. This seems appropriate especially in light of finite-size effects that inevitably plague numerical simulations. Such effects have not been reported in Giulini and Großardt, and have been dealt with by Meter via a clever, albeit artificial, choice of boundary conditions. In this regard, it is perhaps worthwhile to remember that the first reported numerical results concerning \(m_c\) were very crude [50], and it took about five years after [50] for a coherent and more reliable picture to emerge. Such caution is particularly important when one takes into account the non-linear and non-local nature of Eq. (1). Reassuringly, although by no means surprisingly, our results confirm those of [26, 54].

We finish this section with an important conceptual consideration. We find that the regime where the gravitational interaction becomes relevant involves large values of mass, as indicated in Fig. 5. Such values of mass can only occur in large systems, such as the type of heavy molecules previously referred to, requiring a description of multi-particle systems. On the other hand, we are considering the dynamics of a single mass distribution self-interacting via Eq. (1). In order to make these two ideas compatible, one has to work with systems where the multi-particle dynamics can be reduced to that of its center-of-mass, with the latter described by Eq. (1). The dynamics of the center-of-mass has been analyzed by Giulini and Großardt in [28]. They found that if the extent of an object is small in comparison to the uncertainty in localisation of its centre of mass, then Eq. (1) provides a good approximation for the dynamics of the multi-particle system, and our results should be understood in this context. See also the discussion in [1].

1.1 Background

In this section we describe some of the reasoning leading to Eq. (1), and specialists may want to skip this section.

As it is well-known, reconciling Einstein’s General Theory of Relativity and Quantum Mechanics is one of the main open problems in contemporary Theoretical Physics. Whether this has to be achieved by finding a suitable theory of quantum gravity or by modifying Quantum Mechanics in a way that would allow it to be consistently coupled to classical General Relativity (or some other, more general, theory of gravity), is still an open question, although many plausible arguments point in the direction of the former alternative, which also seems to be the consensus among the majority of the physics community. See, e.g., [10, 52, 55, 56] for a general discussion.

If one assumes that gravity is quantized, then the outstanding success of General Relativity implies that quantum aspects pertaining to the gravitational field are negligible in most phenomena presently accessible to observation. However, it is expected that quantum gravity becomes important in the presence of extremely strong gravitational fields, such as those believed to exist in the interior of black hole event horizons and in the early universe. Unfortunately, despite much progress over the decades,Footnote 1 we still lack a robust theory of quantum gravity that could be used to make indirect, yet sufficiently accurate, predictions to match with observations. And due to the energy scales involved, direct measurements of quantum gravitational phenomena seem to be out of the question for generations to come, if not forever.

Given such a state of affairs, it seems natural to use some sort of semi-classical approximation to try to understand general features of how gravity interacts with matter. One hopes to be able to construct semi-classical models that retain enough features of the underlying full-fledged quantum description as a way of extracting information that can help us to constrain current theories of quantum gravity. For instance, despite being derived in a semi-classical formalism, the Hawking radiation [32] is an important guide to quantum theories of gravity, so that the ability to reproduce it becomes one of the first tests to any theory that attempts to quantize the gravitational field.

In such semi-classical approaches, one describes the gravitational field classically, but imposes quantum behavior on the matter present in the model in question. Specifying how matter interacts gravitationally is one of the main aspects of the model, and different choices are available. The most common choices rely on Quantum Field Theory on Curved Backgrounds, where one performs field quantization in a space-time background that satisfies Einstein’s equations. The matter-gravity interaction is described via a minimal coupling, with backreaction effects on the metric tensor generally (but not always) ignored. See the monographs [7, 23, 39, 55] for an in-depth exposition, or the introduction in [17] for a short discussion. While this approach is very natural, it is technically challenging, thus one may ask if there are further simplifications that can lead to more tractable models.

One further simplification would be to consider non-relativistic models. As the non-relativistic limits of Quantum Field Theory and General Relativity are, respectively, ordinary Quantum Mechanics and Newtonian gravity, one possibility is to consider quantum matter that interacts gravitationally via a Newtonian gravitational potential. This is the situation described by the Schrödinger–Newton equation. Its physical content is as follows. Consider a spatially extended mass distribution that self-interacts gravitationally according to Newton’s law of gravity. Denoting the total mass by m and the mass density function by \(\rho _m\), the gravitational potential \(\varPhi \) satisfies

$$\begin{aligned} \varDelta \varPhi = 4\pi G \rho _m, \end{aligned}$$

where G is Newton’s constant. Suppose that the matter in the above equation is described quantum mechanically, in the sense that the time-evolution determines only the probability of finding an amount \(\delta m\) of the mass within a volume \(\delta V\) in space, and that such probability is given by the ordinary Schrödinger equation. In this case, the probability density, \(\varrho \), is given by

$$\begin{aligned} \varrho = |\varPsi |^2, \end{aligned}$$

where the probability density and the mass density are linked through mass by \(\varrho \, m = \rho _m\) and \(\varPsi \) is a wave-function satisfying Schrödinger’s equation with a potential given by \(\varPhi \),

$$\begin{aligned} i \hbar \frac{\partial \varPsi }{\partial t} = \left( - \frac{\hbar ^2}{2m} \varDelta + m\varPhi \right) \varPsi . \end{aligned}$$

Combining the above, we obtain the Schrödinger–Newton equation Eq. (1).

Several remarks are in place. In considering the above reasoning, a specific assumption about the behavior of gravity coupled to matter in the quantum-mechanical realm is made. It consists of a modification of Schrödinger’s equation, allowing for self-interaction inasmuch a massive particle is present, regardless of the existence of an external potential. In particular, the concept of free particles no longer exists. Such a modification, it should be noticed, is not excluded by current theoretical and experimental understandings of quantum mechanics because (a) we do not know precisely what quantum gravity is, and, therefore, the details of its possible semi-classical regimes are open to exploration; and (b) consequently, the limits where a semi-classical description is valid are not known either. Furthermore, it should come as no surprise that the Schrödinger equation is replaced by a more accurate description of quantum mechanical behavior once a more fundamental theory of gravity coupled to matter is formulated. The real question is whether this modification looks like Eq. (1) in some appropriate limit. Regarding this question, what makes the Schrödinger–Newton equation interesting is that the scales involved are far more modest than those of full-fledged theories of quantum gravity, even in the face of the aforementioned gap between known values of \(m_c\) and current technology. This is one of the reasons why Eq. (1) has recently attracted considerable attention (see, e.g., [11, 26, 28, 50, 54] and references therein). Recently, electromagnetic interactions have been added to the Schrodinger–Newton equation to include relativistic effects [37].

Although above we have given a motivation to study Eq. (1) on the basis of a semi-classical analysis, it ought to be stressed that the Schrödinger–Newton equation also plays an important role in some approaches that assume that gravity remains fundamentally classical at all levels, and in Penrose’s discussion of the “collapse” of the wave-function [42, 43] (although it should be stressed that the meaning of the term “collapse” here has nothing to do with the infamous “collapse of the wave-function”). We refer the reader to [28, 54] for a more detailed discussion of the theoretical underpinnings of Eq. (1). In passing, we point out that Eq. (1) does follow from certain non-relativistic limits of the Einstein-Klein-Gordon or Einstein-Dirac systems [27].

We finish this section pointing out that the study of equation Eq. (1) has also spurred a great deal of activity in the mathematical community. See, e.g., [6, 12, 14, 15, 18, 24, 25, 3436, 38, 47, 51].

2 Analytic estimate of the critical mass

In this section, we provide analytic estimates for the critical mass. We proceed essentially as in section 3.1 of [26], making the necessary adaptations for our initial condition. Unfortunately, some of the computations are extremely long, involving several page-long expressions (although they are by no means difficult, consisting essentially of derivatives and algebraic manipulations). Hence, we will describe the procedure, omitting some of the explicit expressions. The final expression for \(m_c\) is given in the “Appendix”.

Consider the following wave function for a free particle of mass m (i.e., a solution to the ordinary Schrödinger equation):

$$\begin{aligned} \varPsi _R(t,x) \!= & {} \! \left( \sqrt{ \pi } \left( 1 \!+\! \frac{i \hbar t}{m \sigma ^2} \right) ^\frac{3}{2} \sqrt{ \sigma ( 2 R \sigma e^{-\frac{R^2}{\sigma ^2}}\!+\! \sqrt{ \pi }(2R^2 \!+\! \sigma ^2)( 1 \!+\! {\text {Erf}}(R/\sigma ) ) ) } \right) ^{-1} \nonumber \\&\times e^{-\frac{(r-R)^2}{ 2\sigma ^2 \left( 1 + \frac{i \hbar t}{m \sigma ^2}\right) }}, \end{aligned}$$
(2)

where \({\text {Erf}}\) is the error function and \(r^2 = |x|^2\). Thus, \(\varPsi _R(0,x)\) has the form described in the introduction, \( e^{-\frac{(r-R)^2}{2\sigma ^2}}\) up to a normalization factor. Naturally, \(\varPsi _R\) reduces to the usual free-particle solution with a Gaussian initial condition when \(R=0\). In fact, the calculations below reduce to those in section 3.1 of [26] when \(R=0\).

Define

$$\begin{aligned} \varPsi (t,x) = A~( \varPsi _0(t,x) + \varPsi _R(t,x ) ), \end{aligned}$$

where A is a normalization constant. By the linearity of the Schroödinger equation, \(\varPsi \) is a solution with initial condition \(\varPsi (0,x) \sim e^{-\frac{r^2}{2\sigma ^2}} + e^{-\frac{(r-R)^2}{2\sigma ^2}}\). The radial probability density associated with \(\varPsi (t,x)\) is

$$\begin{aligned} \varrho _{\text {rad}}(t,r) = 4 \pi r^2 |\varPsi (t,x)|^2. \end{aligned}$$

In principle, differentiating \(\varrho _{\text {rad}}\) with respect to r, we can find the critical points of \(\varrho _{\text {rad}}\), and in particular the value of r for which \(\varrho _{\text {rad}}\) has a global maximum, henceforth called \(r_p\) (it can be inspected graphically that such a maximum in fact exists). It is a function of t, i.e., \(r_p = r_p(t)\). It also depends, of course, on the other parameters of the problem, \(\sigma \), R, m, and \(\hbar \), but this dependence will be omitted in order to simplify the notation.

Unfortunately, the expressions involved in this procedure are too complicated and it is unlikely that \(r_p\) can be found explicitly as a function of t, as in the case where \(R=0\). Since the acceleration of the density’s peak, \(\ddot{r}_p\), is needed to estimate \(m_c\), we will proceed as follows (recall that we are following the ideas of [26]).

Because \(r_p\) is a critical point of \(\varrho _{\text {rad}}\), it satisfies

$$\begin{aligned} \frac{\partial \varrho _{\text {rad}} (t,r_p) }{\partial r } = 0, \end{aligned}$$
(3)

where the above means \(\frac{\partial \varrho _{\text {rad}} (t,r) }{\partial r }\) evaluated at \(r_p\), i.e., \(\left. \frac{\partial \varrho _{\text {rad}} (t,r) }{\partial r }\right| _{r = r_p}\). Differentiating Eq. (3) twice in time, and solving for \(\ddot{r}_p\), we obtain \(\ddot{r}_p\) as a function of t, \(\dot{r}_p\) and \(r_p\), which we write symbolically as

$$\begin{aligned} \ddot{r}_p(t) = f( t, \dot{r}_p(t), r_p(t) ). \end{aligned}$$

Evaluating this expression at zero and using that \(\dot{r}_p(0) = 0\), we obtain \(\ddot{r}_p(0)\) as a function of \(r_p(0)\), which again we write symbolically as

$$\begin{aligned} \ddot{r}_p(0) = f( r_p(0) ). \end{aligned}$$

Next, we need \(r_p(0)\). Here, as before, we need to solve Eq. (3) for r, except that now we also set \(t=0\), and, again, we find that the resulting expression is too complicated to allow for an explicit solution. However, for the case \(t=0\), Eq. (3) simplifies considerably, so that \(r_p(0)\) can be found numerically. We remark that Eq. (3), at \(t=0\), has more than one solution, but once the solutions are found it is easy to directly verify which one is the global maximum. Notice that \(r_p(0)\), and therefore \(\ddot{r}_p(0)\), is a function of \(\sigma \), R, m, and \(\hbar \), although this fact is obscured by the numerical procedure necessary to determine \(r_p(0)\).

To find \(m_c\), we set \(\ddot{r}_p(0)\) equal to the Newtonian gravitational attraction given by \(Gm/(r_p(0))^2\),

$$\begin{aligned} \ddot{r}_p(0) = \frac{G m}{ (r_p(0))^2 }. \end{aligned}$$
(4)

As \(\ddot{r}_p(0)\) and \(r_p(0)\) are functions of \(\sigma \), R, m, and \(\hbar \), we can solve for m in terms of the remaining parameters. Doing this for different values of R yields the first graph in Fig. 5.

We remark that this procedure gives an explicit expression for \(m_c\) in terms of \(\sigma \), R, m, \(\hbar \), and \(r_p(0)\), which is given in the “Appendix 3”. Indeed, from the above procedure, we obtain explicitly \(\ddot{r}_p(0)\) in terms of \(r_p(0)\), and thus Eq. (4) produces \(m_c\) as just stated after solving for m. It is only to find \(r_p(0)\) that a numerical solution is needed. In particular, because such explicit expressions are available and a numerical algorithm had to be implemented solely to determine \(r_p(0)\), we still characterize this argument as an “analytic” estimate of \(m_c\). In fact, the procedure yielding \(r_p(0)\) consists of a simple root finding for an algebraic equation. We thus reserve the the term “numerical” for methods that fully deserve this name, as the numerical solutions to Eq. (1) investigated in the next section.

Finally, it should be pointed out that the procedure described in this section involves the crucial approximation of eventually evaluating all the time-dependent quantities at time zero. While this is sensible, since one would expect that if gravity is going to confine \(\varPsi (t,x)\), this should happen at earlier times before the wave-function has spread out considerably, it is an approximation nonetheless. For the case \(R=0\), more accurate analytic tools are available in [26], but it seems that it would be far too complicated to implement them in our case. We believe that this approximation combined with the convoluted form of our initial conditions is responsible for the discrepancy between the analytic and numerical methods observed in Fig. 5.

3 Numerical methods

To study the time behavior of the Schrdinger-Newton equation, we solve equation Eq. (1) numerically using the initial conditions described above and illustrated in Fig. 1. We follow the numerical method used in [26], with the difference that units of mass, time and space are expressed in terms of G, \(\hbar \) and \(\sigma \), as in the reference [54]. The initial value of the width of the wave-package \(\sigma \) is the distance unit, \(t_r= (\sigma ^5/G \hbar )^{1/3}\) is the unit of time, and \(m_r= (\hbar ^2/G \sigma )^{1/3}\) is the unit of mass. The convenience of these units is that they exhibit the scale-invariance of the Schrodinger–Newton equation and reduce the problem to a single parameter, namely, the mass. If one replaces the mass by \(m=m_r ~ \tilde{m}\), \(t=t_r~\tilde{t}\) and \(r=\sigma ~ \tilde{r}\), the Eq. (1) can be written as

$$\begin{aligned} i \frac{\partial \varPsi ({\tilde{t}, \tilde{r}})}{\partial \tilde{t}} = \left( - \frac{1}{2\tilde{m}} \varDelta + \tilde{m}\varPhi ({\tilde{t}, \tilde{r}}) \right) \varPsi (\tilde{t}, {\tilde{r}}), \end{aligned}$$
(5)

with

$$\begin{aligned} \varDelta \varPhi = 4 \pi \tilde{m} |\varPsi ({\tilde{t}, \tilde{r}})|^{2}. \end{aligned}$$
(6)

Hereafter, all results will be given in terms of the reduced units. The details necessary to reproduce our numerical results are explained in “Appendix 1”. Here, we will discuss only two important aspects of the numerical solutions: the boundary and initial conditions

3.1 Boundary conditions

The system under study lives in an infinite space. However, the simulation box has to be finite and the outer boundary should be far enough from the origin in order to avoid reflection of the wave packet at the boundary. To solve this problem, Meter [54] uses Neumann boundary conditions in his simulations. This means that the part of the packet that touches the outer boundary is eliminated from the simulation domain. Here, we keep the system size big enough to avoid such reflections. In practice, a big system of size L is chosen and the simulation is run during an interval of time such that the wave packet calculated at the boundary is negligible. Explicitly:

$$\begin{aligned} \varrho _{\text {rad}}(\tilde{r}=L) \le 10^{-8}. \end{aligned}$$

3.2 Initial conditions

We use two different shapes for the wave packet as initial conditions. The first one initializes the system with a spherically symmetric Gaussian packet:

$$\begin{aligned} \varPsi _0( \tilde{t}=0, \tilde{r}) = ( \pi )^{-3/4} ~exp \left( -\frac{\tilde{r}^2}{2} \right) . \end{aligned}$$
(7)

This condition was introduced in previous works, [26, 50, 54], and here it is used to compare our simulations with previous numerical results. In references [26, 54], a difference of about 20 % in the critical mass was found. We will comment on this in “Appendix 2”.

The second initial condition is a combination of two Gaussians in the radial variable, as illustrated in Fig. 1:

$$\begin{aligned} \begin{aligned} \varPsi ( \tilde{t}=0, \tilde{r})= & {} \varPsi _0( \tilde{r})+\varPsi _R( \tilde{r})\;, \quad \text{ with } \\ \varPsi _0( \tilde{r})= & {} A_1 ~exp{\left( -\frac{\tilde{r}^2}{2}\right) \;, } \quad \text{ and } \\ \varPsi _R( \tilde{r})= & {} A_2 ~exp \left( -\frac{(\tilde{r}-R)^2}{2} \right) \;, \end{aligned} \end{aligned}$$
(8)

where \(A_{0}\) is such that \(\int _0^{\infty } \tilde{r}^2 |\varPsi _{0}|^2 d\tilde{r}= A\), \(A_2\) is such that \(\int _0^{\infty } \tilde{r}^2 |\varPsi _R|^2 d\tilde{r} = A\) and A such that \(\int _0^{\infty } \tilde{r}^2 |\varPsi |^2 d\tilde{r} = 1\). Note that when \(R=0\) it reduces to the case with only one Gaussian, Eq. (7).

In the next section we describe the evolution of the system using this initial condition and compare the critical mass measured numerically with the result obtained analytically in Sect. 2.

4 Results

In this section we present the study of the evolution of the Schrödinger–Newton equation (5) using as initial condition the distribution expressed by Eq. (8). The distance, R, between the Gaussian peaks is left as a free parameter to explore the corresponding values for the critical mass, \(m_c\). Also in this section, we describe the criterion to define the critical mass and compare the numerical results with the analytical predictions discussed in Sect. 2.

Figure 2 describes the system behavior as a function of time for the particular initial condition with a single Gaussian centered at the origin, Eq. (7). Two kinds of measures are shown in this figure: (a) the value of the peak \(r_p\) of the radial probability density, \(\varrho _{\text {rad}}(\tilde{r})\), as a function of time, and (b, c) plot of \(\varrho _{\text {rad}}(\tilde{r})\) as function of the position \(\tilde{r}\) at different times for different values of m. These figures show clearly that there are two distinct regimes: one at small m where the packet diffuses and the other, at larger m when it oscillates. We can then define \(m_c\) as being the value of mass for which the system changes its behavior from diffusive to oscillatory. Nevertheless, from a numerical point of view this definition is hard to fulfill. We then need a practical criterion that is defined in the following.

Fig. 2
figure 2

Results for the case of a single gaussian \(R=0\). a The peak of the radial probability density \(\varrho _{\text {rad}}(\tilde{r})=4\pi \tilde{r}^2 | \varPsi |^2\), \({r}_p\), plotted against time for several masses, whose values are shown in the figure. Vertical dashed lines indicate the interval of time \(\varDelta T\) used to measure the average of \(r_p\) and define \(\langle r_p \rangle \). b Radial probability density \(\varrho _{\text {rad}}\) as a function of \(\tilde{r}\) at different times (specified in the legend) for \(m=1.14 m_r\). c Radial probability density \(\varrho _{\text {rad}}\) as a function of \(\tilde{r}\) at different times for \(m=1.24 m_r\). Each color represents a different instant of time, which is specified in the legend of the figure. Distances are in units of \(\sigma \) and time in units of \(t_r\). The system size is \(L=10^4\sigma \), \(\varDelta r=0.01\sigma \) and \(\varDelta t=0.01t_r\) (numeric parameters \(\varDelta r\) and \(\varDelta t\) are defined in “Appendix 1”) (color figure online)

The quantitative definition of the critical mass \(m_c\). The critical mass \(m_c\) is defined as the smallest value of mass for which the solution is oscillatory. We stress that this definition has some arbitrariness, because it depends on the interval of time \(\varDelta T\) during which one defines the average value of \(r_p\), \(\langle r_p \rangle \). Also, it has to be assured that the system size is big enough to avoid interference with the wave reflecting at the outer border of the system. Quantitatively, we chose the simulation time such that \(\varrho _{\text {rad}}(\tilde{r}=L)<10^{-8}\). For the system size used in the case of Figs. 2 and 3, \(L=10^4 \sigma \), one can safely use \(\varDelta T \in [2000t_r-4000t_r]\). This interval is represented in the figures by the vertical lines.

Similar results as the ones shown in Fig. 2 were already reported in the references [26, 54]. As already pointed out, these results differ by about 20 %, a point to which we return in “Appendix 2”. The novelty here is that we go further and search \(m_c\) for other initial conditions. An example is given in Fig. 3, where we do the same analysis as in Fig. 2 but with an initial condition where there is one Gaussian at the origin and a second one at a distance \(R=5\sigma \). Qualitatively we observe the same behavior in both cases. The difference is quantitative: the critical mass \(m_c\) changes depending on the value of R.

Fig. 3
figure 3

Same measures as shown in Fig. 2, but here the results are for the initial condition with \(R=5\sigma \). a The peak of the radial probability density \(\varrho _{\text {rad}}=4\pi \tilde{r}^2 | \varPsi |^2\), \(r_p\), plotted against time for several masses. Vertical dashed lines indicate the interval of time \(\varDelta T\) used to measure the average of \(r_p\). b Radial probability density \(\varrho _{\text {rad}}\) as a function of r at different times (specified in the legend) for \(m=0.85 m_r\). c Radial probability density \(\varrho _{\text {rad}}\) as a function of \(\tilde{r}\) at different times for \(m=1.20 m_r\). Each color represents a different instant of time, which is specified in b. Distances are in units of \(\sigma \) and time in units of \(t_r\). The system size is \(L=10^4\sigma \), \(\varDelta r=0.01\sigma \), and \(\varDelta t=0.01t_r\). The time evolution of the radial probability density for the masses exemplified in b and c are also shown in an animation in the supplementary material for time above \(t=100t_r\) (color figure online)

Fig. 4
figure 4

Average value of \(r_p\) versus m for various initial conditions [different values of R, Eq. (8)]. The dashed blue line is the threshold limit to define \(m_c\). \(\langle r_p \rangle \) and R are measured in units of \(\sigma \) and m in units of \(m_r\). \(\langle r_p \rangle \) is defined during the interval \(\varDelta T \in [2000t_r, 4000t_r ]\), which is indicated in Figs. 2 and 3 by the vertical dashed lines (color figure online)

Fig. 5
figure 5

The critical mass \(m_c\) as a function of the position R of the peak second Gaussian [Eq. (8)] for the theoretical computation in a and from numerical simulations in b. Note that the scale of distances are not the same in both figures because the agreement between the simulations and the analytical estimate is only qualitative, as discussed in the text. Masses are in units of \(m_r\) and R in units of \(\sigma \)

We further simulate various initial conditions (in other words, different values of R) and, for each R, scan several values of mass. We then measure the average value of \(r_p\) during the interval \(\varDelta T\) and plot it as a function of m for different values of R. This is shown in Fig. 4. Note that in Fig. 4 we plot not only the cases where the solution is oscillatory but also when it is diffusive. It is clear that when the solution is diffusive, \(\langle r_p \rangle \) is not representative of the behavior of the packet, since it keeps increasing during the measurement interval. For all R simulated, Fig. 4 shows a smooth behavior of \(\langle r_p \rangle \) with m. In other words, there is no clear modification in the behavior of \(\langle r_p \rangle \) for a given m that could suggest a clear definition of \(m_c\). This pushes us to adopt the following criterion to quantitatively define \(m_c\): For all R simulated we analyze the curves of \(r_p\) as a function of time, as exemplified in Figs. 2a and 3a, and we observed that, regardless the initial condition, when \(r_p\) passes a distance of \(\approx 20\sigma \) from the origin, the packet does not come back again. We then define \(m_c\) as a value of mass at which \(\langle r_p \rangle =20\sigma \). We emphasize that this criterion involves a degree of arbitrariness, and it can depend on the system size or on the parameters (\(\varDelta t\), \(\varDelta r\)) of the simulation. However, the parameters (L,\(\varDelta t\), \(\varDelta r\)) have been varied and the essence of the wave packet behavior remained robustly the same.

Figure 5 summarizes the main result of this work. It shows how the value of the critical mass \(m_c\) varies when R changes from 0 to \(\infty \) in an analytic estimate (Fig. 5a) and in the numerical simulations (Fig. 5b). Since the analytical estimate uses the initial shape of the wave packet, ignoring its dynamical evolution, it is not surprising that it has no quantitative agreement with the numerical simulations. However, the results from the theory and from the numerical simulations do share three interesting aspects: (i) the behavior of \(m_c\) is non-monotonic with R; (ii) in particular, \(m_c\) has a minimum value, denoted as \(m_c^{min}\), that is smaller than \(m_c\) at \(R=0\); and (iii) when both Gaussians are very far from each other (limit of large R), the \(m_c\) is the same as in the case with only one Gaussian (\(R=0\)). The point (iii) can be understood as a “decoupling” of the system into two weakly interacting subsystems when R is large. This is intuitively reasonable from the fact that the gravitational interaction decays with the distance, and it can be shown directly from Eq. (1). It is discussed in detail in “Appendix 4”.

We stress that the main conclusions are robust against the exact criterion used to define \(m_c\). We defined \(m_c\) as a value of mass at which \(\langle r_p \rangle =20\sigma \). From Fig. 4, we observe that we could have chosen a slightly different threshold to define \(m_c\), as, for example, \(\langle r_p \rangle =10\sigma \) or \(\langle r_p \rangle =30\sigma \). This would for sure change the absolute value of \(m_c\), but it would not change the conclusions (i), (ii), and (iii) commented above.

5 Appendix 1: Details of the numerical methods

In this section we explain all the details necessary to solve the Eq. (5) numerically. We stress that the procedure used here is very similar to the one developed in the reference [26], but with the reduced variables used in [54]. We keep the same notation as in [26] as much as we can. We also note that, equation Eq. (5) is written in terms of the reduced variables \(\tilde{m}\), \(\tilde{t}\) and \(\tilde{r}\). However, to simplify the notation, in this section we replace \(\tilde{m} \rightarrow m\), \(\tilde{t} \rightarrow t\); and \(\tilde{r} \rightarrow r\).

For concreteness, we rewrite Eq. (5) in a suitable way:

$$\begin{aligned} i \frac{\partial \varPsi }{\partial t}= & {} H \varPsi , \end{aligned}$$
(9)
$$\begin{aligned} H= & {} \frac{1}{2{m}} \varDelta _r + {m}\varPhi . \end{aligned}$$
(10)

The formal solution of this equation is given by \(\varPsi (t,r) = e^{-iHt}\varPsi (0,r)\). We observe that this is the solution when H does not depends on time t. In the case of Eq. (1), H depends on \(\varPsi \) which depends on t. However, for the numerical solution we assume that the time discretization is small enough to consider \(\varPsi \) as constant during it. The first step to solve Eq. (9) numerically is to discretize it in time and in space: time is written as \({t}=n\varDelta t\) and the space as \({r}=j \varDelta r\), where both j and n are integers, \(\varDelta t\) is the time discretization, and \(\varDelta r\) is the space discretization. The wave function is then denoted as \(\varPsi _j^n=\varPsi ( {r} = j\varDelta r,{t}=n\varDelta t)\). A numerical method to solve this equation needs to be both stable and unitary (required because the probability of finding a particle in the whole space should be constant). It is known that the Cayley’s form for the finite-difference representation of \(e^{-iHt}\), given by \(\displaystyle \simeq \frac{1-iH\varDelta t/2}{1+iH\varDelta t/2}\), meets these two criteria [46]. This form leads to the following representation of the solution of the above equation:

$$\begin{aligned} \left( 1+ \frac{1}{2}iH\varDelta t \right) \varPsi _j^{n+1} = \left( 1- \frac{1}{2}iH\varDelta t \right) \varPsi _j^{n} ~. \end{aligned}$$
(11)

With algebraic manipulation, Eq. (11) can be expressed as

$$\begin{aligned} \varPsi ^{n+1} = \left[ 2 \left( I + i\frac{\varDelta t}{2} H \right) ^{-1} - I \right] \varPsi ^{n}. \end{aligned}$$
(12)

We then define the matrix Q,

$$\begin{aligned} Q = \frac{1}{2}\left( I + i \frac{\varDelta t}{2} H \right) , \end{aligned}$$
(13)

and Eq. (12) can be simplified to

$$\begin{aligned} \varPsi ^{n+1} = \left( ~Q^{-1} - I~ \right) ~ \varPsi ^n =\chi ^n - \varPsi ^n, \end{aligned}$$
(14)

where

$$\begin{aligned} \chi ^n = Q^{-1} ~ \varPsi ^n. \end{aligned}$$
(15)

5.1 Setting up the matrix Q

To build the matrix Q, we need to handle the two terms of the hamiltonian represented by the Eq. (10): the Laplacian operator and the nonlinear term \(\varPhi \).

The radial component of the Laplacian, \(\varDelta _r\), in spherical coordinates is given by

$$\begin{aligned} \varDelta _r=\left\{ \begin{array}{ll} \displaystyle {\frac{\partial ^2}{\partial r^2}+\frac{2}{r}\frac{\partial }{\partial r}}&{}\quad \text{ if }\quad r > 0\\ \displaystyle {3\frac{\partial ^2}{\partial r^2}} &{}\quad \text{ if }\quad r=0, \end{array}\right. \end{aligned}$$
(16)

which can be discretized in the following way,

$$\begin{aligned} \varDelta \varLambda _j^n =\left\{ \begin{array}{ll} \displaystyle {\frac{1}{(\varDelta r)^2}\left( \frac{j-1}{j}\varLambda _{j-1}^n-2\varLambda _j^n + \frac{j+1}{j} \varLambda _{j+1}^n \right) }&{}\quad \text{ if }\quad j > 0\\ \displaystyle {\frac{1}{(\varDelta r)^2} (-6 \varLambda _0^n + 6\varLambda _1^n) } &{}\quad \text{ if }\quad j=0. \end{array}\right. \end{aligned}$$
(17)

The second term of the Hamiltonian, \(\varPhi \), is the solution of the Possion equation, Eq. (6). We apply the Green method to solve it [54]:

$$\begin{aligned} \phi = - \int _0^{\infty } G(\mathbf {r}, \mathbf {r}')|\varPsi (r',t)|^2 dV', \end{aligned}$$

where \(G(\mathbf {r}, \mathbf {r}')\) is the Green function,

$$\begin{aligned} G(\mathbf {r}, \mathbf {r}') = \frac{1}{4 \pi ~ \text{ max }(r,r')}. \end{aligned}$$
(18)

The result is given by

$$\begin{aligned} \phi = - 4 \pi m \left[ \frac{1}{r} \int _0^r |\varPsi (r',t)|^2 r'^2 dr' + \int _r^{\infty } |\varPsi (r',t)|^2 r' dr' \right] . \end{aligned}$$
(19)

The expression above can be discretized as

$$\begin{aligned} \phi _{j}^{n} = -4\pi m (\varDelta r)^2 v_{j}^{n}, \end{aligned}$$
(20)

where

$$\begin{aligned} v_j^n = \frac{1}{j}\sum _{i=0}^{j-1}|\varPsi _i^n|^2 i^2 + \sum _{i=j}^{N-1} |\varPsi _i^n|^2 i. \end{aligned}$$
(21)

Since we now have all the parts of the Hamiltonian discretized, we can write the matrix Q in the following way:

$$\begin{aligned} Q=\left( \begin{array}{rrrrr} b_0&{}c_0&{}0&{}0&{} \; \cdots \\ a_1&{}b_1&{}c_1&{}0&{}\cdots \\ 0&{}a_2&{}b_2&{}c_2&{}\\ \vdots &{}&{}&{}\vdots &{} \cdots \\ 0&{}\cdots &{}&{}a_{N-1}&{}b_{N-1} \end{array}\right) , \end{aligned}$$
(22)

with

$$\begin{aligned} a_j = \beta \,\frac{j-1}{j} \quad \text{ for }\quad 0 < j \le N-1, \end{aligned}$$
(23)
$$\begin{aligned} b_0 = \frac{1}{2} - 6\beta - \gamma v_0 \text{, } \quad \quad b_j = \frac{1}{2} - 2\beta - \gamma v_j \quad \text{ for }\quad 0 < j \le N-1,\quad \end{aligned}$$
(24)
$$\begin{aligned} c_0 = 6\beta \quad \quad c_j = \beta \, \frac{j+1}{j} \quad \text{ for }\quad 0 < j < N - 1, \end{aligned}$$
(25)

and \(\gamma \) and \(\beta \) being defined as

$$\begin{aligned} \beta = - \frac{i}{8m}\frac{\varDelta t}{(\varDelta r)^2} \quad \quad \text{ and } \quad \quad \gamma = i \pi m^2(\varDelta t)(\varDelta r)^2. \end{aligned}$$
(26)

Since matrix Q (22) is tridiagonal, the numerically efficient Thomas algorithm may be used do solve the system of equations (15). Then, the system evolution is obtained iterating expression (14).

6 Appendix 2: Comparison between our results and the literature

For the initial condition with only one Gaussian, Eq. (7), there are at least other two works that evaluated numerically the critical mass \(m_c\). These two works, references [26, 54], found a discrepancy of 20 % in its value. In this appendix we compare our results with the results of the reference [54] and show that our estimate for \(m_c\) agrees with this. We also discuss the possible sources of the discrepancy with the work [26].

In reference [54], Meter studies numerically four different values of mass. In Fig. 4 of [54] he shows the temporal evolution of \(r_p\), from which we extract the average value \(\langle r_p \rangle \). Figure 6 compares Meter’s result with the value of \(\langle r_p \rangle \) found for \(R=0\) in this work and shows a good agreement.

We cannot do the same comparison with the reference [26] because the authors did not compute the long time behaviour of the wave packet. In our Figs. 2 and 3, we observe that a long time series is necesssary to obtain the stationary solution. We believe that this is the source of the discrepancy between both results. Another possible source of difference are finite size effects. The authors in reference [26] do not specify their system size, but it is possible that they used a relatively small system. As already noted in the main text, for the kind of boundary conditions we are employing – which is the same as the authors of the reference [26] – it is important to keep the system size big enough in order to avoid reflections of the wave packet at the system boundary.

Fig. 6
figure 6

Average value of \(r_p\) versus m for the initial condition with a single Gaussian. \(\langle r_p \rangle \) is defined during the interval \(\varDelta t \in [2000t_r, 4000t_r ]\). \(\langle r_p \rangle \) is measured in units of \(\sigma \), m in units of \(m_r\), and the system size used in this simulation is \(L=10^4 \sigma \). The diamond symbols are extracted from Fig. 4 of reference [54]

7 Appendix 3: The critical mass

The expression for the critical mass obtained via the method of section 2 is

$$\begin{aligned} m_c = \left( \frac{A}{4B}\right) ^\frac{1}{3}, \end{aligned}$$

where

$$\begin{aligned} A= & {} \hbar ^2 r^3_p(0) \Big ( 8 e^\frac{r^2_p(0) + R^2}{\sigma ^2} R \sigma ^4(2r^4_p(0) - 7 r^2_p(0)\sigma ^2 + 3\sigma ^4) \\&+\,4e^\frac{r^2_p(0) + 2R^2}{\sigma ^2} \sqrt{\pi } \sigma ^3(2R^2+\sigma ^2)(2r^4_p(0)-7 r^2_p(0)\sigma ^2 + 3\sigma ^4) \\&+\,4e^\frac{(r_p(0)+R)^2}{\sigma ^2}\sqrt{\pi } \sigma ^5 \big ( 2r^4_p(0) - 6r^3_p(0)R - 2R^2\sigma ^2 + 3\sigma ^4 \\&+\,r^2_p(0)(6R^2-7\sigma ^2) + r_p(0)(-2R^3+9R\sigma ^2) \big ) \\&+\,4 e^\frac{r^2_p(0)+2R^2}{\sigma ^2}\sqrt{\pi } \sigma ^3 (2R^2 + \sigma ^2)(2r^4_p(0) - 7r^2_p(0)\sigma ^2 + 3\sigma ^4){\text {Erf}}(R/\sigma ) \\&+\,e^\frac{2r^2_p(0)+ 2r_p(0)R + 3R^3}{2 \sigma ^2}\pi ^\frac{1}{4} \sqrt{s^3} \big ( -8 r^4_p(0)( R^2-2\sigma ^2) + 12r^3_p(0)(R^3-2R\sigma ^2) \\&+\,r^2_p(0)(-6R^4+32R^2\sigma ^2 -56\sigma ^4) + 2\sigma ^2(R^4-4R^2\sigma ^2+12\sigma ^4) \\&+\,r_p(0)(R^5 - 16R^3\sigma ^2+36R\sigma ^4) \big ) \\&\times \sqrt{ \sigma \big ( 2e^{-\frac{R^2}{\sigma ^2}}R\sigma + \sqrt{\pi }(2R^2+\sigma ^2) + \sqrt{\pi }(2R^2+\sigma ^2){\text {Erf}}(R/\sigma ) \big ) } \,\, \Big ), \end{aligned}$$

and

$$\begin{aligned} B= & {} G \sigma ^6 \Big ( 2e^\frac{r^2_p(0)+R^2}{\sigma ^2} R\sigma ^2 (2r^4_p(0)- 5r^2_p(0)\sigma ^2 + \sigma ^4 ) \\&+\,e^\frac{r^2_p(0)+2 R^2}{\sigma ^2}\sqrt{\pi } \sigma (2 R^2 + \sigma ^2) (2r^4_p(0)- 5r^2_p(0)\sigma ^2 + \sigma ^4 ) \\&+\,e^\frac{(r_p(0)+ R)^2}{\sigma ^2} \sqrt{\pi } \sigma ^3 \big ( 2r^4_p(0) - 4r^3_p(0)R+4r_p(0)R\sigma ^2+\sigma ^4\!+\!r^2_p(0)(2R^2-5\sigma ^2) \big ) \\&+\,e^\frac{r^2_p(0)+2R^2}{\sigma ^2}\sqrt{\pi }\sigma (2R^2+\sigma ^2)(2 r^4_p(0) -5r^2_p(0) \sigma ^2 + \sigma ^4 ) {\text {Erf}}(R/\sigma ) \\&+\,e^\frac{2r^2_p(0)+2r_p(0)R + 3R^2}{\sigma ^2} \pi ^\frac{1}{4} \sqrt{\sigma ^3} \big ( 4r^4_p(0)-4r^3_p(0)R + 4r_p(0)R\sigma ^2 \\&+\,2\sigma ^4 + r^2_p(0)( R^2-10\sigma ^2 ) \big ) \\&\times \,\sqrt{ \sigma \big ( 2e^{-\frac{R^2}{\sigma ^2}}R\sigma + \sqrt{\pi }(2R^2+\sigma ^2) + \sqrt{\pi }(2R^2+\sigma ^2){\text {Erf}}(R/\sigma ) \big ) } \,\, \Big ). \end{aligned}$$

8 Appendix 4: The limit \(R \gg 1 \sigma \)

In order to investigate the behavior of the system when R is large, we first consider the situation where the initial condition is a single radially symmetric wave package at R, \(\varPsi (0,x) = \varPsi _R(0,x)\). In this situation, it is easy to analytically estimate \(m_c\) exactly as in section 3.1 of [26]. Thus, we consider the free particle solution Eq. (2); find the value \(r_p(t)\) of r for which the radial probability density \(\varrho (t,r) = 4 \pi r^2 |\varPsi (t,x)|^2\) has a global maximum; compute the peak’s acceleration at time zero, \(\ddot{r}_p(0)\); and set it equal to \(G m/ (r_p(0))^2\). This produces the following expression for the critical mass:

$$\begin{aligned} m_c = \frac{ ( R^2 \hbar ^2 + 2\hbar ^2 \sigma ^2 + \hbar ^2 R \sqrt{R^2 + 4\sigma ^2})^\frac{1}{3} }{ (G\sigma ^2)^\frac{1}{3} (R^2 + 4\sigma ^2)^\frac{1}{6} }, \end{aligned}$$
(27)

which of course reduces to \(m_r = \left( \frac{\hbar ^2}{G \sigma } \right) ^\frac{1}{3} \) when \(R=0\). From the above expression it is clear that \(m_c\) increases with R. In particular, one cannot decrease the critical mass using only a single Gaussian as the initial condition.

We now show that the system “decouples” into two weakly interacting subsystems when R is large.

Suppose that \(\varPsi \) is a solution to Eq. (1) with initial condition \(\varPsi (0,x) = \varPsi _1(0,x) + \varPsi _2(0,x)\). For simplicity, assume that \(\varPsi _1(0,x)\) and \(\varPsi _2(0,x)\) have compact and disjoint supports, i.e.,

$$\begin{aligned} {\text {supp}} (\varPsi _i (0)) \equiv \{ x \in \mathbb {R}^3 \, | \, \varPsi _i (0,x) \ne 0 \} \subset B_{r_i}(0), \end{aligned}$$

\(i=1,2\), where \(B_{r_i}(0)\) is the ball of radius \(r_i >0\) centered at the origin (strictly speaking, the support is the closure of the above set, but this will make no difference in our argument), and

$$\begin{aligned} {\text {supp}} (\varPsi _1 (0)) \cap {\text {supp}} (\varPsi _2 (0)) = \varnothing , \, \, |{\text {supp}} (\varPsi _1 (0)) - {\text {supp}} (\varPsi _1 (0)) | \gg 1. \end{aligned}$$

In our case the supports are not disjoint, but since we are interested in the case \(R\gg 1\), this will be approximately the case. As \(\varPsi \) is generally a well-behaved function of t, we can also assume that for small t, \(\varPsi \approx \varPsi _1 + \varPsi _2\), where \(\varPsi _i(t,x)\) is close to \(\varPsi _i(0,x)\) if t is very small. Thus we can also assume that

$$\begin{aligned} {\text {supp}} (\varPsi _1 (t)) \cap {\text {supp}} (\varPsi _2 (t)) = \varnothing , \, \, |{\text {supp}} (\varPsi _1 (t)) - {\text {supp}} (\varPsi _1 (t)) | \gg 1 \end{aligned}$$
(28)

(again, this will hold only as an approximation, but it suffices for our purposes). We restrict our attention to small t, because if confinement is going to happen due to the gravitational interaction, this should happen at earlier times before the wave-function has spread out considerably.

For \(\varPsi = \varPsi _1 + \varPsi _2\), Eq. (1) reads

$$\begin{aligned} i \hbar \frac{\partial \varPsi _1}{\partial t} + i \hbar \frac{\partial \varPsi _2}{\partial t}= & {} \left( - \frac{\hbar ^2}{2m} \varDelta - Gm^2 \int _{\mathbb {R}^3} \frac{|\varPsi _1|^2}{|x-y|} \,dy \right) \varPsi _1 \\&+ \left( - \frac{\hbar ^2}{2m} \varDelta - Gm^2 \int _{\mathbb {R}^3} \frac{|\varPsi _2|^2}{|x-y|} \,dy \right) \varPsi _2 \\&- Gm^2 \left( \int _{\mathbb {R}^3} \frac{|\varPsi _1|^2}{|x-y|} \,dy \right) \varPsi _2 - Gm^2 \left( \int _{\mathbb {R}^3} \frac{|\varPsi _2|^2}{|x-y|} \,dy \right) \varPsi _1 \\&- Gm^2 \left( \int _{\mathbb {R}^3} \frac{ \varPsi _1 \varPsi _2^* + \varPsi _1^* \varPsi _2 }{|x-y|} \,dy \right) (\varPsi _1 + \varPsi _2). \end{aligned}$$

The cross terms \(\varPsi _1 \varPsi _2^* \) and \(\varPsi _1^* \varPsi _2\) vanish, because for any given y, either \(\varPsi _1(t,y)\) or \(\varPsi _2(t,y)\) vanishes in light of expression (28).

If \(x \in {\text {supp}} (\varPsi _1 (t))\), then the term

$$\begin{aligned} - Gm^2 \left( \int _{\mathbb {R}^3} \frac{|\varPsi _1|^2}{|x-y|} \,dy \right) \varPsi _2 \end{aligned}$$

vanishes, and the term

$$\begin{aligned} - Gm^2 \left( \int _{\mathbb {R}^3} \frac{|\varPsi _2|^2}{|x-y|} \,dy \right) \varPsi _1 \end{aligned}$$

will be very small since

$$\begin{aligned} \int _{\mathbb {R}^3} \frac{|\varPsi _2|^2}{|x-y|} \,dy = \int _{{\text {supp}} (\varPsi _2 (t))} \frac{|\varPsi _2|^2}{|x-y|} \,dy, \end{aligned}$$

and x and \({\text {supp}} (\varPsi _2 (t))\) are very far away, so that \(|x-y|\) is very large.

A similar statement holds if \(x \in {\text {supp}} (\varPsi _2 (t))\) instead, and obviously all terms vanish if \( x \notin {\text {supp}} (\varPsi _1 (t)) \cup {\text {supp}} (\varPsi _2 (t))\). Thus

$$\begin{aligned} i \hbar \frac{\partial \varPsi _1}{\partial t} + i \hbar \frac{\partial \varPsi _2}{\partial t}\approx & {} \left( - \frac{\hbar ^2}{2m} \varDelta - Gm^2 \int _{\mathbb {R}^3} \frac{|\varPsi _1|^2}{|x-y|} \,dy \right) \varPsi _1 \\&+ \left( - \frac{\hbar ^2}{2m} \varDelta - Gm^2 \int _{\mathbb {R}^3} \frac{|\varPsi _2|^2}{|x-y|} \,dy \right) \varPsi _2, \end{aligned}$$

i.e., under the above circumstances, the evolution behaves as if the system were linear. Thus if one considers the situation where \(\varPsi _1(0,x) \approx \varPsi _0(0,x)\) and \(\varPsi _2(0,x) \approx \varPsi _R(0,x)\), with \(R \gg 1\), a very large value of m is necessary to confine the part of the wave function associated with \(\varPsi _2\) [recall Eq. (27)]. Hence, confinement happens essentially due to \(\varPsi _1\), which in this case corresponds roughly to the system with a single Gaussian centered at the origin. This explains why \(m_c\) approaches \(m_r\) when R is very large.