In this chapter, we shall introduce attempts to base the calculus on a different concept from that of Weierstraß.

Built on Very Thin Ice: Cantor’s Diagonal Argument

The Presentation of Evidence

It is quite astonishing to discover which precarious arguments are sometimes accepted as being mathematically sound.

An outstanding example thereof is the (time and time again, by experts and non-experts alike) celebrated “proof for the existence of the uncountable infinity”. This bizarre “proof” runs as follows (there should be added some technical details in case the number ends with infinitely many nines, but this is just technical and does not touch the essential reasoning; therefore, it can be omitted):

Suppose we could enumerate the numbers between 0 and 1, written as decimals.

If this is done, we arrive at a list

$$\displaystyle \begin{aligned} \begin{array}[t]{r@{}c@{}c@{}c@{}c} b_1 = 0.a_{11} & a_{12} & a_{13} & a_{14}\ldots \\ b_2 = 0.a_{21} & a_{22} & a_{23} & a_{24}\ldots \\ b_3 = 0.a_{31} & a_{32} & a_{33} & a_{34}\ldots \\ \vdots &&&&\end{array} \end{aligned}$$

The akn are digits from 0 to 9.—Possibly, we have

$$\displaystyle \begin{aligned}b_{59284}=0.5000\ldots\end{aligned}$$

According to our assumption, the list includes each number between 0 and 1.

The surprising “argument” that should supply the contradiction to our supposition is as follows:

  • Take the number c = 0.c1c2c3c4…, defined by c1 = any number, but not a11; c2 = any number, but not a22; c3 = any number, but not a33; etc.

  • This number c differs obviously from all bm. (This c deviates in the m-th digit from bm, as this digit of bm is = amm ≠ cm.)

  • Consequently, the list of the bm does not contain c.

  • But this contradicts the initial assumption that we had a list bm of all numbers between 0 and 1.

  • Contradiction. Presumed end of proof.

Every modern book on the mathematical infinite presents this “proof”, including the book Proofs from THE BOOK from Martin Aigner and Günter M. Ziegler from 1998 as well (pp. 92f). And certainly the authors are excited: star writer e.g. called this proof “both ingenious and beautiful—a total confirmation of art’s compresence in pure math” (p. 255). An appraisal such as “ingenious” calls for attention, no doubt about that!

The Impotence of This Reasoning

Let us pause for thought and ask: does this actually convince anyone?

Let us work out the details. At once we realise:

The above “proof” is not necessarily valid.

(This is no news. I only repeat a well-known reasoning.) —Therefore, I ask two questions.

(1) How many decimal numbers with n digits in between 0 and 1 do exist?

The answer is straightforward:

  1. 1.

    If n = 1, we have exactly 10, to wit: 0.0; 0.1; 0.2; … 0.9.

  2. 2.

    If n = 2, we have exactly 100 = 102, to wit: 0.00; 0.01; 0.02; … 0.99.

    etc.

  3. n.

    Consequently, in the general case n, we have 10n numbers.

And now my second question:

(2) How many digits of c are safely determined by this so-called proof?

The answer to this question is evident: exactly n—one for each digit under consideration. But what is the consequence? Plainly this: the number c defined by this professed proof certainly differs from the first n numbers in the list—but this does not demonstrate conclusively that it is not included in the list at all; instead, it only proves: \(c=b_{m'}\) as well as m′ > n, nothing else.

To be practical, let n = 2. We write down the list of the 102 = 100 numbers in between 0 and 1 with two digits after the decimal point:

$$\displaystyle \begin{aligned} \begin{array}[t]{c} b_1=0.00\,, \end{array}\qquad \begin{array}[t]{c} b_2=0.01 \\ b_3=0.02 \\ \ldots\,, \\ \end{array}\qquad \begin{array}[t]{c} b_{11}=0.10 \\ b_{12}=0.11 \\ \ldots\,, \\ \end{array}\qquad \ldots\,, \qquad \begin{array}[t]{c} b_{100}=0.99\,. \\ \end{array}\end{aligned} $$

Let us take c = 0.12. Then we know c ≠ b1 and c ≠ b2; but we have c = b13. c is not to be found among the first n = 2 numbers but later on it is included in the list.

This reasoning imputes a (most) extremely strange hidden assumption to this “both ingenious and beautiful” “proof”. It is this one:

$$\displaystyle \begin{aligned}\pmb{10^\infty=\infty\,.}\end{aligned}$$

For in fact it is proclaimed: The number c is not contained in the list of the bm. What is true, however, is just this: The number c is not among the first infinitely many (“”) numbers of the list, but only subsequently.

However, who is demanding to calculate with infinity in this way? It is true, e.g. Weierstraß authorized this—i.e. 10 = —but at the same time he demanded to exclude infinity (“ ”) from analysis (p. 12).

But Euler did use infinity in his calculations. He even assumed i > i − 1 (p. 9) and moreover

$$\displaystyle \begin{aligned}{\mathbf{10}}^{\mathbf{i}}>\mathbf{i}\end{aligned}$$

for infinitely large i.

There is no obligation to think like Euler and also none to think like Weierstraß.

In mathematics we do not deal with dogmas but with arguments.

From Chap. 5 we know:

There exists no conclusive or ultimate reasoning for dealing with the actual infinite in analysis.

The Origin of the “Diagonal Argument”

The “proof”, which is described above, was invented by . Cantor demonstrated thus in 1891 for the third time:

Theorem. There exist infinite sets which cannot be related one-to-one to the set of natural numbers.

Cantor designed his proof slightly differently. (He took only two different digits instead of ten as we did.—Strikingly, this proof, given for binary numbers instead of decimal numbers, is not only not conclusive but also wrong—why exactly? And: is it permissible to make a proof about the real numbers dependent on the mode of their representation?)

The True Understanding of the “Diagonal Argument”

It was Bertrand Russell who unearthed the rational essence of Cantor’s proof.

The true theorem of Cantor’s proof is this:

Theorem. The number of subsets of a set is larger than the number of the elements of this set.

Let us take an example. The given set may contain two elements, a and b:

$$\displaystyle \begin{aligned} M\;=\; \{a\,,\;b\}\,. \end{aligned}$$

Then M has the following 22 = 4 subsets, first of all the empty set { } = ∅:

$$\displaystyle \begin{aligned} \emptyset\,,\quad \{a\}\,,\quad \{b\}\,,\quad \{a\,,\;b\} \,.\end{aligned}$$

Hence, there are more subsets than elements: 4 > 2.

Now we faithfully present Russell’s general proof of his theorem (changed, however, are the names of the sets—which Russell calls “classes”). Like Cantor, Russell starts with a list of subsets signified by the elements. Russell calls this list a “one-one correlation R ”.

When a one-one correlation R is established between all the members of M and some of its sub-classes, it may happen that a given member x is correlated with a sub-class of which it is a member; or, again, it may happen that x is correlated with a sub-class of which it is not a member.

We interrupt Russell and illustrate his set-theoretical construction with the help of our example, and that twice.

  1. 1.Illustration.

    The list of the correspondences of the elements and the subsets of M = {a , b} might be

    $$\displaystyle \begin{aligned} a \mapsto \{a\}\;,\qquad b\mapsto \{a\;, b\}\,. \end{aligned}$$

    Then we have for both elements a and b that they correspond to a subset that contains them as an element.

  2. 2.Illustration.

    A second list might be:

    $$\displaystyle \begin{aligned} a\mapsto \emptyset\;,\qquad b\mapsto \{b\}\,.\end{aligned}$$

    In this case, only element b corresponds to a subset, which contains itself as an element, but element a does not.

Now we continue with Russell’s proof.

Let us form the whole class, N say, of those members x which are correlated with sub-classes of which they are not a member.

In our first illustration, we have N = ∅; in the second, it is N = {a}. We continue with the proof. Now follows the decisive sentence (I give it in italics):

This is a sub-class of M, and it is not correlated with any member of M.

For taking first the members of N, each of them is (by the definition of N) correlated with some sub-class of which it is not a member, and is therefore not correlated with N.

Taking next the terms that are not members of N, each of them (by the definition of N) is correlated with some sub-class of which it is a member, and therefore again is not correlated with N.

Thus no member of M is correlated with N.

Since R was any one-one correlation of all members with some sub-classes, it follows that there is no correlation of all members with all sub-classes.

It is extremely impressive how a few words can express a really entangled reasoning. Those who are able to understand this demonstration from Russell may be pleased with their capability of abstract thought.

If you love formalism, you may write down this proof in the following manner (usually “2M” denotes the set of all subsets of M; “M ∖ N” indicates the difference of the sets M and N):

In this way, Russell established:

Theorem. 2n is always larger than n—even if n is infinite.

This implies that the infinite cardinal numbers do not have a maximum.

The Significance of the “Diagonal Argument”

That is why the so-called Diagonal Argument from Cantor is a method of proof in set-theory. It gives the means of proving e.g. that there always exist larger “cardinalities”, even in the infinite.

The “numbers” in set-theory are “cardinal” numbers. In analysis they do not feature (at least not in traditional analysis, in calculus). Calculus needs “numbers for calculations”.

Therefore, Russell’s proof that 2n > n (which holds in set-theory) does not contradict Weierstraß’ principle that 10 = . Quite the contrary: if one accepts Cantor’s Diagonal Argument as permissible reasoning in analysis, you must also accept Weierstraß’ dictum that 10 = , for otherwise the Diagonal Argument fails! This is explained above.

So, if you think analysis (in this sense) and set-theory as one thing, you will be convinced that 10 =  (in analysis) and that 2n > n for all n, including infinite numbers (in set-theory).

Some will like this … But in some vague sense, this does not look altogether consistent. Luckily, one is not forced to think in this style.

For the record: on the case of numbers used for calculating, i.e. in analysis, you may accept the Diagonal Argument. Or you may not. In any case, it is not conclusive there. And if you want to relate it to the facts connected to the decimal numbers (for which it claims validity), it is even less conclusive. For we have seen:

The Diagonal Argument demands the acceptance of an argument regarding infinity, which is invalid in any finite case.

As we know, Leibniz once argued in the same way, thereby surprising Johann Bernoulli (p. 5).

It might cross one’s mind to interpret this argument as a proof by induction:

  • Start: choose c1 ≠ a11.

  • Inductive step: Let cn ≠ ann. Then choose cn+1 ≠ an+1,n+1. —Finished.

An objection to this kind of reasoning is obvious: it is presupposed that the first part of each number 0.c1c2cn formed in this way is contained in the list of the bm: \(0.c_1c_2\ldots c_n=b_{m'}\); of course, it is m′ > n. So what? If any first part of c is contained in the list of the bm—why should this hold no longer for c itself?

If you do not accept the Diagonal Argument in analysis (to repeat: this is as legitimate as is the opposite), you are accepting the validity of

$$\displaystyle \begin{aligned} 10^i>i \end{aligned}$$

for infinite numbers used in calculations. Finally, we shall demonstrate where this kind of thinking will lead to.

Paradox I: Conditionally Convergent Series

Let us finish our detour into set-theory and return to analysis. There are certain curiosities that have arisen during the last three and a half centuries that caught the attention of some analysts.

$$\displaystyle \begin{aligned} \textstyle 1-\frac{{1}}{{2}}+\frac{{1}}{{3}}-\frac{{1}}{{4}}+\frac{{1}}{{5}}-+\ldots = \ln 2\,.\end{aligned}$$

Let us now tackle some of these curiosities in more detail, using the title “paradoxes”. The creator of this name will be uncovered later.

In order to prepare for this, we now return to Riemann. We know: Riemann is always one for offering a surprising perspective.

A Mathematical Monstrosity: The Riemann Theorem on Rearrangements of 1854

Riemann proved the following oddity, nowadays called the “Riemann Theorem on Rearrangements”, in his habilitation thesis of 1854:

Theorem. If a convergent series is no longer convergent when its terms are made positive, the terms can be rearranged in such a way that the series converges on any arbitrarily chosen value.

The proof seems to be straightforward. One assumes a convergent series of numbers:

$$\displaystyle \begin{aligned} \sum_{k=1}^\infty a_k = s\,.\end{aligned}$$

Infinitely many terms have the sign + , and infinitely many have the sign −. We call the positive terms of the series b, the negative ones c:

$$\displaystyle \begin{aligned} \sum_{i=1}^\infty b_i = \sum_{\genfrac{}{}{0pt}{}{a_k>0~\text{and}}{k=1}}^{k=\infty} a_k \qquad \text{as well as}\qquad \sum_{i=1}^\infty c_i = \sum_{\genfrac{}{}{0pt}{}{a_k<0~\text{and}}{k=1}}^{k=\infty} a_k \,. \end{aligned}$$

So we have bi > 0 as well as ci < 0. Neither of those series converges:

$$\displaystyle \begin{aligned} \sum_{i=1}^\infty b_i = \sum_{\genfrac{}{}{0pt}{}{a_k>0~\text{and}}{k=1}}^{k=\infty} a_k = \infty \qquad \text{as well as}\qquad \sum_{i=1}^\infty c_i = \sum_{\genfrac{}{}{0pt}{}{a_k<0~\text{and}}{k=1}}^{k=\infty} a_k = -\infty\,. \end{aligned}$$

If both series had finite sums, then the series \(\sum |a_k|\) would also converge, which contradicts our assumption. If only one of these two series converges, the initial series could not converge. That is why none of our partial series converges.

Now Riemann’s reasoning: let D, say D > 0, be an arbitrarily given value. Then we approach D gradually—at first from below, then from above; and so on, alternating. That is to say, we choose the smallest number n1 with

$$\displaystyle \begin{aligned} \sum_{k=1}^{n_1} b_k > D\,; \quad \text{thereafter, the smallest number }m_1\text{ with} \quad \sum_{k=1}^{n_1} b_k + \sum_{k=1}^{m_1} c_k < D\,, \end{aligned}$$

etc. The difference to D will never be more than the absolute value of the term that is the one before the last sign change. But \(\sum a_k\) converges, and therefore, bk as well as |ck| are decreasing with increasing k below any given quantity; and so does the difference to D—which seems to prove all.

This theorem is a monster. Weierstraß obviously did not like it, and he indeed mentions the notion of conditional convergence (see p. 19), but he ignores this concept in his own analysis.

Why is this theorem a monster? Because it assumes that a mathematician is able to decide infinitely many times at will to change between the partial series of the bi and ci back and forth. “An infinite number of arbitrary choices is an impossibility” says Russell. Infinitely many actual decisions have nothing in common with reality.

Note: A proof may be easy, but the proven statement may be absurd.

Mitigation

A theory of infinitely many single choices in analysis certainly is an extreme. An essential mitigation is to think about a regular change of infinitely many terms of a series. We might have more chance of acceptance if we concentrate on regular mathematical constructions—although Weierstraß might still have taken a different view.

Let us take an example. We start with a series A, halve every term, add both series term-by-term, and regain the initial series, although slightly rearranged.

We have:

$$\displaystyle \begin{aligned} A=\quad \ln 2 &= \textstyle 1-\frac{{1}}{{2}}+\frac{{1}}{{3}}-\frac{{1}}{{4}}+\frac{{1}}{{5}}-\frac{{1}}{{6}}+\frac{{1}}{{7}}-\frac{{1}}{{8}}+-\ldots \\ {} \text{Therefore},\qquad \;\,\, \textstyle \frac{{1}}{{2}}\ \ln 2 &= \textstyle \quad \;\;\;\frac{{1}}{{2}}\quad \;\;\,-\frac{{1}}{{4}}\quad \;\;\, +\frac{{1}}{{6}}\quad \;\;\,-\frac{{1}}{{8}}+-\ldots\,, \\ {} \text{and consequently} \qquad \textstyle B=\frac{{3}}{{2}}\ \ln 2 &= \textstyle 1\quad \;\;\;+\frac{{1}}{{3}}-\frac{{1}}{{2}}+\frac{{1}}{{5}}\quad \;\;\;+\frac{{1}}{{7}}-\frac{{1}}{{4}}++-\ldots \\ &= \quad \text{[rearrangement of]} \quad \ln 2.{} \end{aligned} $$

Do we have A = B, i.e. \(\;\ln 2=\frac {{3}}{{2}}\ \ln 2\)?

According to Weierstraß, this conclusion is a “fallacy” (p. 11). Why?—Is the Riemann Theorem on Rearrangements an answer?—Is this answer satisfactory?

Paradox II: Methods of Summation

For divergent series like

$$\displaystyle \begin{aligned} A= 1-1+1-1+1-1+-\ldots \end{aligned} $$

sometimes other summation methods are in use, e.g. the so-called method of C-1 summation:

$$\displaystyle \begin{aligned} S^{C1}= \lim_{n\to\infty} \frac{s_1+s_2+s_3+\ldots+s_n}{n}\,. \end{aligned}$$

In case of series A, the partial sums are s1 = 1; s2 = 1 − 1; s3 = 1 − 1 + 1; s4 = 1 − 1 + 1 − 1 etc., that is:

$$\displaystyle \begin{aligned} \\ {} s_1&=s_3=s_5=\ldots\ =1\,, \\ s_2&=s_4=s_6=\ldots\ =0\,, \end{aligned} $$

and consequently, the terms of SC 1 are

$$\displaystyle \begin{aligned} p_1&=\textstyle\frac{s_1}{1}=\textstyle 1\,, \\ p_2&=\textstyle\frac{s_1+s_2}{2}=\textstyle \frac{1}{2}\,, \\ p_3&=\textstyle\frac{s_1+s_2+s_3}{3}=\textstyle \frac{2}{3}\,, \\ p_4&=\textstyle\frac{s_1+s_2+s_3+s_4}{4}=\textstyle \frac{2}{4}=\frac{1}{2}\,, \\ p_5&=\textstyle\frac{s_1+s_2+s_3+s_4+s_5}{5}=\textstyle \frac{3}{5}\,, \\ {} & \ldots \\ \mbox{i.e.} \qquad \quad p_{2n-1}&=\textstyle\frac{{n}}{{2n-1}}\,, \\ p_{2n}&=\textstyle \frac{{n}}{{2n}}=\frac{{1}}{{2}}\,, \\ \text{all together:} \qquad \quad \;\;\; S^{C~1}&=\lim_{n\to\infty} p_n=\textstyle \frac{{1}}{{2}}\,. \end{aligned} $$

With this method, we do get:

$$\displaystyle \begin{aligned} A=1-1+1-1+1-1+-\ldots= \textstyle\frac{{1}}{{2}}\;!?\end{aligned}$$

Paradox III: The Convergence of Function Series

Let us take the geometric series

$$\displaystyle \begin{aligned} s(x)=1-x+x^2-x^3+x^4-+\ldots\end{aligned}$$

The well-known way to calculate the sum s(x) is to add sn(x) and x ⋅ sn(x):

and

The consequence is

Yet, “does not exist” is not a beautiful result of a calculation.

Paradox IV: The Term-by-Term Integration of Series

Let us take the series of functions

$$\displaystyle \begin{aligned} f_n(x)=\frac{{n x}}{{1+n^2 x^4}}\qquad \text{with}\quad 0\leqq x\leqq1\,.\end{aligned}$$

As \(\lim \limits _{n\to \infty }\frac {{nx}}{{1+n^2x^4}}=0\; \) if \(\;0\, {\scriptstyle \leqq }\, x\, {\scriptstyle \leqq }\, 1\;\), we usually conclude for the limit n →:

$$\displaystyle \begin{aligned} f(x) = \lim_{n\to\infty} f_n(x) = \lim_{n\to\infty}\frac{{nx}}{{1+n^2x^4}} =0\,. {} \end{aligned} $$
(‡‡)

But if \(x_n=\frac {{1}}{{\sqrt {n}}}\) (and consequently \(\lim \limits _{n\to \infty } x_n=\lim \limits _{n\to \infty }\frac {{1}}{{\sqrt {n}}}=0\)), the value of f(x) at the value \(0=\lim \limits _{n\to \infty } \frac {{1}}{{\sqrt {n}}}\) is

$$\displaystyle \begin{aligned} \textstyle\lim\limits_{n\to\infty} f_n\left(\frac{{1}}{{\sqrt{n}}}\right)= \textstyle\lim\limits_{n\to\infty}\frac{{\sqrt{n}}}{{1+\frac{{n^2}}{{n^2}}}}= \textstyle\lim\limits_{n\to\infty}\frac{\sqrt{n}}{2}=\infty\;! \end{aligned}$$

We remember: in these cases Cauchy concludes to be a value of the function f(x) at the value x = 0 (p. 10). However, present-day analysis handles things differently and demands:

$$\displaystyle \begin{aligned} f(0)= \lim_{n\to\infty} f_n(0)=\lim_{n\to\infty}\frac{{n\cdot0}}{{1+n^2\cdot0}} = \frac{{0}}{{1}}=0\,,\end{aligned}$$

unambiguous and crystal-clear.

If we have, following ‡‡, f(x) = 0 for all \(0\;{\scriptstyle \leqq }\; x\;{\scriptstyle \leqq }\; 1\), integration is easy:

$$\displaystyle \begin{aligned} \int\limits_0^1f(x){\mathit{\,dx}\,}= \int\limits_0^1 0\cdot{\mathit{\,dx}\,}=0\,.\end{aligned}$$

But let us integrate term-by-term:

$$\displaystyle \begin{aligned} \int\limits_0^1 f_n(x){\mathit{\,dx}\,} = \int\limits_0^1 \frac{{nx}}{{1+n^2x^4}}{\mathit{\,dx}\,}\,.\end{aligned}$$

Substituting z = nx2 gives dz  = 2nx dx and we arrive at

With the help of \(f(x)=\lim f_n(x)\), we conclude

$$\displaystyle \begin{aligned} \int_0^1 f(x){\mathit{\,dx}\,}= \lim_{n\to\infty}\;\int_0^1 f_n(x){\mathit{\,dx}\,} = \lim_{n\to\infty} \textstyle\frac{{1}}{{2}}\cdot\arctan n =\textstyle\frac{\pi}{4}\,. \end{aligned}$$

This contradicts the previous result. Consequently, the first equality in the last line is wrong—the others being indubitable. So, for this series, it is not possible to interchange integration and limit.

Conclusion: We need a theorem to tell us, in which cases the interchange of integration and limit is allowed. Naturally, it would be more attractive if the calculation itself would make the problem visible. Some analysts enjoy calculating more than hunting for theorems that allow them to calculate.

Is an Analysis Without Those Paradoxes Possible?

A Source from the Years 1948–53

was an eminent mathematician. In 1934, he became a professor at the Technical University of Darmstadt, where he was awarded a chair in mathematics in 1937, and in 1957/58 he was the rector of this university. In a manuscript dating from 1948–53, Schmieden commented on examples of this kind (all emphases added):

Such examples could be given endlessly; the deeper one penetrates into mathematics, the more such paradoxes emerge—even if one completely disregards set-theory.

The most thrilling aspect of this kind of mathematics is that in spite of those paradoxes one always arrives at a “true” result for each concrete problem. Consequently, it does not surprise that some mathematicians decide at some point that the naive intuition about infinity does no longer suffice in some way or other, whereas working along definite rules in regard to infinity makes it possible to tame those quantities.

Nevertheless, there remains a feeling of discomfort. Also the question remains, whether it is not possible after all to construct analysis in such a way that it works alongside fixed rules as it has been doing so far, and this in such a way that the kinds of aforementioned paradoxes do not emerge. Thus our naive intuition which according to Goethe, more often than not wishes “to represent the infinite with the help of the finite”[Footnote 1], would be granted .

Schmieden, being “fully aware of the imperfections of this attempt” stated some foundational principles along those lines. Among them was the following that is obviously inspired by Johann Bernoulli and Euler but which also surpasses both conceptually :

Reason forcibly demands that there must exist infinitely large numbers, which can only be meaningfully defined as follows:

An infinitely large natural number is a natural number, which cannot be grasped by an unlimited continuation of the process of counting, for counting is necessarily bound to the degree of finiteness.

Consequently, to arrive at such a “number”, it requires, figuratively speaking, a jump, or as a metaphor: infinitely large numbers form the “horizon of finiteness”.

The idea is:

If we conceive an infinitely large number called Ω (capital omega), reached by such a jump, obviously nothing speaks against but everything for it that we are allowed to calculate with such a “number” in just the same way as with a usual number.

Basically, Johann Bernoulli and Euler had shared this opinion. A little later Schmieden gives the following definition:

\(\omega =\frac {{1}}{{\Omega }}\) (small omega) is an infinitely small number which is smaller than each positive rational number and which is defined by the level of finite numbers. (ω is a “zero number”.)

In the same way as for any infinitely large number, there exists one of a larger order[Footnote 2], and there exists for any zero number one of a smaller order[2].

Therefore, our ordinary zero known from conventional analysis is only allowed to be used as another notation of the identity a ≡ a.

The latter would have pleased Frege (p. 10), and Weierstraß, too, would have given his approval in regard to such precision in relation to equality.

Finally, there is Schmieden’s foundational equation:

We define

$$\displaystyle \begin{aligned} \lim_{n\to\infty} n = \Omega\;\;; \end{aligned}$$

each other limit has to be related to this expression.

(The sign “” is not a good choice here and should have been avoided altogether.)

It is fairly evident that it is possible to construct from infinitely large natural numbers suitable rational numbers, e.g. \(\frac {{1}}{{\Omega }}\) or \(1+\frac {{1}}{{\Omega }}\) etc. The essential question is: Is this enough for analysis? After all, we now have “infinitely near” numbers, and this also in the neighbourhood of each rational number. In Schmieden’s words:

Around each ordinary rational number there exists an ω-sphere which does not include any other ordinary rational number, but which already contains infinitely many rational numbers of the second class [i.e. level] in the lowest ω-level.

With the help of the enlarged notion of rational numbers, which inevitably follows from the introduction of Ω, we reach the seemingly paradoxical result that analysis, if founded on Ω-numbers, gets along with these [Ω-]rational numbers, and even more—that no other numbers can occur, because each number defined via the usual limit arises in our system as an Ω-rational number.

Here we have the essence of Schmieden’s idea: instead of the real numbers (following Müller or Bertrand or Cantor or Dedekind), we take “Ω-rational” numbers, e.g. \(q\pm \frac {{1}}{{\Omega }}\) or \(q\pm \frac {{5\Omega ^2-7\Omega +3}}{{\Omega ^3-\frac {3}{4}}}\), etc. These can be used in calculations in the usual manner. The “Ω-rational” numbers should be enough because—as we know since Euler—analysis consists essentially of calculations, especially of calculations with the infinite.

Schmieden Dissolves the Paradoxes

It was the aim of Schmieden to dissolve such “paradoxes” as shown above with the help of more precise calculations—i.e. with his Ω-numbers. Schmieden states:

We especially inspect those instances where analysis puts up signs with the warning: here it is forbidden to calculate as in the finite case!

In what follows, representative examples of his method are shown.

Resolving paradoxI:

(p. 227).

Take the series

$$\displaystyle \begin{aligned} \ln 2 = \textstyle 1-\frac{{1}}{{2}}+\frac{{1}}{{3}}-\frac{{1}}{{4}}+\frac{{1}}{{5}}-\frac{{1}}{{6}}+\frac{{1}}{{7}}-\frac{{1}}{{8}}+-\ldots\end{aligned}$$

Schmieden now takes care in regard to the infinite (“+ −…”) and calculates precisely:

$$\displaystyle \begin{aligned} \ln 2 = \sum_{n=1}^\Omega \textstyle \frac{(-1)^{n-1}}{n} = \textstyle 1-\frac{{1}}{{2}}+\frac{{1}}{{3}}-\frac{{1}}{{4}}+\frac{{1}}{{5}}-\frac{{1}}{{6}}+\frac{{1}}{{7}}-\frac{{1}}{{8}}+-\ldots +\frac{{1}}{{\Omega-1}}-\frac{{1}}{{\Omega}} =: A\,.\end{aligned}$$

Here Ω is assumed to be even. But careful: “+ −…” has different meanings in the last two lines of formulae. (a) In the first line, “+ −…” indicates “and so on” without last term. (b) In the second line, “+ −…” marks a gap, like in \(1+\frac {{1}}{{2}}+\frac {{1}}{{4}}+\ldots +\frac {{1}}{{2^n}}\).

Now Schmieden rearranges the series A and supposes that Ω divided by 4 leaves no remainder. (Why not? If it is useful! At a pinch, make Ω = 4 Ω.) He then realizes the sum formula of the rearranged series:

We see:

It is true, A and B have the same positive terms (all odd denominators, including Ω − 1), but the second half of the negative terms of A (even denominators) are missing in B, up to Ω:

and we have

$$\displaystyle \begin{aligned} A=B-C\,. \end{aligned}$$

There you are! That is why we have B > A!

B is not at all a rearrangement of A, for it has infinitely fewer (negative) terms.

Schmieden even calculates the value of this difference C. This he does by integration—using the fact that in case of a continuous function and an infinitely small interval the estimated value of the area under this function only differs from the integral by an infinitely small amount. We denote this just like we did in the case of Johann Bernoulli (p. 4) with ≈ and observe that \(\omega =\frac {{1}}{{\Omega }}\) is infinitely small. We are going to calculate

Schmieden defines z = 2, consequently dz  = 2ω (as n always grows by 1), and therefore \(\frac {\omega }{2n\omega } = \frac {1}{2} \frac {\mathit {\,dz}\,}{z}\). But \(\frac {{1}}{{2n\omega }}\) is continuous (in the sense of ε-δ). Lower limit: \(z=2n\omega =2\left (\frac {\Omega }{4} +1\right )\cdot \omega = \frac {1}{2} + 2\omega \), upper limit: \(z=2n\omega =2\left (\frac {\Omega }{2}\right )\cdot \omega =1\), and therefore:

$$\displaystyle \begin{aligned} C \approx \textstyle \frac{{1}}{{2}}\cdot\int\limits_{z=\frac{{1}}{{2}}+2\omega}^1\textstyle \frac{\mathit{\,dz}\,}{z} \approx \textstyle \frac{{1}}{{2}}\cdot(\ln1-\ln\frac{{1}}{{2}}) = \textstyle \frac{{1}}{{2}}(\ln1 -\ln1+\ln2) = \textstyle \frac{{1}}{{2}}\ln2\,. \end{aligned}$$

So we get in fact

$$\displaystyle \begin{aligned} B-C&\approx \textstyle \frac{{3}}{{2}}\ln2-\frac{{1}}{{2}}\ln2=\ln2=A \\ \mathit{resp.}\qquad A-B&=-C\approx\textstyle -\frac{{1}}{{2}}\ln2\,, \end{aligned} $$

as we expected. Nothing of the kind \(\frac {{3}}{{2}}\ln 2=\ln 2\)!

We record:

The investigation of regular “rearrangements” of conditionally convergent series leads to comparable results: seemingly paradoxical equations of conventional analysis are explained in calculations with Ω-numbers through the omission of infinitely many terms with a finite sum.

Besides, the concept of “conditional” (i.e. not “absolute”) convergence is dispensable. Instead, all series are treated equally, like finite sums. (This would have pleased Weierstraß.)

Interestingly, in their publication from 1958, Schmieden and Laugwitz hint at the fact that “nevertheless, the difference between absolute and conditional convergence remains important for technical calculations, insofar as absolutely convergent series do allow arbitrary rearrangements without further considerations”. Obviously, this sentence is a concession by the authors and is probably due to the pressure exerted by the editors of the journal—because essentially it contradicts the foundational view of the authors: in their way of precisely calculating, all series are treated equally.

Resolving paradox II:

(p. 11).

Schmieden considers the C-1 Summation of the series

$$\displaystyle \begin{aligned} A= 1-1+1-1+1-1+-\ldots \end{aligned}$$

from p. 11. There we examined the expression

$$\displaystyle \begin{aligned} S^{C~1}= \lim_{n\to\infty} \frac{s_1+s_2+s_3+\ldots+s_n}{n}\,. \end{aligned}$$

Schmieden presents this series more conveniently via Ω instead of \(\lim \) and :

$$\displaystyle \begin{aligned} A^{C~1}= \frac{s_1+s_2+s_3+\ldots+s_\Omega}{\Omega}\,. \end{aligned}$$

This fraction he writes down in more detail. Thereby he pays attention to the following: (i) The first 1 of series A is included in all partial sums sk; this amounts to the first term being \(\frac {\Omega }{\Omega }\). (ii) The second 1 of series A is only included in Ω − 1 of the partial sums sk (i.e. not in s1); this gives the second term as \(\frac {\Omega -1}{\Omega }\); etc.

$$\displaystyle \begin{aligned} \\ {} A^{C~1}&= \textstyle \frac{\Omega}{\Omega}-\;\;\frac{\Omega-1}{\Omega}\quad \!\!+\quad \frac{\Omega-2}{\Omega}\quad -+ \ldots +(-1)^\Omega\ \frac{\Omega-(\Omega-2)}{\Omega} +\frac{(-1)^{\Omega+1}}{\Omega} \\ &=1-(1-\omega)+(1-2\omega)-+\ldots +(-1)^\Omega\cdot2\omega\quad \;\;+(-1)^{\Omega+1}\cdot1\omega \end{aligned} $$

He then considers the difference of A − AC 1 step-by-step, going from term to term, as follows:

All this is very accurate. Note the sum A has the value \(\frac {{1}}{{2}}\), but we have \(A^{C~1}=-\frac {{1}}{{2}}\) if Ω is even, if not we have \(A^{C~1}\approx \frac {{1}}{{2}}\). The sum A always has one of the values 1 or 0.—No paradox whatsoever.

The following question presents itself:

What constitutes the sum A C 1 really?

Schmieden just calculates. For the general

$$\displaystyle \begin{aligned} A= a_0+a_1+a_2+\ldots+a_{\Omega-1} \end{aligned}$$

he takes, as we have already seen,

$$\displaystyle \begin{aligned} A^{C~1}={\textstyle\frac{{1}}{{\Omega}}}\sum_{n=0}^{\Omega-1} s_n ={\textstyle\frac{{1}}{{\Omega}}}\left(\Omega a_0+(\Omega-1)a_1+(\Omega-2)a_2+\ldots+(\Omega-(\Omega-1))a_{\Omega-1}\right)\,. \end{aligned}$$

Especially, if

$$\displaystyle \begin{aligned} A(x)= 1+x+x^2+\ldots+x^{\Omega-1}, \end{aligned}$$

he substitutes an = xn and gets

Therefore, we have

$$\displaystyle \begin{aligned} s_\Omega(x)-A^{C~1}(x)=\omega\sum_{n=0}^{\Omega-1}(-1)^nnx^n =K\,. \end{aligned}$$

K generates the convergence of sΩ(x) to AC 1(x), but only in the infinite: As the factor ω in K shows, a finite number of terms of K will have sum ≈ 0.

Resolving paradox III :

(p. 224).

From the formula

$$\displaystyle \begin{aligned} s_n(x) = \frac{{1+(-1)^{n-1}x^n}}{{1+x}} \qquad \text{if}\qquad x\neq-1\,, \end{aligned}$$

Schmieden directly concludes

$$\displaystyle \begin{aligned} s_\Omega(x)=\textstyle \frac{{1}}{{1+x}}\left(1+(-1)^{\Omega-1} x^\Omega\right)\,. {} \end{aligned} $$
(§§)

1. If |x| < 1 and |x|≉1, we get xΩ ≈ 0, and consequently,

$$\displaystyle \begin{aligned} s_\Omega(x)\approx\textstyle \frac{{1}}{{1+x}}\,.\end{aligned}$$

2. If x = 1 , we get for an even or odd number  Ω of terms:

$$\displaystyle \begin{aligned} s_\Omega(1)=\textstyle \frac{{1}}{{2}}\left(1+(-1)^{\Omega-1}\right) = \begin{cases} 1\quad \text{if }\Omega\text{ is odd} \\ 0\quad \text{if }\Omega\text{ is even.} \end{cases} \end{aligned}$$

3. If x < 1 and x ≈ 1, the tricky xΩ is calculated by means of a trick. Like Euler, Schmieden takes for ν > 0:

$$\displaystyle \begin{aligned}x:=1-\xi\omega^\nu\end{aligned}$$

where ξ is finite and positive. He gets (remember from p. 9: instead of \(\lim \limits _{n\to \infty }\left (1+\frac {x}{n}\right )^n = e^x\), Schmieden has of course \(\left (1+\omega x\right )^\Omega \approx e^x\)) with the help of the following subsidiary calculation:

$$\displaystyle \begin{aligned} x^\Omega = \left(1-\xi\omega^\nu\right)^\Omega= \left(1-\omega\cdot\xi\omega^{\nu-1}\right)^\Omega \approx e^{-\xi\omega^{\nu-1}} \begin{cases} \approx1 & \text{if } \nu>1\,, \\ = e^{-\xi} & \text{if } \nu=1\,, \\ \approx 0 & \text{if } \nu<1\,, \end{cases} \end{aligned}$$

and finally for §§ (as 1 + x ≈ 2), the result

$$\displaystyle \begin{aligned} s_\Omega(x)\approx \begin{cases} \frac{{1}}{{2}}\left(1+(-1)^{\Omega-1}\right) & \text{if } \nu>1\,, \\ \frac{{1}}{{2}}\left(1+(-1)^{\Omega-1} e^{-\xi}\right) & \text{if } \nu=1\,, \\ \frac{{1}}{{2}} & \text{if } 0<\nu<1\,. \end{cases} \end{aligned}$$

The case |x| > 1 is judged by Schmieden as being “nonsensical” (Fig. 14.1).

Fig. 14.1
figure 1

The function sΩ(x) = 1 − x + x2 − +… in [0, 1], at right for

Everything is exactly explained: if ν = 1 and:

  1. (i)

    Ω is odd, the function sΩ(x) increases from the value \(\frac {{1}}{{2}}\) near x = 1 as e-function \(\frac {{1}}{{2}}\left (1+e^{-\xi }\right )\) up to the value 1 for x = 1 (i.e. up to ξ = 0).

  2. (ii)

    Ω is even, the function decreases from the value \(\frac {{1}}{{2}}\) near x = 1 as e-function \(\frac {{1}}{{2}}\left (1- e^{-\xi }\right )\) up to the value 0 for x = 1.

In other words: depending on whether Ω is odd or even, the value of the function near x = 1 decreases from \(\frac {{1}}{{2}}\) to 0 or increases from \(\frac {{1}}{{2}}\) to 1. This is by no means indeterminate! Of course, some calculation is needed—in mathematics, you do not get anything without work. The distinction between Ω being odd or even shows the different possibilities of the results of your calculation.

Resolving paradox IV:

(p. 13).

Only Eq. ‡‡ is thorny. Schmieden considers yet again the details and studies

$$\displaystyle \begin{aligned} f_\Omega(x)=\frac{{\Omega x}}{{1+\Omega^2 x^4}}\,. \end{aligned}$$

As before, he takes x = ξων and gets

$$\displaystyle \begin{aligned} f_\Omega\left(\xi\omega^\nu\right)= \frac{{\xi\omega^{\nu-1}}}{{1+\xi^4\omega^{4\nu-2}}} = \frac{{\xi\Omega^{1-\nu}}}{{1+\xi^4\Omega^{2-4\nu}}} \approx \begin{cases} 0 \qquad \text{if } \nu>1\,, \\ \xi \qquad \text{if } \nu=1\,, \\ \xi^{-3}\Omega^\alpha\,,(0<\alpha<2)\,, \;\;\;\text{if } \frac{{1}}{{3}}<\nu<1,\\ \qquad \qquad \text{as}\;\;\frac{{\xi\Omega^{1-\nu}}}{{1+\xi^4\Omega^{2-4\nu}}}\sim \Omega^{1-\nu-2+4\nu} \\ \qquad \qquad \qquad \qquad \qquad \quad =\Omega^{3\nu-1} \\ \xi^{-3} \,\quad \text{if } \nu=\frac{{1}}{{3}}\,, \\ 0 \qquad \text{if } 0<\nu<\frac{{1}}{{3}}\,. \end{cases} \end{aligned} $$

Therefore near 0 in the case \(\frac {{1}}{{3}}\leqq \nu \leqq 1\), we unambiguously obtain

$$\displaystyle \begin{aligned} f_\Omega \neq f\;! \end{aligned}$$

Schmieden also deals with delta-functions (indeed, if you know Ω-numbers, you have true delta-functions!). This would lead us astray, but one example of an elementary delta-function will nevertheless be shown below.

The First Formal Version of a Nonstandard-Analysis in the Year 1958

In 1954, the then young was handed a manuscript by Carl Friedrich von Weizsäcker (1912–2007) to report on it in his seminar in Göttingen. After Laugwitz’ affirmative reaction to it, Weizsäcker brought him into contact with the author Curt Schmieden. This was the beginning of a very fruitful cooperation, and in 1962, at the age of 29, Laugwitz took up a position as chair in mathematics at the University of Darmstadt.

Four years earlier, in 1958, and after some quarrels behind the scenes, the renowned journal Mathematische Zeitschrift had published an article by Schmieden and Laugwitz in its volume 69. Later it turned out to be the first publication in a field nowadays called “nonstandard-analysis”.

In it, the two authors expressively emphasize that:

It should be noticed that not only a rebuilding or a mere modification of the conventional analysis emerges here, but that a true enlargement is also produced.

Right at the start, they showed that their new analysis incorporates the species of ’s delta-functions as true “functions”, without the necessity of modifying the notion of a function. As an example of a delta-function, they presented

$$\displaystyle \begin{aligned} \delta(x)= \frac{{1}}{{\pi}}\cdot\frac{\Omega}{1+\Omega^2x^2}\,. \end{aligned}$$

The typical features of a delta-function are easily confirmed: δ(x) ≈ 0, except when x ≈ 0—where δ(x) is infinitely large, and \(\int \limits _{-\infty }^{+\infty }\delta (x){\mathit {\,dx}\,}=1\).

The Foundation in the Year 1958

Laugwitz formulated some of Schmieden’s principal ideas in the language of the then fashionable algebra. He referred to the definitions of Cantor. Cantor had defined convergent sequences of rational numbers as novel numbers, and he had named these sequences “limits” (p. 5). The duo Schmieden/Laugwitz proposed to take the totality of sequences of rational numbers and to define each of these sequences as an individual “number”—independent of their convergence, just: all these sequences! As a name for the sequences, they chose “Limit” (as Cantor did, but with a capital “L” as a distinction) as well as “Ω-number”.

$$\displaystyle \begin{aligned} (a_n)_n&= a_\Omega = \operatorname*{\mathrm{Lim\hspace{.3ex}}}_{n=\Omega} a_n\,, \\ \text{especially}\qquad (n)_n &= \Omega\,,\quad \textstyle (\frac{{1}}{{n}})_n=\omega \quad \text{and also}\quad \left((-1)^n\right)_n=(-1)^\Omega\,. \end{aligned} $$

That is to say, Schmieden’s “jump” from the process of counting 1, 2, 3, … to Ω (see p. 15) now obtains as its final destination the complete sequence Ω = (n)n! Even this divergent sequence defines an Ω-number, of course an infinitely large one.

The preceding number to Ω, i.e. Ω − 1, is the sequence (n − 1)n = (0, 1, 2, 3, …); its successor, i.e. Ω + 1, is the sequence (2, 3, 4, …), etc.

The discombobulating facts in favour of conventional analysis (that the new numbers comprise zero divisors and therefore do not constitute a field—and consequently division is not always possible—as well as the missing of their “total order”) were clearly articulated in this chapter, but not a single word of valuation was added.

For example, we have

$$\displaystyle \begin{aligned} (0,1,0,1,0,1,\ldots)\ \cdot\ (1,0,1,0,1,0,\ldots)\ =\ 0\,,\end{aligned}$$

although neither factor being 0 = (0, 0, 0, …). And, moreover, we are not able to call one of these two factors greater than the other—although obviously they are different.

Further Peculiarities of the New Analysis in the Year 1958

The authors also clearly identified those facts that appear in this new “enlargement” of analysis differently from in conventional analysis:

  • There are three different equivalence relations as orders of magnitude: 1. “finitely equal”, 2. “having the same magnitude” and 3. “having the same order of magnitude”.

  • Some theorems are stated and proved, which are wrong in conventional analysis but true in the new enlargement. 1. Every limit exists. This is just the new definition of “number”. In detail (in the case of sequences of Ω-numbers, we write their “index” above): The limit of a sequence of Ω-numbers \((a_\Omega ^{(p)})_p\) is defined for each “positive integer” as well as for each “infinitely large”Ω-number gΩ, i.e. gΩ = (gn)n where gn > gk > 0 if n > k. The limit

    $$\displaystyle \begin{aligned} \\ {} \operatorname*{\mathrm{Lim\hspace{.3ex}}}\limits_{p=g_\Omega} a_\Omega^{(p)} = a_\Omega^{(g_\Omega)}& = b_\Omega \end{aligned} $$

    is defined as the sequence of components

    $$\displaystyle \begin{aligned}\\ {} b_n&=a_n^{(g_n)}\,, \end{aligned} $$

    i.e. by the “diagonal sequence” defined from the sequence \((a_\Omega ^{(p)})_p\) of Ω-numbers \(a_\Omega ^{(p)}\).—Examples: (a) In case of the constant sequence \(a_\Omega =\left (\frac {{1}}{{n}}\right )_n = \omega \), we have \(\lim \limits _{p=\Omega } a_\Omega ^{(p)} =\left (\frac {{1}}{{n}}\right )_n = a_\Omega = \omega \), as it should be; and \(\lim \limits _{p=2\Omega } a_\Omega ^{(p)} =\left (\frac {{1}}{{2n}}\right )_n = \frac {{1}}{{2}}\omega \). (b) The sequence \(b_\Omega ^{(p)}\) with the pth term \(b_\Omega ^{(p)}=\left (\frac {{1}}{{p\cdot n}}\right )_n\) is \(\lim \limits _{p=\Omega } b_\Omega ^{(p)}=\left (\frac {{1}}{{n\cdot n}}\right )_n = \frac {{1}}{{\Omega \cdot \Omega }} = {\omega }^2\) and \(\lim \limits _{p=2\Omega } b_\Omega ^{(p)}=\left (\frac {{1}}{{2n\cdot n}}\right )_n = \frac {{1}}{{2\Omega \cdot \Omega }} = \frac {{1}}{{2}}{\omega }^2\)—right?

    2. Any two limits are interchangeable. The essential reason for this is that limits are only indicated but not calculated. In detail: Given a double sequence \((a_\Omega ^{p,q})_{p,q}\) of Ω-numbers \(a_\Omega ^{g_n,h_n}\), the limit is given by the sequence of the components \(a_n^{g_n,h_n}\). Example:

    $$\displaystyle \begin{aligned}a^{p,q}=\frac{{1}}{{1+\frac{p}{q}}}\,.\end{aligned}$$

    (The given Ω-numbers are constant sequences of rational numbers with the constant components .) Conventionally, we have

    $$\displaystyle \begin{aligned} \lim_{q\to\infty}\ \lim_{p\to\infty}\ a^{p,q}\ =\ 0\,,\qquad \text{but}\qquad \lim_{p\to\infty}\ \lim_{q\to\infty}\ a^{p,q}\ =\ 1\,. \end{aligned}$$

    However, for Ω-numbers, independently from the order, we have

    $$\displaystyle \begin{aligned} \operatorname*{\mathrm{Lim\hspace{.3ex}}}_{\genfrac{}{}{0pt}{}{p=g_\Omega}{q=h_\Omega}}\ a^{p,q} \ = \ \frac{{1}}{{1+\frac{g_\Omega}{h_\Omega}}}\,.\end{aligned}$$

    If you choose p = gΩ (conventionally: p →) for taking the first limit, you obtain

    $$\displaystyle \begin{aligned} \textstyle \frac{{1}}{{1+\frac{g_\Omega}{q}}}\,,\end{aligned}$$

    i.e. an infinitely small number for each finite q = 1, 2, 3, … Yet, if you then choose for the other limit hΩ = gΩ, you nevertheless get the finite value

    $$\displaystyle \begin{aligned} \textstyle \frac{{1}}{{1+\frac{g_\Omega}{h_\Omega}}} = \frac{{1}}{{1+\frac{g_\Omega}{g_\Omega}}} = \frac{1}{1+1} = \frac{1}{2}\,, \end{aligned}$$

    which says that the infinitely many small numbers \(\frac {{1}}{{1+\frac {g_\Omega }{q}}}\) nevertheless have a finite limit of the value \(\frac {{1}}{{2}}\).

    3. Divergent series are of equal rank to the others, as in the new version of analysis the calculations are the same: in the finite and in the infinite. As an example, the famous series:

    $$\displaystyle \begin{aligned} \sum_{p=1}^\Omega \frac{{1}}{{p(p+1)}} = \sum_{p=1}^\Omega \left(\frac{{1}}{{p}} - \frac{{1}}{{p+1}}\right) = \sum_{p=1}^\Omega \frac{{1}}{{p}} - \sum_{p=2}^{\Omega+1}\frac{{1}}{{p}} = 1 - \frac{{1}}{{\Omega+1}} \approx 1\,. \end{aligned}$$

    A convergent series on the left is split into the difference of two divergent series, and then this difference is calculated to be a finite value. Conventionally, an absolute no-go!

    4. The limit function of a sequence of continuous functions is continuous. The authors’ comment: “This theorem has no simple analogy in conventional analysis”. There is no reference made to the “Cauchy Sum Theorem” (p. 23). As an example, the sequence of functions xn, for the closed interval \(0\;{\scriptstyle \leqq }\; x\;{\scriptstyle \leqq }\; 1\), is given (already treated above on p. 26).

  • The Weierstraß Approximation Theorem was also obtained for the new version of analysis.

  • Differentiation and integration were briefly mentioned. However, a “differentiable” function does not need to have a derivative for each value! (“The ‘derivative’ might be ‘irrational’”.)

  • The Mean Value Theorem is proved (for “normal” functions).

Naturally, the notion of “function” is precarious. Of course, complete freedom in the sense of Bolzano (p. 12) could not be permitted. After all the starting point of Schmieden/Laugwitz is conventional analysis (which they call “usual”)—and not a completely new, extravagant shape of analysis. That is why they restrict themselves to such functions that are “already completely defined, if their values are given for the rational numbers of the domain”. Sensibly, the description of the values of the function has to be given in a standardized mathematical language.

Finale

was a practitioner of calculating. As a consequence, he had the idea of practicing analysis by calculating: no complicated general theorems that state the rules of calculation or rule out certain techniques of calculation—just calculate and, besides, treat the infinite (i.e. the specifically analytical) in just the same way as the finite. (He expressed it this way: “calculating analysis”.) The result of the calculation has to be assessed according to the requirements of the problem at the end of that calculation.

That this is permissible was shown to him by his calculations (“exactly as in the finite”—what should be wrong with this?) as well as by his results. To anchor his technique of “Omega Analysis”—an enlarged version of “Value Analysis”—reliably within the current foundational principles of mathematics, he left to others.

His enthusiastic colleague and co-worker gained credit for his precise algebraic foundation of Schmieden’s ideas. This was first published in the most general version in an article from 1958, which was signed by both of them.

In 1978, Laugwitz published a stronger algebraic construction, developed from Schmieden’s essential ideas, and in 1986, he presented a new version, which incorporates some features of formal logic.

Laugwitz was always aware that Schmieden’s Ω-numbers are not a field because one cannot divide by all non-zero numbers. This seemed to be a problem, especially at a time, when all mathematics was dominated by ’s idea of structure. Subsequently, in 1978, Laugwitz constructed, following others, a field \(^*\!K\) by algebraic means. Later, in 1986, Laugwitz pretended to, but did not clearly define a set \(^\Omega \!K\) with Ω-numbers. Instead of defining this, he changed to a formal language and developed a theory where “the rules of an ordered field” are valid.

These attempts to represent Schmieden’s ideas within the actual methodology of calculus could not do justice to Schmieden’s fundamental concerns. In a field that includes Ω-numbers, it must be settled, whether is a natural number or not. The mathematician might lack the means to decide which it is—but it will definitely be one of the two. But this does not fit to Schmieden’s essential idea. Schmieden really left open both possibilities—and if it were necessary, he distinguished both cases. Methodical stipulations of this kind—even more if they were bound to be unknown in every detail—were surely not his aim.

Foundational Problems

This first publication of a nonstandard analysis by these two men was soon followed by versions from other authors.

The first was by , a model theorist who also coined the name “nonstandard-analysis”. This approach to nonstandard analysis requires the knowledge of logics as a precondition to deal with continuous functions.

confined himself to algebraic constructions, essentially relying on ultra-filters for his definition of a field with “infinite” numbers.

In 1977, published a first axiom system for nonstandard analysis. Other versions followed.

This means that another mathematical theory (logics, universal algebra) is always needed to provide nonstandard analysis with suitable “numbers”. This is not satisfactory and is also a severe obstacle for an easy acceptance of this novel approach to analysis, unless you simply confine yourself to an axiom system.

Axiomatics

When inspecting the actual textbooks of analysis, it appears that there is no adequate consideration paid to the foundations. Generally, the real numbers are not introduced constructively, as was taught by Müller, and Dedekind or by Cantor and Heine, but instead, following Hilbert, axiomatically. This helps to save time.

Well, possibly, this fits excellently to the new role of mathematics as a security police for all sciences, which was promulgated by Hilbert in 1917 (p. 26) in order to teach the unconditional (at least provisional) acceptance of arbitrary axiom systems detached from any informal substantial justifications.

If it is so, there is then no argument to be put forward against the idea of taking an axiom system for nonstandard numbers (today usually called “hyper-real numbers”)—possibly apart from the fact that the elaboration of a nonstandard analysis differs in some respects from standard analysis. (This has been shown with the help of some “paradoxes” as well as generalized: see from p. 25.) This challenges one’s independent thought as well as impedes the usage of textbooks: you always have to check which version of analysis the chosen author likes. Standardization of the curriculum facilitates lecturing, especially if many students have to be educated. In small circles of specialists, subtle discussions will be eased.

A Path to Independency

Schmieden’s intention had been to rescue conventional analysis from its “paradoxes” as well as to give it a better form: fewer theorems, more computations.

An opposite development started in the 1980s. It continued for some decades in small circles. Under the (for the laymen astonishing) title “constructive nonstandard-analysis”, the theory was developed further as a special field in its own right in which more advanced methods (like sheafs and topoi) were implemented. This leads to a departure from the first origins of nonstandard analysis as well as from classical calculus, but it undoubtedly created some marvellous mathematics.

After the discovery of Weierstraß’ construction of the real numbers in 2016, it may well be possible to develop a new kind of approach to nonstandard analysis.

Nonstandard-Analysis and the History of Analysis

Not Schmieden and Laugwitz, but was the one who immediately started to relate the new theory to historical texts. The theorem nowadays called the “Cauchy Sum Theorem” (p. 23) came up fairly early for discussion. In his book from 1963, gave an interpretation of Cauchy’s writings. His result was: Cauchy’s theorem is correct if we add one of the two additional assumptions: (a) the series is uniformly convergent or (b) the family (sn(x))n of partial sums is equicontinuous in the interval.

’s idea of translating historical mathematical texts into the language of Nonstandard-analysis impressed the philosopher of science Imre Lakatos (1922–74). In a lecture, which appeared in print only after his death, he claimed that “Cauchy’s theorem was true and his proof as correct as an informal proof can be”. According to , the Cauchy Sum Theorem does therefore not need any additional assumptions to gain validity. Thus contradicted .—The further development (i.e. that was right in quite another sense from what he thought) was traced on pp. 15f .

A Satisfying Finish

It was the aim of my dissertation of 1981 to substantiate ’ thesis. In contrast to him, I based my arguments on the Darmstadt version of Nonstandard-analysis. Today I am aware that my approach (see p. 18) was mistaken.

My doctor father Laugwitz was intrigued by my work, and it inspired him to undertake his own detailed studies of former analysts, esp. of Cauchy’s works, thereby producing quite a few articles on the topic.

My studies of Bolzano’s mathematics, which started in 1986, showed me the inadequacy of ’ methodology in regard to the understanding of historical mathematical writings. As a consequence, I had to change my approach. In winter 1990, I turned again to the study of Cauchy’s analysis, this time from a new perspective. Prof Dr Laugwitz tried to hinder my work with a maximum of vigour. After he failed, I lost his favour and consequently all local academic support (and sadly, more than that). The Darmstadt University showed itself incapable of an unbiased clarification of this issue until today.

If Prof Dr Laugwitz had been factually successful at his time (not only institutionally), this book could not have been written.