Keywords

2010 Mathematics Subject Classification

1 The Alternate Scaling Algorithm

A positive matrix  is a matrix with positive coordinates. A nonnegative matrix  is a matrix with nonnegative coordinates. Let \(D = \text{ diag }(x_1,\ldots , x_n)\) denote the \(n\times n\) diagonal matrix with coordinates \(x_1,\ldots , x_n\) on the main diagonal. The diagonal matrix D is positive if its coordinates \(x_1,\ldots , x_n\) are positive. If \(A = (a_{i,j})\) is an \(m\times n\) positive matrix, if \(X = \text{ diag }(x_1,\ldots , x_m)\) is an \(m\times m\) positive diagonal matrix, and if \(Y= \text{ diag }(y_1,\ldots , y_n)\) is an \(n\times n\) positive diagonal matrix, then \(XA = (x_ia_{i,j})\), \(AY = (a_{i,j} y_j)\), \(XAY = (x_ia_{i,j} y_j)\) are \(m\times n\) positive matrices.

Let \(A = (a_{i,j})\) be an \(n \times n\) matrix. The ith row sum of A is

$$ \text{ rowsum }_i(A) = \sum _{j=1}^n a_{i,j}. $$

The jth column sum of A is

$$ \text{ colsum }_j(A) = \sum _{i=1}^n a_{i,j}. $$

The matrix A is row stochastic if it is nonnegative and \( \text{ rowsum }_i(A) = 1\) for all \(i \in \{1,\ldots , n\}\). The matrix A is column stochastic if it is nonnegative and \( \text{ colsum }_j(A) = 1\) for all \(j \in \{1,\ldots , n\}\). The matrix A is doubly stochastic if it is both row stochastic and column stochastic.

Let \(A = (a_{i,j})\) be a nonnegative \(n \times n\) matrix such that \( \text{ rowsum }_i(A) > 0\) and \( \text{ colsum }_j(A)>0\) for all \(i,j \in \{1,\ldots , n\}\). Define the \(n \times n\) positive diagonal matrix

$$ X(A) = \text{ diag } \left( \frac{1}{ \text{ rowsum }_1(A)}, \frac{1}{ \text{ rowsum }_2(A)},\ldots , \frac{1}{ \text{ rowsum }_n(A)} \right) . $$

Multiplying A on the left by X(A) multiplies each coordinate in the ith row of A by \( 1/ \text{ rowsum }_i(A)\), and so

$$ \left( X(A) A\right) _{i,j} = \frac{a_{i,j}}{ \text{ rowsum }_i(A)} $$

and

$$\begin{aligned} \text{ rowsum }_i\left( X(A) A\right)&= \sum _{j=1}^n (X(A) A)_{i,j} = \sum _{j=1}^n \frac{a_{i,j}}{ \text{ rowsum }_i(A)} \\&= \frac{ \text{ rowsum }_i(A)}{ \text{ rowsum }_i(A)} = 1 \end{aligned}$$

for all \(i \in \{1,2,\ldots , n\}\). The process of multiplying A on the left by X(A) to obtain the row stochastic matrix X(A)A is called row scaling. We have \(X(A) A = A\) if and only if A is row stochastic if and only if \(X(A) = I\). Note that the row stochastic matrix X(A)A is not necessarily column stochastic.

Similarly, we define the \(n \times n\) positive diagonal matrix

$$ Y(A) = \text{ diag } \left( \frac{1}{ \text{ colsum }_1(A)}, \frac{1}{ \text{ colsum }_2(A)},\ldots , \frac{1}{ \text{ colsum }_n(A)} \right) . $$

Multiplying A on the right by Y(A) multiplies each coordinate in the jth column of A by \(1/ \text{ colsum }_j(A)\), and so

$$ \left( AY(A) \right) _{i,j} = \frac{a_{i,j}}{ \text{ colsum }_j(A)} $$

and

$$\begin{aligned} \text{ colsum }_j(AY(A))&= \sum _{i=1}^n (A Y(A))_{i,j} = \sum _{i=1}^n \frac{a_{i,j}}{ \text{ colsum }_j(A)} \\&= \frac{ \text{ colsum }_j(A)}{ \text{ colsum }_j(A)} = 1 \end{aligned}$$

for all \(j \in \{1,2,\ldots , n\}\). The process of multiplying A on the right by Y(A) to obtain a column stochastic matrix AY(A) is called column scaling. We have \(AY(A) = A\) if and only if \(Y(A) = I\) if and only if A is column stochastic. The column stochastic matrix AY(A) is not necessarily row stochastic.

Let A be a positive \(n \times n\) matrix. Alternately row scaling and column scaling the matrix A produces an infinite sequence of matrices that converges to a doubly stochastic matrix This result (due to Brualdi, Parter, and Schnieder [1], Letac [3], Menon [4], Sinkhorn [7], Sinkhorn–Knopp [8], Tverberg [9], and others) is classical.

Nathanson [5, 6] proved that if A is a \(2\times 2\) positive matrix that is not doubly stochastic but becomes doubly stochastic after a finite number L of scalings, then L is at most 2, and the \(2\times 2\) row stochastic matrices that become doubly stochastic after exactly one column scaling were computed explicitly. An open question was to describe \(n \times n\) matrices with \(n \ge 3\) that are not doubly stochastic but become doubly stochastic after finitely many scalings. Ekhad and Zeilberger [2] discovered the following row-stochastic but not column stochastic \(3\times 3\) matrix, which requires exactly one column scaling to become doubly stochastic:

$$\begin{aligned} A = \left( \begin{matrix} 1/5 &{} 1/5 &{} 3/5 \\ 2/5 &{} 1/5 &{} 2/5 \\ 3/5 &{} 1/5 &{} 1/5 \end{matrix} \right) . \end{aligned}$$
(1)

Column scaling A produces the doubly stochastic matrix

$$ A Y(A)= \left( \begin{matrix} 1/6 &{} 1/3 &{} 3/6 \\ 2/6 &{} 1/3 &{} 2/6 \\ 3/6 &{} 1/3 &{} 1/6 \end{matrix} \right) . $$

The following construction generalizes this example. For every \(n \ge 3\), there is a two parameter family of row-stochastic \(n\times n\) matrices that require exactly one column scaling to become doubly stochastic

Let \(A = \left( \begin{matrix} a_{i,j} \end{matrix} \right) \) be an \(m \times n\) matrix. For \(i=1,\ldots , m\), we denote the ith row of A by

$$ \text {row}_i(A) = \left( \begin{matrix} a_{i,1}, a_{i,2}, \ldots , a_{i,n} \end{matrix} \right) . $$

Theorem 1

Let k and \(\ell \) be positive integers, and let \(n > \max (2k, 2\ell )\). Let x and z be positive real numbers such that

$$\begin{aligned} 0< x+z < \frac{1}{k} \qquad \text {and}\qquad x+z \ne \frac{2}{n} \end{aligned}$$
(2)

and let

$$\begin{aligned} y = \frac{x+z}{2} \qquad \text {and}\qquad w = \frac{1-k(x+z)}{n-2k}. \end{aligned}$$
(3)

The \(n \times n\) matrix A such that

$$ \text {row}_i(A) = {\left\{ \begin{array}{ll} (\underbrace{x,x,\ldots , x}_{k} \underbrace{w,w,\ldots , w}_{n-2k}\underbrace{z,z,\ldots , z}_{k} &{} \text { if } i \in \{1,2,\ldots , \ell \}\\ &{} \\ (\underbrace{y,y,\ldots , y}_{k} \underbrace{w,w,\ldots , w}_{n-2k}\underbrace{y,y,\ldots , y}_{k} &{} \text { if } i \in \{\ell +1, \ell + 2,\ldots , n -\ell \}\\ &{} \\ (\underbrace{z,z,\ldots , z}_{k} \underbrace{w,w,\ldots , w}_{n-2k}\underbrace{x,x,\ldots , x}_{k} &{} \text { if } i \in \{n - \ell +1,n - \ell + 2,\ldots , n \} \end{array}\right. } $$

is row stochastic but not column stochastic. The matrix obtained from A after one column scaling is doubly stochastic.

Proof

If

$$ i \in \{1,2,\ldots , \ell \} \cup \{n - \ell +1,n - \ell + 2,\ldots , n \} $$

then

$$ \text{ rowsum }_i(A) = k(x+z) + (n-2k)w = 1. $$

If

$$ i \in \{ \ell +1, \ell +2,\ldots , n - \ell \} $$

then

$$ \text{ rowsum }_i(A) = 2ky + (n-2k)w = 1. $$

Thus, the matrix A is row stochastic.

If

$$ j \in \{1, 2,\ldots ,k \} \cup \{ n-k+1, n- k + 2, \ldots , n\} $$

then

$$ \text{ colsum }_j(A) = \ell x + (n-2\ell ) y + \ell z =ny = \frac{n}{2}(x+z) \ne 1. $$

If

$$ j \in \{ k +1, k +2,\ldots , n - k\} $$

then

$$ \text{ colsum }_j(A) = nw \ne 1. $$

Thus, matrix A is not column stochastic.

The column scaling matrix for A is the positive diagonal matrix

$$\begin{aligned} Y(A)&= \text{ diag }\left( \underbrace{ \frac{1}{ny}, \ldots , \frac{1}{ny}}_{k}, \underbrace{ \frac{1}{nw}, \ldots , \frac{1}{nw} }_{n-2k}, \ \underbrace{ \frac{1}{ny}, \ldots , \frac{1}{ny}}_{k} \right) . \end{aligned}$$

For the column scaled matrix AY(A), we have the following row sums. If

$$ i \in \{1,2,\ldots , \ell \} \cup \{n - \ell +1,n - \ell + 2,\ldots , n \} $$

then

$$ \text{ rowsum }_i(AY(A)) = \frac{kx}{ny} + \frac{(n-2k)w}{nw} + \frac{kz}{ny} = \frac{k(x+z)}{ny} + 1 - \frac{2k}{n} = 1. $$

If

$$ i \in \{ \ell +1, \ell +2,\ldots , n - \ell \} $$

then

$$ \text{ rowsum }_i(A) = \frac{2ky}{ny} + \frac{(n-2k)w}{nw} = \frac{2k}{n} + 1 - \frac{2k}{n} = 1. $$

Thus, the matrix AY(A) is row stochastic. This completes the proof.    \(\square \)

For example, let \(k = \ell = 1\) and \(n = 3\), and let wxyz be positive real numbers such that

$$ 0< x+z < 1, \qquad x+z \ne \frac{2}{3} $$
$$ y = \frac{x+z}{2} \qquad \text {and}\qquad w = 1- x - z. $$

The matrix

$$\begin{aligned} A = \left( \begin{matrix} x &{} w &{} z \\ y &{} w &{} y \\ z &{} w &{} x \end{matrix} \right) , \end{aligned}$$
(4)

is row stochastic but not column stochastic. By Theorem 1, column scaling A produces a doubly stochastic matrix. Choosing \(x = 1/5\) and \(z = 3/5\), we obtain the matrix (1).

Here is another example. Let \(k = 2\), \(\ell = 3\), and \(n = 7\). Choosing

$$ x = \frac{1}{4}, \quad y = \frac{3}{16}, \quad z = \frac{1}{8}, \quad w = \frac{1}{12} $$

we obtain the row but not column stochastic matrix

$$ A = \left( \begin{matrix} 1/4 &{} 1/ 4 &{} 1/12 &{} 1/12 &{} 1/12 &{} 1/8 &{} 1/8 \\ 1/4 &{} 1/ 4 &{} 1/12 &{} 1/12 &{} 1/12 &{} 1/8 &{} 1/8 \\ 1/4 &{} 1/ 4 &{} 1/12 &{} 1/12 &{} 1/12 &{} 1/8 &{} 1/8 \\ 3/16 &{}3/16 &{} 1/12 &{} 1/12 &{} 1/12 &{} 3/16 &{} 3/16 \\ 1/8 &{} 1/8 &{} 1/12 &{} 1/12 &{} 1/12 &{} 1/4 &{} 1/ 4 \\ 1/8 &{} 1/8 &{} 1/12 &{} 1/12 &{} 1/12 &{} 1/4 &{} 1/ 4 \\ 1/8 &{} 1/8 &{} 1/12 &{} 1/12 &{} 1/12 &{} 1/4 &{} 1/ 4 \end{matrix} \right) . $$

Column scaling produces the doubly stochastic matrix

$$ AY(A) = \left( \begin{matrix} 4/21 &{} 4/21 &{} 1/7 &{} 1/7 &{} 1/7 &{} 2/21 &{} 2/21 \\ 4/21 &{} 4/21 &{} 1/7 &{} 1/7 &{} 1/7 &{} 2/21 &{} 2/21 \\ 4/21 &{} 4/21 &{} 1/7 &{} 1/7 &{} 1/7 &{} 2/21 &{} 2/21 \\ 1/7 &{} 1/7 &{} 1/7 &{} 1/7 &{} 1/7 &{} 1/7 &{} 1/ 7 \\ 2/21 &{} 2/21 &{} 1/7 &{} 1/7 &{} 1/7 &{} 4/21 &{} 4/21 \\ 2/21 &{} 2/21 &{}1/7 &{} 1/7 &{} 1/7 &{} 4/21 &{} 4/21 \\ 2/21 &{} 2/21 &{} 1/7 &{} 1/7 &{} 1/7 &{} 4/21 &{} 4/21 \end{matrix} \right) . $$

Theorem 2

Every \(n\times n\) matrix A constructed in Theorem 1 satisfies \(\det (A) = 0\).

Proof

There are three cases.

If \(k > 1\) or \(n-2k > 1\), then A has two equal columns and \(\det (A) = 0\).

If \(\ell > 1\) or \(n-2 \ell > 1\), then A has two equal rows and \(\det (A) = 0\).

If \(k = \ell = 1\) and \(n = 3\), then

$$ A = \left( \begin{matrix} x &{} w &{} z \\ y &{} w &{} y \\ z &{} w &{} x \end{matrix} \right) $$

and

$$ \det (A) = w(x-z)(x+z-2y) = 0. $$

This completes the proof.    \(\square \)

Theorem 2 is of interest for the following reason. Let \(A = \left( \begin{matrix} a_{i,j} \end{matrix} \right) \) be an \(n \times n\) matrix. If \(\det (A) \ne 0\), then the system of linear equations

$$\begin{aligned} a_{1,1} t_1 + a_{2,1}t_2 + \cdots + a_{n,1}t_n&= 1 \\ a_{1,2} t_1 + a_{2,2}t_2 + \cdots + a_{n,2}t_n&= 1 \\ \vdots&\\ a_{1,n} t_1 + a_{2,n}t_2 + \cdots + a_{n,n}t_n&= 1 \end{aligned}$$

has a unique solution. Equivalently, if \(\det (A) \ne 0\), then there exists a unique \(n \times n\) diagonal matrix \(T = \text{ diag }(t_1,\ldots , t_n)\) such that the matrix \(B = TA\) is column stochastic.

Suppose that the matrix A is positive and row stochastic. If \(t_i > 0\) for all \(i \in \{1,\ldots , n\}\), then T is invertible and \(B = TA\) is a positive column stochastic matrix. Setting \(X = T^{-1}\), we have \(XB = A\). Moreover, X is the row scaling matrix associated to B. Thus, if A is a row stochastic matrix such that column scaling A produces a doubly stochastic matrix, then we have pulled A back to a column stochastic matrix B, and we have increased by 1 the number of scalings needed to get a doubly stochastic matrix.

Unfortunately, the matrices constructed in Theorem 1 have determinant 0.

2 Open Problems

  1. 1.

    Does there exist a positive \(3\times 3\) row stochastic but not column stochastic matrix A with nonzero determinant such that A becomes doubly stochastic after one column scaling?

  2. 2.

    Let A be a positive \(3\times 3\) row stochastic but not column stochastic matrix that becomes doubly stochastic after one column scaling. Does \(\det (A) = 0\) imply that A has the shape of matrix (4)?

  3. 3.

    Here is the inverse problem: Let A be an \(n \times n\) row-stochastic matrix. Does there exist a column stochastic matrix B such that row scaling B produces A (equivalently, such that \(X(B) B= A\))? Compute B.

  4. 4.

    Modify the above problems so that the matrices are required to have rational coordinates.

  5. 5.

    Determine if, for positive integers \(L \ge 3\) and \(n \ge 3\), there exists a positive \(n \times n\) matrix that requires exactly L scalings to reach a doubly stochastic matrix.

  6. 6.

    Classify all matrices for which the alternate scaling algorithm terminates in finitely many steps.