Abstract
The distinguished econometrician Ragnar Frisch (1895–1973) also played an important role in optimization theory. In fact, he was a pioneer of interior-point methods. This note reconsiders his contribution, relating it to history and modern developments.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
This note considers just one contribution of Ragnar Frisch, Nobel Laureate in Economic Sciences 1969. It is, however, a most important one, belonging not to core economics, but rather to those parts of optimization theory that are most applied in economics.
The underlying issue goes broadly as follows. In optimization, linear equality constraints may help to reduce the number of variables. Not so, if constraints come as inequalities —as indeed, they often do. Variable elimination then becomes hard. So, why not incorporate sign restrictions by way of some barrier method? This was Frisch’s pioneering idea.Footnote 1 To enforce acceptable choice of nonnegative variables, [18] used the logarithmic potential ”as a guiding device—a sort of radar—to prevent us from hitting the boundary.”
In hindsight, his approach appears most natural. Yet it put him on a novel path, rich in consequences, but—at that time—distinctly beside the dominant orientation of optimization theory.
To date, his approach has impressed few economists,Footnote 2 and it took some time before it attracted wide attention in mathematics.Footnote 3 However, from the late 1980s onwards, his idea has been pursued by many—with much effort, great success, and due references.Footnote 4 Thus emerged an important, new class of optimization techniques, called interior-point methods, all somewhat tempered insofar as keeping each inequality strictly satisfied throughout the procedure.
Fiacco and McCormick [11] were chief inventors.Footnote 5 But certainly, [16, 17, 18, p.13] came first.Footnote 6 The purpose of this note is to set his contributions right, at the origin of barrier methods.
In reviewing parts of his approach, and briefly mentioning some subsequent developments, this note aims at reaching diverse readers. Included are economists and mathematicians, interested in applicable optimization and its history.Footnote 7 The material below is self-contained. It provides a general view, and is intended to be accessible with few prerequisites.
2 Frisch and optimization theory
Frisch’s interests revolved around econometrics [12] and many of his studies dealt with the dynamics, organization and planning of national development. In all, being well trained in mathematics, he saw optimization mainly as a supplement, especially useful to input-output analysis [31]. This is seen most clearly in his papers from 1955–1956. He had though, a strong and lasting interest in optimization per se.
Already his 1934 paper in Econometrica—a bit too long—on “national organization of a commodity and service exchange” featured a convex, linear-quadratic problem. Ahead of Stiegler (1945), [14] had a first rudimentary go at the diet problem, as reviewed by [40].Footnote 8 It appears that Frisch knew neither Kantorovich’s [26] work on production planning nor Hitchcock’s [24] study of product distribution. However, during the late 1940s and early 1950s he followed closely the development of linear programming—and its impacts on economic analysis—unfolding in [4, 8], and [30].Footnote 9
In parallel—during intense, fairly short spells—he worked on intimately related issues. To wit, from mid October 1953 until early May 1954 he issued ten memoranda at his Oslo institute of economics, all in Norwegian, all on linear programming; see [2]. In three early ones—dated Oct. 18, Nov. 13 and Jan. 14—he applied the ” logarithmic potential” to keep each nonnegative decision variable strictly positive.Footnote 10
By adding such highly nonlinear terms to the problem’s objective—whence inviting continuous optimization—he distinctly deviated from the linear programming of those days, dominated by Dantzig’s simplex algorithm.
As is well known, that algorithm has a discrete nature. It proceeds on the boundary of the polyhedron by jumping from one vertex to a better one—if any.Footnote 11 However, in the early 1970s, after long having enjoyed its practical efficiency, the optimization community became intrigued by the huge difference between its common and worst-case performances. An example of [29] unleashed a quest for solution procedures whose running times were polynomial functions of the number of bits needed to store the problem data. The first constructive steps in that direction were taken by [28] and [27]—neither of whom mentioned Frisch or barrier methods.Footnote 12 Their papers inspired development and practical use of polynomial-time methods. Central in that regard is Frisch’s logarithmic potential.
3 The problem and barrier methods
To review the originality of Frisch’s contributions it is convenient to let this section frame a suitable setting by way of the problem:
\(A\) is here a linear operator—alias a matrix—mapping a finite-dimensional vector space \(\mathbb {X}\) onto another such space \(\mathbb {Y}\). Both spaces are equipped with inner products \( \left\langle \cdot ,\cdot \right\rangle \). By standing assumption, the criterion \(c:\mathbb {X\rightarrow R}\) and the constraint functions \(c_{j}: \mathbb {X\rightarrow R}\) are concave and differentiable. The number \(\left| J\right| \) of inequalities is finite.
For interpretation, the vector \(x\in \mathbb {X}\) records inputs to a linear production process \(x\mapsto Ax,\) bound to bring out a prescribed vector \( y\in \mathbb {Y}\)—without violating any constraint \(c_{j}(x)\ge 0,\) \(j\in J\).
Henceforth suppose problem (1) be strictly feasible in that
Under that premise, fix a positive barrier parameter \(\beta \) and consider the modified problem:
Problem (3) remains concave and differentiable—and most important: it enforces strict satisfaction of the inequality constraints. Footnote 13 The Karush-Kuhn-Tucker optimality conditions for (3) are necessary and sufficient, and they read:
Here \(A^{*}\) is the transposed operator, mapping output prices \( y^{*}\) to input prices \(A^{*}y^{*}\).Footnote 14 Fix any pair \((x,y^{*})\) that satisfies conditions (4). Let
be the Lagrange function of the original problem (1). Introduce multipliers \(\lambda _{j}:=\beta /c_{j}(x)\) in (4) to see that, besides \(Ax=y,\) there holds
Looking at problem (1) anew, stationarity condition (5) says that \(x\) is a best response to price vectors \(\lambda ,y^{*}\). Moreover, for that same problem, the triple \((x,\lambda ,y^{*})\) provides an overestimate of the duality gap
The upshot is that an optimal solution to the “surrogate problem” (3) produces a primal-dual feasible choice for the original problem (1) as nearly optimal as desired.
Iterative restarts with smaller \(\beta \) entail a steadily decreasing duality gap. Versions of such so-called barrier methods were developed, analyzed and popularized in the 1960s by [10]. Their classic 1968 book, which referred to Frisch, was widely read, and their software much used; see [37] on their Sequential Unconstrained Minimization Technique.Footnote 15 Shanno [42] justly credits them as major inventors of interior-point primal methods.
These methods were long believed to suffer from ill-conditioning of Hessian matrices—a feature which may obstruct steepest ascent or quasi-Newton procedures.Footnote 16 Dire consequences have, however, rarely been documented. In particular, when stated in primal-dual form, ill-conditioning of interior-point methods is quite benign; see [49], or [38]. To see how that form enters, it is fitting to recall next how Frisch dealt with linear instances of problem ( 1).
4 Linear programming with the logarithmic potential
This section reviews how Frisch handled linear programming. It uses his format, subsumed by the more general problem (1).
Henceforth let \(\mathbb {X=R}^{J}\), \(c(x)=\left\langle x^{*},x\right\rangle ,\) \(c_{j}(x)=x_{j}\), and \(\mathbb {Y=R}^{I}\) for some finite set \(I\). Consequently, using standard inner products, problem (1 ) specializes to a linear program in canonical form:
\(A\) is an \(I\times J\) matrix with linearly independent rows. Hence the system \(Ax=y\) is solvable for some maximal subvector \(x_{B},\) \(B\subseteq J,\) \(\left| B\right| =\left| I\right| ,\) in terms of the remaining part \(x_{N}\), \(N:=J\diagdown B\). Thus emerges an equivalent equation system of the form
Substituting for \(x_{B}\) in the objective gives a reduced but affine criterion
Frisch [17] started with tableau (7), (8), referring to variables \(x_{B}\) as dependent, variables \(x_{N}\) as independent, and he called \(\left| N\right| \) the number of degrees of freedom.Footnote 17
In linear programming jargon, \(x_{B}\) comprises basic variables, \( x_{N}\) nonbasic variables, and \(r_{N}\) reduced prices. Note that \(x=(x_{B},0)\) becomes feasible for (6) if \(\chi \ge 0\). If moreover, \(r_{N}\le 0,\) the vertex \((x_{B},0)\) is optimal, yielding best value \(v\). Along the same line, still with \(\chi \ge 0,\) if some reduced price \(r_{n}\) remains positive, one may increase \(x_{n}\) beyond \(0\)—whence improve the objective—until some \(x_{b}\) (if any) hits its minimal level \( 0\). Thus a nonbasic \(n\) comes to replace a basic \(b\).Footnote 18
Frisch [17] proceeded, however, differently. Justified by strict feasibility (2) and equation (7) he considered the logarithmic potential
and its reduced gradient \(V^{\prime }(x_{N})\). If construed as a tentative direction of improvement, \(V^{\prime }(x_{N})\) competes with the reduced price vector \(r_{N}\). So, Frisch sought to strike “a wise compromise between these two directions”, setting
for some positive \(\beta \). Thus, in hindsight, modulo variable elimination, he would
and dealt thereby with a most important instance of problem (3). As vehicle he applied the reduced-form gradient (9). When the latter has a positive “projection” \( \left\langle r_{N},g_{N}\right\rangle \) onto the reduced price vector \(r_{N}, \) he recommended that \(x_{N}>0\) be updated with suitable stepsize along \( g_{N}\)—and that \(x_{B}>0\) thereafter be adjusted by (7). Clearly, whenever halted, the method furnishes a feasible solution alongside an overestimate of the duality gap.
Frisch saw the early days of computation and programming. At that time, he had no direct access to electronic computers. A main concern of his was to identify active constraints. Accordingly, his implementation—by requiring repeated, on-line judgements—now appears somewhat ad hoc. And his convergence analysis is incomplete. Moreover, he did not apply Newton methods. Nonetheless, he opened the door to subsequent developments as indicated next.
5 Enter Newton methods
By strict feasibility (2), primal problem (6) and its dual
are both feasible—with equal optimal values. The optimality conditions for (6) read:
For comparison, and in view of (4), the corresponding conditions for (10) take the simpler form:
it being understood that \(x>0\) is an implicit constraint.
Until the late 1980s, interior-point procedures for (6) used Newton methods to solve (11), letting \(\beta \) dwindle stepwise towards zero. However, around 1990, setting \(\lambda _{j}:=\beta /c_{j}(x)= \beta /x_{j}\), conditions (11) were reformulated equivalently, but more expediently, as
with the tacit understanding that all \(\lambda _{j},x_{j}>0\). Solving the last equations—called a primal-dual system —amounts to find a root of the function
For that purpose, Newton methods proceeds by iteratively updating the currently proposed point \((x,\lambda ,y^{*})=:z\) along a direction \( (\Delta x,\Delta \lambda ,\Delta y^{*})=:\Delta z\) that solves the linearized equation \(F(z+\Delta z)\approx F(z)+F^{\prime }(z)\Delta z=0\). Implementations go under the heading of (central) path-following methods; see [9, 23, 34], or [44]. Fine, but crucial details deal with the linear solver, step computation, and updating of the barrier parameter \(\beta \); conf. Vanderbei [38, 47].
6 Concluding remarks
In applied economics, what are the most important or frequent problems? Paul Samuelson—also a Nobel laureate in economics—surmised it is the efficient allocation of scarce resources towards competing ends. For those who agree, both in theory and practice, optimization becomes indispensable. Especially so if optimization produces endogenous prices—alias dual variables—to guide economic decisions or valuations.
Similarly, in applied mathematics, what are the most important or frequent problems? Most practitioners would contend that these amount to solve an equation system—or equivalently, to find a root of some specified function. For such purposes, Newton methods stand out as highly efficient. It is most appropriate therefore that such methods have come to occupy center stage in a large class of interior-point, optimization techniques.
It’s noteworthy that Sir Isaac Newton and Ragnar Frisch provided chief inputs to these businesses. No less noteworthy was the confluence of economics with mathematics during the 1950–1960s. Central economists included Arrow, Debreu, Dorfman, Frisch, Hurwicz, Koopmans, Leontief, Samuelson, Solow and others. Among the important mathematicians were Dantzig, Gale, Kuhn, Nash, Shapley, Tucker and von Neumann—to name a few.
During that period, linear programming was a focal point—and a great source of inspiration. In his marvellous little book, [22] expressed the belief that linear economic models would become central to economists’ training. We regret that he was wrong. The subject has largely dropped out of the curriculum at most economics departments. Yet, major economic issues fit within the frames of linear programs. Included are activity planning, complementarity, cooperative production games, financial hedging, input-output analysis, matching, security valuation, shadow pricing, substitution, theory of comparative advantages, two-person zero-sum games etc.—and a plethora generic applications. Moreover, the field is one where the economic problem remains master.
No wonder that Frisch immediately saw the great potential. Yet more of a wonder that, after his time, so much of economics have taken a largely narrative turn.
Notes
Many optimizers cite Frisch with an extra initial letter M—probably transmitted from the many papers or presentations in France of Monsieur Frisch [11] somewhat mysteriously cited Frisch, a misnomer which has propagated through citation trees.
After Frisch’s time, optimization has grown impressively, in theory and applications. But alas, it is now less emphasized in economists’ training.
See also [6].
Frisch, also a trained goldsmith and passionate bee-keeper (1952), was editor of Econometrica where the 1949 paper of Dantzig appeared.
The discrete nature of the simplex method has taken some hold. However, [5], Sect. 7.2, pages 156-160] also stressed its continuous nature, seen as a (reduced) gradient procedure.
For nice presentations of the Khachiyan and Karmarkar methods, as well as for historical remarks, see [41].
The logarithm also enters stochastic programming. There, constraints often come as reliability requirements-say, in the form \(\Pr \left\{ c(x)\ge \underline{c}\right\} \ge p>0\) where \(\underline{c}\) is a random vector. Then, it’s very convenient if that vector has log-concave distribution, to the effect that \(\ln \Pr \left\{ c(x)\ge \underline{c}\right\} \) becomes concave; see [39].
Henceforth, to assist economic intuition, the symbol \(*\) signals a price transform or variable.
Sonnevend [43] first proposed these specially for linear programs.
The editor of a conservative Oslo magazine felt compelled to inform Frisch, who had socialist inclinations, that there were competing concepts of freedom.
References
Arrow, K.J.: The work of Ragnar Frisch, econometrician. Econometrica 28(2), 175–192 (1960)
Bjerkholt, O.: Some unresolved problems of mathematical programming. In: Basu, D. (ed.) Economic models–methods, theory and applications, pp. 3–19. World Scientific, Singapore (2009)
Carroll, C.W.: The created response surface technique for optimizing nonlinear restrained systems. Oper. Res. 9(2), 169–185 (1961)
Dantzig, G.B.: Programming of interdependent activitities: II mathematical model. Econometrica 17, 200–211 (1949)
Dantzig, G.B.: Linear programming and extensions. Princeton University Press, Princeton (1963)
Dantzig, G.B.: The diet problem. Interfaces 20(4), 43–47 (1990)
Dikin, I.: Iterative solution of problems of linear and quadratic programming. Sov. Math. Dokl. 8, 674–675 (1967)
Dorfman, R., Samuelson, P.A., Solow, R.M.: Linear programming and economic analysis. McGraw-Hill, New York (1958)
Ferris, M., Mangasarian, O. L., Wright, S. J.: Linear programming with MATLAB, MPS-SIAM series on optimization (2007)
Fiacco, A.V., McCormick, G.P.: Computational algorithm for the sequential unconstrained minimization technique for nonlinear programming. Manag. Sci. 10(4), 601–617 (1964)
Fiacco, A.V., McCormick, G.P.: Nonlinear programming, sequential unconstrained minimization techniques, Research Analysis Corporation (1968) and SIAM (1990)
Frisch, R.: Problems and methods of econometrics. In: Bjerkholt, O., Dupont-Kieffer, A. (eds.) The Poincaré lectures 1933. Routledge, London (2009)
Frisch, R.: Circulation planning: proposal for a national organization of a commodity and service exchange. Econometrica 2, 258–336 and 422–435 (1934)
Frisch, R.: Introduction. In: Wold, K. (ed.) Kosthold og levestandard, en økonomisk undersøkelse (Nutrition and standard of living, an economic investigation), pp. 1–13. Fabritius og Sønners Forlag, Oslo (1941)
Frisch, R.: Das Ausleseproblem in der Bienenzüchtung. Zeitschrift für Bienenforschung 1, 7 (1952)
Frisch, R.: Principles of linear programming—with particular reference to the double gradient form of the logarithmic potential method. Memorandum from the Institute of Economics, University of Oslo, Oslo (1954)
Frisch, R.: The logarithmic potential method of convex programming with particular application to the dynamics of planning for national development. Memorandum from the Institute of Economics, University of Oslo, Oslo (1955)
Frisch, R.: La résolution des problèmes de programmes linéaires par la méthode du potential logarithmique, Cahiers du Séminaire D’Économetrie, No 4—Programme linéaire—Agrégation et nombre indices 7–23 (1956a)
Frisch, R.: Formulazione di un piano di sviluppo nazional come problema di programmazione convessa, L’industria (1956b)
Frisch, R.: Macroeconomics and linear programming. In: 25 Economic Essays in Honour of Erik Lindahl, Ekonomisk Tidskrift, pp. 38–67 Stockholm (1956c).
Frisch, R.: The multiplex method for linear programming. Sankhya 18(3&4), 329–362 (1957)
Gale, D.: The theory of linear economic models. The University of Chicago Press, Chicago (1960)
Gondzio, J., Terlaky, T.: A computational view of interior-point methods for linear programming. In: Beasley, J. (ed.) Advances in linear and integer programming, Ch. 3, pp. 103–144. Oxford University Press, Oxford (1996)
Hitchcock, F.L.: The distribution of a product from several sources to numerous localities. J. Math. Phys. 20, 224–230 (1941)
Huard, P.: Resolution of mathematical programming with nonlinear constraints by the method of centers. North-Holland, Amsterdam (1967)
Kantorovich, L.V.: A new method of solving some classes of extremal problems. Dokl. Akad Sci USSR 28, 211–214 (1940)
Karmarkar, N.: A new polynomial-time algorithm for linear programming. Combinatorica 4(4), 373–395 (1984)
Khachiyan, L.G.: A polynomial algorithm for linear programming. Sov. Math. Dokl. 20, 191–194 (1979)
Klee, V., Minty, G.J.: How good is the simplex algorithm, in inequalities III, pp. 159–172. Academic Press, New York (1972)
Koopmans, T.C. (ed.): Activity analysis of production and allocation, monograph 13 Cowles commision for research in economics. Wiley, Hoboken (1951)
Leontief, W.W.: Input-output economics, vol. 2. Oxford University Press, New York (1986)
Lootsma, F.A.: Hessian matrices of penalty functions for solving constrained optimization problems. Philips Res. Rep. 24, 322–331 (1969)
Luenberger, D.: Introduction to linear and nonlinear programming. Addison-Wesley, Boston (1984)
Marsten, R., Subramanian, R., Saltzman, M., Lustig, I., Shanno, D.: Interior point methods for linear programming: just call Newton, Lagrange, and Fiacco and McCormick. Interfaces 20(4), 105–116 (1990)
Meggido, N.: Pathways to the optimal set in linear programming. In: Meggido, N. (ed.) Progress in mathematical programming: interior-point and related methods, Chap 8, pp. 131–158. Springer, Berlin (1989)
Murray, W.: Analytical expressions for the eigenvalues and eigenvectors of the Hessian matrices of barrier and penalty functions. J. Optim. Theory Appl. 7, 189–196 (1971)
Nash, S.G.: SUMT (revisited). Oper. Res. 46, 763–775 (1998)
Nocedal, J., Wright, S.J.: Numerical optimization. Springer, Berlin (2006)
Prékopa, A.: Contributions to the theory of stochastic programming. Math. Program. 4, 202–221 (1973)
Sandmo, A.: Ragnar Frisch on the optimal diet. Hist. Polit. Econ. 25, 313–327 (1993)
Schrijver, A.: Theory of linear and integer programming. Wiley, New York (1986)
Shanno, D.: Who invented the interior-point method? Documenta mathematica, optimization stories, 21st ISMP Berlin, pp. 55–64 (2012)
Sonnevend, G.: An “analytic center” for polyhedrons and new classes of global algorithms for linear (smooth, convex) programming. In: Prekopa, A., Szelezsan, J., Strazicky, B.: (eds.) System modelling and optimization, Proceedings 12th IFIP Conference Lecture Notes in Control and Information Sciences 84, pp. 866–876 Springer, Berlin (1985)
Sporre, G.: On some properties of interior point methods for optimization, Doctoral Thesis, Royal Institute of Technology, Stockholm (2003)
Stiefel, E.: Note on Jordan elimination, linear programming and Tchebycheff approximation. Numerische Mathematik 2, 1–17 (1960)
Stiegler, G.J.: The cost of subsistence. J. Farm Econ. 27(2), 303–314 (1941)
Vanderbei, R.J.: Linear programming. Kluwer, Boston (1996)
van Eijk, C.J., Sandee, J.: Quantitative determination of an optimum economic policy. Econometrica 27, 1–13 (1959)
Wright, M.H.: Ill-conditioning and computational error in interior methods for nonlinear programming. SIAM J. Optim. 9, 84–111 (1999)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Bjerkholt, O., Flåm, S.D. Ragnar Frisch and interior-point methods. Optim Lett 9, 1053–1061 (2015). https://doi.org/10.1007/s11590-014-0807-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11590-014-0807-x