##
The Dual of the Least-Squares Method

The least-squares method was firmly established as a scientific approach by Gauss, Legendre and Laplace within the space of a decade, at the beginning of the nineteenth century. Legendre was the first author to name the approach, in 1805, as “méthode des moindres carrés,” a “least-squares method.” Gauss, however, is credited to have used it as early as 1795, when he was 18 years old. He, subsequently, adopted it in 1801 to calculate the orbit of the newly discovered planet Ceres. Gauss published his way of looking at the least-squares approach in 1809 and gave several hints that the least-squares algorithm was a minimum variance linear estimator and that it was derivable from maximum likelihood considerations. Laplace wrote a very substantial chapter about the method in his fundamental treatise on probability theory published in 1812. Surprisingly, there still remains an unexplored aspect of the least-squares method: since the traditional
formulation is stated as minimizing the sum of squared deviations subject to the linear (or nonlinear) specification of a regression model, this mathematical programming problem
must have a dual counterpart. This note fills this gap and shows that the least-squares estimates of unknown parameters and deviations can be obtained by maximizing the net value of sample information.

Publication Type:

Working or Discussion Paper

PURL Identifier:

http://purl.umn.edu/121693

JEL Codes:

C20; C30

Series Statement:

Working Papers

12-001

Record created 2017-04-01, last modified 2017-05-27