Gaussian_elimination
Gaussian_elimination
Using row operations to convert a matrix into reduced row echelon form is sometimes called Gauss–
Jordan elimination. In this case, the term Gaussian elimination refers to the process until it has reached
its upper triangular, or (unreduced) row echelon form. For computational reasons, when solving systems of
linear equations, it is sometimes preferable to stop row operations before the matrix is completely reduced.
Row operations
There are three types of elementary row operations which may be performed on the rows of a matrix:
Echelon form
For each row in a matrix, if the row does not consist of only zeros, then the leftmost nonzero entry is
called the leading coefficient (or pivot) of that row. So if two leading coefficients are in the same column,
then a row operation of type 3 could be used to make one of those coefficients zero. Then by using the row
swapping operation, one can always order the rows so that for every non-zero row, the leading coefficient
is to the right of the leading coefficient of the row above. If this is the case, then matrix is said to be in row
echelon form. So the lower left part of the matrix contains only zeros, and all of the zero rows are below
the non-zero rows. The word "echelon" is used here because one can roughly think of the rows being
ranked by their size, with the largest being at the top and the smallest being at the bottom.
For example, the following matrix is in row echelon form, and its leading coefficients are shown in red:
It is in echelon form because the zero row is at the bottom, and the leading coefficient of the second row
(in the third column), is to the right of the leading coefficient of the first row (in the second column).
A matrix is said to be in reduced row echelon form if furthermore all of the leading coefficients are equal
to 1 (which can be achieved by using the elementary row operation of type 2), and in every column
containing a leading coefficient, all of the other entries in that column are zero (which can be achieved by
using elementary row operations of type 3).
The table below is the row reduction process applied simultaneously to the system of equations and its
associated augmented matrix. In practice, one does not usually deal with the systems in terms of equations,
but instead makes use of the augmented matrix, which is more suitable for computer manipulations. The
row reduction procedure may be summarized as follows: eliminate x from all equations below L1, and
then eliminate y from all equations below L2. This will put the system into triangular form. Then, using
back-substitution, each unknown can be solved for.
The second column describes which row operations have just been performed. So for the first step, the x is
3
eliminated from L2 by adding 2 L1 to L2. Next, x is eliminated from L3 by adding L1 to L3. These row
operations are labelled in the table as
Once y is also eliminated from the third row, the result is a system of linear equations in triangular form,
and so the first part of the algorithm is complete. From a computational point of view, it is faster to solve
the variables in reverse order, a process known as back-substitution. One sees the solution is z = −1,
y = 3, and x = 2. So there is a unique solution to the original system of equations.
Instead of stopping once the matrix is in echelon form, one could continue until the matrix is in reduced
row echelon form, as it is done in the table. The process of row reducing until the matrix is reduced is
sometimes referred to as Gauss–Jordan elimination, to distinguish it from stopping after reaching echelon
form.
History
The method of Gaussian elimination appears – albeit without proof – in the Chinese mathematical text
Chapter Eight: Rectangular Arrays of The Nine Chapters on the Mathematical Art. Its use is illustrated in
eighteen problems, with two to five equations. The first reference to the book by this title is dated to
179 AD, but parts of it were written as early as approximately 150 BC.[1][2][3] It was commented on by Liu
Hui in the 3rd century.
According to Grcar[4] solution of linear equations by elimination was invented independently in several
cultures in Eurasia starting from antiquity and in Europe definite examples of procedure were published
already by late Renaissance (in 1550's). It is quite possible that already then the procedure was considered
by mathematicians elementary and in no need to explanation for professionals, so we may never learn its
detailed history except that by then it was practiced in at least several places in Europe.
The method in Europe stems from the notes of Isaac Newton.[5][6] In 1670, he wrote that all the algebra
books known to him lacked a lesson for solving simultaneous equations, which Newton then supplied.
Cambridge University eventually published the notes as Arithmetica Universalis in 1707 long after
Newton had left academic life. The notes were widely imitated, which made (what is now called) Gaussian
elimination a standard lesson in algebra textbooks by the end of the 18th century. Carl Friedrich Gauss in
1810 devised a notation for symmetric elimination that was adopted in the 19th century by professional
hand computers to solve the normal equations of least-squares problems.[7] The algorithm that is taught in
high school was named for Gauss only in the 1950s as a result of confusion over the history of the
subject.[8]
Some authors use the term Gaussian elimination to refer only to the procedure until the matrix is in
echelon form, and use the term Gauss–Jordan elimination to refer to the procedure which ends in reduced
echelon form. The name is used because it is a variation of Gaussian elimination as described by Wilhelm
Jordan in 1888. However, the method also appears in an article by Clasen published in the same year.
Jordan and Clasen probably discovered Gauss–Jordan elimination independently.[9]
Applications
Historically, the first application of the row reduction method is for solving systems of linear equations.
Below are some other important applications of the algorithm.
Computing determinants
To explain how Gaussian elimination allows the computation of the determinant of a square matrix, we
have to recall how the elementary row operations change the determinant:
Swapping two rows multiplies the determinant by −1
Multiplying a row by a nonzero scalar multiplies the determinant by the same scalar
Adding to one row a scalar multiple of another does not change the determinant.
If Gaussian elimination applied to a square matrix A produces a row echelon matrix B, let d be the product
of the scalars by which the determinant has been multiplied, using the above rules. Then the determinant
of A is the quotient by d of the product of the elements of the diagonal of B:
Computationally, for an n × n matrix, this method needs only O(n3) arithmetic operations, while using
Leibniz formula for determinants requires operations (number of summands in the formula times the
number of multiplications in each summand), and recursive Laplace expansion requires O(n 2n) operations if
the sub-determinants are memorized for being computed only once (number of operations in a linear combination
times the number of sub-determinants to compute, which are determined by their columns). Even on the fastest
computers, these two methods are impractical or almost impracticable for n above 20.
To find the inverse of this matrix, one takes the following matrix augmented by the identity and row-
reduces it as a 3 × 6 matrix:
By performing row operations, one can check that the reduced row echelon form of this augmented matrix
is
One can think of each row operation as the left product by an elementary matrix. Denoting by B the
product of these elementary matrices, we showed, on the left, that BA = I, and therefore, B = A−1. On the
right, we kept a record of BI = B, which we know is the inverse desired. This procedure for finding the
inverse works for square matrices of any size.
where the stars are arbitrary entries, and a, b, c, d, e are nonzero entries. This echelon matrix T contains
a wealth of information about A: the rank of A is 5, since there are 5 nonzero rows in T; the vector space
spanned by the columns of A has a basis consisting of its columns 1, 3, 4, 7 and 9 (the columns with
a, b, c, d, e in T), and the stars show how the other columns of A can be written as linear combinations
of the basis columns.
All of this applies also to the reduced row echelon form, which is a particular row echelon format.
Computational efficiency
The number of arithmetic operations required to perform row reduction is one way of measuring the
algorithm's computational efficiency. For example, to solve a system of n equations for n unknowns by
performing row operations on the matrix until it is in echelon form, and then solving for each unknown in
reverse order, requires n(n + 1)/2 divisions, (2n3 + 3n2 − 5n)/6 multiplications, and
(2n3 + 3n2 − 5n)/6 subtractions,[10] for a total of approximately 2n3/3 operations. Thus it has a
arithmetic complexity (time complexity, where each arithmetic operation take a unit of time, independently
of the size of the inputs) of O(n3).
This complexity is a good measure of the time needed for the whole computation when the time for each
arithmetic operation is approximately constant. This is the case when the coefficients are represented by
floating-point numbers or when they belong to a finite field. If the coefficients are integers or rational
numbers exactly represented, the intermediate entries can grow exponentially large, so the bit complexity
is exponential.[11] However, Bareiss' algorithm is a variant of Gaussian elimination that avoids this
exponential growth of the intermediate entries; with the same arithmetic complexity of O(n3), it has a bit
complexity of O(n5), and has therefore a strongly-polynomial time complexity.
Gaussian elimination and its variants can be used on computers for systems with thousands of equations
and unknowns. However, the cost becomes prohibitive for systems with millions of equations. These large
systems are generally solved using iterative methods. Specific methods exist for systems whose
coefficients follow a regular pattern (see system of linear equations).
Bareiss algorithm
The first strongly-polynomial time algorithm for Gaussian elimination was published by Jack Edmonds in
1967.[12]: 37 Independently, and almost simultaneously, Erwin Bareiss discovered another algorithm, based
on the following remark, which applies to a division-free variant of Gaussian elimination.
In standard Gaussian elimination, one subtracts from each row below the pivot row a multiple of
by where and are the entries in the pivot column of and respectively.
Bareiss variant consists, instead, of replacing with This produces a row echelon form that
has the same zero entries as with the standard Gaussian elimination.
Bareiss' main remark is that each matrix entry generated by this variant is the determinant of a submatrix
of the original matrix.
In particular, if one starts with integer entries, the divisions occurring in the algorithm are exact divisions
resulting in integers. So, all intermediate entries and final entries are integers. Moreover, Hadamard's
inequality provides an upper bound on the absolute values of the intermediate and final entries, and thus a
bit complexity of using soft O notation.
Moreover, as an upper bound on the size of final entries is known, a complexity can be obtained
with modular computation followed either by Chinese remaindering or Hensel lifting.
As a corollary, the following problems can be solved in strongly polynomial time with the same bit
complexity:[12]: 40
Numeric instability
One possible problem is numerical instability, caused by the possibility of dividing by very small numbers.
If, for example, the leading coefficient of one of the rows is very close to zero, then to row-reduce the
matrix, one would need to divide by that number. This means that any error which existed for the number
that was close to zero would be amplified. Gaussian elimination is numerically stable for diagonally
dominant or positive-definite matrices. For general matrices, Gaussian elimination is usually considered to
be stable, when using partial pivoting, even though there are examples of stable matrices for which it is
unstable.[13]
Generalizations
Gaussian elimination can be performed over any field, not just the real numbers.
Computing the rank of a tensor of order greater than 2 is NP-hard.[14] Therefore, if P ≠ NP, there cannot
be a polynomial time analog of Gaussian elimination for higher-order tensors (matrices are array
representations of order-2 tensors).
Pseudocode
As explained above, Gaussian elimination transforms a given m × n matrix A into a matrix in row-
echelon form.
In the following pseudocode, A[i, j] denotes the entry of the matrix A in row i and column j with the
indices starting from 1. The transformation is performed in place, meaning that the original matrix is lost
for being eventually replaced by its row-echelon form.
while h ≤ m and k ≤ n
/* Find the k-th pivot: */
i_max := argmax (i = h ... m, abs(A[i, k]))
if A[i_max, k] = 0
/* No pivot in this column, pass to next column */
k := k + 1
else
swap rows(h, i_max)
/* Do for all rows below pivot: */
for i = h + 1 ... m:
f := A[i, k] / A[h, k]
/* Fill with zeros the lower part of pivot column: */
A[i, k] := 0
/* Do for all remaining elements in current row: */
for j = k + 1 ... n:
A[i, j] := A[i, j] - A[h, j] * f
/* Increase pivot row and column */
h := h + 1
k := k + 1
This algorithm differs slightly from the one discussed earlier, by choosing a pivot with largest absolute
value. Such a partial pivoting may be required if, at the pivot place, the entry of the matrix is zero. In any
case, choosing the largest possible absolute value of the pivot improves the numerical stability of the
algorithm, when floating point is used for representing numbers.[15]
Upon completion of this procedure the matrix will be in row echelon form and the corresponding system
may be solved by back substitution.
See also
Fangcheng (mathematics)
Gram–Schmidt process - another process for bringing a matrix into some canonical form.
Fourier–Motzkin elimination - an algorithm for eliminating variables of a system of linear
inequalities, rather than equations.
References
1. "DOCUMENTA MATHEMATICA, Vol. Extra Volume: Optimization Stories (2012), 9-14" (http
s://www.emis.de/journals/DMJDMV/vol-ismp/10_yuan-ya-xiang.html). www.emis.de.
Retrieved 2022-12-02.
2. Calinger 1999, pp. 234–236
3. Timothy Gowers; June Barrow-Green; Imre Leader (8 September 2008). The Princeton
Companion to Mathematics (https://archive.org/details/princetoncompanio00gowe). Princeton
University Press. p. 607. ISBN 978-0-691-11880-2.
4. Grcar 2011a, pp. 169–172
5. Grcar 2011a, pp. 169–172
6. Grcar 2011b, pp. 783–785
7. Lauritzen, p. 3
8. Grcar 2011b, p. 789
9. Althoen, Steven C.; McLaughlin, Renate (1987), "Gauss–Jordan reduction: a brief history",
The American Mathematical Monthly, 94 (2), Mathematical Association of America: 130–142,
doi:10.2307/2322413 (https://doi.org/10.2307%2F2322413), ISSN 0002-9890 (https://search.
worldcat.org/issn/0002-9890), JSTOR 2322413 (https://www.jstor.org/stable/2322413)
10. Farebrother 1988, p. 12
11. Fang, Xin Gui; Havas, George (1997). "On the worst-case complexity of integer Gaussian
elimination" (https://scholar.archive.org/work/2htta67odrfilhxegzjqfratyy). Proceedings of the
1997 international symposium on Symbolic and algebraic computation. ISSAC '97. Kihei,
Maui, Hawaii, United States: ACM. pp. 28–31. doi:10.1145/258726.258740 (https://doi.org/1
0.1145%2F258726.258740). ISBN 0-89791-875-4.
12. Grötschel, Martin; Lovász, László; Schrijver, Alexander (1993), Geometric algorithms and
combinatorial optimization (https://books.google.com/books?id=hWvmCAAAQBAJ&pg=PA28
1), Algorithms and Combinatorics, vol. 2 (2nd ed.), Springer-Verlag, Berlin, doi:10.1007/978-
3-642-78240-4 (https://doi.org/10.1007%2F978-3-642-78240-4), ISBN 978-3-642-78242-8,
MR 1261419 (https://mathscinet.ams.org/mathscinet-getitem?mr=1261419)
13. Golub & Van Loan 1996, §3.4.6
14. Hillar, Christopher; Lim, Lek-Heng (2009-11-07). "Most tensor problems are NP-hard".
arXiv:0911.1393 (https://arxiv.org/abs/0911.1393) [cs.CC (https://arxiv.org/archive/cs.CC)].
15. Kurgalin, Sergei; Borzunov, Sergei (2021). "Algebra and Geometry with Python" (https://doi.o
rg/10.1007/978-3-030-61541-3). SpringerLink. doi:10.1007/978-3-030-61541-3 (https://doi.or
g/10.1007%2F978-3-030-61541-3). ISBN 978-3-030-61540-6.
Works cited
Atkinson, Kendall A. (1989), An Introduction to Numerical Analysis (2nd ed.), New York: John
Wiley & Sons, ISBN 978-0471624899.
Bolch, Gunter; Greiner, Stefan; de Meer, Hermann; Trivedi, Kishor S. (2006), Queueing
Networks and Markov Chains: Modeling and Performance Evaluation with Computer Science
Applications (2nd ed.), Wiley-Interscience, ISBN 978-0-471-79156-0.
Calinger, Ronald (1999), A Contextual History of Mathematics, Prentice Hall, ISBN 978-0-02-
318285-3.
Farebrother, R.W. (1988), Linear Least Squares Computations, STATISTICS: Textbooks and
Monographs, Marcel Dekker, ISBN 978-0-8247-7661-9.
Lauritzen, Niels, Undergraduate Convexity: From Fourier and Motzkin to Kuhn and Tucker.
Golub, Gene H.; Van Loan, Charles F. (1996), Matrix Computations (3rd ed.), Johns Hopkins,
ISBN 978-0-8018-5414-9.
Grcar, Joseph F. (2011a), "How ordinary elimination became Gaussian elimination", Historia
Mathematica, 38 (2): 163–218, arXiv:0907.2397 (https://arxiv.org/abs/0907.2397),
doi:10.1016/j.hm.2010.06.003 (https://doi.org/10.1016%2Fj.hm.2010.06.003),
S2CID 14259511 (https://api.semanticscholar.org/CorpusID:14259511)
Grcar, Joseph F. (2011b), "Mathematicians of Gaussian elimination" (https://www.ams.org/not
ices/201106/rtx110600782p.pdf) (PDF), Notices of the American Mathematical Society, 58
(6): 782–792
Higham, Nicholas (2002), Accuracy and Stability of Numerical Algorithms (2nd ed.), SIAM,
ISBN 978-0-89871-521-7.
Katz, Victor J. (2004), A History of Mathematics, Brief Version, Addison-Wesley, ISBN 978-0-
321-16193-2.
Kaw, Autar; Kalu, Egwu (2010). "Numerical Methods with Applications: Chapter 04.06
Gaussian Elimination" (http://mathforcollege.com/nm/mws/gen/04sle/mws_gen_sle_txt_gaus
sian.pdf) (PDF) (1st ed.). University of South Florida. Archived (https://web.archive.org/web/2
0120907224427/http://mathforcollege.com/nm/mws/gen/04sle/mws_gen_sle_txt_gaussian.p
df) (PDF) from the original on 2012-09-07.
Lipson, Marc; Lipschutz, Seymour (2001), Schaum's outline of theory and problems of linear
algebra, New York: McGraw-Hill, pp. 69–80, ISBN 978-0-07-136200-9
Press, WH; Teukolsky, SA; Vetterling, WT; Flannery, BP (2007), "Section 2.2" (https://web.arc
hive.org/web/20120319193237/http://apps.nrbook.com/empanel/index.html?pg=46),
Numerical Recipes: The Art of Scientific Computing (3rd ed.), New York: Cambridge
University Press, ISBN 978-0-521-88068-8, archived from the original (http://apps.nrbook.co
m/empanel/index.html?pg=46) on 2012-03-19, retrieved 2011-08-08
External links
Interactive didactic tool (http://pnp.mathematik.uni-stuttgart.de/igt/eiserm/lehre/Gael/)