www.egwald.com Egwald Web Services

Egwald Web Services
Domain Names
Web Site Design

Egwald Website Search JOIN US AS A FACEBOOK FAN Twitter - Follow Elmer WiensRadio Podcasts - Geraldos Hour

 

Statistics Programs - Econometrics and Probability Economics - Microeconomics & Macroeconomics Operations Research - Linear Programming and Game Theory Egwald's Mathematics Egwald's Optimal Control
Egwald HomeMathematics HomeLinear Algebra HomePolynomialsVectorsMatricesLinear EquationsLinear Differential EquationsNonlinear Dynamics HomeReferences & Links
 

Egwald Mathematics: Linear Algebra

Systems of Linear Equations

by

Elmer G. Wiens

Egwald's popular web pages are provided without cost to users.
Please show your support by joining Egwald Web Services as a Facebook Fan: JOIN US AS A FACEBOOK FAN
Follow Elmer Wiens on Twitter: Twitter - Follow Elmer Wiens

linear equations | matrix representation | non-homogeneous system | homogeneous system
general solution | inconsistent system | echelon algorithm | singular value algorithm
overdetermined systems | least squares solution | normal equations | svd least squares
square A matrix | A inverse solution | P*A = L*U solution | A = Q*R solution | svd solution | references

Systems of Linear Equations.

A system of linear equations can be expressed as:

a1,1 x1 + a1,2 x2   + . . . +   a1,n xn = b1

a2,1 x1 + a2,2 x2   + . . . +   a2,n xn = b2

                  . . . . . . . . . . .

am,1 x1 + am,2 x2   + . . . +   am,n xn = bm

Example.

6 * x1 + 2 * x2 + 2 * x3 = 6
4 * x1 + -3 * x2 + 5 * x3 = -4
1 * x1 + 6 * x2 + 4 * x3 = 3



Matrix Representation.

The same problem expressed in matrix and vector form is:

A *x   =   b

where A = [ai, j] is the matrix of coefficients, xT = (x1, x2, . . . . xn) is the vector of unknowns, and bT = (b1, b1, . . . . bm) is the vector of RHS (right hand side) coefficients.

If all the coefficients of the vector b are zero, the system A * x = b is called a homogeneous system.

Otherwise, the system is called nonhomogeneous.


Example: Nonhomogeneous System.

A =
622
4-35
164
  x =
x1
x2
x3
  b =
6
-4
3

The columns (and rows) of the matrix A are linearly independent, as confirmed by the following diagram, where I use the standard (x, y, z) coordinate system.

Please note that I distinguish between the axis x, and the vector xT = (x1, x2, x3). Also, note that while the vector xT is a row vector, and the vector x is a column vector, the diagrams below do not distinguish between column and row vectors.

Graphically, the matrix A, its column vectors, and the column vector b can be represented

Column Space of a Matrix A

Since the columns of A span R3, one can express any three dimensional vector, such as b in the example, as a linear combination of the columns of A. The solution to the linear equation problem, the vector

xT = (x1, x2, x3) = ( 1, 1, -1)

provides the coefficients of the linear combination of the columns of A that yield b.

A*x=b
622
4-35
164
*
1
1
-1
=
6
-4
3


Example: Homogeneous System.

Since the columns of A span R3, with the RHS vector bT = (0, 0, 0) the only solution is xT = (0, 0, 0), called the zero solution.

Any homogeneous system, (ie a homogeneous system with arbitrary matrix A and bT = oT the zero vector), always has at least one solution, the zero or trivial solution where xT = oT.

However, if the columns of A are not linearly independent, its homogeneous system has nontrivial solutions.


General Solution.

If xp is a particular solution of the nonhomogeneous system, and xh is a solution of the homogeneous system, then:

A * (xp + xh) = b + o = b,

proving that (xp + xh) is also a solution of the nonhomogeneous system.


The Image of a Matrix.

Let A be an m by n matrix. Write A = {a1, a2, ... an} as the set of m by 1 column vectors of A, ie A = [a1, a2, ... an].

The image of A, im(A), is the span of the columns of A. Formally:

im(A) = span(A) = {y in Rm | y = A * x,  for some x in Rn.

Sometimes this is referred to as the column space of the matrix A, denoted by R(A).


The Kernel of a Matrix A.

The kernel of A, ker(A), is the set of all vectors x in Rn for which A * x = o, the zero vector in Rm.  Formally:

ker(A) = {x in Rn | A * x = 0  in Rm}

Sometimes this is referred to as the null space of the matrix A, denoted by N(A).


The matrix A, and its transpose matrix AT (represented by A' in the diagrams below) generate four important subspaces: im(A), ker(A), im(AT), and ker(AT).


Example: General Solution with Linearly Dependent Columns (and Rows).

The matrix of the next linear equation problem has linearly dependent columns (and rows).

A*x=b
823
246
123
*
x1
x2
x3
=
4
-6
-3

The following diagram displays this linear equation problem. The image of the matrix A, Im(A), is the two dimensional yellow subspace of R3 generated by the columns of A.

Image of the Matrix A in Yellow

Because the vector bT = (4, -6, -3) lies within the Im(A), it can be expressed as a linear combination of the columns of A. The particular solution to the linear equation problems is xpT = (1, -2, 0), since:

A*xp=b
823
246
123
*
1
-2
0
=
4
-6
-3

The ker(A) is the line (also a subspace) through the vector xhT = (0, -6, 4), since xh is a solution of the homogeneous system for the matrix A:

A*xh=o
823
246
123
*
0
-6
4
=
0
0
0

Now let us look at AT, the transpose matrix of the matrix A.

The image of the matrix AT, Im(AT), is the two dimensional yellow subspace of R3 generated by the rows of A (the columns of AT).

Image of the Matrix A transpose in Yellow

As suggested by the diagram, ker(A) — the line through the vector xh — is perpendicular to the Im(AT) subspace.

Similarly, the kernel of the matrix AT, ker(AT) is the line through the vector yhT = (0, -3, 6), since yh is a solution of the homogeneous system for the matrix AT:

AT*yh=o
821
242
363
*
0
-3
6
=
0
0
0

The subspace ker(AT) is perpendicular to the subspace Im(A), as one can see two diagrams above.

Example: General Solution.

The general solution to the linear equation problem A * x = b is x = xp + s * xh for any scalar s. If x is defined by:

x=xp+ s * xh
x1
x2
x3
=
1
-2
0
+ s *
0
-6
4

then

A * x = A * xp + s * A * xh = b + s * o = b.


Example: Inconsistent System.

Consider the system with the same matrix A as in the example above, but with a RHS vector b that lies outside of the Im(A).

A*x=b
823
246
123
*
x1
x2
x3
=
4
-6
3

The following diagram displays this linear equation problem. The image of the matrix A, Im(A), is the same two dimensional yellow subspace of R3 generated by the columns of A.

Inconsistent System A*x = b

However, the vector bT = (4, -6, 3) lies outside the Im(A) subspace. Therefore, it cannot be expressed as a linear combination of the columns of A. Thus, this linear equation problem has no particular solution, although its homogeneous system has solutions consisting of each vector on the line through the vector xhT = (0, -6, 4).


Solutions: Inconsistent System.

In the diagram above, I projected the vector bT = (4, -6, 3) onto the Im(A) subspace, yielding the vector pT = (4, -3.6, -1.8). Recall that the point of the vector p is the closest point on Im(A) to the point of the vector b. Consequently, the vector (b - p) is perpendicular to the vector p. To verify this, note that the dot product pT * (b - p) = (4, -3.6, -1.8)T * (0, -2.4, 4.8) = (0, 0, 0)T.

A case can be made that the vector pT = (4, -3.6, -1.8) provides a "solution" to the inconsistent system A * x = b, if one solves the system A * x = p, instead.

In least-squares problems, with more equations than unknowns, the solution provided by the algorithms is the vector x such that A * x = p. That is A * x equals the projection of the vector b onto Im(A).

Writing the matrix A = [a1, a2, a3], one can see that a3 = 1.5 * a2, ie a2 and a3 are linearly dependent. Dropping the vector a2 yields the least-squares problem B * z = b given by:

B*z=b
83
26
13
*
z1
z2
=
4
-6
3

The least-squares solution is zT = (0.8286, -0.8762) since

B*z=p
83
26
13
*
0.82857
-0.87619
=
4
-3.6
-1.8

A least-squares problem always has a solution for B * z = p, since p is in the Im(B). Furthermore, if the columns of B are linearly independent, the solution vector z is unique.

To get back to the inconsistent system, write wT = (0.82857, 0, -0.87619). Then, w is a solution of A * x = p since

A*w=p
823
246
123
*
0.82857
0
-0.87619
=
4
-3.6
-1.8

However, the vector x = w + s * xh, where s is a scalar and xh is a solution of A * x = o, given by:

x=w+ s * xh
x1
x2
x3
=
0.8286
0
-0.8762
+ s *
0
-6
4

is also a solution to A * x = p,  because  A * (w + s * xh) = A * w + s * A * xh = p + s * o = p.


If small changes in the coefficients of the matrix A of a least-squares problem produce a matrix  with linearly dependent columns, the columns of A are "almost" linearly dependent. The estimates of the coefficients of the solution vector are tentative, since small changes in the data matrix A result in a system with a matrix  having many solutions.


Solution: Inconsistent System - Minimum Length.

One might choose as "optimal" the solution vector xT = (x1, x2, x3)T = w + s * xh with minimum length. The length (norm) of x = norm(w + s * xh) is:

norm(x) = sqrt(x21 + x22 + x23) = sqrt[841/1225+36*s^2+(-92/105+4*s)^2],

having a minimum for s+ = 0.0674. The solution of minimum length to the inconsistent system is:

x+ = w + s+ * xh = (0.8286, -0.4044, -0.6066).

The singular value decomposition algorithm provides x+ = A+ * b = (0.8286, -0.4044, -0.6066), where A+ is the pseudoinverse of A.


Echelon Algorithm.

The Echelon Algorithm is similar to the Gauss-Jordan method and the Gaussian method. The Echelon Algorithm permits one to calculate the general solution of a system of linear equations having more columns than rows. Begin by writing the matrix A and the vector b side by side as shown in Tableau0 below. This matrix arrangement is called the augmented matrix.

Using the following row operations on the tableaux of the augmented matrix, convert the left hand matrix to echelon form.

1. exchange two rows to get the largest entry (in absolute value) onto the diagonal.

2. divide a row by the diagonal entry.

3. add a multiple of one row to another row.

Begin in the upper left hand corner of the left hand matrix. Proceed along the diagonal converting each column into the appropriate column of the identity matrix as shown in the sequence of tableaux. Exchange rows with the lower row that has the largest entry (in absolute value) for the column. If all entries on and below the diagonal of the column are zeros, proceed to the next column.


Tableau0
 Ab
R01
R02
R03
6121-5
48106
1232
7
14
4

Tableau1
 Ãß
R11 = R01 / 6
R12 = R02 - (4) * R11
R13 = R03 - (1) * R11
120.1667-0.8333
009.33339.3333
002.83332.8333
1.1667
9.3333
2.8333

Tableau2
 Ãß
R21 = R11 - (0.1667) * R22
R22 = R12 / 9.3333
R23 = R13 - (2.8333) * R22
120-1
0011
0000
1
1
0

The last tableau contains the information one needs to obtain the general solution to the original linear equation problem A * x = b.

Echelon Tableau
Ã*x=ß
120-1
0011
0000
*
x1
x2
x3
x4
=
1
1
0

The solution to the Echelon Tableau yields the particular solution, xp, and the entire set of solutions of the associated homogeneous system, ie all vectors in the Ker(A).

Vectors in the Ker(A) are scalar multiples of the basis vectors, xh1, xh2, of the Ker(A) subspace.

x=xp + s1 * xh1 + s2 * xh2
x1
x2
x3
x4
=
1
0
1
0
+ s1 *
-2
1
0
0
+ s2 *
1
0
-1
1


Use the form below to solve some other linear equation problem.

Specify the dimensions of A (m <= n, 2 <= m,n) and fill in / change the numbers in the upper left hand corner, and then click "Solve".

Linear Equations Echelon Algorithm.

Matrix Ab
Row dimension of A:
m:  
Column dimension of A:
n:  

If you add a constant to an entry of the b vector, the system might be inconsistent. If you add a constant to an entry of a basis vector of Ker(A), you can create a matrix with more linearly independent columns.


Singular Value Decomposition Algorithm.

Consider the same linear equation problem A*x = b solved with the Echelon Algorithm.

A*x=b
6121-5
48106
1232
*
x1
x2
x3
x4
=
7
14
4

Using singular value decomposition yields the orthogonal matrices U and V, and the diagonal matrix S of singular values of A:

A * x = U * S * VT * x = b,

USVTx=b
-0.6580.75230.0342
-0.7268-0.6225-0.2903
-0.1971-0.21580.9563
17.9819000
010.800500
0000
-0.3922-0.7844-0.4737-0.0815
0.16740.3348-0.5667-0.734
-0.90450.4020.1005-0.1005
0-0.33330.6667-0.6667
x1
x2
x3
x4
=
=
=
=
7
14
4

Since the matrix S has 2 positive singular values, the rank of the matrix is 2. Furthermore, the columns of U and V provide orthonormal bases for the following subspaces:

SubspaceOrthonormal Bases
Im(A)first 2 columns of U
Ker(AT)last 1 columns of U
Im(AT)first 2 columns of V
Ker(A)last 2 columns of V

The Pseudoinverse is A+ = V * S+ * UT

A+ = V*S+*UT
0.0260.00620.001
0.0520.01240.0019
-0.02210.05180.0165
-0.04810.04560.0156
=
-0.39220.1674-0.90450
-0.78440.33480.402-0.3333
-0.4737-0.56670.10050.6667
-0.0815-0.734-0.1005-0.6667
*
0.055600
00.09260
000
000
*
-0.658-0.7268-0.1971
0.7523-0.6225-0.2158
0.0342-0.29030.9563

Both the columns and rows of A are linearly dependent. Setting x+ = A+ * b yields p = A * x+, the projection of b onto the Im(A) subspace.

x+ = A+ * b

x+=A+* b= 
x1
x2
x3
x4
=
=
=
=
0.0260.00620.001
0.0520.01240.0019
-0.02210.05180.0165
-0.04810.04560.0156
7
14
4
=
=
=
=
0.2727
0.5455
0.6364
0.3636

p = A * x+

p=A* x+= 
p1
p2
p3
=
=
=
6121-5
48106
1232
0.2727
0.5455
0.6364
0.3636
=
=
=
7
14
4

The RHS vector, b, lies inside the Im(A). Consequently, the system A * x = b is consistent. Therefore, p = b.

 


Overdetermined Systems
With # Rows m >= # Columns n

Suppose one wants to derive a linear equation model of the form A * x = b from a data set with a design (data) matrix A having more rows than columns. The columns of A and the column vector b represent measurable factors of the phenomena of interest. Each row of A represents one observation of the measured factors of the phenomena.

Since the matrix A has more rows than columns, one expects that A * x = b will be inconsistent, ie that b will lie outside the Im(A). Consequently, b will not be represented exactly as a linear combination of the columns of A.

I discuss least squares methods, also called multiple regression methods, on the Statistical web pages. Here I derive methods that statisticians and econometricians use to evaluate the least square (multiple regression) estimates for overdetermined systems of linear equations.


Least Squares (Multiple Regression) Solution.

With an inconsistent system, the least squares solution obtains a vector x+ from the data such that the vector p given by:

p = A * x+

is the projection of the vector b onto Im(A).

To be the projection of the vector b onto Im(A), p must be the vector in Im(A) that minimizes the distance between p and b, ie:

minimize sqrt{ (b1 - p1)2 + (b2 - p2)2 + .   .   . + (bn - pn)2 }

giving the name "least squares" to this solution.

As the number of observations increases, one hopes that ones understanding of the phenomena improves. This could be the case if the least squares estimated values of each parameter of the vector x+ converge as the number of observations increases.

Furthermore, one might deduce that a model of the form:

z = a1* x+ 1 + a2* x+ 2 +   .   .   + an*x+ n,

for a given instance of factors represented by the vector

aT = (a1, a2, .   .   .   an)

could be used to predict the measured value of the factor represented in the b column, if the a vector was observed.


Normal Equations.

The normal equations provide a direct method for obtaining the least squares parameters:

x+T = (x+1, x+2, .   .   .   x+n)

to the linear equations system A * x = b.

To reduce the dimensions of the system, multiply the equation A * x = b by the matrix AT, the transpose of A. The resulting equations are called the normal equations:

AT * A * x = AT * b.

Example.

As an experimenter, I want to summarize the data set with an equation in R3 space with standard coordinates (x, y, z):

z = x+1 * x + x+2 * y

described by the linear equations system A * x = b. The matrix A has two columns representing the independent variables (factors) of my model, x and y. The RHS vector b represents the dependent variable, z. I want to estimate the coefficients x+1 and x+2 using least squares. Perhaps, I want to use the equation to predict values of the dependent variable z for as yet unknown values of the independent variables x and y.

A*x=b
3.2-0.4
-6.60.9
8-2.9
1.4-6.6
-6.4-6.7
3.5-2.8
-1.76.9
3.2-3.9
*
x1
x2
=
0.5
-0.95
4.65
-2
-8.35
-1.75
4.4
-1.55

The normal equations system AT * A * x = AT * b is:

AT*A*x=AT*b
3.2-6.681.4-6.43.5-1.73.2
-0.40.9-2.9-6.6-6.7-2.86.9-3.9
*
3.2-0.4
-6.60.9
8-2.9
1.4-6.6
-6.4-6.7
3.5-2.8
-1.76.9
3.2-3.9
*
x1
x2
=
3.2-6.681.4-6.43.5-1.73.2
-0.40.9-2.9-6.6-6.7-2.86.9-3.9
*
0.5
-0.95
4.65
-2
-8.35
-1.75
4.4
-1.55

Multiplying by the matrix AT yields the reduced system with 2 equations and 2 unknowns:

(AT*A)*x=(AT*b)
186.1-30.79
-30.79168.49
*
x1
x2
=
77.145
95.91

The matrix (AT*A) is symmetric. If the data matrix A has linearly independent columns, (AT*A) has an inverse. The normal equations provide a system with a square coefficient matrix, solvable by standard methods.

Proceeding with the inverse of (AT*A):

x+=(AT*A)-1*(AT*b)= 
x+1
x+2
=
0.00550.001
0.0010.0061
*
77.145
95.91
=
0.5246
0.6651

The least squares solution using the normal equations is x+ = ( 0.5246, 0.6651 ).

The projection of b onto Im(A) is p = A * x+. Comparing p and b and computing b - p:

 a1a2 b-p=(b-p)
1
2
3
4
5
6
7
8
3.2
-6.6
8
1.4
-6.4
3.5
-1.7
3.2
-0.4
0.9
-2.9
-6.6
-6.7
-2.8
6.9
-3.9
 
0.5
-0.95
4.65
-2
-8.35
-1.75
4.4
-1.55
-
1.4126
-2.8636
2.2678
-3.6552
-7.8134
-0.0263
3.6974
-0.9152
=
-0.9126
1.9136
2.3822
1.6552
-0.5366
-1.7237
0.7026
-0.6348

The distance from Im(A) to b — the length of the vector (b-p) — is 4.131, found by computing the square root of the sums of the squares of the (b-p) column. (Im(A) is a subspace of R8, while b and p are vectors in R8)

For each row i of the data matrix A = [a1, a2], the triples (ai,1, ai,2, bi) and (ai,1, ai,2, pi) are points in R3 with standard coordinates (x, y, z). Furthermore, the least squares solution can be represented by the plane:

z = 0.5246 * x + 0.6651 * y

The plane is displayed below. If the point (ai,1, ai,2, bi) lies above the plane, a blue line joins it to the point (ai,1, ai,2, pi) on the plane. If (ai,1, ai,2, bi) lies below the plane, a red line joins it to (ai,1, ai,2, pi).

Least Squares System A*x = b

By clicking the refresh button on your browser, you can generate a new system of linear equations, with the new solution with least squares parameters x+ = (x+1, x+2) displayed above.

The functional form,

z = x+1 * x + x+2 * y,

ensures that the solution plane passes through the origin.

In the example of the next section, the estimated solution plane is not restricted to pass through the origin.


Singular Value Decomposition and Least Squares.

The "normal equations" least squares method of solving systems of linear equations computes the matrix (AT * A). For systems with many observations, the data matrix A has many rows. Multiplying two matrices using a computer incurs the risk of roundoff errors distorting the components of the matrix (AT * A) from their true values. Furthermore, the matrix (AT * A) may be ill-conditioned if the columns of A are nearly linearly dependent. Experimenters often include as many factors as possible in their design matrix A, hoping to improve the "explanatory power" of their model. The probability that subtle inter dependencies among the columns of A will render A nearly linearly dependent increases with the number of included factors.

The singular value decomposition method controls for such errors. Firstly, it controls roundoff errors by decomposing the data matrix A using Householder and Jacobi transformations, known for their stability. This decomposition is obtained without computing (AT * A). Secondly, singular value decomposition controls for linear dependence (and roundoff error) by setting very small singular values to zero.

Example.

The matrix A has two columns representing the independent variables (factors) of my model, x and y. The RHS vector b represents the dependent variable, z. As an experimenter, I want to summarize the data set with an equation in R3 space with standard coordinates (x, y, z):

z = x+1 + x+2 * x + x+3 * y

described by the linear equations system A * x = b. I need to estimate three coefficients x+1, x+2, and x+3 since I do not require the solution plane to pass through the origin. To obtain a z-intercept coefficient given by x+1, I append a column of ones to the front of the matrix A

 

A*x=b
17.52.1
14.3-3
1-2.2-7.4
12.95
1-2.11.8
1-7.71.3
1-5.53.7
1-2.8-7.6
*
x1
x2
x3
=
4.6
4.05
-4.4
3.55
2.85
-0.9
-2.1
-3.9

Singular value decomposition yields the orthogonal matrices U and V, and the diagonal matrix S of singular values of A:

A * x = U * S * VT * x = b,

USVTx=b
-0.52160.17610.4514
-0.14150.38580.3792
0.40860.40830.2789
-0.356-0.21550.4273
0.0604-0.21070.3516
0.4119-0.41430.2909
0.1918-0.48360.3319
0.45180.39640.2713
14.248300
012.51260
002.7843
0.0355-0.8472-0.5302
0.00340.5306-0.8476
0.99940.02830.0217
x1
x2
x3
=
=
=
4.6
4.05
-4.4
3.55
2.85
-0.9
-2.1
-3.9

Since the matrix S has 3 positive singular values, the rank of the matrix A is 3. Computing A+ = V * S+ * UT, the least squares solution to A * x = b is:

x+ = A+ * b.

x+=A+* b= 
x1
x2
x3
=
=
=
0.16080.13590.10120.15240.12630.10530.11950.0986
0.04310.0286-0.00420.0164-0.009-0.0391-0.0285-0.0073
0.011-0.0179-0.04070.03120.01480.0150.0282-0.0415
4.6
4.05
-4.4
3.55
2.85
-0.9
-2.1
-3.9
=
=
=
1.0152
0.4885
0.3991

The least squares solution using singular value decomposition is x+ = ( 1.0152, 0.4885, 0.3991 ).

The projection of b onto Im(A) is p = A * x+. Comparing p and b and computing b - p:

 a1a2a3 b-p=(b-p)
1
2
3
4
5
6
7
8
1
1
1
1
1
1
1
1
7.5
4.3
-2.2
2.9
-2.1
-7.7
-5.5
-2.8
2.1
-3
-7.4
5
1.8
1.3
3.7
-7.6
 
4.6
4.05
-4.4
3.55
2.85
-0.9
-2.1
-3.9
-
5.5168
1.9183
-3.0127
4.4272
0.7078
-2.2271
-0.1947
-3.3856
=
-0.9168
2.1317
-1.3873
-0.8772
2.1422
1.3271
-1.9053
-0.5144

The distance from Im(A) to b — the length of the vector (b-p) — is 4.281, found by computing the square root of the sums of the squares of the (b-p) column. (Im(A) is a subspace of R8, while b and p are vectors in R8)

For each row i of the data matrix A = [a1, a2, a3], the triples (ai,2, ai,3, bi) and (ai,2, ai,3, pi) are points in R3 with standard coordinates (x, y, z). Furthermore, the least squares solution can be represented by the plane:

z = 1.0152 + 0.4885 * x + 0.3991 * y

The plane is displayed below. If the point (ai,2, ai,3, bi) lies above the plane, a blue line joins it to the point (ai,2, ai,3, pi) on the plane. If (ai,2, ai,3, bi) lies below the plane, a red line joins it to (ai,2, ai,3, pi).

 

Least Squares System A*x = b

By clicking the refresh button on your browser, you can generate a new system of linear equations, with the new solution with least squares parameters x+ = (x+1, x+2, x+3) displayed above.



A Square Matrix.

If A is a square matrix of dimension n, the system of linear equations problem has of n equations in n unknowns. If the rank of A, rank(A), equals n, its columns and rows are linearly independent and the problem has a unique solution.

The variables x = [xi] that solve the system of linear equations A * x = b can be obtained by various methods.

You can adjust the matrix A and the RHS vector b in the form below to see how the different methods solve the problem. The available methods are based on the techniques for decomposing a matrix that I discussed on the matrices web page. These methods are the A inverse solution using the Gauss-Jordan method,   P*A = L*U solution using the Gaussian method, A = Q*R solution using Householder transformations, and the svd solution using singular value decomposition.

Solve A * x = b
*x1 + *x2 + *x3 + *x4 + *x5=
*x1 + *x2 + *x3 + *x4 + *x5=
*x1 + *x2 + *x3 + *x4 + *x5=
*x1 + *x2 + *x3 + *x4 + *x5=
*x1 + *x2 + *x3 + *x4 + *x5=
Square dimension of A:


The Gauss-Jordan Inverse Method

The Gausss-Jordan inverse method uses row operations on the initial tableau [ A  |  I ] to convert the RHS matrix to the identity matrix I and the LHS matrix to the inverse matrix A-1.

Solve A * x = b with A-1.

x=A-1* b= 
x1
x2
x3
x4
x5
=
=
=
=
=
2-310-0
-36-410
1-46-41
01-46-2.8
-001-2.81.64
310
325
356
405
475
=
=
=
=
=
1
1
1
1
1

 

Solution vector xT
11111


The Gaussian Decomposition Method

The Gaussian decomposition method factors the matrix A into the product of a lower triangular matrix L and an upper triangular matrix U, keeping track of row switches to construct a permutation matrix P so that P * A = L * U.

Solve A * x = b using P * A * x = L * U * x = P * b.

 

LUx=Pb
10000
0.73331000
0.74670.65100
0.78670.350.538510
0.86670.1250.19230.35711
758090105125
0-2.6667-7-12-16.6667
00-0.65-1.6-2.5
000-0.5385-1.1538
0000-0.3571
x1
x2
x3
x4
x5
=
=
=
=
=
00001
10000
01000
00100
00010
310
325
356
405
475

The solution vector x is obtained in two steps. First, solve for y in L * y = P * b by forward substitution:

Ly=P * b
10000
0.73331000
0.74670.65100
0.78670.350.538510
0.86670.1250.19230.35711
y1
y2
y3
y4
y5
=
=
=
=
=
475
310
325
356
405

Solution vector yT
475-38.33333-4.75-1.69231-0.35714

Next, solve for x in U * x =y by backward substitution:

Ux=y
758090105125
0-2.6667-7-12-16.6667
00-0.65-1.6-2.5
000-0.5385-1.1538
0000-0.3571
x1
x2
x3
x4
x5
=
=
=
=
=
475
-38.3333
-4.75
-1.6923
-0.3571

Solution vector xT
11111



The Householder QR Method

The Householder decomposition method factors the matrix A into the product of an orthonormal matrix Q and an upper triangular matrix R, such that A = Q * R.

Solve A * x = b using A * x = Q * R * x = b.

QRx=b
-0.3939-0.730.5585-00
-0.4011-0.3366-0.7228-0.4509-0
-0.42260.0152-0.27820.81060.2945
-0.46560.30460.06970.0751-0.8246
-0.53720.51070.2886-0.36610.483
-139.6138-146.626-161.0443-183.6639-215.7022
02.41436.51311.229515.5124
000.55851.51122.4155
000-0.4509-0.993
0-0000.2945
x1
x2
x3
x4
x5
=
=
=
=
=
310
325
356
405
475

Multiplying by QT gives R * x = QT * b.

Rx=QT * b
-139.6138-146.626-161.0443-183.6639-215.7022
02.41436.51311.229515.5124
000.55851.51122.4155
000-0.4509-0.993
0-0000.2945
x1
x2
x3
x4
x5
=
=
=
=
=
-846.6501
35.6692
4.4853
-1.444
0.2945

These equations are solved for x by back substitution yielding:

Solution vector xT
11111


The Singular Value Decomposition Method

The singular value decomposition method factors the matrix A into the product A = U * S* VT, where U consists of the eigenvectors of A * AT, V consists of the eigenvectors of AT * A, and S consists of the square roots of the eigenvalues of A * AT (which equal the eigenvalues of AT * A).

Solve A * x = b using A * x = U * S * VT * x = b.

USVTx=b
-0.36310.642-0.52340.3778-0.1982
-0.38160.44040.1746-0.60170.5176
-0.41960.10060.6398-0.0213-0.6356
-0.4791-0.27080.25180.61430.5063
-0.5629-0.5572-0.472-0.3427-0.1801
384.06740000
010.1448000
000.562300
0000.14880
00000.0767
-0.3631-0.3816-0.4196-0.4791-0.5629
0.6420.44040.1006-0.2708-0.5572
-0.52340.17460.63980.2518-0.472
0.3778-0.6017-0.02130.6143-0.3427
-0.19820.5176-0.63560.5063-0.1801
x1
x2
x3
x4
x5
=
=
=
=
=
310
325
356
405
475

 

Since the matrix S has 5 positive singular values, the rank of the matrix A is 5. Computing A+ = V * S+ * UT, the solution to A * x = b is:

x=A+* b= 
x1
x2
x3
x4
x5
=
=
=
=
=
2-310-0
-36-410
1-46-41
-01-46-2.8
-001-2.81.64
310
325
356
405
475
=
=
=
=
=
1
1
1
1
1

Solution vector xT
11111




References.

  • Ayres, Frank Jr. Matrices. New York: Schaum McGraw-Hill, 1962.
  • Ayres, Frank Jr. Modern Algebra. New York: Schaum McGraw-Hill 1965.
  • Bretscher, Otto. Linear Algebra with Applications. Upper Saddle River: Prentice Hall, 1997.
  • Burden, Richard L. and J. Douglas Faires. Numerical Analysis. 6th ed. Pacific Grove: Brooks/Cole, 1997.
  • Cohn, P. M. Linear Equations. London: Routledge, 1964.
  • Demmel, James W. Applied Numerical Linear Algebra. Philadelphia: Siam, 1997.
  • Dowling, Edward T. Mathematics for Economists. New York: Schaum McGraw-Hill, 1980.
  • Lipschutz, Seymour. Linear Algebra. New York: Schaum McGraw-Hill, 1968.
  • Mathews, John H. and Kurtis D. Fink. Numerical Methods Using MATLAB. 3rd ed. Upper Saddle River: Prentice Hall, 1999.
  • Press, William H., Brian P. Flannery, Saul A. Teukolsky, and William T. Vetterling. Numerical Recipes: The Art of Scientific Computing. Cambridge: Cambridge UP, 1989.
  • Strang, Gilbert. Linear Algebra and Its Applications. 3d ed. San Diego: Harcourt, 1976.
  • Varah, James. Numerical Linear Algebra: Computer Science 402 Lecture Notes. Vancouver: University of B.C., 2000.
  • Watkins, David S. Fundamentals of Matrix Computations. New York: John Wiley, 1991.

 

 

 

   

      Copyright © Elmer G. Wiens:   Egwald Web Services       All Rights Reserved.    Inquiries