Nipissing University
TUTORIALS
TUTORIALS HOME

GENERAL MATH
NOTATION & METHODS OF PROOF
INDUCTION
COMPLEX NUMBERS
POLYNOMIALS

LINEAR ALGEBRA
VECTORS
SYSTEM OF LINEAR EQUATIONS
MATRICES
EIGENVALUES & EIGENVECTORS
ORTHOGONALITY
VECTOR SPACE
DISTANCE & APPROXIMATION

MAIN
HOME
TESTS
TUTORIALS
SAMPLE PROBLEMS
COMMON MISTAKES
STUDY TIPS
GLOSSARY
APPLICATIONS
MATH HUMOUR

SYSTEMS OF LINEAR EQUATIONS TUTORIAL

Introduction

Linear equations are equations of the form a1x1 + a2x2 + ... + anxn = b, where a1...an are constants called coefficients, x1...xn are variables, and b is the constant term. This equations is a linear equation in n variables. The variables are also often seen as x, y, z, etc. Linear equations involve only variables of degree 1 (no exponents or roots of the variables), and they involve only sums and constant multiples of variables, without any products of variables.

The following equations are linear:

The following equations are NOT linear:

A solution of a linear equation is a vector composed of values [s1,...,sn], where when we can substitute x1 = s1, ..., xn = sn to obtain a correct equation. An equation in one variable (example, 2x = 6) has only one solution (a point on the real number line, in this case x = 6). A linear equation in two variables has solutions which geometrically form a line in the plane. A linear equation in three variables has solutions which form a plane. For linear equations of more variables, the geometric interpretations aren't as clear, but they still have an infinite set of solutions.

Systems of linear equations are groups of more than one linear equation. Solving the system refers to finding a solution common to all of the linear equations in the systems. There are only three possibilities:

1) The system has one unique solution (a consistent system)
2) The system has infinitely many solutions (also a consistent system)
3) The system has no solutions (an inconsistent system)

Geometrically, one solution can be interpreted geometrically as the point where the various lines, planes, etc. defined by the linear equations intersect. Infinitely many solutions occur when the equations define lines and/or planes that intersect in a line or plane, such as the intersection of two planes or two equal lines. An inconsistent system can be geometrically interpreted as lines and/or planes which never intersect, such as parallel lines in the plane, or parallel lines or planes in 3 dimensions. In 3 dimensions, we can also have skew lines, which are not parallel, but never intersect.


Methods of solving systems

When presented with a system of linear equations, we are almost always interested in solving that system. Here, various methods of solving these systems are looked at, with important definitions and theorems presented along the way. What is important to remember is that solving systems takes practice, and several examples will be given.

Back substitution

Back substitution is the simplest method of finding the solution to the system, and is also the last step in the more powerful methods of solving systems that we see later. Back substitution involves isolating a single variable, and substituting it back into another equation to isolate another variable, and continuing until we have solved for all variables.

To use back substitution, we need a system like above where the a variable is already isolated, or where we can express one variable in terms of another, as in the following example.

It is not often that our system of equations is already in a state where we can proceed directly to back substitution. We need to get our system of equations to a point where we have isolated variables. We can do this using elementary row operations. Elementary row operations can be used to change the system to one that we can use back substitution on, without changing the solution. If we denote row 1 by R1, row 2 by R2, etc., and k is a scalar, the three elementary row operations are as follows:

1)Swap two rows (equations), denoted R1↔R2 (this would swap rows 1 and 2)
2)Multiply a row by a nonzero scalar, denoted kR1 (this would multiply row 1 by k)
3)Add a multiple of a row to another row, denoted R1 + kR2 (this would add k times row 2 to row 1, and replace the previous row 1)

Elementary row operations should be familiar from solving systems in high school math courses. By introducing the concept of an augmented matrix, we can simplify our work on the system, and combine this with elementary row operations and back substitution to define another method of solving systems.

(If matrices are not familiar to you, or you'd like to review the topic, see the matrices tutorial) An augmented matrix is a matrix which contains the coefficients and constant terms from our system of linear equations. By putting the systems in this matrix, we only need to concentrate on what is important, these constants, since the variables themselves are never operated on. An augmented matrix is as follows,

Gaussian Elimination

Gaussian elimination is a three step method for solving systems of linear equations:

1) Write out the system as an augmented matrix (use zeroes for a particular variable of the system if it doesn't appear in one of the equations)
2) Use elementary row operations to reduce the matrix to row echelon form (see below)
3) Use back substitution to finally solve the system

Row echelon form occurs when a matrix is in a form as follows:

1) The first entry in each row (called the leading entry) has only zeroes in the column below it.
2) When a row has a leading entry, the leading entry in all rows above it are in columns to the left of the leading entry.

The following matrices are in row echelon form:

Let's take the system below and perform the steps involved in Gaussian elimination.

After putting the system into an augmented matrix, we perform elementary row operations on this system until we get it into row echelon form as described above. This allows us to perform back substitution, beginning with the bottom row in the system, in order to find the solution. In this way, Gaussian elimination provides us with an algorithm of sorts to solve a system of linear equations. Below, we solve the above system using Gaussian elimination.

Gauss-Jordan Elimination

We can improve the above method by adding another condition to Gaussian elimination. Gauss-Jordan elimination involves the same steps as Gaussian elimination, but we continue elementary row operations until we get the augmented matrix into row-reduced echelon form (rref), as defined next.

A matrix is in row-reduced echelon form if:

1) The leading entry in each row (if any) is a one.
2) There are no entries in the column above or below any leading entry.
3) Any leading entries in a row is to the right of a leading entry in a row above.

With Gauss-Jordan elimination, we avoid the need for back substitution. When our matrix is in rref, our leading 1's specify the variable in question, with our last column giving us the value for that variable. We specify a theorem about rref before performing Gauss-Jordan elimination on a system below.

Theorem: The row-reduced echelon form of a matrix A, denoted by rref(A), is unique.

The rank of a matrix A, rank(A), is equal to the number of nonzero rows in rref(A). Using this information, we can state the following theorem: The number of free variables in a system of linear equations in n variables with augmented matrix A is found by the formula

number of free variables = n - rank(A)

Free variables are those not associated with a leading entry in the row echelon form of A, while those with a leading entry are leading variables. When we have a consistent system with free variables, it means geometrically that our lines or planes intersect in infinitely many places (a line or plane), rather than a single point. The free variables can be set to anything, and the leading variables depend on those free variables.

Let's see an example of this idea of free variables.

Homogeneous Systems

Homogeneous systems of linear equations are systems where the constant term in each equations is zero. Therefore, they have the form [A|0], where A is the matrix of coefficients of the variables in the system of equations. Systems of this type always have a solution. There is always the trivial solution where [x1, x2, ..., xn] = [0,0,...0]. This can be interpreted as a point at the origin of the space Rn. As well, the system may have infinitely many solutions. Let's look at homogeneous systems.

In a homogeneous system, if the number of equations is less than the number of variables in the system, the system will always have infinitely many solutions. This is because a homogeneous system has to be consistent, and our formula for free variables guarantees at least one free variable. Together, this means that our homogeneous system will have infinitely many solutions, as shown below.

Linear Independence

A set of vectors is said to be linearly dependent if each vector in the set can be expressed as a linear combination of the other ones. This is equivalent to saying that there is some nontrivial linear combination (not all scalars are zero) of this set of vectors that equals 0. If the vectors are not linearly dependent, they are otherwise called linearly independent. Solving questions involving linear dependence involves putting our vectors into an augmented system, and then solving that system. Some examples are below.

We can see from above that if a system of linear equations is put in augmented form [A|b], then it has a solution if b is a linear combination of columns of A. Also, if we create a matrix B, where the rows of B are vectors in row form, these vectors are linearly dependent iff (if and only if) Rank(B)<m, where m is the number of vectors. Let's take our above examples and show how this works.

Finally, any set of vectors in Rn is linearly dependent if the number of vectors (m) is greater than the dimension of the space (n), that is, m>n.

Spanning Sets

Closely linked to the idea of linear dependence is the idea of spanning sets. Say we have a set of vectors in Rn, where S = {v1, v2, ..., vk}. The set of all possible linear combinations of the vectors in this set S is called the span of S. It is denoted span(v1, v2, ..., vk) or span(S). If span(S) = Rn, then S is called a spanning set for Rn. S can also span subsets of the whole space, such as a line, plane, etc. Let's look at an example of how to test to see if S is a spanning set.

Solving systems of linear equations can take some practice. Learning which methods are best and learning how to correctly solve systems takes time, and more questions are provided in the sample problems section of this site. You can also go to the test section to see if and where you need to practice.

| Top of Page |

FACULTY HOMEPAGES
Alex Karassev

NIPISSING LINKS
Nipissing University
Nipissing Email
Web Advisor

E-MAIL
Murat Tuncali
Alex Karassev
Vesko Valov
Andrew Dean

QUESTIONS/COMMENTS

Please forward any questions, comments, or problems you have experienced with this website to Murat Tuncali.