Definition 8.1: Null Space If $\mx{A}$ is an $m\times n$ matrix (i.e., with $m$ rows and $n$ columns), then
the entire set of solutions to $\mx{A}\vc{x}=\vc{0}$ is called the null space of $\mx{A}$.
Definition 6.2: Row and Column Vectors from a Matrix The $i$:th row vector of an $r \times c$ matrix, $\mx{A}$, is denoted by $\vc{a}_{i,}^\T$, and it has
$c$ scalar elements in it.
The $i$:th column vector of $\mx{A}$ is denoted by $\vc{a}_{,i}$, which has $r$ scalar elements in it.
Using vectors, a matrix can thus be written in the following two ways,
\begin{equation}
\mx{A} = \bigl(\vc{a}_{,1} \,\,\, \vc{a}_{,2} \,\,\,\dots\,\,\, \vc{a}_{,c}\bigr)
= \left(
\begin{array}{c}
\vc{a}_{1,}^\T\\
\vc{a}_{2,}^\T\\
\vdots \\
\vc{a}_{r,}^\T\\
\end{array}
\right).
\end{equation}
Definition 2.5: Column Vector Notation Given a basis, a column vector, $\vc{v}$, in $n$ dimensions (we have used $n\in [1,2,3]$) is a column of $n$
scalar values. These scalar components, sometimes called vector elements,
of the vector can either be numbered, i.e., $v_1$, $v_2$, and $v_3$,
or we can use $x$, $y$, and $z$ as subscripts when that is more convenient. The notation is:
\begin{gather}
\underbrace{ \vc{u} =
\begin{pmatrix} u_x
\end{pmatrix} =
\begin{pmatrix} u_1
\end{pmatrix}}_{\text{1D vector}},
\spc\spc
\underbrace{ \vc{v} =
\begin{pmatrix} v_x \\
v_y
\end{pmatrix} =
\begin{pmatrix} v_1 \\
v_2
\end{pmatrix}}_{\text{2D vector}},
\spc\spc
\\
\underbrace{ \vc{w} =
\begin{pmatrix} w_x \\
w_y \\
w_z
\end{pmatrix} =
\begin{pmatrix} w_1 \\
w_2 \\
w_3
\end{pmatrix}}_{\text{3D vector}},
\end{gather}
where $\vc{u} = u_x \vc{e}_1$, $\vc{v} = v_x \vc{e}_1 + v_y \vc{e}_2$,
and $\vc{w} = w_x \vc{e}_1 + w_y \vc{e}_2 + w_z \vc{e}_3$.
Definition 5.1: Linear Combination The vector, $\vc{u}$, is a linear combination of the vectors, $\vc{v}_1,\dots,\vc{v}_n$ when
$\vc{u}$ is expressed as
\begin{equation}
\vc{u} = k_1\vc{v}_1 + k_2\vc{v}_2 + \dots + k_n \vc{v}_n = \sum_{i=1}^{n} k_i \vc{v}_i,
\end{equation}
where $k_1, k_2, \dots, k_n$ are scalar values.
Definition 3.1: Dot Product The dot product between two vectors, $\vc{u}$ and $\vc{v}$ is denoted $\vc{u}\cdot \vc{v}$, and
is defined as the scalar value
\begin{equation}
\vc{u}\cdot \vc{v} = \left\{
\begin{array}{ll}
\ln{\vc{u}}\ \ln{\vc{v}} \cos[\vc{u},\vc{v}], & \text{if } \vc{u}\neq \vc{0} \text{ and } \vc{v}\neq \vc{0},\\
0, & \text{if } \vc{u}=\vc{0} \text{ or } \vc{v}=\vc{0}.
\end{array}
\right.
\end{equation}
Definition 3.8: A Parameterized Plane A plane, parameterized by $t_1\in\R$ and $t_2\in\R$, can be described by a starting point, $S$, and two
direction vectors, $\vc{d}_1$ and $\vc{d}_2$.
All points, $P(t_1,t_2)$, on the plane can be described by
\begin{equation}
P(t_1, t_2) = S + t_1 \vc{d}_1 + t_2\vc{d}_2.
\end{equation}
If one of $\vc{d}_1$ or $\vc{d}_2$ is $\vc{0}$ or if $\vc{d}_1$ and $\vc{d}_2$ are parallel (either in
the same direction, or opposite), then this degrades to a parameterized line equation.
Theorem 8.2: Neither of the operations of Gaussian elimination () changes
the row space of an $m \times n$ matrix $\mx{A}$ after applying the operation.
Theorem 8.3: Given a system $\mx{A}\vc{x}=\vc{0}$, which after Gaussian elimination is on row echelon form $\mx{R}\vc{x}=\vc{0}$, i.e.,
$\mx{A}\vc{x} = \vc{0} \Longleftrightarrow \mx{R}\vc{x} = \vc{0}$, then
the non-zero rows of $\mx{R}$ form a basis for the row space of $\mx{R}$ and due to also for $\mx{A}$, and
the column vectors with the first non-zero element of the row vectors form the column space
basis of $\mx{R}$, while the column space for $\mx{A}$ is the column vectors from $\mx{A}$ with the column numbers
as the columns for the column space for $\mx{R}$.
Theorem 8.8: Dimension Theorem For an $m\times n$ matrix $\mx{A}$, i.e., with $n$ columns, it holds that
\begin{equation}
\rank(\mx{A}) + \nullity(\mx{A}) = n.
\end{equation}
Definition 3.2: Orthogonal Projection If $\vc{v}$ is a non-zero vector, then the orthogonal projection of $\vc{u}$ onto $\vc{v}$ is denoted $\proj{\vc{v}}{\vc{u}}$,
and is defined by
\begin{equation}
\proj{\vc{v}}{\vc{u}} = \frac{\vc{u} \cdot \vc{v}}{ \ln{\vc{v}}^2 } \vc{v}.
\end{equation}
Note that if $\ln{\vc{v}}=1$, i.e., $\vc{v}$ is normalized, then the expression for projection gets simpler:
$\proj{\vc{v}}{\vc{u}} = (\vc{u} \cdot \vc{v})\vc{v}$.
Definition 7.8: Minor Assume that $\mx{A}$ is an $n \times n$ matrix. The minor, denoted $D_{ij}$, is the determinant of the $(n-1) \times (n-1)$
matrix that you obtain after removing row $i$ and column $j$ in $\mx{A}.$
Theorem 8.9: The rank of a matrix $\mx{A}$ (of size $m\times n$) is the largest integer $r$ for
which some $r\times r$-minor is $\neq 0$.
Theorem 7.8: If $\mx{A}$ is a square matrix and $\adj(\mx{A})$ is its adjoint matrix, then
\begin{equation}
\mx{A} (\adj(\mx{A})) = (\adj(\mx{A})) \mx{A} = \det(\mx{A}) \mx{I} .
\end{equation}
If $\det(\mx{A}) \neq 0$ then $\mx{A}$ is invertible and the inverse is given by
\begin{equation}
\mx{A}^{-1} = \frac{1}{\det \mx{A}} \adj \mx{A} .
\end{equation}
Theorem 8.6: Rank of Product The rank of a product $\mx{A} = \mx{B} \mx{C}$ is less than or equal to the rank of the terms, i.e.,
\begin{equation}
\rank \mx{A} \leq \rank \mx{C},
\end{equation}
and
\begin{equation}
\rank \mx{A} \leq \rank \mx{B}.
\end{equation}
This chapter is about a concept called the rank, which is a property of a matrix.
In fact, it reveals several aspects of a matrix. Some of the examples here may be shown
for square matrices but everything in this chapter works for rectangular matrices, say of size $m \times n$.
We start with an example from the application of structure from sound.
Example 8.1:Structure from Sound
There are many examples from applied mathematics, which involves the concept of rank. One such example is the so called structure from sound problem. In this example, sounds are recorded at a number of stationary, but unknown, microphone positions. Unknown sounds are emitted at unknown positions and at unknown times. It turns out that it is still possible to calculate (i) where the microphones are, (ii) how the sound source has moved, and (iii) when the sound sources were emitted, just using the measured sound signals from the microphones. Here rank plays an important role.
Figure 8.1:
Illustration of the measurement situation. The sound source (held in the left hand) is moving, while the eight microphones are stationary.
Figure 8.1:
Illustration of the measurement situation. The sound source (held in the left hand) is moving, while the eight microphones are stationary.
Figure 8.1 shows the measurement situation. Eight microphones capture the sound as a single sound source is moving in the room. The eight sound files are used to calculate the microphone positions and the sound source movement. See also the following YouTube clip.
The 3D sound source path and the microphone locations that are computed are also visualized in Interactive Illustration 8.2.
Interactive Illustration 8.2:
The blue circles represent where the eight microphones are located. However, these
are unknown and so is the moving position of the sound source. Press/click forward to see the sound source path.
Interactive Illustration 8.2:
The red sound source path and blue microphone positions were calculated from 30 seconds of sound recordings. Press the rotation icon in the top left corner to rotate the point cloud, or right-hand click plus drag to see the model from any orientation.
Interactive Illustration 8.2:
The red sound source path and blue microphone positions were calculated from 30 seconds of sound recordings. Press the rotation icon in the top left corner to rotate the point cloud, or right-hand click plus drag to see the model from any orientation.
We start by explaining the null space, column space, and row space of a matrix. As will be shown these three spaces are themselves linear vector spaces. They are examples of linear subspaces, which are explained next.
A linear subspace of a vector space is a subset of the vectors that is itself a linear space. To show that a subset of a linear space is a linear subspace one only has to check three things. This is captured by the following theorem.
Theorem 8.1:
Let $V$ be a linear vector space over the scalars $F$. (Here we will typically only use $F=\R $ or $F=\mathbb{C}$).
A subset $W \subset V$ is a linear vector space, i.e. a subspace, if and only if $W$ satisfies the following three conditions:
The zero vector $\vc{0} \in W$.
If $\vc{v}_1 \in W$ and $\vc{v}_2 \in W$ then $\vc{v}_1+\vc{v}_2 \in W$.
If $\vc{v} \in W$ and $\lambda \in F$ then $\lambda \vc{v}\in W$.
First, we prove that $W$ is a vector space if the conditions hold. By condition 1, we see that $W$ is nonempty.
The conditions 2 and 3 ensure that the result of adding two vectors or multiplying a vector with a scalar does not take us outside the subset. Since the elements of $W$ are already part of vector space $V$, we then know that all the other rules
of Theorem 2.1 hold.
Second, we prove that if the subset $W$ is a vector space then the three conditions must hold.
Since $W$ is a vector space, then the properties 2 and 3 are satisfied. Since $W$ is non-empty, there is at least one element $\vc{w}$. We can then form $ \vc{0} = 0 \vc{w} \in W$. Thus property 1 holds.
$\square$
Example 8.2: Let $\mx{A}$ be an $m \times n$ matrix. Let $V$ be the linear vector space $\R^n$. Let $W \subset V$ be the subset of vectors $\vc{v} \in V$ that fulfills $\mx{A}\vc{v} = \vc{0}$. Then $W$ is a linear vector space. This is true since
The zero vector $\vc{0} \in W$ since $\mx{A}\vc{0} = \vc{0}$.
If $\vc{v}_1 \in W$ and $\vc{v}_2 \in W$ then $\mx{A}(\vc{v}_1+\vc{v}_2) = \mx{A}\vc{v}_1+\mx{A}\vc{v}_2 = \vc{0}+\vc{0}$. Therefore $\vc{v}_1+\vc{v}_2 \in W$.
If $\vc{v} \in W$ and $\lambda \in F$ then $\mx{A}(\lambda \vc{v}) = \lambda \mx{A} \vc{v}= \lambda\vc{0} =\vc{0}$. Therefore $\lambda \vc{v}\in W$.
This means that the solutions to a homogeneous system is a linear space.
Example 8.3: Let $\mx{A}$ be an $m \times n$ matrix. Let $V$ be the linear vector space $\R^m$. Let $W \subset V$ be the subset of vectors $\vc{v} \in V$ that can be generated as $\vc{v}= \mx{A}\vc{u}$ for some $\vc{u} \in \R^n$. Then $W$ is a linear vector space. This is true since
The zero vector $\vc{0} \in W$ since $\vc{0}\in\R^m$ can be generated from $\vc{0} \in \R^n$ by $\vc{0} =\mx{A}\vc{0}$.
If $\vc{v}_1 \in W$ and $\vc{v}_2 \in W$ then there exists $\vc{u}_1 \in \R^n$ and $\vc{u}_2 \in \R^n$ so that
$\vc{v}_1 = \mx{A}\vc{u}_1$ and $\vc{v}_2 = \mx{A}\vc{u}_2$. But then $\vc{v}_1+\vc{v}_2 \in W$, since
$(\vc{v}_1+\vc{v}_2) = \mx{A}\vc{u}_1+\mx{A}\vc{u}_2 = \mx{A}(\vc{u}_1 + \vc{u}_2)$.
If $\vc{u} \in W$ and $\lambda \in F$ then there exists $\vc{u} \in \R^n$ so that $\vc{v} = \mx{A}\vc{u}$ . But then
$\lambda \vc{v}\in W$ since
$\lambda \vc{v} = \lambda \mx{A}(\vc{u}) = \mx{A} (\lambda \vc{u})$.
This means that the set of linear combinations of a set of vectors is a linear space.
As we have seen in Section 5.5,
a system of equations such as $\mx{A}\vc{x} = \vc{0}$ is called homogeneous. We will now show that the solutions to such systems
reveal certain useful properties of $\mx{A}$. Let us start with an example.
Assume we have a matrix
and that we want to solve for $\mx{A}\vc{x}=\vc{0}$, where $\vc{x} = (x_1, x_2, x_3)$ is a column vector.
This is the same example as in Equation (5.34), i.e.,
As can be seen, the solution is a line which
is dependent only on one varying parameter $t$. Sometimes the solution is denoted by $\vc{x}_\mathrm{h}$ to indicate
that it is a solution to the homogeneous system $\mx{A}\vc{x}=\vc{0}$.
This entire set of solutions is called the null space of $\mx{A}$ and in this particular case,
the dimension of the null space is 1, since it only depends on one parameter, namely $t$.
We also use the term null vector for $(1/16, -5/8, 1)$.
Next, this is formalized into a definition.
Definition 8.1:Null Space
If $\mx{A}$ is an $m\times n$ matrix (i.e., with $m$ rows and $n$ columns), then
the entire set of solutions to $\mx{A}\vc{x}=\vc{0}$ is called the null space of $\mx{A}$.
By definition, the null space must be a subspace of $\R^n$, that is, $\nullity(\mx{A}) \leq n$.
It is a linear subspace, (see Example 8.2). The dimension of this subspace is important and has its own definition.
Definition 8.2:Nullity
The dimension of the null space is denoted $\nullity(\mx{A})$.
Since the null space is a subspace of $\R^n$, we have $\nullity(\mx{A}) \leq n$. But how many dimensions lower is the null space? This is captured loosely by the number of constraints in $\mx{A}\vc{x}=\vc{0}$, however, here we must be careful. Just because there are $m$ equations in $\mx{A}\vc{x}=\vc{0}$ it does not automatically mean that $\nullity(\mx{A}) = n - m$. How many constraints there are in $\mx{A}$ is precisely what is captured by the concept of rank. What we are later going to show is that $\nullity(\mx{A}) = n - r$, i.e., that the dimension is lowered by $r$ (i.e., the rank of $\mx{A}$) using the constraints in $\mx{A}\vc{x}=\vc{0}$.
Definition 8.3:Null Vectors
Let $\vc{p}_1, \ldots, \vc{p}_k$ be a basis for the null space.
A solution, $\vc{x}$, is on the form $\vc{x}(t_1,\dots t_k) = t_1\vc{p}_1 + \dots + t_k\vc{p}_k$,
where $k$ is the dimension of the null space. The basis vectors $\vc{p}_i$ are sometimes called null vectors, although sometimes any vector in the null space is called a null vector.
Using Gaussian elimination, we obtain a basis for the null space and thereby also the dimension of the null space.
It is the number of independent parameters in the solution to $\mx{A}\vc{x}=\vc{0}$ that
dictates the dimension of the null space.
Example 8.4:Null Space Here, we will again look at the solution of a system of equations
on the form $\mx{A}\vc{x}=\vc{0}$.
The following system of equations
where the step above subtracted row 1 from row 2. As can be seen, row 2 and 3 are now identical, which
means that we will obtain a row of zeroes when they are subtracted, i.e.,
As we saw in Section 5.5, the procedure in cases such as this is now to
set one parameter to being variable, for example, $x_5=t$. Using row 3, we get $x_4+2x_5=0$, that is,
$x_4=-2t$.
For row 2, we need to introduce one more variable parameter, say, $x_3=s$, which
gives us $x_2 = -4s + 3t$. Finally, using row 1, we can get $x_1 = -\frac{1}{2}s-\frac{5}{2}t$.
The solution is then
Per Definition 8.1, the solution above is the null space of $\mx{A}$
and since it depends on two variables $s$ and $t$, we know that the dimension of the null space is 2,
that is, $\nullity(\mx{A})=2$.
Furthermore, we also say that the vectors $\vc{m}$ and $\vc{n}$ span the null space, i.e., they
are basis vectors for the null space.
For an $m\times n$ matrix $\mx{A}$, recall from Definition 6.2 how
the rows $\vc{a}_{i,}^\T$, where $i\in{1,\dots,m}$, and the columns, $\vc{a}_{,j}$, are accessed, i.e.,
In addition, we repeat that the notation of column and row vectors in this book, i.e.,
all vectors are column vectors by default and if you want to use a row vector,
you take a column vector, $\vc{x}$, and you transpose it to get a row vector, $\vc{x}^\T$.
Also, $(x_1,x_2,x_3)$ is also a column vector (note the commas), while $(x_1 \spc x_2 \spc x_3)$ is a row vector,
which is described below Definition 2.5.
Now that the null space is known, two related concepts, namely the column space and the row space, can be introduced.
Definition 8.4:Column Space
Let $\mx{A}= \left(\vc{a}_{,1} \vc{a}_{,2} \dots \vc{a}_{,n}\right)$ be an $m\times n$ matrix.
The column space of $\mx{A}$ is then the set of all linear combinations of the
column vectors, $\vc{a}_{,i}$.
Now, we note that matrix-vector multiplication, $\mx{A}\vc{x}$, can be expressed as
that is, as a linear combination (see Definition 5.1)
of the column vectors of $\mx{A}$. So, the column space is therefore all the vectors
that can be created by taking a constant times the first column vector plus
another constant times the second column vector and so on.
Note that the column space is a subspace of $\R^m$ by definition, since the number of elements in
all column vectors of $\mx{A}$ is $m$.
Definition 8.5:Row Space
Let $\mx{A}= \left(\begin{array}{c} \vc{a}_{1,}^\T\\ \vc{a}_{2,}^\T\\ \vdots \\\vc{a}_{m,}^T \end{array}\right)$ be an $m\times n$ matrix.
The row space of $\mx{A}$ is then the set of all linear combinations of the
row vectors, $\vc{a}_{i,}^\T$.
Similar to the column space, we can write this as a vector-matrix multiplication, but this time we use
a row vector $\vc{x}^\T$ with $m$ elements and multiply from the right, i.e.,
So, this is also a linear combination (see Definition 5.1), but this time using
the row vectors of $\mx{A}$.
So, the row space is therefore all the vectors
that can be created by taking a constant times the first row vector plus
another constant times the second row vector and so on.
Note that the row space is a subspace of $\R^n$ by definition, since the number of elements in
all row vectors of $\mx{A}$ is $n$.
Example 8.5:Column Space 1 In this example, we continue with the matrix from Example 8.4. Recall
that we transformed the system of equations into
The null space for this example was calculated in the previous example and here, we will show
how the column space is obtained.
As it turns out, the basis vectors for the column space can be directly "fetched" from the matrix above,
where we have used Gaussian elimination to find the solution. The column vectors that form the column space
are the ones with the first non-zero element of each row.
Hence, we obtain the following vectors,
Since there are three basis vectors for the column space, the dimension of $\mx{A}$'s column space is also 3.
The reason why this method works will be explained later in this chapter.
Since one can verify that $\vc{a}_{,3} = 2\vc{a}_{,1}-3\vc{a}_{,2}$, the column space is at most a linear
combination of two vectors. This means that if $\vc{a}_{,1}$ is parallel to $\vc{a}_{,2}$ then the column space
is just dependent on one vector and otherwise, it is dependent on two. If $\vc{a}_{,1}$ is parallel to $\vc{a}_{,2}$,
then $\vc{a}_{,1} \cdot \vc{a}_{,2}= \ln{\vc{a}_{,1}} \ln{\vc{a}_{,2}}$ (see Definition 3.1).
Since $\vc{a}_{,1} \cdot \vc{a}_{,2} = 1\cdot 4 + 2\cdot 1 + 3\cdot(-3) = -3$ and
$\ln{\vc{a}_{,1}} \ln{\vc{a}_{,2}}\geq 0$, we draw the conclusions that they are not parallel and that the column
space is a combination of two vectors.
For example, the column space can be expressed as
where $s$ and $t\in \R$.
i.e., the column space is a linear combination of $\vc{a}_{,1}$ and $\vc{a}_{,2}$.
As we will see later, this means that $\rank(\mx{A})=2$, since the column space is
a linear combination of two vectors.
Geometrically, this means that the column space is a plane
(see Definition 3.8), which also is clear from
Equation (8.16).
Definition 8.6:Row and Column Rank
Let $\mx{A}$ be a matrix. The row rank of $\mx{A}$, denoted $\rowrank(\mx{A})$, is the maximum number of linearly
independent row vectors of $\mx{A}$. Similarly, the column rank, denoted $\colrank(\mx{A})$, is the maximum number of
linearly independentcolumn vectors of $\mx{A}$.
In the following, we will go through a number of steps that will lead up to a point where we easily can obtain
the column/row vectors of the column/row space of a matrix.
We start with the following theorem.
Theorem 8.2:
Neither of the operations of Gaussian elimination (Theorem 5.2) changes
the row space of an $m \times n$ matrix $\mx{A}$ after applying the operation.
The rows of $\mx{A}$ are denoted $\vc{a}_{i,}^{\T}$ where $i\in\{1,\dots, m\}$ as usual. In this proof, we use
the following notation for the operations of Theorem 5.2
$(i)$: $\vc{a}_{i,}^{\T} \leftrightarrow \vc{a}_{j,}^{\T}$ to indicate that the order of two rows are swapped,
$(ii)$: $\vc{a}_{i,}^{\T} \rightarrow k\vc{a}_{i,}^{\T}$ to indicate that a row is multiplied by a non-zero constant, $k$, and
$(iii)$: $\vc{a}_{i,}^{\T} \rightarrow \vc{a}_{i,}^{\T} + \vc{a}_{j,}^{\T}$ for addition of one row to another.
$(i)$: It is clear that when the first rule is applied only the order of the rows change and thus, the row space stays the same.
$(ii)$: For this rule, the matrix before and after applying the rule are the same except on row $i$.
Now, recall from Equation (8.12) that the row space is expressed as
Since we can think of $k \vc{a}_{i,}^{\T}$ as a "new" row vector and $k_i/k$ as an arbitrary constant before the new row vector,
it is clear that everything that can be expressed in the first line of Equation (8.18) can also be expressed in the last line of Equation (8.18), and vice versa. Hence, the second rule
preserves the row space.
$(iii)$: With the third rule $\vc{a}_{i,}^{\T} \rightarrow \vc{a}_{i,}^{\T} + \vc{a}_{j,}^{\T}$, we rewrite
Equation (8.17) as
Here, $\vc{a}_{i,}^{\T} + \vc{a}_{j,}^{\T}$ is the updated row vector and as can be seen, the constant
before $\vc{a}_{j,}^{\T}$ has changed to $k_j - k_i$, and hence any value can be obtained by changing $k_j$ appropriately.
As a conclusion, we now know that the row space of $\mx{A}$ is in the row space of matrix obtained after using rule $(iii)$.
However, we need to show the rule in the other direction $(\leftarrow)$ as well. Since
by similar arguments, the row space of a matrix with rule $(iii)$ applied is in the row space of the original matrix.
This concludes the proof.
$\square$
Note that it is important that Theorem 8.2 only works for the row space and
not for the column space.
In the following, it will be useful to have a specific term for the matrix that is obtained after Gaussian elimination.
This is described in the following definition.
Definition 8.7:Row Echelon Form of a Matrix
A matrix $\mx{A}$ is said to be on row echelon form if its
rows with only zeroes (if any) are placed at the bottom, and the first non-zero coefficient, $a_{ij}$, on
each row with number $i$ is placed such that all other rows with a first non-zero coefficient, $a_{i'j'}$,
fulfills $i' < i$ (the row $i'$ is above row $i$) and $j' < j$.
Note that a matrix on row echelon form can be obtained by applying Gaussian elimination until no more reductions can be made.
This gives a triangular-like structure for matrices on row echelon form, e.g.,
Therefore, this type of matrices are sometimes said to be a "staircase equivalent" matrix of $\mx{A}$.
The leading non-zero coefficients, underlined in the previous equation,
are important. These elements of the matrix are often called pivot element.
Some texts require that these leading non-zero coefficients are made to ones. This can be achieved
by dividing the entire row by the leading non-zero coefficient. However, we do not require that here.
Next, we present a theorem that says how the row spacebasis vectors can be found and how the dimension of the
row space is found.
Theorem 8.3:
Given a system $\mx{A}\vc{x}=\vc{0}$, which after Gaussian elimination is on row echelon form $\mx{R}\vc{x}=\vc{0}$, i.e.,
$\mx{A}\vc{x} = \vc{0} \Longleftrightarrow \mx{R}\vc{x} = \vc{0}$, then
the non-zero rows of $\mx{R}$ form a basis for the row space of $\mx{R}$ and due to Theorem 8.2 also for $\mx{A}$, and
the column vectors with the first non-zero element of the row vectors form the column spacebasis of $\mx{R}$, while the column space for $\mx{A}$ is the column vectors from $\mx{A}$ with the column numbers
as the columns for the column space for $\mx{R}$.
To show that the non-zero rows are a basis we need to show
(i) that they span the space and (ii) that they are independent.
(i) That any combination of the rows are also combination of the non-zero rows is clear.
Therefore the non-zero rows span the row space.
(ii) To show that they are independent. No nonzero row of the row echelon matrix is a combination of
the other rows since in the column of its leading 1, the other rows are 0. Hence the nonzero rows are independent.
$\square$
Note that the number of basis vector of the row/column space is the row/column rank.
Next, we will show that the dimension of the row space (i.e., the row rank) is the same as the dimension of the column space (i.e., the
column rank). After the following theorem, we will give an example of how the row and column spacebasis can be found.
Theorem 8.4:Column and Row Rank Size
The row rank is equal to the column rank of a matrix $\mx{A}$.
Let us say that the row rank of $\mx{A}$ is $k$ and that $\{ \vc{b}_{1,}^\T,\dots,\vc{b}_{k,}^\T \}$
is a row spacebasis for $\mx{A}$.
The $i$:th row of $\mx{A}$ is $\vc{a}_{i,}^\T = \left(a_{i1}, a_{i2},\dots,a_{in} \right)^\T$
and it can also be written using the row spacebasis vectors as $\vc{a}_{i,}^\T = \sum_{r=1}^k c_{ir} \vc{b}_{r,}^\T$, for
some set of constants, $c_{ir}$. The $j$:th element of $\vc{a}_{i,}^\T$ is then
An interpretation of this is that every column of $\mx{A}$ is a linear combination of $k$
vectors, which means that the $\colrank(\mx{A})$ can be at most $k$.
Expressed differently, we have $\colrank(\mx{A}) \leq \rowrank(\mx{A})$.
If we apply this to the transpose of $\mx{A}$ then we have $\colrank(\mx{A}^\T) \leq \rowrank(\mx{A}^\T)$
which is the same as $\rowrank(\mx{A}) \leq \colrank(\mx{A})$, and since we had
$\colrank(\mx{A}) \leq \rowrank(\mx{A})$ to start with, it follows that
$\rowrank(\mx{A}) = \colrank(\mx{A})$.
$\square$
Hereafter, we do not need to distinguish between $\rowrank(\mx{A})$ and $\colrank(\mx{A})$, and instead, we refer to it
as just rank, denoted $\rank(\mx{A})$. Note also that the theorem above can be stated as $\rank(\mx{A})= \rank(\mx{A}^{\T})$.
This means that if you want to find the column space vectors of a matrix $\mx{A}$, then you can transpose it to $\mx{A}^{\T}$
and then use Theorem 8.3.
Next, we show an example of how the column spacebasis can be found.
i.e., you look at each row and find the first non-zero element and take that column vector where this element resides.
The column space for $\mx{A}$ is obtained using the column vectors with the same column number as the column vectors for for column space for
$\mx{R}$, i.e.,
Since the column space is spanned by the three vectors above, we have $\rank(\mx{A})=3$ also when using
the column space. This is to be expected since $\colrank(\mx{A})=\rowrank(\mx{A})$.
Theorem 8.5:
The rank of a matrix is the number of pivot elements after Gaussian elimination.
The row rank of a matrix $\mx{A}$ is equal to the dimension of the row space of the reduced matrix $\mx{R}$ after Gaussian elimination. Since a basis of the row space of $\mx{R}$ consists of those rows that have a pivot element, the dimension and thereby the rank is the number of pivot elements.
$\square$
Next, we present a theorem that provides an upper bound of the rank of a matrix-matrix product.
Theorem 8.6:Rank of Product
The rank of a product $\mx{A} = \mx{B} \mx{C}$ is less than or equal to the rank of the terms, i.e.,
As we showed earlier, each row of the product $\mx{A}$ is a linear combination of the rows in $\mx{C}$. This means that all rows of $\mx{A}$ lie in the rowspace of $\mx{C}$, which means that $\rank \mx{A} \leq \rank \mx{C}$. Thus we have shown that the rank of a product is less than or equal to the rank of its rightmost term.
The other equality follows from studying the rows of $\mx{A}$ that are linear combination of the rows of $\mx{B}$. Alternatively the other equality can be proved using $\mx{A}^T = \mx{C}^T \mx{B}^T$. Here the rightmost term is $\mx{B}^T$. So $\rank \mx{A}^T \leq \rank \mx{B}^T$. But then we have
$ \rank \mx{A} = \rank \mx{A}^T \leq \rank \mx{B}^T = \rank \mx{B} $.
$\square$
Theorem 8.7:
Let $\vc{x}_\mathrm{h}$ be the solution to $\mx{A}\vc{x}=\vc{0}$ and let $\vc{x}_\mathrm{p}$ be the solution to
a particular system of equations $\mx{A}\vc{x}=\vc{y}$, in which case the "entire" solution to $\mx{A}\vc{x}=\vc{y}$ is
Now, we know that $\vc{x}_h$ is in the null space of $\mx{A}$ and that $\vc{y}$ must be in the column space of $\mx{A}$,
otherwise it is not a solution to $\mx{A}\vc{x}=\vc{y}$. In Interactive Illustration 8.3,
we show how these two solutions may interact.
$\vc{x}_h$
$\vc{x}_p$
$\vc{x}_{\mathrm{tot}}$
Interactive Illustration 8.3:
Assume that we have a homogeneous system $\mx{A}\vc{x}=\vc{0}$ and we call the solutions to
this $\vc{x}_h$. In this case, the set of solutions is a line as we can see above. Pull the slider to change
the location of $\vc{x}_h$. Then click/touch Forward to proceed.
Interactive Illustration 8.3:
Furthermore, we assume that we want a particular solution to $\mx{A}\vc{x} = \vc{y}$ and we call this
solution $\vc{x}_p$, which is also visualized above.
Interactive Illustration 8.3:
In this final step, we show the total solution $\vc{x}_{\mathrm{tot}} = \vc{x}_p + \vc{x}_h$, and we
can see that the line described by $\vc{x}_h$ is translated by $\vc{x}_p$ to form $\vc{x}_{\mathrm{tot}}$,
which also is a line. Recall that $\vc{x}_h$ spans the null space, while $\vc{x}_p$ lies in the column space of $\mx{A}$.
Interactive Illustration 8.3:
Assume that we have a homogeneous system $\hid{\mx{A}\vc{x}=\vc{0}}$ and we call the solutions to
this $\hid{\vc{x}_h}$. In this case, the set of solutions is a line as we can see above. Pull the slider to change
the location of $\hid{\vc{x}_h}$. Then click/touch Forward to proceed.
The dimension theorem below is very handy since if you know either the dimension of the null space or the
rank then you can find the other one of these (assuming you know the number of rows in the matrix, which is a
reasonable assumption).
Theorem 8.8:Dimension Theorem
For an $m\times n$ matrix $\mx{A}$, i.e., with $n$ columns, it holds that
\begin{equation}
\rank(\mx{A}) + \nullity(\mx{A}) = n.
\end{equation}
If $\rank(\mx{A})=n$, this is the same as being able to invert the matrix since its determinant will be non-zero.
There are therefore no other solutions to $\mx{A}\vc{x}=\vc{0}$ than $\vc{x}=\vc{0}$,
i.e., $\nullity(\mx{A})=0$, which means $\rank(\mx{A})+0=n$. Hence, the theorem holds for that case.
Next, we assume that $\rank(\mx{A}) < n$. In this case, there must be $n-r$ free variables in the solution
to $\mx{A}\vc{x}=\vc{0}$. By free variables we mean $s$ and $t$ in Example 8.4,
for example. Let us call the free variables $s_1$, $\dots$, $s_{n-r}$.
There are $n-r$ linearly independent null vectors, which we denote by $\vc{x}_1$, $\dots$, $\vc{x}_{n-r}$.
As we know, the solutions to $\mx{A}\vc{x}=\vc{0}$ are dictated by
Hence, $\vc{x}_1$, $\dots$, $\vc{x}_{n-r}$ is a basis for the null space and $\nullity(\mx{A}) = n-r$,
which proves the theorem.
$\square$
Recall that $\nullity(\mx{A})$ dictates the number of parameters needed to describe the
solution to the homogeneous system $\mx{A}\vc{x}=\vc{0}$ and that $\rank(\mx{A})$
is the number of linearly independent right-hand sides $\vc{y}$ when $\mx{A}\vc{x}=\vc{y}$ has a solution.
Finally, the sum of these two numbers is $n$, i.e., the number of columns of $\mx{A}$.
Example 8.8:Dimension Theorem In Example 8.4 and Example 8.7,
we had a $4\times 5$ matrix $\mx{A}$, i.e., $m=4$ and $n=5$. We first saw that
The Dimension Theorem 8.8 states that $2+3=n$, where $n$ is the
number of columns of the matrix. As can be seen, this is correct since $m=5$ is the number of columns of
the matrix.
Note that $\mx{A}$ can be thought of as a transform or linear mapping (see Chapter 9),
and that $n$ is the dimension of the domain of $\mx{A}$.
Given this, we can think of $\rank(\mx{A})$ as the number of dimensions that "survive" after applying $\mx{A}$
and $\nullity(\mx{A})$ is the number of dimensions that are "collapsed" by $\mx{A}$.
Example 8.9:Orthogonal Projection of a Point onto a Plane - revisited In Example 3.10, a point $P$ was projected onto a plane defined by a normal vector $\vc{n}$
and a point on the plane $S$, which we in this example assume is the origin, $O$.
We found that $P$ projected orthogonally down to the plane can be expressed as
\begin{equation}
Q = P - \proj{\vc{n}}{\vc{v}},
\end{equation}
(8.36)
where $Q$ is the projected point and $\vc{v}=P-O$.
The orthogonal projection formula (Definition 3.2) is
$\proj{\vc{n}}{\vc{v}} = \left((\vc{v} \cdot \vc{n})/\ln{\vc{n}}^2\right) \vc{n} $
and as a result, the expression for $Q$ can be rewritten as
In order to operate with matrices, we need to use vectors instead of points, so we introduce
$\vc{q} = \overrightarrow{OQ}$ and we already have
$\vc{v} = \overrightarrow{OP}$.
This leads to
The dimension of the domain is 3, since we have three-dimensional vectors in $\R^3$ as input.
Now since only vectors in the plane "survive", we know immediately that $\rank(\mx{A})=2$ and since we collapse one
dimension, we have $\nullity(\mx{A})=1$.
$P$
$O$
$\vc{n}$
$\vc{v}$
$\proj{\vc{n}}{\vc{v}}$
$-\proj{\vc{n}}{\vc{v}}$
$Q$
$\vc{q}$
Interactive Illustration 8.4:
This illustration will show how a point, $P$ (gray
circle), is projected orthogonally down on to
a plane surface defined by
a normal vector, $\vc{n}$, and a starting point, $O$, which is the origin. The point $P$ can be moved around in this illustration.
Click/touch Forward to commence the illustration.
Interactive Illustration 8.4:
First, a vector from $O$ to the point, $P$, is created, i.e., $\vc{v}=P-O$.
Interactive Illustration 8.4:
The vector $\vc{v}=P-O$ is then projected onto the normal, $\vc{n}$, i.e.,
$\proj{\vc{n}}{\vc{v}}$ is created.
Interactive Illustration 8.4:
The projected vector $\proj{\vc{n}}{\vc{v}}$ is then subtracted from $\vc{v}$, which creates a point in the plane.
This point will appear in the next step of this illustration. Press/touch Forward.
Interactive Illustration 8.4:
Finally, the point $P$ has been projected orthogonally to the plane and the projected point, $Q$,
is shown as a red circle. Thus, we have $\vc{q} = Q - O$. Note that we derived a matrix, $\mx{A}$, above
that immediately calculates $\vc{q}$ as $\vc{q} = \mx{A}\vc{v}$. The points $\vc{v}$ are three-dimensional,
i.e., $\vc{v} \in \R^3$. Note though that the projected points $\vc{q}$ always reside in the light blue plane,
that is, the set of all points $\vc{q}$ is two-dimensional, which in turn implies that
$\rank(\mx{A}) = 2$. The nullity of the matrix is 1, i.e., $\nullity(\mx{A})=1$, since we "remove"
one dimension with the transform, $\mx{A}$.
Interactive Illustration 8.4:
Finally, the point $\hid{P}$ has been projected orthogonally to the plane and the projected point, $\hid{Q}$,
is shown as a red circle. Thus, we have $\hid{\vc{q} = Q - O}$. Note that we derived a matrix, $\hid{\mx{A}}$, above
that immediately calculates $\hid{\vc{q}}$ as $\hid{\vc{q} = \mx{A}\vc{v}}$. The points $\hid{\vc{v}}$ are three-dimensional,
i.e., $\hid{\vc{v} \in \R^3}$. Note though that the projected points $\hid{\vc{q}}$ always reside in the light blue plane,
that is, the set of all points $\hid{\vc{q}}$ is two-dimensional, which in turn implies that
$\hid{\rank(\mx{A}) = 2}$. The nullity of the matrix is 1, i.e., $\hid{\nullity(\mx{A})=1}$, since we "remove"
one dimension with the transform, $\hid{\mx{A}}$.
Now, let us explore the null space of $\mx{A}$. We know from Section 8.2 that
nullity can be computed by performing Gaussian elimination on $\mx{A}\vc{x}=\vc{0}$. Without loss of generalization,
we can assume that $\vc{n} = (0,0,1)$, i.e., $f=1$, and $S=O=(0,0,0)$.
This results in
that is, the result is already on row echelon form and since the last row has only zeroes, we have $\rank(\mx{A})=2$.
We can set $p_z=t$, where $t\in\R$. This leaves the first two rows unaffected, which means that the
null space is spanned by $(0,0,t)$, i.e., $\nullity(\mx{A})=1$.
This makes sense since the normal is $\vc{n} = (0,0,1)$, and all vectors are collapsed by $\mx{A}$ to a plane $z=0$.
The column spacebasis vectors are the first and second column vectors, i.e.,
where $u\in\R$ and $v\in\R$, which also means that $\rank(\mx{A})=2$.
Finally, we note that Theorem 8.8 states that $\rank(\mx{A}) + \nullity(\mx{A}) = n$, which indeed is true
also in this case since $2+1=3$.
Example 8.10: This is another example showing how to compute the rank, nullity, etc. We start with
a $4\times 6$ matrix $\mx{A}$, which is
We will perform Gaussian elimination in order to get the matrix on row-echelon form.
First, for row 2, 3, and 4, we take the top row minus the row itself, which generates
and let us now use this in $\mx{R}\vc{x}=\vc{0}$.
We immediately get $2 x_6 = 0$ from the bottom row, i.e., $x_6=0$.
Using row 3, we get $x_4+x_6=0$, i.e., $x_4=0$ as well.
Row 2 gives us $2x_3 + 0 + 2x_5 + 0 = 0$, i.e., $x_3+x_5=0$.
Therefore, we set $x_5=t$, which gives us $x_3=-t$.
The top row says $x_1 + x_2 + x_3 + x_4 + x_5 + x_6 = x_1+x_2 -t + 0 +t +0 =0$,
i.e., $x_1+x_2=0$. This time, we set $x_2=s$, i.e, $x_1 = -s$.
In total, the solution to the homogeneous system is $\mx{A}\vc{x}=\vc{0}$
This means that $\nullity(\mx{A})=2$ since there are two variables ($s$ and $t$) in this expression.
The null vectors are $\vc{z}^1$ and $\vc{z}^2$.
Since $n=6$ and $\nullity(\mx{A})=2$, we have $\rank(\mx{A})= 6-2=4$.
The row vectors spanning the row space are all the 4 row vectors of $\mx{R}$ but
also all the row vectors of $\mx{A}$ since Gaussian elimination does not change row space.
The index of the column vectors spanning the column space are identified in $\mx{R}$,
i.e, column 1, 3, 4, and 6, since those have the first element on each respective row.
Hence, the column vectors spanning the column space are the column vectors in
$\mx{A}$ with column number 1, 3, 4, and 6.
Note that $\mx{A}\vc{z}^1=\vc{0}$ and $\mx{A}\vc{z}^2=\vc{0}$ (try performing
the matrix-vector multiplications) and finally, if we have a system
$\mx{A}\vc{x}=\vc{b}$, then the solution will lie in the row space of $\mx{A}$.
In Chapter 7, different ways to compute the determinant of a square matrix were presented.
Using minors (see Definition 7.8), one can also find the rank of a matrix, as the subsequent theorem describes.
Theorem 8.9:
The rank of a matrix $\mx{A}$ (of size $m\times n$) is the largest integer $r$ for
which some $r\times r$-minor is $\neq 0$.
This proof is divided into two parts, namely showing that the rank of the matrix is at least $r$
and the second part shows that the rank is at most $r$, which together complete the proof.
$i)$
Let us call an $r\times r$ submatrix by $\mx{S}$ whose determinant, $d$, is $\neq 0$,
i.e., $d=|\mx{S}|\neq 0$.
As we know, such a minor must be obtained by crossing out $m-r$ rows and $n-r$ columns.
Next, we form a matrix $\mx{B}$ consisting of the rows that were not crossed out,
which then consists of these $r$ rows of $\mx{A}$. The matrix elements of $\mx{S}$ are also included in $\mx{B}$,
i.e., the minor $d$ can also be obtained by further crossing some columns from $\mx{B}$,
and since $d\neq 0$, the columns of $\mx{S}$ must be linearly independent.
Since $\mx{S}$ is contained in $\mx{B}$ there are at least $r$ linearly independentcolumn vectors in
$\mx{B}$ and we come to the conclusion that the rank of $\mx{B}$ is at least $r$.
However, $\mx{B}$ has only $r$ rows and therefore the rank must be exactly $r$ and the
row vectors of $\mx{B}$ must be linearly independent.
Therefore, $\mx{A}$ must also have at least $r$ linearly independentrow vectors,
i.e., $\rank(\mx{A})\geq r$.
$ii)$
The purpose of this step is to show that the rank of $\mx{A}$ is at most $r$.
Let us assume that $\mx{A}$ has $t$ linearly independent rows.
We create a matrix $\mx{C}$ consisting of these rows and this matrix must have rank $t$,
which means there are $t$ linearly independent columns in $\mx{C}$ as well.
Next, we take these $t$ linearly independent columns by crossing out the other ones.
This results in a $t\times t$ matrix, whose determinant is $\neq 0$ since its columns
are linearly independent. This minor is also a minor of $\mx{A}$.
Hence, the rank of $\mx{A}$ cannot be $> r$, since there then need to more more than $r$
linearly independent rows in $\mx{A}$.
$\square$
As a consequence of Theorem 8.9 and Theorem 7.8,
where the latter says that the inverse of a matrix only is defined if its determinant is non-zero,
we see that if a square $n\times n$
matrix $\mx{A}$ is to be invertible, then we must have $\rank(\mx{A})=n$.
Example 8.11: Assume we want to compute the rank of the following $4\times 4$ matrix,
The determinant of $\mx{A}$ is zero (this is left as an exercise) so the rank must be less than 4.
Next, we calculate one of the $3\times 3$ minors, e.g.,
and since $D_{11}=0$, we do not gain any new knowledge and as it turns out,
we also have $D_{12}=D_{13}=D_{14}=0$ as well.
However, continuing to strike out the second row and the first column, we obtain
and
since at least one of the $3\times 3$ minors is not zero (in this case, $D_{21} \neq 0$),
we can conclude that $\rank(\mx{A})=3$, since we already know that
the only $4\times 4$ determinant is 0, all due to Theorem 8.9.
Example 8.12:The Structure from Sound Problem
In Interactive Illustration 8.2, we studied the problem of determining both sound source positions and microphone positions from the measured sound. In this setting, it turns out that signal processing can be used to find so called time-difference-of-arrival measurements. To make the presentations slightly simpler we will present the same problem, for the case of so called time-of-arrival problems.
In the example we had $m=8$ microphones. Let $\vc{r}_i$, $ i = 1, \dots, m$ be the column vector representations of the 3D positions of these eight microphones. Study the sound position at some of the time instances $t_1, \dots, t_n$, where we have chosen $n$ time instances of the sound recordings. Let $\vc{s}_j$, $ j = 1, \dots, n$ be the spatial coordinates
of these $n$ 3D positions at which the sounds were transmitted.
Assume that we measure the difference in time $t_{ij}$ between the emitted and the received sound for each combination of a time instance $j$ and for each microphone $j$. After multiplying this with the speed of sound $v$, we obtain measured distances
\begin{equation}
d_{ij} = v t_{ij} = {\ln{\mathbf{r}_i - \mathbf{s}_j}}.
\end{equation}
(8.51)
The problem is now to calculate both $\vc{r}_i$ and $\vc{s}_j$ from the distance measurements $d_{ij}$.
as the relative vectors between the first receiver position and the rest and between the first sender position and the rest.
Let these be the column vectors of matrices $\mathbf{R}$ and $\mathbf{S}$ respectively, so that
Here we see that the matrix $\mathbf{R}^\T$ has three columns, so the rank can be at most 3. Similarily
$\mathbf{S}$ has three rows so the rank can be at most 3.
From Theorem 8.6, we see that the rank that is equal to or
lower than the rank of $\mathbf{R}$ and that of $\mathbf{S}$.
Thus the rank of $\mathbf{B}$ is at most 3.
In the last example, we found that a matrix $\mathbf{B}$, which can be formed from measurements, is
related to the parameters that we seek by $\mathbf{B} = \mathbf{R}^\T \mathbf{S}$. How can we calculate $\mathbf{R}$
and $\mathbf{S}$ from $\mathbf{B}$? In other words, how can we factorize the matrix $\mathbf{B}$ and write it as a product
of two factors $\mathbf{R}$ and $\mathbf{S}$? This is part of the topic of factorization and these are useful tools in many applications.
However, this is unfortunately out of scope of this book.
Popup-help:
A basis (bases in plural) is a set of linearly independent (basis) vectors, such that each vector in the space
can be described as a linear combination of the basis vectors.
If the basis is $\{\vc{e}_1,\vc{e}_2,\vc{e}_3\}$, then any vector $\vc{v}$ in the space can be described by
three numbers $(x,y,z)$ such that $\vc{v} = x\vc{e}_1 +y\vc{e}_2 +z\vc{e}_3$. Often the basis is implicit,
in which case we write $\vc{v} = (x,y,z)$, or to make it possible to distinguish between vector elements from
one vector $\vc{v}$ to another $\vc{u}$, we may use $\vc{v}=(v_x,v_y,v_z)$ and $\vc{u}=(u_x,u_y,u_z)$ as notation.
Popup-help:
Let $\mx{A}= \left(\vc{a}_{,1} \vc{a}_{,2} \dots \vc{a}_{,n}\right)$ be an $m\times n$ matrix.
The column space of $\mx{A}$ is then the set of all linear combinations of the
column vectors, $\vc{a}_{,i}$.
Popup-help:
An $n$-dimensional column vector, $\vc{v}$, is represented with respect to a basis and consists of a column of $n$ scalar values.
The vector elements are sometimes denoted $v_1$, $v_2$,..., $v_n$. For two- and three-dimensional vectors, we sometimes
also use $v_x$, $v_y$, and $v_z$.
The notation is
where $\vc{u} = u_x \vc{e}_1$, $\vc{v} = v_x \vc{e}_1 + v_y \vc{e}_2$,
and $\vc{w} = w_x \vc{e}_1 + w_y \vc{e}_2 + w_z \vc{e}_3$.
Note that $\vc{e}_i$ are the basis vectors.
In our text, we also use the shorthand notation $\vc{w} = \bigl(w_1,w_2,w_3\bigr)$, which means the same as above
(notice the commas between the vector elements).
A row vector does not have any commas between, though.
Popup-help:
A mapping $F$ is a rule that is written as $F: N \rightarrow M$,
where for every item in one set $N$, the function $F$
provides one item in another set $M$. Here $N$ is the domain.
Popup-help:
Gaussian elimination is the process where a system of linear equations reduces the number of free variables
on each line by one until the bottom most row says $kx_n=l$, i.e., $x_n=l/k$. After that, the rest of the
free variables can recovered by substitution.
Popup-help:
A homogeneous system of equations is one such that $\mx{A}\vc{x}=\vc{0}$.
Popup-help:
On parameterized form, a line is defined by a starting point, $S$, and a line direction $\vc{d}$, and a scalar
parameter, $t$, such that any point on the line can be obtained from the line equation, i.e.,
$P(t) = S + t\vc{d}$. This representation works in any dimension.
On implicit form, a line is defined by a starting point, $S$, and a normal vector, $\vc{n}$. For any point, $P$,
on the line, it holds that $\vc{n}\cdot (P-S) = 0$. This representation works only in two dimensions.
Popup-help:
A systems of equations is called linear if it only contains polynomial terms of the zero:th and first order,
that is, either constants or first-order terms, such as $9x$, $-2y$, and $0.5z$.
Popup-help:
The vector, $\vc{u}$, is a linear combination of the vectors, $\vc{v}_1,\dots,\vc{v}_n$ if
$\vc{u}$ is expressed as $\vc{u} = k_1\vc{v}_1 + k_2\vc{v}_2 + \dots + k_n \vc{v}_n$.
Popup-help:
The set of vectors, $\vc{v}_1,\dots,\vc{v}_n$, is said to be linearly independent, if the equation
only has a single solution which is $k_1 = k_2 = \dots = k_n =0$.
If there is at least one other solution, then the set of vectors is linearly dependent.
Popup-help:
A mapping $F$ is linear if it satisfies
Popup-help:
Assume that $\mx{A}$ is an $n \times n$ matrix. The minor, denoted $D_{ij}$, is the determinant of the $(n-1) \times (n-1)$
matrix that you obtain after removing row $i$ and column $j$ in $\mx{A}$.
Popup-help:
A normal, $\vc{n}$, is part of the description of the two-dimensional line and the three-dimensional plane on implicit form.
A plane equation is
where $\vc{n}$ is a vector that is orthogonal to the plane, $P$ is any point on the plane,
and $S$ is a known point on the plane.
Similarly, a line is $\vc{n}\cdot (P-S)=0$, where $\vc{n}$ is a vector orthogonal the line, $P$ is any point on the line, and
$S$ is a particular point on the plane. The plane can also be written as $ax+by+cz+d=0$ and a line can be written
as $ax+by+c=0$.
Popup-help:
If $\mx{A}$ is an $m\times n$ matrix (i.e., with $m$ rows and $n$ columns), then
the entire set of solutions to $\mx{A}\vc{x}=\vc{0}$ is called the null space of $\mx{A}$.
The dimension of the null space is denoted $\nullity(\mx{A})$.
The solution, $\vc{x}$, is on the form $\vc{x}(t_1,\dots t_k) = t_1\vc{p}_1 + \dots + t_k\vc{p}_k$,
where $k$ is the dimension of the null space and the vectors $\vc{p}_i$ are called null vectors.
Popup-help:
Loosely speaking, orthogonal projection is such that a right angle (orthogonal) is formed between the
projected object and the vector from the projected object to the object itself.
More strictly, if $\vc{v}$ is a non-zero vector, then the orthogonal projection of $\vc{u}$ onto $\vc{v}$ is denoted $\proj{\vc{v}}{\vc{u}}$,
and is defined by
Note that if $\ln{\vc{v}}=1$, i.e., $\vc{v}$ is normalized, then the expression for projection gets simpler:
$\proj{\vc{v}}{\vc{u}} = (\vc{u} \cdot \vc{v})\vc{v}$.
Popup-help:
On parameterized form (also called explicit form), a plane equation is given by $P(t_1, t_2) = S + t_1 \vc{d}_1 + t_2\vc{d}_2$,
where $S$ a point in the plane and $\vc{d}_1$ and $\vc{d}_2$ are two non-parallel non-zero vectors
on the plane. Using the parameters $t_1$ and $t_2$, any point $P(t_1,t_2)$ can be reached.
On implicit form, a plane equation is given by $\vc{n}\cdot(P-S)=0$, where $\vc{n}$ is the normal of
the plane and $S$ is a point on the plane. Only for points, $P$, that lie in the plane is the
formula equal to $0$.
Popup-help:
Let $\mx{A}$ be a matrix. The row rank of $\mx{A}$, denoted $\rowrank(\mx{A})$, is the maximum number of linearly
independent row vectors of $\mx{A}$. Similarly, the column rank, denoted $\colrank(\mx{A})$, is the maximum number of
linearly independentcolumn vectors of $\mx{A}$. It can be shown that $\rowrank(\mx{A}) = \colrank(\mx{A})$,
and therefore, we can simply use $\rank(\mx{A})$. This also means that $\rank(\mx{A})= \rank(\mx{A}^{\T})$.
Popup-help:
A matrix $\mx{A}$ is said to be on row echelon form if its
rows with only zeroes (if any) are placed at the bottom, and the first non-zero coefficient, $a_{ij}$, on
each row with number $i$ is placed such that all other rows with a first non-zero coefficient, $a_{i'j'}$,
fulfills $i' < i$ (the row $i'$ is above row $i$) and $j' < j$.
A matrix on row echelon form can be obtained by applying Gaussian elimination until no more reductions can be made.
This gives a triangular-like structure for matrices on row echelon form, e.g.,
Therefore, this type of matrices are sometimes said to be a "staircase equivalent" matrix of $\mx{A}$.
Popup-help:
Let $\mx{A}= \left(\begin{array}{c} \vc{a}_{1,}^\T\\ \vc{a}_{2,}^\T\\ \vdots \\\vc{a}_{m,}^T \end{array}\right)$ be an $m\times n$ matrix.
The row space of $\mx{A}$ is then the set of all linear combinations of the
row vectors, $\vc{a}_{i,}^\T$.
Popup-help:
A row vector is column vector that has been transposed so that the column now is a row.
In this book, column vector are the default vectors, and a row vector can thus be produced
from a column vector, $\vc{v}$, simply by transposing, i.e., $\vc{v}^\T$.
A row vector is thus $\vc{v}^\T = \bigl(v_1\ \ v_2 \ \ v_3\bigr)$. Note the spaces between the vector elements.
This is in contrast to a column vector,
$\vc{w} = \begin{pmatrix} w_1 \\ w_2 \\ w_3 \end{pmatrix}$.
Popup-help:
A system of equations with $n$ equations is called triangular if row $i$, where $i\in\{1,\dots,n\}$,
has $n-i+1$ free variables.
Popup-help:
The zero vector, $\vc{0}$, has length 0, and can be created from one point, $P$, as $\vc{0}=\overrightarrow{PP}.$