This section presents some of the most common concepts from linear algebra that have direct application to robotics and autonomous systems. The interested reader is referred to the numerous excellent texts on this topic, such as [18], for a detailed discussion. This book will restrict its attention to linear algebra over thevector space of
‐tuples of real numbers
. The operations ofvector addition andmultiplication of vectors by scalars are defined intuitively in terms of component‐wise operations. If
,
, and
, for addition,
Multiplication of a vector
by a scalar
is likewise defined component‐wise as
One of the implications of this definition is that for a subset
to be a vector subspace of
, thezero vector must be contained in
. Since the scalar
can always be chosen, the requirement that
be closed with respect to multiplication of vectors by a scalar means that
for any vector
. The most common way to define vector subspaces is by considering thespan of a a fixed finite set of vectors.
The definition above asserts that the span of a set of vectors is in fact a vector subspace. This is not difficult to see. A vector subspace must be closed with respect to the operations of vector addition and multiplication of a vector by a scalar. Suppose that
are two arbitrary vectors in the span of the vectors
. By definition there must be two sets of coefficients
and
such that the finite linear combinations of these vectors may be defined as
and
However, if this is the case, the sum can be written as
and it follows that the vector
lies in the span of this set of vectors. Similarly, if
lies in the span of the set of vectors
and
is any real number, it can be written that
Thus, the vector
also lies in the span of these vectors. The span of the set of vectors is therefore closed with respect to vector addition and multiplication of vectors by scalars, so it is a vector subspace of
.
The reader should carefully note that the above definition states that the range and nullspace of a matrix
are vector subspaces of
and
, respectively. Since the range of a matrix
is defined to be the span of its column vectors, it is closed under the operations of vector addition and multiplication of vectors by scalars, following the reasoning after Definition A.2.
The definitions of the range and nullspace of a matrix
have been introduced because they are immediately useful in the analysis of numerous problems in robotics. The following theorem shows that the range of a matrix
and the nullspace of the transposed matrix
induces anorthogonal decomposition. This decomposition is useful in applications where a matrix equation must be solved; it enables a geometric interpretation.
This theorem is interpreted as follows. For any vector
in
,
can be written in the form
where
lies in the range of
,
lies in the nullspace of
, and
.
Now consider the matrix equation
where
is an
matrix,
is an
vector, and
is an
vector. With the fundamental definition of the range of a matrix given in the previous section, an alternative way to express the fact that matrix Equation (A.2) has at least one solution may be defined.
Theorem A.2 summarized the fact that the existence of a solution to the matrix Equation (A.2) can be interpreted as a requirement that vector
lies in the range of the matrix
. The next theorem shows that multiple solutions to the matrix equation exist if and only if the nullspace of the matrix
contains some non‐zero vector.
It has been shown that the existence of a solution of the matrix Equation (A.2) can be couched in terms of the range of the matrix
in Theorem A.2. The characterization of multiple solutions of this equation is given in terms of the nullspace of the matrix
in Theorem A.3. While these two theorems provide a geometric interpretation of the solution of the matrix Equation (A.2), it is also common to discuss solutions of this equation in terms ofrank conditions on the matrix
. The definition of the rank of a matrix
is prepared for by introducing the notion oflinear independence of a set of vectors and abasis for a vector subspace of
.
If a set of vectors
is linearly dependent it is possible to express at least one of the vectors in terms of the remaining vectors in the set. To see why this is the case, suppose that the set of vectors is not linearly independent. This means that there is some set of coefficients, not all equal to zero, such that
Suppose, without loss of generality, that the coefficient
in this summation. In this case
can be solved for in the form
Collections of vectors that are linearly independent play a critical role in determining bases for vector subspaces.
The definition of a basis for a vector subspace, as well as its dimension, will make it possible to give a precise description the existence and multiplicity of solutions to the matrix Equation (A.2). This description will be given in terms of the rank of a matrix
.
The column rank of a matrix
is the dimension of the range of
. The following theorem notes that these two values are equal for a given matrix.
One implication of the above theorem is that the largest possible dimension of any subspace
contained in
is the dimension
of the ambient space.
It has been shown that the existence and uniqueness of solutions to the linear matrix equations
can be described in terms of the range and nullspace of the matrix
when
. When
is a square matrix, the existence and uniqueness of the solutions to this equation can be discussed in terms of theinvertibility of
.
The following theorem summarizes several well known facts relating the inverse, rank, range, nullspace, and determinant of the matrix
to the existence and uniqueness of solutions to the linear matrix equation
.
The proof of all of these properties can be carried out using the definitions and theorems developed thus far, but in some cases they can be lengthy. With the use of thesingular value decomposition, introduced in Section A.1.4, it is straightforward to show the equivalence of these properties. Therefore, the proof of this theorem is deferred until the next section.
Some of the common, fundamental theorems that describe the matrix Equation (A.2) have been discussed in Sections A.1.1 and A.1.2. It is often the case that the matrix equation
cannot be solved exactly. It may be that the information given isinconsistent with the form of the matrix equation. This often occurs if the data in the coefficient matrix
or the right hand side
is obtained via experiment. Even if the data is known precisely or analytically, it can also happen that the solution
to this equation is desired if it exists, but it is sufficient if it is merely “close to satisfying” the equation. This occurs often in the synthesis of control laws, for example.
No matter the motivation, vectors
that only approximately solve the matrix Equation (A.2) can be found by replacing Equation (A.2) with an optimization problem. A vector
should minimize the norm of theresidual ordefect
The norm
is greater than or equal to zero for all possible choices of
, and it can only equal zero when
equals zero. Thus, if the matrix Equation (A.2) has an exact solution, then it exactly solves the optimization problem in Equation (A.3). However, it is always possible to solve Equation (A.3), even if Equation (A.2) does not have an exact solution. This fact enables us to define approximate solutions of the matrix Equation (A.2).
The solution of the alternative optimization problem in Equation (A.3) uses a property of orthogonal matrices derived in Chapter . The norm of a vector does not change when acted upon by arotation matrix. For the problem at hand, consider the norm of the residual
when this vector is premultiplied by some orthogonal matrix
,
The minimization in Equation (A.3) is equivalent then to the minimization problem
The orthogonal matrix
can be chosen to simplify the computational task.
In practice there are many reasonable choices for the construction of the matrix
. Methods based on Givens transformations, Householder transformations, and Jacobi rotations are just a few of the orthogonal matrices that can be used in this approach. The interested reader should consult [18] for a complete summary. This book will only discuss one such method based on the singular value decomposition of a matrix.
Suppose that
, so that the problem isoverdetermined orexactly determined. Partition the matrix
so that
corresponds to the first
columns of the matrix
and
corresponds to the next
columns of
in
and conformally partition the matrix
into the
diagonal matrix
and the
zero matrix
This decomposition is appled to theoptimization problem in Equation (A.4) by choosing
to be equal to
.
The term
is independent of the choice of vector
. The first term, however, can be solved using the following theorem.
Next, the proof of Theorem A.7 will be reconsidered using the singular value decomposition of the matrix
.
The derivation of theimage based visual servo control law required a solution to the matrix equation
for the velocity of the origin and angular velocity of thecamera frame in theinertial frame. The matrix
was shown to have dimension
where
is the number ofsystem feature points. This matrix is not square in general, which complicates the discussion of the solution of the associated matrix equation. Rewrite the equation above in the form
where
is an
matrix,
is an
vector of unknowns, and
is a
vector of given right hand side values.
Thealgebraic eigenvalue problem appears in many contexts in the study of dynamics and control of robotic systems. It is used to definespectral decompositions that are important in understanding the stability of linear systems, and it is used to construct diagonalizations of real symmetric matrices. This latter construction is important in the definition ofprincipal axes for theinertia matrix of a rigid body.
It is essential to note that the definition of the eigenvector requires that
is a non‐trivial vector. In other words,
cannot identically equal zero. This definition also shows that an eigenvector is a direction that is invariant under the application of the matrix
. It is also evident that if
is an eigenvector of the matrix
, then so is
for any scalar
. The eigenvectors of
are said to be determined only up to a scalar multiple.
One of the most common descriptions of solutions to the algebraic eigenvalue problem can be inferred by applying the theory from the last few sections about the solvability of linear systems.
This theorem shows that the eigenvalues of the
matrix
are defined from the roots of the characteristic polynomial
. Even if
is a real matrix, so that the coefficients in the characteristic polynomial are real, the roots of the characteristic polynomial can still be complex. Also, while every
matrix has
eigenvalues, it is not always the case that there are
distinct eigenvectors. This phenomenon is studied in more detail in Section A.2.2.
For an important class of matrices, those matrices that areself‐adjoint, a full set of
eigenvectors can be found. A matrix
is self‐adjoint whenever
where
is the complex conjugate of the matrix
. In the applications under consideration, only real matrices will be considered, and a real matrix is self‐adjoint when it issymmetric. The following theorem shows that it is possible to construct an orthonormal basis for
in terms of the eigenvectors of
in this case. This result is a special case of thespectral theory associated with self‐adjoint matrices.
The spectral theory for self‐adjoint matrices in the last section provides a way to diagonalize a self‐adjoint matrix in terms of its eigenvectors. As already noted, not all
matrices have a full set of
linearly independent eigenvectors. A nearly diagonal factorization that is qualitatively similar to the spectral factorization can be achieved by employing theJordan canonical rorm of a general
matrix. This factorization is particularly useful for the study of the stability of linear systems, as studied in Chapter utilizing feedback linearization to compensate for nonlinear terms.
Note that the columns of the matrix
, in contrast to the matrix
for the self‐adjoint case, is not necessarily constructed from the eigenvectors of the matrix
. The columns of
are sometimes denoted the generalized eigenvectors of the matrix
. The interested reader is referred to [18] for a discussion.
It is often the case in the study of the dynamics of robotic systems that a matrix system of equations that have the form
must be solved, where
is an
matrix,
is an
vector of unknowns, and
is a known
vector. In some cases, these equations must be solved symbolically, while in other cases a numerical solution suffices. Several conditions in Sections A.1.3, and A.1.4 guarantee the existence and uniqueness of solutions to systems of equations having the form shown in Equation (A.15). In particular, it has been shown that if there is a unique solution
to Equation (A.18), the solution of these equations is equivalent to the calculation of the inverse
of the matrix
. The proof of Theorem A.7 following Theorem A.9 shows that the singular value decomposition can be used to calculate the inverse of a matrix when it exists. Unfortunately, the singular value decomposition is computationally expensive compared to several alternative numerical algorithms; thus, other methods are often preferred in robotic system applications. This section will discuss one such choice that uses sequences of Gauss transformations to derive the inverse of a matrix
, or equivalently, to solve equations having the form shown in Equation (A.15).
The strategy underlying the use of Gauss transformations to solve Equation (A.15) is easy to describe. Suppose two sequences can be found such that
and
of
invertible matrices such that premultiplication of
by the product of the
and postmultiplication of
by the product of the
is equal to the identity matrix, as in the equation
If this is the case, the inverse if the matrix
can be written as
Note that it is critical in this construction that each of the matrices
or
in the sequence is invertible for each
and
. The principle conclusion discussed in this section is that such a product can be constructed provided certainpivot elements of the matrix
are not equal to zero. The construction is carried out usingGauss transformations.
A Gauss transformation is a matrix that is equal to the identity matrix except for a single off‐diagonal entry. Such a matrix has either the lower triangular form shown in Figure (A.1) where
is the off‐diagonal entry located at row
and column
, or it has the upper triangular form shown in Figure (A.2) where
is the off‐diagonal entry located at row
and column
.
Figure A.1 Gauss transformation, lower triangular form.
Figure A.2 Gauss transformation, upper triangular form.
It is clear that each of these matrices has an inverse, and it is simple to calculate the inverse as
or
A Gauss transformation has several properties that can be attributed to its highly structured form that are crucial in applications. Lower triangular matrices
will be discussed first and several properties of these matrices will be derived. It will be shown how to select such a matrix so that premultiplication of
by the matrix will zero out a specific entry in
below the main diagonal. If a matrix
is premultiplied by a judiciously selected sequence of such matrices, it is possible to make all of the entries below the main diagonal of the resulting matrix equal to zero. In other words, the product
can be made upper triangular.
Then, it will be argued that similar constructions based on the upper triangular matrices
can be achieved. Postmultiplication of the matrix
by a carefully constructed
can be made to zero out a specific entry above the main diagonal. Postmultiplication of
by a suitably crafted collection of these matrices will result in all entries above the main diagonal set equal to zero. In other words, the product
can be made lower triangular.
Finally, both types of matrices will be used together to achieve the
decomposition of a matrix
into the product of a lower triangular and an upper triangular matrix. Consider first the Gauss transformation
. It is crucial to the understanding of Gauss transforms to observe that premultiplication of a matrix
by
only changes row
and leaves the rest of the matrix
unchanged. In fact, it is straightforward to observe that the only terms that are modified by premultiplication by
are denoted by
for
in the product below
Each newly modified term in row
denoted
in Equation (A.18) is given by
for
. Consider this equation for the choice of
,
If the diagonal term
, it is possible to choose
so that the Guass transformation
introduces a zero at the entry
of the matrix
. The condition
can be inforced by choosing
since
The diagonal entry
is known as thepivot term, and the success of this strategy relies on the fact that
is not equal to zero. The observations that premultiplication by
affects only row
, and that
can be chosen so that
, leads to a general procedure for constructing a sequence of matrices
that drives the matrix
to a lower triangular form. The matrices
can be successively chosen to zero out entries while progressing down each column below the main diagonal. That is, entries are zeroed out, as depicted in Figure (A.3). It is left as an exercise to show that this algorithm is guaranteed to zero out all the sub‐diagonal entries of the matrix
, provided that the pivot elements
for
are non‐zero. These steps are summarized in Figure (A.4).
Figure A.3 Progression of zeros below diagonal.
Figure A.4
Gauss transformations to zero sub‐diagonal entries of
.
All of the observations made thus far regarding the lower triangular Gauss transformations
can be extended to the upper triangular Gauss transformations
with some minor modifications. Postmultiplication of
by
modifies only the entries of the column
. Equation (A.21) depicts the elements
for rows
in column
that are modified when we compute the product
As in the case of
, it is possible to introduce a zero at location
provided the pivot term
on the diagonal is not zero. Each term
for rows
in column
satisfies the equation
For the specific choice of
,
If
is defined as
,
Analogous to the strategy summarized in Figure (A.4), the sequence of upper triangular Gauss transformations can be used to make the product
lower triangular. Figures (A.5) and (A.6) summarize the sequence of entries that are successively set to zero in this strategy.
Figure A.5 Progression of zeros above diagonal.
Figure A.6
Gauss transformations for upper triangular factorization of
.