Matrices as mappings on vector spaces

Consider a matrix images. When we multiply a vector images by this matrix, we get a vector images. This suggests that a matrix is more than just an arrangement of numbers, but it can be regarded as an operator mapping images to images:

images

Given the rules of matrix algebra, it is easy to see that this mapping is linear, in the sense that:

images

This means that the mapping of a linear combination is just a linear combination of the mappings. Many transformations of real-world entities are (at least approximately) linear.

If we restrict our attention to a square matrix images, we see that this matrix corresponds to some mapping from images to itself; in other words, it is a way to transform a vector. What is the effect of an operator transforming a vector with n components into another vector in the same space? There are two possible effects:

  1. The length (norm) of the vector is changed.
  2. The vector is rotated.

In general, a transformation will have both effects, but there may be more specific cases.

If, for some vector v, the matrix A has only the first effect, thus means that, for some scalar images, we have

images

A trivial case in which this happens is when A = λI, i.e., the matrix is a diagonal of numbers equal to λ:

images

Actually, this is not that interesting, but we will see in Section 3.7 that the condition above may apply for specific scalars (called eigenvalues) and specific vectors (called eigenvectors).

If a matrix has the only effect of rotating a vector, then it does not change the norm of the vector:

images

This happens if the matrix A has an important property.

DEFINITION 3.2 (Orthogonal matrix) A square matrix images is called an orthogonal matrix if PTP = INote that this property also implies that P−1 = PT.

To understand the definition, consider each column vector pj of matrix P. The element in row i, column j of PTP is just the inner product of pi and pj. Hence, the definition above is equivalent to the following requirement:

images

In other words, the columns of P is a set of orthogonal vectors. To be more precise, we should say orthonormal, as the inner product of a column with itself is 1, but we will not be that rigorous. Now it is not difficult to see that an orthogonal matrix is a rotation matrix. To see why, let us check the norm of the transformed vector y = Pv:

images

where we have used Property 3.1 for transposition of the product of matrices. Rotation matrices are important in multivariate statistical techniques such as principal component analysis and factor analysis.


Comments

Leave a Reply

Your email address will not be published. Required fields are marked *