Consider the dot product of two vectors and .
Now consider the transformed vectors and . The dot product
of the transformed vectors is:
So when looking at a matrix product , know that it will
carry out the transformation and then dot product of any two vectors if they
are placed on either side of the product.
Special case: length² of transformed vector
Let . What is the of ? Let's make this a function
that takes in any vector and returns the length-squared of
the transformed vector . This function is simply .
We can see as being the implementation of this function.
acts similarly, but with an inverse transformation.Again,
consider the dot product of two vectors and . Now
consider the transformed vectors and , where we are
now inverting the transformation . The dot product of the transformed
vectors is:
So when looking at a matrix product , know that when this
matrix is inverted, , it will carry out the inverse
transformation and then dot product of any two vectors if they are placed on
either side of the product.
for orthonormal
When is orthonormal, , so it's easy to see
that:
So orthonormal could be thought of as preloading the inverse
then dot product of two vectors, even without the matrix inverse operation
appearing in the formula. However, more pertinently, we have
, so is the identity matrix. And so we
have arrived at the intuitive idea that a rotation transformation, ,
doesn't change the length of a vector or the angle between two vectors.
when not full rank (projection)
If has say 2 orthonormal columns in , then
it doesn't have an inverse, but we can ask for the projection of a vector
onto the column space of by computing . Here,
projects onto the column space of , and then
rehydrates the projection so that it's expressed in the
original space.
Example
Consider a random vector distributed
as a standard normal distribution, . We can
derive a random variable for some matrix . Using the
change of variable formula for distributions, we can derive the distribution
of .
If and were 1 dimensional random variables, with
, the change of variable formula would be:
The multidimensional version, with , of this formula is:
and applying this to the standard normal distribution, we get:
where is the covariance matrix of (aka the cross-covariance matrix .
So while the covariance matrix of is specified to
parameterize the multivariate normal distribution, under the hood, this
matrix is preloading the dot product of two vectors that are brought back
to the original space by inverting the transformation .
Note: the fact that is the covariance matrix of is
another separate result that is a special case of the more general theorem:
for which there is another page.