Appendix A — A General Introduction to Matrices
A.1 Matrices, Vectors, and Scalars
A.1.1 Definition of a Matrix
A rectangular array is a set of numbers ( or symbols representing numbers) where the numbers are arranged in rows and columns where every column has the same number of rows and every row has the same number of columns.
A matrix is a rectangular array where each entry is called an element of the matrix. A matrix is often denoted by a capital letter, e.g.,
A.1.2 The Size or Dimension of a Matrix
The size or dimension of a matrix
Alternative Notation: dim(
Examples:
Matrix
Define matrix B
Matrix
A.1.3 Row and Column Vectors
When a matrix only has one row it may be called a row vector with dimension
- Think of the row vector as a row in a spreadsheet that contains all the values of the variables/attributes for one case or observation.
When a matrix only has one column, it may be called a column vector with dimensions
- Think of a column vector as a column in a spreadsheet with the values of the observations for one variable/attribute.
A.1.4 Scalars
A scalar is a matrix of size
ex.
In the programming language R, a
A.1.5 General Notation for a Matrix
If matrix
or, in shorthand,
At times you may see the dimension of matrix indicated using a subscript so
A.2 Special Matrices
A.2.1 A Square Matrix
A square matrix is a matrix with the same number of columns and rows, i.e., of size
A.2.1.1 Determinants of Square Matrices
- Given a square matrix, one can calculate a scalar number known as the determinant of the matrix. The determinant of matrix
is denoted in several ways: or for a matrix as
The determinant “determines” or describes how the matrix structure may affect other matrices when used in operations.
The determinant of a matrix can be calculated using only the numbers in the matrix.
For a matrix of order 2, the calculation is straightforward:
- For higher order matrices, the calculations get more complicated.
A.2.2 A Symmetric Matrix
A symmetric matrix
A.2.3 A Diagonal Matrix
A diagonal matrix is a square matrix
A.2.4 The Identity Matrix
The Identity matrix is a diagonal matrix where every diagonal value is equal to 1 and all off-diagonal elements are 0.
A.2.5 The and Vectors
A column vector of all 1s of size
A column vector of all 0s of size
A.3 Matrix Operations
A.3.1 Equality of Matrices
Definition: Two matrices
For two matrices to be equal they must be of the same size (dimension).
Example: If
then
A.3.2 Addition of Matrices
To add two matrices
The addition of two matrices,
Example: if
then
A.3.3 Multiplying a Matrix by a Scalar
Let
Multiplying a matrix by a scalar results in a matrix where every element has been multiplied by the scalar.
Example for matrix
Scalar multiplication is commutative:
A.3.4 Multiplying a Matrix by another Matrix
A.3.4.1 Definition
Let
To multiply two matrices
Multiplying
where
Example:
Here dim(
Therefore,
A.3.4.2 Properties
Matrix multiplication is not commutative! If dim(
) = and dim( ) = , the product does not exist when . If the and are both square with the same dimension then may exist but it is not necessarily true that .Matrix multiplication is distributive under matrix addition such that for three matrices of the correct sizes,
.
A.3.4.3 Exercises
A.3.4.3.1 Is matrix multiplication commutative?
Compute
and
In matrix multiplication, order matters! Often
A.3.4.3.2 Is matrix multiplication distributive?
Check if
A.3.4.4 Matrix Multiplication Terminology
Depending upon the sizes of
A.4 The Transpose of a Matrix
A.4.1 Definition
Let
The transpose of
If
If
Then
Example:
The transpose of a matrix always exists.
It is common in print to see column vectors represented as their transpose e.g.,
A.4.2 Properties of a Transpose
- Identity:
- Transpose under Multiplication
Notice when taking the transpose of a product, we switch the order when multiplying the product of the transposes.
Example:
so
Then
So
This property is often used, going both ways, in regression. You should be familiar with it.
A.5 More on the Determinant of A Square Matrix
- If
is any square matrix that contains a row (or column) or zeros, then = 0.
Example: Let
Then
- If
is an diagonal matrix, then is the product of the entries on the main diagonal, i.e.,
Example: Let
- If
and are square matrices of the same size, then
A.6 The Inverse of a Matrix
A.6.1 Definition
A matrix
If
A.6.2 Remarks
- Only square matrices can have inverses.
- Not All square matrices have inverses.
- If
exists for matrix , then the inverse is unique. - If a matrix
has an inverse, we say is invertible. - Matrix
is invertible, if and only if .
Proof:
If
where the identity matrix
- When
that means at least one row of the matrix can be calculated as a linear combination of other rows in the matrix. If , then we say the matrix is singular or non-invertible. - If
is invertible, then the matrix equation has a unique solution.
A.6.3 Calculating the Inverse of a Matrix
Calculating the inverse of a matrix of order 2 is straightforward:
Since the
When calculating the determinant (or inverse) of a matrix using a computer, especially when the matrix is large or has values that differ by several orders of magnitude, special techniques are required to minimize the risk of getting values of 0 or close to 0 just due to the limited precision of a computer.
Example:
Define matrix
Then
and
A.6.4 Properties of the Inverse
Assuming an
Notice when taking the inverse of a product, we switch the order when multiplying the product of the inverses.
A.6.4.1 Orthogonality
- If
is an matrix, is said to be orthogonal if . - An
matrix is orthogonal if and only if . - The determinant of an orthogonal matrix is either 1 or -1.
- Orthogonal matrices have nice properties such as enabling numerical stability in computer-based linear regression algorithms.
A.7 Linear Independence and the Rank of a Matrix
A.7.1 Definition of Linear Independence
Let
The vectors
If the vectors
Practical Interpretation: If a set of vectors is linearly dependent, then at least one of the vectors is redundant (does not add any new information to the other vectors).
Example:
If a set of vectors
Practical Interpretation: If a set of vectors is linearly independent, then all vectors contribute new information.
Either row vectors or column vectors can be linearly dependent.
- Row vectors: Cases can be repeated so have identical information.
- Column vectors: Variables contain similar or redundant information.
A.7.2 Definition of the Rank of a Matrix
The rank of a matrix is the maximum number of linearly independent columns (rows).
A.7.3 Properties
- The inverse of a
matrix exists if an only if . We say is of full rank. - For an
matrix , . - if
, then .
A.8 Probability Results for Random Vectors
A.8.1 Definitions for a Random Vector.
Let
A vector
Since each random variable
The joint distribution describes how the
Given a random vector
- The expected value of
is the mean vector of and represents the center of the joint distribution of . This is denoted as:
- Given a
-dimensional random variable , there exists a symmetric matrix, , called the variance-covariance or covariance matrix of . It has the form:
is equivalent to the variance of just the random variable itself. provides us information on how the random variables are related (distributionally).If
, then are uncorrelated.
Example: Contour and surface plots of the joint density function of two normally distributed random variables with a joint distribution.
- In the first case, the two variables are independent, i.e., covariance is 0. The contour plot makes it easy to see
. - In the second case, the covariance is greater than zero. The contour plot shows how the values of
and are related and that creates an angle in the joint density function.
A.8.2 Definition of a Multivariate Normal Distribution
A random vector
where:
The joint Normal distribution of
A.8.3 Properties
- If
, then the marginal distribution of is . - If
then
and
A.9 Matrices and the Classic Multiple Linear Regression Model
Let there be
Assume the true model is:
The Equation A.3 model can also be written as a set of
One can then convert the
Then define
One can combine Equation A.4 and Equation A.5 to get the matrix form of the true model:
The multiplication of
In the Equation A.6 form of the linear model, the matrix
A.10 A Geometric Perspective on Matrices and Matrix Operations
The previous sections discuss matrices from an analytical perspective. This section will look at matrices from a geometric perspective. This is based heavily on ideas and content in the YouTube series by 3Blue1Brown called Essence of linear algebra.
A.11 Vectors, Points, Spans, and Bases in
A.11.1 Vectors with One Element:
Given a vector with one element:
We can multiply this vector by any scalar value and get another value on the number line.
This is known as scaling the vector.
A.11.1.1 The 1-d Unit Vector
We can represent any point on the number line by scaling a vector of length 1,
The red vector in Figure A.2 above is a unit vector
A.11.1.2 Linear Combinations of 1-d Vectors
We can also create linear combinations of 1-d vectors by adding them together.
The set of all linear combinations of a vector, say
In 1-d, the span of
Thus
Since there is only 1 vector,
The set of vectors
Here,
A vector space of dimension
- One can do addition and scalar-multiplication operations
- Those operations are commutative and distributive
- The subset contains the zero vector 0 (the origin).
- If the subset contains
then it contains for every scalar . - If subset contains
and , then it contains .
A.11.2 Vectors with Two Elements:
Let’s consider a vector with two elements:
With two elements, this can be thought of as representing a point on a 2-dimensional
- The element values are the distance from 0 in the
direction and then the direction. - We use the following notation to denote the
and elements of the 2 dimensional vector.
We can still scale a two dimensional vector by multiplying it by a scalar.
However, we cannot create every point in the 2-d space with one vector, only those points in the span of the original vector.
A single vector does not have a span that generates a 2-d space.
To generate the 2-d space, we need a second vector that is not in the span of the first vector.
A.11.2.1 Unit Vectors in 2-d
Let’s define two unit vectors,
A.11.2.2 Scaling and Vector Addition in 2-d
We can consider the vector
The orange line in Figure A.4 represents the addition of two vectors as discussed in Section A.3.2.
Geometrically, it can be seen as moving the tail of the second vector to the head of the first vector. Their sum is the new location of the head of the second vector.
Let’s add
- We plot the two vectors with tails at the origin and then move the second (red) vector so its tail is at the head of the first vector.
- The result is the head of the shifted second vector, here
A.11.2.3 Linear Combinations and Linear Independence
With two 2-d vectors,
If
Assuming
If two vectors are linearly dependent, it means they share the same span. Thus the dimension of the span of the set
Consider
- We can find scalars
such that .
- Thus
and are linearly dependent and share the same span as seen here.
Reducing the span by one dimension from 2-1 is equivalent to converting the vector space from a 2-d plane to a 1-d number line.
To generate a 2-d space, we need a second vector that is not in the span of the first vector i.e., is linearly independent of the first vector.
The unit vectors
Thus the set
The set
A.11.2.4 Basis Vectors
Is the set
Any set of two linearly independent vectors
- The vectors do not even have to be orthogonal.
Changing the basis is equivalent to changing the reference coordinate system for the space.
- Any time we interpret the values of the elements in a vector, we are implicitly using the bases vectors to shape our interpretation.
Thus we normally assume we are using
A.11.3 Three dimensions and Higher:
The same concepts from
A.11.3.1 Vector Elements
For three dimensions
For
A.11.3.2 Span and Bases in
With two linearly independent vectors in 3-d vector space, their span is still in 2-d space.
- In 3-d space, graphing the result of every linear combination of 2 linearly independent vectors, e.g.,
, creates a 2-d plane in 3-d space, centered on the origin. - In
, the result of every linear combination of linearly independent vectors creates an -dimensional hyper-plane in space, centered on the origin.
If we add a third vector in
- Add the first two vectors by moving the tail of the second to the head of the first.
- Then, move the tail of the third vector to the head of the second.
- The new location of the head of the third vector is the result.
This is equivalent to:
The span of the three 3-d vectors is the set
- The dimension of the
is the number of linearly independent vectors in . - If one or more of the vectors in
is linearly dependent on one or more of the others, (it is in the span of one or their linear combination), the dimension of set is still the number of linearly independent vectors in , so may be 1 or 2. - If the three vectors are all linearly independent of each other, the span of
has dimension 3, the three vectors can generate , and the set can serve as a basis for . - This can be thought of as taking the span of the first two vectors (a horizontal
plane) and using the third vector to move it from in the dimension.
In
- The dimension of a set of vectors
is the number of linearly independent vectors in . - If one or more of the vectors in
is linearly dependent on one or more of the others, (it is in the span of one or a linear combination of others), the dimension of set is still the number of linearly independent vectors in , so may from range from 1 to . - If the
vectors are all linearly independent of each other, the span of has dimension , the vectors can generate , and the set can serve as a basis for .
In
A.12 Matrices as Linear Transformations of Vectors
A.12.1 Linear Transformations in General
A transformation is a function that maps an input value to an output value.
We are interested in using a function to map a vector to a different vector in
- The function can be considered the rules for reshaping and moving the input vector to look like the output vector.
- This is equivalent to applying a function to a point in
to produce another point in .
To make the transformation linear, we have to add two constraints to the function.
- It must preserve the linearity of lines - all input lines must be output as lines.
- The origin must not be shifted.
These are equivalent to transformations that keep all the gird lines on the plan as parallel and evenly spaced.
A linear transformation of any
An example in
- Start with a vector
If we use transformation that moves the vector to
If our transformation moved
We can write this transformation in terms of any input as
We can make this even more general.
We can describe any linear transformation in
- Two for the vector where
lands after the transformation and - Two for the vector where
lands after the transformation.
We combine these four numbers (two vectors) into a 2x2 matrix as
Where the first column describes the new location of
So, to apply a transformation matrix to any vector in
Given a 2x2 matrix
Or in general.
Putting the matrix on the left of the vector is equivalent to using
We have previously seen
A square matrix of size
- The linear transformation is completely described by
numbers which describe the new locations of the unit vectors in the basis of the space. - We can put these numbers into the columns of a square matrix to describe where each of the
unit vectors winds up after the transformation. - Matrix multiplication can be interpreted as reshaping the space to move the input vector into a new position in
.
A.12.2 Executing Multiple Transformations
We often want to execute a sequence of transformations, sometimes called creating a composition of transformations.
This is equivalent to multiplying multiple matrices.
Since each multiplication results in a new matrix describing where each of the
When multiplying a vector by two matrices, we write it in the form from left to right as
where we execute as
As noted in Section A.3.4.3, compositions of matrix transformations (multiple multiplications) are not commutative.
That means, that except for special cases,
A.12.3 Determinants
These linear transformations by matrices are reshaping the space of the input vector.
This means they are often either stretching the space or shrinking the space, with or without some rotations in one or more dimensions.
In
Consider the transformation matrix
- It scales
by a factor of 3 and by a factor of 2. - This means that the 1 x 1 square formed by
and now is a 3 x 2 rectangle so has an area of 6. - We can say this linear transformation scaled the area by a factor of 6.
Now, consider a linear transformation known as a shear transformation.
- It has the transformation matrix
- This leaves
in place but moves over to . - This means the 1 x 1 square is now a parallelogram but it still has area 1.
This scaling factor of a linear transformation matrix is called the determinant of the transformation matrix as seen back in Section A.2.1.1.
This scaling factor applies to any area defined by vectors in
.For
, the matrix will scale the volumes instead of the area.For
, it is the volume of the parallelepiped created by the three unit vectors along each axis.If
is a transformation matrix with determinant 3, it scales up all the areas by a factor of 3.If
is a transformation matrix with determinant 1/2, it shrinks down all the areas by a factor of 1/2.
If a 2-d transformation matrix has a determinant of 0, it shrinks all of the space into a 1-d line or even a single point (0,0).
This is equivalent to the columns of the matrix being linearly dependent.
This can be useful as we will see later on to simplify the representation of a transformation as a matrix.
What does it mean to have a determinant that is
- This is known as a transformation that flips the coordinate reference system or “invert the orientation of space”.
- This can also be visualized as
moving from the usual position to the right of to now being on the left of . - The Absolute Value of the determinant still shows how much the areas have been scaled (increased or shrunk down).
- In
, this means the orientation has been inverted from a “right hand rule” to a “left-hand rule”.
To take a general view, consider the matrix
If
- The transformation creates a new rectangle out of the unit vector 1x1 square as seen in Figure A.7.
-
is the factor for how much is stretched. -
is the factor for how much is stretched -
is the area of the new rectangle compared to the original 1x1 square.
If either
- The transformation creates a new parallelogram out of the unit vector 1x1 square as seen in Figure A.8.
-
is still the area of the new parallelogram, with base and height , compared to the original 1x1 square.
If
- The transformation creates a new parallelogram out of the unit vector 1x1 square and stretches (shrinks) it in the diagonal direction.
When multiplying two matrices,
A.13 Matrices and Systems of Linear Equations
A system of linear equations is a set of equations that describes the relationships among
We want to solve the system of linear equations and transformation matrices can help us do that.
If you have a system of linear equations you can organize it with all of the variables and their coefficients on the left and all of the results (scalars) on the right of the equal sign.
- You may need to make
or coefficients explicit so you have the same number of coefficients in each equation.
Now, convert the system of equations into matrix form by converting the variables and results to vectors and the coefficients into a matrix.
We can label each part as follows
and now rewrite in matrix vector form:
We can interpret the system of equations then as a matrix transformation (or function) of the input vector
Our interest is in figuring out what input vector
There are two cases
- The matrix
squishes the space down by a dimension, i.e., - The matrix
squishes the space in a way that preserves the dimension of the space, i.e.,
A.13.0.1 Solutions with
When
We can find this by reversing the transformation. This reverse of the transformation is its own transformation matrix which is called the inverse of
As an example, if
So in case 2, where
Now we can solve for
This can be interpreted as using
A.13.0.2 Solutions with
When the system of equations has a transformation matrix
However, there can still be a solution to the system of equations if the solution exists in the lower dimensional space.
As an example, if a 2-d matrix squishes down to a line, it could be true that
In
- It would much hard to find a solution for the single line than for the plane even though both have
.
To differentiate the different types of output in higher dimensions, we use the term Rank.
- If the transformation matrix has an output of 1-d, the matrix has Rank = 1.
- If the transformation matrix has an output of 2-d, the matrix has Rank = 2.
- and so on.
The set of all possible outputs of
- You can think about it as the number of columns in
where each represents the effect on the unit vectors. - The span of these vectors is all possible outputs, which by definition is the column space.
- So, Rank is also the number of dimensions of the column space.
- If a matrix has a Rank = the number of columns, which is as high as it could be, the matrix has Full Rank.
If a matrix has full rank, then the only vector that transforms to the origin is the 0 vector.
If a matrix has less than full rank, so it squishes down to a smaller dimension, then you can have a lot of vectors that transform to 0.
If a 2-d matrix squishes to a line, there is a second line, in a different direction, where all the vectors on the second line, get squished onto the origin.
This set of vectors (or planes in higher dimensions) that transform to the origin, to 0, are called the Null Space or Kernel of the matrix.
In a system of linear equations, the Null Space of the matrix
A.14 Change of Bases
In Section A.11.2.4, we discussed that there many possible basis vectors for a vector space.
The choice of basis vectors determines how to describe other vectors in terms of the origin, the direction of movement and the unit of distance.
As an example,
This relationship between the numbers and a geographic interpretation as a vector is defined by the Coordinate System.
-
and are part of a “standard” Coordinate system with length one and horizontal and vertical basis vectors of length 1.
A.14.1 Differing Linear Coordinate Systems
Suppose someone else uses a different set of basis vectors,
What the standard system defines as
The original vector has not moved - it is just being described from the perspective of a different coordinate system.
- The origin is the same
- It is the same approach of scaling each basis vector and adding the results.
- However the orientation of the axes and the scaling of the units is different.
- The choice of these is arbitrary and can be changed to provide a more visually or mathematically convenient perspective of a vector.
- Consider a picture taken by a tilted camera and tilting your head a few degrees so it looks like a normal portrait or landscape perspective. The location of objects and the relationships among objects in the picture did not change, but it might be easier to interpret now that your are looking at it from a new perspective.
In this example, the standard coordinate system would describe
In the alternate system,
A.14.2 Translating (Transforming) a Vector Between Coordinate Systems
Given two coordinate systems we can translate a vector from one representation to the other if we can describe the basis vectors in one system in terms of the other.
Assume there is a vector identified as
What would its description be in the standard coordinate system?
We know how to describe the basis vectors in the alternate system using standard coordinates so we can apply the scale from the alternate system to those.
This is just the same as using the basis vectors of the alternate system, as a transformation matrix to describe the change using the standard coordinate system.
This matrix is called the Change of Basis matrix as it changes
That is equivalent to changing the description of the vector based on scaling the basis vectors
- The change of basis matrix allows us to describe a vector from the alternate system in terms of the standard coordinate system basis vectors
and .
To transform from the standard to the alternate just requires using the inverse of the change of basis matrix.
- To translate a standard coordinate vector
, multiply it by the inverse of the change of basis matrix to get the description in terms of scaling the basis vectors and . - We get the result mentioned earlier.
A.14.3 Translating (Transforming) a Matrix Between Coordinate Systems
We can also translate a matrix that describes a transformation for a vector written in one coordinate system so the translated matrix describes the same spatial transformation from the perspective of the basis vectors in the alternate coordinate system.
- This means we can’t just multiply by the coordinate transformation matrix as that would still be describing the transformation in terms of
and .
Assume we have a transformation matrix
The following steps will translate a transformation matrix written in one system so it can describe the same transformation of a vector in an alternate coordinate system by using the change of basis matrix.
- Start with the vector written in the alternate coordinate system.
- Translate it to the standard coordinate system using the change of basis matrix.
- Transform it with the transformation matrix in the standard coordinate system.
- Translate it back to the alternate system using the inverse of the change of basis matrix.
The result of this series of matrix multiplications in the new matrix on the right which will now create the equivalent rotation of
We can now choose to use different coordinate systems, and each system’s set of basis vectors’ and go back and forth in a way that makes it easier to do our analysis.
- This can be thought of as similar to using a log transform to minimize numerical precision errors when multiplying a very large number and and very small small number on a computer.
- We take the log of each number, add the logged values, and then use the anti-log (or exponentiation) to translate back to the original space to get the final result.
A.15 Eigenvectors and Eigenvalues of a Linear Transformation Matrix
A.15.1 Background
Eigenvectors and their associated eigenvalues have been used in mathematics to “simplify” the analysis of linear transformations since the 18th century.
- Euler was using linear transformations in
to analyze the rotation of solid bodies centered on the origin. - He found that the linear transformation matrix describing the rotation also rotated most vectors.
- However, the vector of the axis of rotation was special in that it did not rotate under the transformation.
- Another way of saying this is the axis of rotation vector remains on its own span after the transformation and is not rotated off of it like most vectors.
- Others studied these special vectors and determined that many (but not all) matrices of linear transformations may have one or more of these special vectors, which remain on their original span, and which may be scaled by the transformation but not rotated off the span.
- In 1904, David Hilbert coined the term eigenvector (German for “own”) as each transformation matrix may have one or more of its “own” special vectors.
- The eigenvalues are the scale factors associated with one more eigenvectors for a matrix.
When a matrix has a set of eigenvectors that span the space, changing the coordinate system to use them as the basis vectors (an eigenbasis), greatly simplifies the transformation matrix.
- Translating the transformation matrix into the eigenbasis coordinate system translates the original matrix into a diagonal matrix.
- The columns now represent each eigenvector.
- The values on the diagonals are the eigenvalues (scale factors) for each eigenvector!
A.15.2 Example
Let’s assume we have a transformation matrix
We can see from the matrix that
is special in that it is not rotated off of its span (the axis). It is only scaled by a factor of 3.Any other vector on the
axis is also scaled by a factor of 3.Thus
is an eigenvector with eigenvalue 3.It turns out the vector
is also not rotated off its span and is scaled by a factor of 2.An other vector in its span is also not rotated off the span and is just scaled by a factor of 2.
Thus
is an eigenvector with eigenvalue 2.
Using eigenvectors as the basis provides a way of describing a linear transformation that emphasizes the effects of the transformation without worrying about which coordinate system is being used to describe it.
A.15.3 Derivation
Given a transformation matrix
To find the eigenvectors and eigenvalues of a matrix
We can do some rearranging to put into a matrix form.
The last line in Equation A.9 says we are looking for a vector
- This means that we are looking for a vector
that is squished by to . - That only happens if
.
We are thus looking for a
As an example, if
- If we had chosen another matrix, the eigenvalue might not be 1.
Since
To check if a value of
Using the formula for a 2x2 matrix determinant we can compute
- That gives us the following:
This a quadratic polynomial in
To figure out the eigenvectors with these eigenvalues, substitute the eigenvalue back into the
For eigenvalue 2, we get
- So an eigenvector for
is and any vector in the span of that line, , is a solution for the equation.
Similarly, the eigenvector for
- So an eigenvector for
is and any vector in the span of that line, , is a solution for the equation.
Note not every matrix has eigenvectors.
A rotation of
- The transformation matrix is
and the roots of are the roots of so the roots are the complex numbers . - Having only complex roots means there are no eigenvectors in
.
For a shear transformation in
A.15.4 Eigenbases
Assume we have a transformation matrix in
Putting this into a transformation matrix, we have
Any time we have a diagonal matrix, where all off diagonal values = 0, we can interpret this as the matrix of an eigenbasis where each column is an eigenvector with an eigenvalue of the diagonal element for that column.
Diagonal matrices are often easier to work with than non-diagonal matrices.
- As an example, raising a diagonal matrix to the power
simply requires raising the diagonal values to the power and multiplying the vector.
If you have a transformation that has sufficient eigenvectors to create an eigenbasis, you can change to the eigenbasis to make your computations and then change back as desired as seen in Section A.14.3.
Follow similar steps.
- Take the original transformation
- Put the eigenvectors into a change of basis matrix and put that on the right.
- Put the inverse of the eigenvector change of basis matrix on the left.
The resulting matrix is guaranteed to be diagonal which may make the computations easier.
A.15.5 Beyond Matrices
The concepts of eigenvectors and eigenvalues have been looked at from the perspective of matrices which represent linear transformations.
These matrices are functions that map an input vector to an output vector in
The concepts of eigenvectors and eigenvalues can be extended to operate in other functions that express linear transformation.