Section 11 Important Definitions
11.1 Systems of Equations
- Row operations
The elementary row operations are 1) swap two rows 2) scale a row by a nonzero scalar 3) replace a row by the sum of that row plus a scalar multiple of another row
- Linear combination
A linear combination of a set of vectors \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) is a sum of the form \[ x_1 \mathsf{v}_1 + x_2 \mathsf{v}_2 + \cdots + x_n \mathsf{v}_n \] where the weights \(x_1, x_2, \ldots, x_n\) are real numbers.
- Span
The span of a set of vectors \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) is the set of all possible linear combinations of those vectors, so \[ span(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n) = \{ x_1 \mathsf{v}_1 + x_2 \mathsf{v}_2 + \cdots + x_n \mathsf{v}_n \mid x_1, x_2, \ldots, x_n \in \mathbb{R}\}, \]
- linear independence
- A set of vectors \(\mathsf{v}_1, \mathsf{v}_2,\ldots, \mathsf{v}_n\) are
linearly independent if the only way to write
\[ \mathsf{0} = c_1 \mathsf{v}_1 + c_2 \mathsf{v}_2 + \cdots + c_n \mathsf{v}_n \] is with \(c_1 = c_2 = \cdots = c_n = 0\).
Connection to Matrices: If \(A = [\mathsf{v}_1 \mathsf{v}_2 \cdots \mathsf{v}_n]\) is the matrix with these vectors in the columns, then this is the same as saying that \(A x = \mathsf{0}\) has only the trivial solution. This is true if and only if \(A\) has a pivot in every column so that there are no free variables.
- linear dependence
- Conversely, a set of vectors \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) are linearly dependent if there exist scalars \(c_1, c_2,\ldots, c_n \in \mathbb{R}\) that are not all equal to 0 such that \[ \mathsf{0} = c_1 \mathsf{v}_1 + c_2 \mathsf{v}_2 + \cdots + c_n \mathsf{v}_n \] This is called a dependence relation among the vectors.
Connection to Matrices: If \(A = [\mathsf{v}_1 \mathsf{v}_2 \cdots \mathsf{v}_n]\) is the matrix with these vectors in the columns, then this is the same as saying that \(\mathsf{x} = [c_1, c_2, \ldots, c_n]^{\top}\) is a nontrivial solution to \(A \mathsf{x} = \mathsf{0}\).
11.2 Linear Transformations
- linear transformation
A function \(T: \mathbb{R}^n \to \mathbb{R}^m\) is a linear transformation if the following three properties hold:
- \(T({\bf 0}) = {\bf 0}\).
- \(T(\mathsf{u} + \mathsf{v}) = T(\mathsf{u}) + T(\mathsf{v})\) for all vectors \(\mathsf{u},\mathsf{v} \in \mathbb{R}^n\).
- \(T(c \mathsf{u}) = c T(\mathsf{u})\) for all vectors \(\mathsf{v} \in \mathbb{R}^n\) and all scalars \(c \in \mathbb{R}\).
These properties say that \(T\) sends 0 to 0 and is preserves addition and scalar multiplication.
- onto
- A linear transformation \(T: \mathbb{R}^n \to \mathbb{R}^m\) with matrix \(A\) is onto if
The function is onto if \(A\) has a pivot in every row.
- onee-to-one
- A linear transformation \(T: \mathbb{R}^n \to \mathbb{R}^m\) with matrix \(A\) is one-to-one if
The function is one-to-one if \(A\) has a pivot in every column.
11.3 Vector Spaces
- span
A set of vectors \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) span a vector space \(V\) if for every \(\mathsf{v} \in V\) there exist a set of scalars (weights) \(c_1, c_2, \ldots, c_n \in \mathbb{R}\) such that \[ \mathsf{v} = c_1 \mathsf{v}_1 + c_2 \mathsf{v}_2 + \cdots + c_n \mathsf{v}_n. \] Connection to Matrices: If \(A = [\mathsf{v}_1 \mathsf{v}_2 \cdots \mathsf{v}_n]\) is the matrix with these vectors in the columns, then this is the same as saying that \(\mathsf{x} = [c_1, \ldots, c_n]^{\top}\) is a solution to \(A x = \mathsf{v}\).
- linear independence
A set of vectors \(\mathsf{v}_1, \mathsf{v}_2,\ldots, \mathsf{v}_n\) are linearly independent if the only way to write \[ \mathsf{0} = c_1 \mathsf{v}_1 + c_2 \mathsf{v}_2 + \cdots + c_n \mathsf{v}_n \] is with \(c_1 = c_2 = \cdots = c_n = 0\).
Connection to Matrices: If \(A = [\mathsf{v}_1 \mathsf{v}_2 \cdots \mathsf{v}_n]\) is the matrix with these vectors in the columns, then this is the same as saying that \(A x = \mathsf{0}\) has only the trivial solution.- linear dependence
Conversely, a set of vectors \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) are linearly dependent if there exist scalars \(c_1, c_2,\ldots, c_n \in \mathbb{R}\) that are not all equal to 0 such that \[ \mathsf{0} = c_1 \mathsf{v}_1 + c_2 \mathsf{v}_2 + \cdots + c_n \mathsf{v}_n \] This is called a dependence relation among the vectors.
Connection to Matrices: If \(A = [\mathsf{v}_1 \mathsf{v}_2 \cdots \mathsf{v}_n]\) is the matrix with these vectors in the columns, then this is the same as saying that \(\mathsf{x} = [c_1, c_2, \ldots, c_n]^{\top}\) is a nontrivial solution to \(A \mathsf{x} = \mathsf{0}\).- linear transformation
A function \(T: \mathbb{R}^n \to \mathbb{R}^m\) is a linear transformation when:
- \(T(\mathsf{u} + \mathsf{v}) = T(\mathsf{u}) + T(\mathsf{v})\) for all \(\mathsf{u}, \mathsf{v} \in \mathbb{R}^n\) (preserves addition)
- \(T(c \mathsf{u} ) = c T(\mathsf{u})\) for all \(\mathsf{u} \in \mathbb{R}^n\) and \(c \in \mathbb{R}\) (preserves scalar multiplication). It follows from these that also \(T(\mathsf{0}) = \mathsf{0}\).
- one-to-one
A function \(T: \mathbb{R}^n \to \mathbb{R}^m\) is a one-to-one when:
- onto
A function \(T: \mathbb{R}^n \to \mathbb{R}^m\) is a onto when:
- subspace
A subset \(S \subseteq \mathbb{R}^n\) is a subspace when:
- \(\mathsf{u} + \mathsf{v} \in S\) for all \(\mathsf{u}, \mathsf{v} \in S\) (closed under addition)
- \(c \mathsf{u} \in S\) for all \(\mathsf{u}\in S\) and \(c \in \mathbb{R}\) (closed under scalar multiplication) It follows from these that also \(\mathsf{0} \in S\).
- basis
A basis of a vector space (or subspace) \(V\) is a set of vectors \(\mathcal{B} = \{\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\}\) in \(V\) such that
- \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) span \(V\)
- \(\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\) are linearly independent Equivalently, one can say that \(\mathcal{B} = \{\mathsf{v}_1, \mathsf{v}_2, \ldots, \mathsf{v}_n\}\) is a basis of \(V\) if for every vector \(\mathsf{v} \in V\) there is a unique set of scalars \(c_1, \ldots, c_n\) such that \[ \mathsf{v} = c_1 \mathsf{v}_1 + c_2 \mathsf{v}_2 + \cdots + c_n \mathsf{v}_n. \] (The fact that there is a set of vectors comes from the span; the fact that they are unique comes from linear independence).
- dimension
The dimension of a subspace \(W\) is the number of vectors in any basis of \(W\). This is also the fewest number of vectors required to span the subspace.
11.4 Matrices
- invertible
The square \(n \times n\) matrix \(A\) is invertible when there exists an \(n \times n\) matrix \(A^{-1}\) such that \(A A^{-1} = I = A^{-1} A\). The Invertible Matrix Theorem collects over two dozen equivalent conditions, each of which guarantees that \(A\) is invertible.
- null space
The null space \(\mbox{Nul}(A) \subset \mathbb{R}^n\) of the \(m \times n\) matrix \(A\) is the set of solutions to the homogeneous equation \(A \mathsf{x} = \mathbf{0}\)> We also write this as \[ \mbox{Nul}(A) = \{ \mathsf{x} \in \mathbb{R}^n : A \mathsf{x} = \mathbf{0} \} \] Connection to Linear Transformations: If \(T(\mathsf{x}) = A \mathsf{x}\), then the kernel of \(T\) is the null space of matrix \(A\).
- column space
The column space \(\mbox{Col}(A) \subset \mathbb{R}^m\) of the \(m \times n\) matrix \(A\) is the set of all linear combinations of the columns of \(A\). For \(A = \begin{bmatrix} \mathsf{a}_1 & \mathsf{a}_2 & \cdots & \mathsf{a}_n \end{bmatrix}\), we have \[ \mbox{Col}(A) = \mbox{span} ( \mathsf{a}_1, \mathsf{a}_2, \ldots , \mathsf{a}_n ) \] We also write this as \[ \mbox{Col}(A) = \{ \mathsf{b} \in \mathbb{R}^m : \mathsf{b} = A \mathsf{x} \mbox{ for some } \mathsf{x} \in \mathbb{R}^n \}. \] Connection to Linear Transformations: If \(T(\mathsf{x}) = A \mathsf{x}\), then the range (also called the image) of \(T\) is the column space of matrix \(A\).
- rank
The rank of the \(m \times n\) matrix \(A\) is the dimension of the column space of \(A\). This is also the number of pivot columns of the matrix.
- eigenvalue and eigenvector
For a square \(n \times n\) matrix \(A\), the scalar \(\lambda \in \mathbb{R}\) is an eigenvalue for \(A\) when there exists a nonzero vector \(\mathsf{x} \in \mathbb{R}^n\) such that \(A \mathsf{x} = \lambda \mathsf{x}\). The nonzero vector \(\mathsf{x}\) is the eigenvector for eigenvalue \(\lambda\). The collection of all of these eigenvalues and eigenvectors is called the eigensystem of A.
- diagonalization
A square \(n \times n\) matrix is diagonalizable when \(A = P D P^{-1}\) where \(D\) is a diagonal matrix and \(P\) is an invertible matrix. In this case, the eigenvalues of \(A\) are the diagonal entries of \(D\) and their corresponding eigenvectors are the columns of \(P\).
- dominant eigenvalue
The eigenvalue \(\lambda\) of the square matrix \(A\) is the dominant eigenvalue when \(| \lambda | > | \mu |\) where \(\mu\) is any other eigenvalue of \(A\). The dominant eigenvalue determines the long-term behavior of \(A^t\) as \(t \rightarrow \infty\).