MATH141
1.1 Systems of Linear Equations:
- A system of linear equations can be represented by an augmented matrix.
- An elementary row operation is one of the following:
- Interchanging two rows.
- Multiplying a row by a nonzero constant.
- Adding a multiple of one row to another row.
- A system of linear equations is consistent if it has at least one solution.
- A system of linear equations is inconsistent if it has no solution.
1.2 Row Reduction and Echelon Forms:
- A matrix is in row echelon form if:
- All rows that contain only zeros are at the bottom of the matrix.
- The first nonzero entry in each nonzero row is called a pivot.
- The pivot in each row is to the right of the pivot in the row above it.
- A matrix is in reduced row echelon form if:
- It is in row echelon form.
- Each pivot is 1.
- Each pivot is the only nonzero entry in its column.
- A system of linear equations is equivalent to an augmented matrix in reduced row echelon form.
- A pivot column in an augmented matrix corresponds to a leading variable in the system of linear equations.
1.3 Vector Equations:
- A vector in n-dimensional space is an ordered list of n real numbers.
- The zero vector is the vector consisting of all zeros.
- Scalar multiplication of a vector by a scalar k multiplies each component of the vector by k.
- The sum of two vectors is a vector obtained by adding the corresponding components of the vectors.
1.4 The Matrix Equation Ax = b:
- A matrix is a rectangular array of numbers.
- The product of a matrix A and a vector x is a linear combination of the columns of A.
- The system of linear equations Ax = b is consistent if and only if b is a linear combination of the columns of A.
- The inverse of a square matrix A is a matrix A^-1 such that AA^-1 = A^-1A = I, where I is the identity matrix.
2.1 Matrix Operations:
- The sum of two matrices of the same size is a matrix obtained by adding the corresponding entries of the matrices.
- Scalar multiplication of a matrix by a scalar k multiplies each entry of the matrix by k.
- The product of two matrices A and B is a matrix obtained by multiplying the rows of A by the columns of B.
- The associative property holds for matrix multiplication: (AB)C = A(BC).
- The distributive property holds for matrix multiplication: (A+B)C = AC+BC.
2.2 Matrix Algebra:
- The transpose of a matrix A is a matrix AT obtained by interchanging the rows and columns of A.
- A symmetric matrix is a matrix A such that AT = A.
- A skew-symmetric matrix is a matrix A such that AT = -A.
- The determinant of a 2×2 matrix A is det(A) = ad – bc.
- The determinant of an n x n matrix A can be computed by cofactor expansion along any row or column.
- A matrix A is invertible if and only if det(A) ≠ 0.
2.3 Elementary Matrices and a Method for Finding A^-1:
- An elementary matrix is a square matrix obtained from the identity matrix by performing a single elementary row operation.
- If E is an elementary matrix, then EA performs the same row operation on A as E performs on I.
- If A can be transformed into I by a sequence of elementary row operations, then the product of the corresponding elementary matrices is A^-1.
3.2 Diagonalization:
- A matrix A is diagonalizable if there exists an invertible matrix P and a diagonal matrix D such that A = PDP^-1.
- If A is diagonalizable, then D has the eigenvalues of A on its diagonal and P has the corresponding eigenvectors as its columns.
- A matrix A is diagonalizable if and only if it has n linearly independent eigenvectors.
- A matrix A is diagonalizable if and only if there exists a basis of R^n consisting of eigenvectors of A.
4.1 Vector Spaces and Subspaces:
- A vector space is a set of vectors closed under addition and scalar multiplication.
- A subspace of a vector space is a subset that is itself a vector space under the same operations.
- The span of a set of vectors is the set of all linear combinations of the vectors.
- A set of vectors is linearly independent if no vector in the set can be expressed as a linear combination of the others.
4.2 Null Spaces, Column Spaces, and Linear Transformations:
- The null space of a matrix A is the set of all solutions x to the homogeneous equation Ax = 0.
- The column space of a matrix A is the span of its columns.
- The rank of a matrix A is the dimension of its column space.
- A linear transformation T from R^n to R^m is a function that preserves vector addition and scalar multiplication.
- The matrix of a linear transformation T with respect to the standard bases of R^n and R^m is the matrix whose columns are the images of the basis vectors.
5.1 Inner Product, Length, and Orthogonality:
- An inner product on a vector space V is a function that takes two vectors in V and returns a scalar.
- The dot product of two vectors in R^n is an example of an inner product.
- The length of a vector x is ||x|| = sqrt(x1^2 + x2^2 + … + xn^2).
- Two vectors are orthogonal if their dot product is zero.
- An orthogonal set of vectors is a set of vectors in which every pair is orthogonal.
- An orthonormal set of vectors is an orthogonal set of unit vectors.
5.2 Orthogonal Sets:
- Gram-Schmidt Process: Given a basis {v1, v2, …, vn} for a subspace of R^n, we can construct an orthonormal basis {u1, u2, …, un} using the following steps:
- Set u1 = v1/||v1||.
- For i = 2 to n, let ui = vi – proj_{u1}(vi) – proj_{u2}(vi) – … – proj_{ui-1}(vi), where proj_{u}(v) = ((v.u)/(u.u))u.
- Normalize each ui to obtain an orthonormal set.
6.1 The Gram-Schmidt Process and Orthogonal Bases:
- An orthogonal basis for a subspace is a basis that is also an orthogonal set.
- Any linearly independent set of vectors can be converted into an orthogonal basis by applying the Gram-Schmidt process.
- The orthogonal complement of a subspace V of R^n is the set V⊥ = {x ∈ R^n : x.v = 0 for all v ∈ V}.
- The orthogonal complement of a subspace is itself a subspace.
6.2 Projections:
- The projection of a vector x onto a nonzero vector u is proj_{u}(x) = ((x.u)/(u.u))u.
- The projection of a vector x onto a subspace V is proj_{V}(x) = ∑(x.v/u.v^2)u, where the sum is taken over an orthonormal basis for V.
- The projection matrix P onto a subspace V with orthonormal basis {u1, u2, …, um} is P = [u1|u2|…|um][u1|u2|…|um]T.
7.1 Orthogonal Diagonalization:
- A matrix A is orthogonally diagonalizable if there exists an orthogonal matrix Q and a diagonal matrix D such that A = QDQ^-1.
- If A is orthogonally diagonalizable, then the columns of Q are an orthonormal basis of R^n consisting of eigenvectors of A.
- If A is a real symmetric matrix, then A is orthogonally diagonalizable.
- If A is an n x n matrix with n distinct eigenvalues, then A is orthogonally diagonalizable.