Recursively applying the laplace expansion
|
|
|---|---|
![]() |
![]() |
• For 2×2 matrices, if 𝑨 = 𝑎(( 𝑎)( 𝑎)), recall that the inverse of A is 𝑎()
𝑨3(= 𝑎((𝑎)) − 𝑎()𝑎)( 1 −𝑎)( 𝑎)) −𝑎() 𝑎((• Hence, 𝑨 is invertible if
and only if
𝑎((𝑎)) − 𝑎()𝑎)( ≠ 0
| • This quantity is the determinant of 𝑨 ∈ ℝ)×), i.e., det 𝑨 = 𝑎(( 𝑎)( 𝑎))= 𝑎((𝑎)) − 𝑎()𝑎)( 𝑎() |
|---|
which we have observed in the preceding example.
• For 𝑛 = 3 (known as Sarrus’ rule), |
|---|
• We call a square matrix 𝑻 an upper-triangular matrix if 𝑇𝑖𝑗 = 0 for 𝑖 > 𝑗, i.e., the matrix is zero below its diagonal.
• Analogously, we define a lower-triangular matrix as a matrix with zeros above its diagonal.
• How can we compute the determinant of an 𝑛×𝑛 (𝑛 > 3) matrix?
• We reduce this problem to computing the determinant of 𝑛 − 1 × 𝑛 − 1 matrices. By recursively applying the Laplace expansion, we can compute determinants of an 𝑛×𝑛 matrix by ultimately computing determinants of 2×2 matrices.
−1KLM𝑎KMdet 𝑨𝒌,𝒋
KH(
| • 2. Expansion along row 𝑗 | + | ![]() |
|
|---|---|---|---|
| det 𝑨 = J | |||
| • Let us compute the determinant of | 1
|
2 | ||
|---|---|---|---|---|
| 𝑨 = | 1 | |||
| 0 |
|
|||
|
||||
using Sarrus’ rule:
det 𝑨 = 1 T 1 T 1 + 3 T 0 T 3 + 0 T 2 T 2 − 0 T 1 T 3 − 1 T 0 T 2 − 3 T 2 T 1 = 1 − 6 = −5.
• Adding a multiple of a column/row to another one does not change
det 𝑨• Multiplication of a column/row with 𝜆 ∈ ℝ scales det 𝑨 by 𝜆. In
particular, det 𝜆𝑨 = 𝜆𝒏det 𝑨
• Swapping two rows/columns changes the sign of det 𝑨
• Because of the last three properties, we can use Gaussian elimination
to compute det 𝑨 by bringing 𝑨 into row-echelon form. We can stop
Gaussian elimination when we have 𝑨 in a triangular form where the
elements below the diagonal are all 0. Recall: the determinant of a
triangular matrix is the product of the diagonal elements.


• We can verify this result with the previous example.



| • Matrices characterize linear transformations. | B | B | 1 | A (1,1)T | 1 | ||||||
|---|---|---|---|---|---|---|---|---|---|---|---|
| B | 1 | 1 | |||||||||
| -1 | |||||||||||
| 0 | |||||||||||
| C | -1 |
|
|||||||||
| -1 | 1 | -1 | |||||||||
| C | -1 | ||||||||||
|
|||||||||||
|
|||||||||||
| D | |||||||||||
| 0 | |||||||||||







| C |
|
cos(−45°) | |||||
|---|---|---|---|---|---|---|---|
|
|
sin(−45°) | |||||
| andcos(−45°) sin(−45°) | |||||||
|
|||||||
• Some linear transformations (matrices) are not invertible
• The characteristic polynomial 𝑝𝑨 𝜆 ≔ det 𝑨 − 𝜆𝑰 will allow us to compute eigenvalues and eigenvectors.
• Example
| • 𝑨 = 1 2 | ![]() |
|||
|---|---|---|---|---|
| 𝑝𝑨 𝜆 = det 𝑨 − 𝜆𝑰 = 1 − 𝜆 2 | ||||
• Let 𝑨 ∈ ℝ+×+be a square matrix. Then 𝜆 ∈ ℝ is an eigenvalue of 𝑨 and 𝒙 ∈ ℝ+\ 𝟎 is the corresponding eigenvector of 𝑨 if
𝑨𝒙 = 𝜆𝒙
• rk 𝑨 − 𝜆𝑰+ < 𝑛
• det 𝑨 − 𝜆𝑰 = 0
| • 𝑨 = 1 2 |
|
|
![]() |
|
|---|---|---|---|---|
| 𝑝𝑨 𝜆 = det 𝑨 − 𝜆𝑰 = 1 − 𝜆 2 | ||||
• If 𝜆 is an eigenvalue of 𝑨 ∈ ℝ+×+, then the corresponding eigenspace 𝐸y is the solution space of the homogeneous system of linear equations ( 𝑨 −𝜆𝑰 𝒙 = 𝟎
• Example (The Case of the Identity Matrix)
• Useful properties regarding eigenvalues and eigenvectors
• A matrix 𝑨 and its transpose 𝑨Xpossess the same eigenvalues, but not
necessarily the same eigenvectors
• The eigenspace 𝐸y is the null space of 𝑨 − 𝜆𝑰 since
𝑨𝒙 = 𝜆𝒙 ⟺ 𝑨𝒙 − 𝜆𝒙 = 𝟎
⟺ 𝑨 − 𝜆𝑰 𝒙 = 𝟎 ⟺ 𝒙 ∈ ker 𝑨 − 𝜆𝑰
• Symmetric, positive definite matrices always have positive, real
eigenvalues.
∀𝒙 ∈ 𝑉\ 𝟎 : 𝒙X𝑨𝒙 > 0
4
|
3 − 𝜆 − 2 T 1 |
|---|
giving the roots 𝜆( = 2 and 𝜆) = 5.
• Step 3: Eigenvectors and Eigenspaces. From our definition of the eigenvector
1 3 − 𝜆 𝒙 = 𝟎
• For 𝜆 = 5 we obtain
𝑥𝟏 𝑥𝟐= −1 −2 2

• This eigenspace is one-dimensional as it possesses a single basis vector.
• Analogously, we find the eigenvector for 𝜆 = 2 by solving
• The corresponding eigenspace is given as
𝐸) = span[ 1−1 ]
• In our previous example, the geometric multiplicity of 𝜆 = 5 and 𝜆 = 2 is 1.
• In another example, the matrix 𝑨 = 2 0 1 2has two repeated eigenvalues 𝜆( = 𝜆) = 2. The algebraic multiplicity of 𝜆( and 𝜆) is 2.



• Definition. A square matrix 𝑨 ∈ ℝ+×+is defective if it possesses fewer than 𝑛 linearly independent eigenvectors
• positive semidefinite: 𝒙Š𝑺𝒙 = 𝒙Š𝑨Š𝑨𝒙 = 𝑨𝒙Š𝑨𝒙 ≥ 𝟎
• If rk(𝑨) = 𝑛, then 𝑺 ≔ 𝑨Š𝑨 is positive definite.
so that we obtain the eigenvalues 𝜆( = 1 and 𝜆) = 7, where 𝜆( is a
repeated eigenvalue. Following our standard procedure for computing
eigenvectors, we obtain the eigenspaces
−1 −1 1
𝐸( = span[
• 1 0
,
• 0 1
] , 𝐸Ž = span[



• To construct such a basis, we exploit the fact that 𝒙𝟏, 𝒙𝟐 are eigenvectors
using such linear combinations.
• Therefore, even if 𝒙𝟏and 𝒙𝟐 are not orthogonal, we can apply the Gram-
• which are orthogonal to each other, orthogonal to 𝒙𝟑, and eigenvectors of 𝑨
associated with 𝜆( = 1.





−1KLM𝑎KMdet 𝑨𝒌,𝒋
3
sin 45°

3 − 𝜆 − 2 T 1
1 3 − 𝜆 𝒙 =
𝟎