---
title: Linear Algebra II
tags:
- 2020
- LA
- NYCU
- note
author: Maxwill
lastUpdated: 2020/06/10
description : Linear Algebra II
copyright : "screen shots are mainly from Ming-Hsuan Kang's lecture notes, paste for note clarity and save time only"
---
# Linear Algebra II
* [NCTU Lecture Notes](https://hackmd.io/r-CG8R7xTZ2S3pRa6JD9FA)
* [NCTU 2020 spring](https://hackmd.io/xJD55HhBS-CiMdAiSCNYEg)
## week 1-1 (3/2/2020)
* review
* linear independency
* basis and dimension
* dimension theorem
* direct sum
* matrix representation :
$$
Rep_{\alpha,\beta}(T) = (Rep_\beta(T(\vec{\alpha_1}) ...Rep_\beta(T(\vec{\alpha_i}))
$$
* change of basis
$$
Rep_\alpha(T) = Rep_{\beta,\alpha}(id)Rep_\beta(T)Rep_{\alpha,\beta}(id)
$$
* determinant
* Invariant Subspace
* a subspace $W \subset V\ s.t. \forall w \in W, T(w)\in W$
* direct sum of Invariant Subspace
* Eigenspace
* def: $E(\lambda) = \{v \in V \mid T(v)=\lambda v \}$
* $ker(A-\lambda I)$, $E(\lambda)\ of\ A$
* attendence 10%
* next quiz hint:
* matrix representation + diagonization
* lecturing
* $T:V\rightarrow V$ be linear transformation
* For $\vec{v} \in V$ what is the smallest T-invariant subspace?
* Theorem :
::: info
let $W = span\{v, T(v), T^2(v), T^3(v)...\}$
then
1. $W$ is a T-invariant subspace
2. let $dim(W) = m$, then $$\alpha = \{ \vec{v} , T(\vec{v}), T^2(\vec{v}), ...T^{m-1}(\vec{v})\}$$ is a basis of $W$
4. $T^m(\vec{v}) = \sum a_iT^i(\vec{v})$
5. $Rep_{\alpha}(T)$ = \begin{bmatrix}
0 & 0 & ...& a_0 \\
1 & 0 & ...& a_1 \\
0 & 1 & ...& ...\\
0 & 0 & ...& a_{m-1}
\end{bmatrix}
6. determinant and characristic function by MI
7. $f_{T\mid _W}(x) = ?$
:::
::: info
Proof:
1. By definition : $W = span\{w, T(w), T^2(w), T^3(w)...\}$
2. $\exists \ k\ \ni T^k \in span\{w, T(w), T^2(w), T^3(w)...T^{k-1}(w)\}$ for V is finite-dimentional
3. $T^k(w)$ is linear combination of $\alpha$
4. by induction , $T^{n}$ with $n > k$ is too
5. then $T(\vec{w}) \in W$ is trivial
:::
* asked teaher for concept confirmation
* max linear independent set = generating set in this case. But teacher emphisize the concept is different
* uses $F$ instead of $R$ for generasity
## week 1-2(3/4/2020)
* matrix transformation self review
* http://www.taiwan921.lib.ntu.edu.tw/mypdf/math02.pdf


* isomorphic linear transformation <=> invertible linear transformation
* 
* standard representation
* 
* 
* change of coordinate
* 
* find eigenvalues and vectors
* $A = P^{-1}QP$
* course
* [use cyclic subspace to proof Cayley-Hamilton Theorem](https://ccjou.wordpress.com/2011/01/31/%E5%88%A9%E7%94%A8%E5%BE%AA%E7%92%B0%E5%AD%90%E7%A9%BA%E9%96%93%E8%AD%89%E6%98%8E-cayley-hamilton-%E5%AE%9A%E7%90%86/)
* invariant subspace is useful for analyzing
## week 2-1(March/9th/2020)
* 3/10/2020 ask teacher
* teacher explained in detailed and intuively
* What's the problem?
* Why invariant subspace?
* to create more 0
* Why Annihilator?
* to get blocks of invariant subspaces
* Usage of Cayley-Hamilton Theorem?
* find such a f(x) quickly
* The next will be Jordan form
* what to fill in blocks?
* btw, diagonizable and invertible is independent
* [link](https://yutsumura.com/true-or-false-every-diagonalizable-matrix-is-invertible/)
### !T-invariant subspaces
* to find a basis $\alpha$ $\ni$
* and extend the basis from span W to span V
* then the matrix representation of T : V->V
$$
\begin{pmatrix}
T|_W & A \\
O & B
\end{pmatrix}
$$
* if we get direct sum of T-invariant subspaces
$$
\begin{pmatrix}
T|_{W_1} & O \\
O & T|_{W_2}
\end{pmatrix}
$$
### Annihilator - Ann(T)
* $L(V,V) \rightarrow \ set\ of \ polynomials$
* $for\ T \in L(V,V),\ Ann(T) = \{f(x)\in F[x]\mid f(T)\ is \ 0\ transformation\}$
* a way to get decomposition of $V$ to direct sum of T-invariant subspaces
### !Cayley-Hamilton Theorem
* $f_T(T) \in Ann(T)$
* in this case, the usage is to find a $f(x) \ni f(T)\equiv 0$
* proof:
$\forall v \in V$, let T invariant subspace generated by $v$ be $W$
$let\ V=W+W^{'}$
since $[T]_{W+W^{'}}$ can be express as
$$
\begin{pmatrix}
T|_{W} & A \\
O & B
\end{pmatrix}
$$
$f_{T|_{W+W^{'}}}(T) = g(x)f_{T|_W}(x)$ Note that g(x) is the $det(\lambda I-B)$ part
$f_{T|_W}(v) = 0$ from collagory that followed directly with definition of $T|_W$
$Q.E.D.$
### !(General Eigenspace)Decomposition of V
* goal : $V=ker_\infty(T) \bigoplus Im_\infty(T)$
* goal2 : $V=\bigoplus E_\infty(\lambda)$
* part 1


Note : hint for T-invariance - exchangable ops
* part 2

Note:
in part 2
if m is inf of all ms satisfy prerequisition
then $V = ker(T^{m-1}) \bigoplus Im(T^{m-1})$ does not hold in general
confirmed by teacher
* part 3

* part 4

Note:
* Theorem 4
* case minimal m = 0 => invertible(rank(n)) => trivial
* case minimal m >= 1 => is true
* above notes are from Ming-Hsuan Kang
* [proof of diagonizability](https://math.okstate.edu/people/binegar/4063-5023/4063-5023-l18.pdf)
### Next Jordan Form
* what's in the block exactly
### What Linear Algebra studies
* real world problem, natural functions
* how to express in good basis(change of basis)
* fourier, Laplace...
* meaningful, easy to compute
* linear transformations
* differential
* integral
* etc.
* its quite different from Abstract Algebra by teacher
* whereas I think the way to think is similar
* just LA emphasize more on linearity and dimension etc.
## Week 2-2
### quiz problem
* pA : calculation of T-invariant subspace
* pB : let T : R3->R3 be reflex transformation, show T is diagonizable
* hint $T^2 = I$ + HW
* I was the first one finished
### other material
* [Invariant Subspace]https://math.okstate.edu/people/binegar/4063-5023/4063-5023-l18.pdf
## Week 3-1
### Nilpotent LT
- definition
- T is a nilpotent LT
- $V = ker_{\infty}(T)$
- $T^k = \vec{0} for\ some\ k$
- use similar technique as T-invariant subspace
- find a matrix rep of T
$$
\begin{pmatrix}
0 & 0 & 0 & ...&0 \\
1 & 0 & 0 & ...&0 \\
0 & 1 & 0 & ...&0 \\
0 & 0 & 1 & ...&0 \\
...&...&...&...&0
\end{pmatrix}
$$
### Theorem
## Week 3-2 skipped, 3-3(self study@Saturday)
### !Th. - V can be decomposed with Nilpotent LT(on V)

#### Proof sketch of part 3(2020/3/23, correct by teacher)
:::info
let T be nilpotent LT on V of index k+1
choose $v_i$ s.t. $T^{k}(v_i)$ forms a basis of $Im(T^{k})$
Objective :
$V = W\bigoplus cyclic(v_i)$ for some T-invariant subspace $W$
Proof:
- key : must have **dot diagram** in mind
- Induction on k+1, index of T
- $W$ as the left part removing higher dimension cyclic subspaces
- extension of basis is like finding the current longest given the past longest
- $v_i$ past longest
- $u_i$ current longest
1. when $k = 0, T = 0, T^0 = I$, holds trivially
2. when k holds
- $V = ker(T^k) \bigoplus Fv_i$
- devide V to apply $T^k$ is 0 or non-zero
- $V^{'} = ker(T^k)$, $T^{'} = T|_{V^{'}}$
- $T^{'}$ is nilpotent on $V^{'}$ of index k
- Now we find a basis of $Im({T^{'}}^{k-1})$, a subspace of $ker(T^k)$
- part original:
- from $v_i$ to $T(v_i)$
- ${T^{'}}^{k-1} (T(v_i))$ = $T^k(v_i)$
- l.i. subset of $Im({T^{'}}^{k-1}))$
- part extended:
- $u_i$
- $ker(V) = W_0 \bigoplus cyclic(T(v_i)) \bigoplus cyclic(u_i)$
- by induction hypotesis
- $T(v_i) \bigoplus u_i$ forms basis of $Im({T^{'}}^{k-1})$
- $V = ker(T^k) \bigoplus Fv_i$
- $V = W_0 \bigoplus cyclic(u_i) \bigoplus cyclic(T(v_i)) \bigoplus Fv_i$
- $V = W \bigoplus cyclic(v_i)$
- since $cyclic(u_i)$ is T-invariant
3. by induction, Q.E.D.
- by the proof step can actually see W is cyclic subspaces' direct sum
:::
Illustration Diagram
::: spoiler

:::
## HW3
## Week 4-1(2020/3/23)
### !Jordan Form
#### part 1
- for a LT T
- $V = \bigoplus ker_\infty(T-\lambda_i I)$
- $V = \bigoplus E_\infty(\lambda_i)$
#### part 2
- $T-\lambda I$ is nilpotent on $E_\infty(\lambda_i)$
- $E_\infty(\lambda_i) = \bigoplus cyclic(v_i)$
#### part 3
- these $cyclic(v_i)$s have a representation of
$$
\begin{pmatrix}
0 & 0 & 0 & ...&0 \\
1 & 0 & 0 & ...&0 \\
0 & 1 & 0 & ...&0 \\
0 & 0 & 1 & ...&0 \\
...&...&...&...&0
\end{pmatrix}
$$
#### Conclusion
- Jordan Form
#### Uniqueness?
- General Eigenspace Decomposition (YES)
- Cyclic Subsapce Decomposition (No)
- but the dimensions are unique
- generally, the matrix rep. can be said to be unique
#### Steps to find a Jordan form matrix rep.
1. find $f_A(x)$
2. find **dot diagram** for each $\lambda$
- by observe the nulity of $(T-\lambda I)^k$
ex:
0 <- $T(v_2)$ <- $v_2$
0 <- $v_1$
$$
\begin{pmatrix}
\lambda & 1 & 0 \\
0 & \lambda & 0 \\
0 & 0 & \lambda \\
\end{pmatrix}
$$
#### Steps to find a Jordan basis
1. just solve it from basis of $E\infty(\lambda)$
#### Case Study
1.
$$
\begin{pmatrix}
5 & 7 & 1 & 1 & 5 \\
-2 & -3 & -1 & -1 & -5 \\
-1 & -3 & 1 & 1 & -3 \\
0 & 0 & 3 & 0 & 1 \\
3 & 3 & 1 & 0 & 6
\end{pmatrix}
$$
2.
let V be a subspace of real funcitons spanned by $\alpha = \{x^{-2}e^x, xe^x, e^x\}$
let D be the differential operator
find Jordan form of D and its jordan basis
### Jordan Chevalley Decomposition
$A = P^{-1}(D+N)P = P^{-1}DP + P^{-1}NP$
where
$P^{-1}DP$ is semi-simple part
$P^{-1}DP$ is nilpotent part
- $\forall A \in M_n(\mathbb{C})$ there exist unique $D,N \in M_n(\mathbb{C})$
- A = D + N
- DN = ND
- D is diagonizanle
- N is nilpotent
### Advantage of Jordan Form
#### power of matrix
- by $(D+N)^k$ with the fact that N is nilpotent
## Realse of HW1 Quiz1 Quiz2
- HW1 30/30
- Quiz1 16/20
- Quiz2 20/20
::: spoiler



:::
## Week 4-2
* Hw and test
* approximate Jordan form with diagonizable matrix
---
## Week 5 - next topic - Inner Product Space
---
## Week 5-1 - Inner Prodct Space
- Definition over $\mathbb{R}^N$ and $\mathbb{C}^N$
- Definition of orthogonal
- General definition when $\mathbb{F}$ is $\mathbb{R}$ or $\mathbb{C}$
- Inner Product of
- Continuous Functions
- Discrete Signals
- Discrete Fourier Transformation(DFT)
- meaningful basis
- easy-to-compute basis
- Theorem:
- Every inner product is induced from some basis
---
## Week 5-2 - Normal LT
### Symmetric and Hermitian
### Adjoint
#### definition

- defined on vector space with inner product
- without basis => more inside
#### self-adjoint
1. All eigenvalues of T are real.
2. T admits a set of eigenvectors which forms an orthonormal basis of V. (Especially, T is diagonalizable.)
3. Under an orthonormal basis, the conjugate transpose of the matrix representation of T is equal to the matrix representation of T∗
### Normal and Ker/Im

- proof of 5

### Normal <=> Diagonalizable under orthogonal basis
- proof
- <=
- trivial
- =>
- key : general eigenspaces = eigenspaces
---
## HW5


---
## Week 6 - fill in week 5
- diagonization of symmetric matrix
- Grandsmith and projection
- norm of $C^2$
- 2020/04/07 making up HW5
:::spoiler



:::
**Q: relationship of Rn and Cn
Q: induced innerproduct by basis**
- finite filed > 0 is not well defined
- positive definite
- compare relation
- inner prouct( >= 0)
- only have = 0
[induced inner product](https://math.stackexchange.com/questions/1233384/how-to-choose-an-inner-product-with-respect-to-a-basis-in-such-a-way-that-this-b)
---
## Week 6-1 - Othogonal/Unitary
### Definition
- preserve inner product

### Equivalences

### Theorems revisitted

### Isometric
- $isometric$ + $f(\vec{0}) = \vec{0}$ $\iff$ $orthogonal$
- must be LT

- proof: theorem

- proof: LT

#### general isometry
- affine map(LT + bias)
#### Isometric 2-D case study
- all orthogonal matrice in $\mathbb{R}^2$
- rotation/reflection by parametric method
### Det and Eigenvalues of Orthogomal Matrices
- $det(A) = +-1$ by $det(AA^T) = det(I) = det(A)^2$
- $T$ is ortho LT, $W$ is $T-invariant$ implies $W^\perp$ is too.
- proof is exercise
#### Orthogonal 3-D case
- $det(A) = +-1$
- exist $\lambda = +-1$ => rotate axis or reflecton axis
- by direct sum of T-invariant subspaces(exercise theorem)
- the left part is orthogonal matrice of rank 2 with det = 1, which must be a rotation
### Questions
**Q: relationship of Rn and Cn
Q: induced innerproduct by basis**
**Q: orthogonal - normal - symmetric(A*=A) relations**
---
## HW6
## TA hour by MC kang 2020/4/14, learned a lot
- HW6 3-2
- more analytic way, discuss det=+1,-1, only on 3D
### n-reflections theorem
[reference link](http://faculty.uml.edu/dklain/orthogonal.pdf)
### more on Orientation(advanced topic)
- isometric LT has 2 type!
- continuous isometric (rotate) vs no away (reflect)
- add dimension, what happen
- spin by dimension
- orientation preserving orthogonal LT
- det +-1
- maintain isometric during continuous changing process!
- orientation is 2 for all R^n by 2 reflection = 1 rotation
- may not cover in class :(
---
## Week 7 - self study - Review Concepts
### Inner product
#### Q: induced basis and induced inner product
### self-adjoint, Hermitian
#### conjugate transpose
#### self-adjoint Th
1. All eigenvalues of T are real.
2. T admits a set of eigenvectors which forms an orthonormal basis of V. (Especially, T is diagonalizable.)
#### self-adjoint, Hermitian
- Under an orthonormal basis, the conjugate transpose of the matrix representation of T is equal to the matrix representation of T∗
- A symmetric/Hermitian matrix is a matrix representation of a self-adjoint linear transform under an orthonormal basis.
---
### Normal LT
#### def
T∗ and T commute
#### Properties

#### Theorem
A complex linear transformation is diagonalizable under some
orthonormal basis if and only if it is normal.
---
### Orthogonal and unitary
#### def
T* = T-1
#### n-reflections
## Week 7 quiz
- prove cannot find a continuous family of iometry for reflection\
- first show b is inrelevant
- by cont. compose cont. (det。F~(t))
---
## Week 7 - Real Canonical Form
### analysis of real matrix A
- pairs of conjecate roots of $f_A(X)$
- pairs of eigenvalues
- conclusion
- for a eigenvalue $\lambda=a-bi$, and $\vec{v}$ be the corresponding eigenvector
- let $v = v_1+iv_2$
- $A(v_1+iv_2) = (av_1+bv_2) + i(-bv_1 + av_2)$
- notice that v1, v2 must be l.i.
- else the eigen value is real number
- complex block is
$$
\begin{pmatrix}
a & -b\\
b & a\\
\end{pmatrix}
$$
### analysis of othogonal matrix A
- $\lambda_i = e^{-i\theta_i}=cos\theta_i-isin\theta_i$
- complex block is
$$
\begin{pmatrix}
cos\theta & -sin\theta\\
sin\theta & cos\theta\\
\end{pmatrix}
$$
- thus we can see **orthogonal matrix** as **reflextions + 2D-rotations**
- notice that $\beta$(change of basis) can be chosen to be orthonormal

---
## next topic, quadatic form
---
## Week 7 - Quadratic Form
### Quafratic Form


### Talor Series Revisited

- k-th order approximation
- 1st + 2nd-order can be used to detemine local max/min
### Example

### Case: 2 variable, Binary Quadratic form

### Case: Ternary
calculation example
### General Case: N


### Key
- Quadratic form is Real Symmetric Matrix
- Diagonizable(R) with orthogonal basis
---
## Week 7 - Conic Sections
### Purpose of this chapter
- zero set
### Term
- G: $ax^2 + bxy + cy^2 + dx + ey + f$
- Q: $ax^2 + bxy + cy^2 + dxz + eyz + fz^2$
- H: $ax^2 + bxy + cy^2$
### Zero Sets of binary Quadratic Form
- key: the signs of two eigenvalues
- $sign(\lambda_1) = sign(\lambda_2)$ => {0, 0}
- $sign(\lambda_1) \neq sign(\lambda_2)$ => two lines
- one of them zero => one line

### deal with below quadratic terms

### zero set of non-degenerate ternary quadratic form

#### Conix sections

### Conclusion

- let H be $ax^2 + bxy + cy^2$
- with $Z(G) = Z(Q)\cap Z(z-1)$ ~= $Z(Q)\cap Z(z=0) = Z(H)$
- we can judge G by H if non-degenerate
---
## Week 7 - Equivalent Quadratic Forms, Signature
### Equivalent Quadratic Form
- def, two quadratic form is called Euivalent iff
- can obtain each other by change of basis
- but since we want signature => need not to be orthonormal
- Diagonal Form
- note the simbol usage
- $Q(\vec{x}^t)$
- more common to use row vector to repr. variable
### Review, change of variable to diagonal form

- note that $\vec{y}^tD\vec{y} = \sum_i\lambda_iy_i^2$
#### use orthogonal instead of orthonormal basis

#### Standard form : above is to change diagonal matrix to +-1

### Signature of Real Quadratic Forms

- Note: trace of standard form(i.e., signature) + rank can decide standard form
### Signature <=>(1-to-1) Eq Quadratic Forms
- pf: coming chapters
### Example is trivial
Q: Why signature
---
## HW 7
### Apllications of Quadratic Forms(binary and tenary)
### Taylor expansion and extreme values
### Zero set, discussion on degenerate form
---
## Week 8 - Office hour 2020/4/21

### Taylor Expansion for determining local min/max
- terms of poly of $dx_i$
- higher order terms are dominated by lower order ones
- multiply of infinity smalls
- we use diagonalize technique to make thing easy
- chage of variable so that
- only square terms is non-zero
### about Trace
- $tr(AB) = tr(BA)$ poof by direct calculation
- for more, ex: ABC
- treat AB as D or BC as D and reduce to 2 matrix case
- $tr(ABC) \neq tr(BAC)$ in general
- $tr(A) = PDP^{-1}) = tr(DP^{-1}P) = tr(D)$ in this case
---
## Weel 8 - Positive Definite Quadratic Form
### Equivalences and proof
- Q is positive definite()
- A is positive definite(eigenvalue)
- Q has has unique minimun at 0
#### Case of semi
### Theorem of n1 is
### Sylvester's critirien
#### Induction Proof
### extreme values at 0?
### Example
- Q, not positive definite, by can be semi-positive definite?
- should check higher order?(the conclusion of not local extreme is too fast IMO)
- A(myself): bad Q, this is not discussion for derivatives(Hessian)
---
## Week 8 - Bilinear Form
### Definition
### Matrix Representation
- Note taht tenary up has no matrix repr.
### Coordinate-Free Quadratic Form
- isomorphism between Quadratic form and symmetric bilinear form
### Relation with Quadratic form and inner product
### Non-degenerate
### Multilinear form, Tensorproduct, Theorem(Extra)
---
## Week 9 - 2020/4/27 practice exam+ review
### Midterm hint
- True & False
- Jordan Form
- calculations
- mini poly and possible jordan forms
- Quadratic form
- calculations
- tennary + binary
- local extreme
- zero set discussioni
- positive definite proof
- Symmetric, Normal, etc.
- proof of diagonalizability
- proof of eigenvalues all real(by inner product and real symmetric)
- when have diagonal form(normal, poly with no repitive roots are zero)
- Q: poly
### My Q
- Quantum Computing learning path
- proof of jordan form
- about nilpotent LT's cyclic decomposition
- about general eigenvalue direct sum
---
## Week 9 - individual office hour @ 2020/4/28 15:00-16:00
### What does Quntum Computing study
- Q: is lie Group/Algebra related?
- Yes, bried introduction
- A: to make "good" universal gates
- traditional bit and gate($(1,0)^N \to (1,0)^M$)
- now want universal approximation for wave functions
- Me: and apply them efficiently is the algo part
- Q: What should I study
- a variety of fields
- dont need to be deep, but know the essential concepts
- cause there are many isomorphic realations between fields
- think of the problem on ball and design way to higher dimension
### Main clairifications
- Innerprodect space and basis
- when is inner product well defined
- minimal poly and relation with joran form
- eigenspace decomposition(multiply 0 of blocks)
- restate some inportent fact
- intuition about multilinear form
### Summary Before Midterm
#### Jordan Form
- general eigenspace decomposition of LT
- T-invariant subspace
- cyclic subspace decompositioin of Nilpotent LT
- nilpotent LT
- Cayley-Hamilton Theorem Revisited
- Real Canonical Form
- Topic:
- Generalized kernel
- Eigen Space discussion
- minimal polinomial and possible jordan form
#### Inner Product Space
- Inner Product is defined on vector spaces that
- F is R or C(or non-finitem, to make >= 0 well defined)
- $v^tw^{bar}$
- 3 main concepts, their maxtrix representation, and theorems
- $A^*$, (inner product space)adjoint - (matrix) conjugate transpose
- Self-Adjoint - $A = A^*$
- Normal - $A, A^*$ commute
- Unitary - $A^*A = I$
- Isometric
- Self-adjoint
- Real symmetric or hermitian
- implies:
- real eigenvalues
- diagonalizable under unitary basis
- existence and uniqueness
- Normal
- A complex linear transformation is diagonalizable under some orthonormal basis if and only if it is normal
- proofs are important and interesting
#### Bilinear Form
- Quadratic from
- discussion of zero set
- conic cure
- discussion of extreme value
- Taylor, gradient, Hessian
- EQ quadratic form, signature
- 1-to-1 relation of symmetric bilinear form with Quadratic form
- Inner product as "symmetric" and "positive-definite" "bilinear form"
- Idendity if induced by basis
- multilinear form and tensor product
- example illustration
- positive definite bilinear form
- Sylvester’s critirien
### Trick
- proof v = 0 by <v,v>=0
- proof u-w = 0
- proof space V = W, $W \subset V$ and $dim~W = dim~V$
- proof of practice exam 5.a-b
- induction on dimension
- consider diagonal matrix first!
### Pictures
::: spoiler










:::
### self study + Q for night office hour
- hermitian > normal
* + eigenvalues all real?
- [link: EQ def normal](https://en.wikipedia.org/wiki/Normal_matrix#Equivalent_definitions)
- normal := A* and A commute
- normal <=> A* is poly of A
- normal <=> A and A∗ can be simultaneously diagonalized
- Actually, $A^* = P^tD^{bar}P$ by proof in pdf?
- commute <=> Simultaneous Diagonalizability
- [Thm. 5.1](https://kconrad.math.uconn.edu/blurbs/linmultialg/minpolyandappns.pdf)
- also Thm. 4.11 for equivalence of diagonalizability
### online TA hour
- diagonal represent as poly <=> poly n->n need n-1 degree
- Lagrange construction
- if A want to be represent as poly(B)
- if B position i, j is same, A pos i, j has to be same
- degree smaller
- commute and both diagonalizable <=> Simultaneous Diagonalizability
- + AB reltation stated can <=> A can be poly of B
- T and T* commute iif T* is poly of T is true
- consider after diagonalized(always can do, normal)
- T* is just $T^{bar}$
- then use Lagrange construction
- proof trick
- 5-a

- about projection
- pairwise product is zero transformation
- lecture note is wrong
- sum is idendity
- matrix congruence, quadratic form change of variable
- https://en.wikipedia.org/wiki/Matrix_congruence
- is undser "orthogonal basis"
### Remainning Q
- taylor expansion for extreme value in genreal
- my guess, 0, +-, 0, +- ...
---
## Week 13 - SVD
### Best fit subspace
#### motivation and eq defs
### left singular values
### best fit subspace and SVD
- proof by induction on k
### Note on details
- $rank(A^tA) = rank(AA^t) = rank(A)$
- proof by definition + norm def
- https://math.stackexchange.com/questions/349738/prove-operatornamerankata-operatornameranka-for-any-a-in-m-m-times-n
- $A^tA$ is symmetic
- $A^tA$ is semi-positive definite
- $A^tA$ has **orthonormal eigen decomposition**
### relation with right singular value
- $Av_i = \sqrt{\lambda_i}u_i$
- can proof this will be left eighebasis for A^t
### SVD - the decoomposition
- standard basis => alpha{v} => sinvular value diagonal matrix(m*n) => beta => standard basis
- $U\sum V^t$
- $Av_i = \sqrt{\lambda_i}u_i$
- $A = \sum _{i=1}^r \sqrt{\lambda_i}u_iv_i^t$
- obs: sum of rank one matrices
### compact SVD
- use only $m*r, r*r, r*n$
### confusion
- not famil
### SVD view of best fit k-subspace
### Compression with SVD
### PCA - best fit affine subspace
- max variance subspace
### SVD vs PCA
- care mean or not
---
## HW 9 @ Week 13
- last week take a leaf
- https://hackmd.io/Pcp80W3CT26eKqhUwIxzpw
---
## Week 13 - Spectral Drawing
- Vertex => n-dimensional e vectors
- Now want to project to subspace W
- Minimal "Edge Energy Function"
- define as the sum of square of distance of edges in subspace W
- If want minimal => take eigenspaces of Laplance matrix
- can show by definition trivially
- want each c.c. has the same projection is best => eigenspaces
- used on similarity graph => spectral clustering
- If want minimal + orthogonal to eigenspace of (connected) graph
- this is spectral drawing
## HW 10 - Spectral Drawing
- [LA HW 10 Spectral Drawing](/A0lGBNVlSQ-Ss7A5fluKsQ)
---
## Week 14 - Matrix Exponential
### Motivation
- differential equation solution
### Definition
- $exp(At) = lim_{n\to\infty} \sum_{m=1}^{n}\frac{A^mt^m}{m!}$
### Matrix Limit
- Convergence, Abs Convergence, Complex Convergence
- Def: Matrix Limit is Entry-wise
- Product of Matrix Limit
- by elementwise discussion
### Discussion of Jordan Block (exp(Jt))
### Solution of linear system of DE
---
## Week 14 - Discrete-Time Markov Chain
- the lecture
- capture the eigen-related features of transition matrix
- discuss the asymptotic trend when apply the same P infty time
### Positive Matrix
- defined by entry
### P(Probability) vector and Transition Matrix
- P-vector
- entry sum to 1
- T-matrix
- each column is a P-vector
### Eigenvalues of Positive Transition Matrix abs "<= 1"
- proof

### The eigenvalue "1"
- 1 is the only (in complex number set) eigenvalue with abs 1
- triangular inequality
#### Geometric Multiplicity = 1
- reminder: geo multiplicity = how many Jordan block = rank of eigenspace
- proof
- to make = 1
- required $|v_i| = |v_j|~\forall i,j$
- refer to last section(proof all eigen abs <= 1)
- for replace j with i
- since P is positive
- for take in the abs
- all v_i have same sign
- HW explicitly proof this
- conclusion
- $v = v_i(1,1,1,...,1) = v_i\vec{1}$ is the only eigenvector of abs 1 eigenvalue
#### All Jordabn block of eigenvalue one is of size 1
- reminder: dot diagram
- proof by contradiction

### Asymptotic behavior of $\vec{\pi}^{(k)}$
#### $P^t~and~P$ share the same characteristic poly

#### Decompose Space into Directsum with Jordan form result

#### abs(Eigenvalue) < 1 => go to 0

#### conclusion

### Conclusion
- take kernel of $P-I$, get the only stationary p-vector!
---
## Week 14 : Q on transition matrix
- what happen if the P is "stuck"?
- A, B, C state
- A to A, B to C, C to B
- start A => A
- start B/C => 0.5B + 0.5C

- ans : "positive" = > rank = 1
- will eigenvalue of P be all positive? or 0
- 0 means non-trivial nullspace exist
### HW 11 - if $P^k$ is positive, how about P
- can use positive transition result
- hint: P eigenvalue $\lambda$ => P^k eigenvalue $\lambda^k$
---
## Week 15 - Page Rank
- trivial
---
## Week 15 - Commuting LT
### Eigenspace of T1 is T2-invariant subspace
- proof by $\forall v \in E_{T_1}(\lambda), T_1T_2v = T_2T_1v=T_2\lambda v$
- $E_{T_1}(\lambda)$ is T2 invariant
### Simutaneously Diagonalizable
- restriction of a diagonalizable LT on an invariant subspace is still diagonalizable
### What about Jordan
- nilpotent counter example
### Spectral drawing and symmetric
- $\sigma : V \to V$ be LT that change the vertices is still same graph
- than $L$(the Laplance) and $\sigma$ is commutable
- pick the eigenspaces as a whole => symmetric!
- since the picked eigenspaces are $\sigma -invariant$
---
## Week 15 - Dual Vector Spaces
### Dual Vector Space

### Dual basis

- note that all isomorphisms are based on a certain basis
### Extra structures can be provided by dual space

- Q this

### Dual space of inner product space has natural dual basis
- use f_v(w) = <w, v>
- if v_i forms orthonormal basis
### pullback

- let T be a V to W, f be a W to F in W*
- get a V to F by V to W and W to F
### adjoint

- generalized to for V, W
- $T: V \to W$ to $T^*: W \to V$ by a natural way
---
## Week 15 - HW12 s interesting

### solution
## Reference Book - Invariant Subspaces of Matrices with Applications
---