当前位置:网站首页>Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"

Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"

2022-06-26 23:40:00 No card, no change

The following mainly focuses on the following issues :

  1. The physical meaning of Schmidt orthogonalization
  2. The meaning of characteristic subspace
  3. How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”( a key !)

The physical meaning of Schmidt orthogonalization

The process of Schmidt orthogonalization :

Yes α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization ( Hereinafter referred to as ” Orthogonalization “)

  1. β 1 = α 1 \beta_1 = \alpha_1 β1=α1
  2. β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2(β1,β1)(α2,β1)β1

Understand it as a whole , ( α 2 , β 1 ) (\alpha_2, \beta_1) (α2,β1) Express α 2 \alpha_2 α2 and β 1 \beta_1 β1 Inner product , Is a number ; similarly , ( β 1 , β 1 ) (\beta_1,\beta_1) (β1,β1) It's also a number ; β 1 \beta_1 β1 Namely α 1 \alpha_1 α1.

therefore β 2 \beta_2 β2 It's essentially α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The linear combination of .

Simplify β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2(β1,β1)(α2,β1)β1

β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 = α 2 − ∣ α 2 ∣ ∣ β 1 ∣ c o s θ ∣ β 1 ∣ 2 β 1 = α 2 − ∣ α 2 ∣ c o s θ ∣ β 1 ∣ β 1 = α 2 − ∣ α 2 ∣ β 1 ∣ β 1 ∣ c o s θ \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 \\ =\alpha_2 - \frac{|\alpha_2||\beta_1|cos\theta}{|\beta_1|^2}\beta_1 \\ = \alpha_2 - \frac{|\alpha_2|cos\theta}{|\beta_1|}\beta_1 \\ = \alpha_2 - |\alpha_2|\frac{\beta_1}{|\beta_1|}cos\theta β2=α2(β1,β1)(α2,β1)β1=α2β12α2β1cosθβ1=α2β1α2cosθβ1=α2α2β1β1cosθ
among θ \theta θ by α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The angle between .

After Visualization :

 Insert picture description here

The meaning of characteristic subspace

Simple understanding : For a matrix A A A A characteristic value of λ \lambda λ , The linear independent eigenvector corresponding to the eigenvalue is α 1 \alpha_1 α1 α 2 \alpha_2 α2 、 …… 、 α s \alpha_s αs, this s s s The set of all vectors and zero vectors that can be formed by a linear combination of linearly independent eigenvectors is called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of .

Well defined :

Given n n n Order matrix A A A , Regulations V λ = { α ∈ C n ∣ A α = λ α } V_\lambda = \{\alpha∈C^n|A\alpha=\lambda \alpha\} Vλ={ αCnAα=λα}

(1) V λ V_\lambda Vλ Non empty : 0 ∈ V λ 0 ∈ V_\lambda 0Vλ

(2) V λ V_\lambda Vλ Close to addition :
α , β ∈ V λ ⇒ A α = λ α , A β = λ β ⇒ A ( α + β ) = λ ( α + β ) ⇒ α + β ∈ V λ \alpha,\beta∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha,A\beta=\lambda\beta \\ \Rightarrow A(\alpha+\beta)=\lambda(\alpha+\beta) \\ \Rightarrow \alpha+\beta∈V_\lambda α,βVλAα=λα,Aβ=λβA(α+β)=λ(α+β)α+βVλ
(3) V λ V_\lambda Vλ Logarithmic multiplication closed :
α ∈ V λ ⇒ A α = λ α ⇒ A ( k α ) = k ( A α ) = k ( λ α ) = λ ( k α ) ⇒ k α ∈ V λ \alpha∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha\\ \Rightarrow A(k\alpha)=k(A\alpha)=k(\lambda\alpha) = \lambda(k\alpha) \\ \Rightarrow k\alpha∈ V_\lambda αVλAα=λαA(kα)=k(Aα)=k(λα)=λ(kα)kαVλ

(4) α 1 , α 2 , . . . , α s ∈ V λ ⇒ k 1 α 1 + k 2 α 2 + . . . + k s α s ∈ V λ , k i ∈ C \alpha_1,\alpha_2,...,\alpha_s∈ V_\lambda \Rightarrow k_1\alpha_1+k_2\alpha_2 + ... + k_s\alpha_s∈V_\lambda,k_i∈ C α1,α2,...,αsVλk1α1+k2α2+...+ksαsVλ,kiC

(5) V λ V_\lambda Vλ yes n n n Subspace of dimensional complex vector space

λ \lambda λ yes A A A The eigenvalues of the ⇔ \Leftrightarrow V λ ≠ 0 V_\lambda ≠{0} Vλ=0

V λ V_\lambda Vλ be called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of

How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”

The reason for this thinking comes from a class of linear algebra problems :

Known matrix A = ? A=? A=?( Real symmetric matrix )
(1) Find invertible matrix P P P, bring P − 1 A P P^{-1}AP P1AP It's a diagonal matrix
(2) Find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix

The students who have reviewed everything must be able to calculate it skillfully .
First question : All linearly independent eigenvectors corresponding to each eigenvalue are arranged together ;
Second questions : take The same eigenvalue After Schmidt orthogonalization of all corresponding linear independent eigenvectors , Normalize all eigenvectors , Then put all the linearly independent eigenvectors together .
  \space  
We all know that the eigenvectors corresponding to different eigenvalues of a real symmetric matrix are not only linearly independent , But also orthogonal ( This is easier to prove ); So we only need to orthogonalize all eigenvectors corresponding to the same eigenvalue .
But you may have never seen a class of problems ,“ Prescribed matrix A A A Not a real symmetric matrix , It is required to find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix ”, because In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized .
  \space  
Why is that ? The reason will be explained below .

“ In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized ” The fundamental reason is that Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector .

For example, two-dimensional vector , hypothesis A = ( 1 − 1 0 2 ) A=\left(\begin{matrix} 1 & -1 \\ 0 & 2 \end{matrix}\right) A=(1012), Its eigenvalue λ 1 = 1 、 λ 2 = 2 \lambda_1=1、\lambda_2 = 2 λ1=1λ2=2, The corresponding eigenvectors are α 1 = ( 1 0 ) 、 α 2 = ( 1 − 1 ) \alpha_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\alpha_2=\left(\begin{matrix} 1 \\ -1 \end{matrix}\right) α1=(10)α2=(11), By orthogonalizing them, we get β 1 = ( 1 0 ) 、 β 2 = ( 0 1 ) \beta_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\beta_2=\left(\begin{matrix} 0 \\ 1 \end{matrix}\right) β1=(10)β2=(01). But clearly , A β 2 ≠ λ 2 β 2 A\beta_2≠\lambda_2\beta_2 Aβ2=λ2β2, That is to say, after orthogonalization β 2 \beta_2 β2 The corresponding characteristic value is no longer λ 2 \lambda_2 λ2 了 .

Why? “ Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector ”

Visualize the above example :
 Insert picture description here

After Schmidt orthogonalization, the two vectors are indeed orthogonal , But according to the following properties of eigenvalues and eigenvectors , because β 2 \beta_2 β2 Not by α 2 \alpha_2 α2 Linear representation , therefore β 2 \beta_2 β2 Not eigenvalues λ 2 \lambda_2 λ2 Eigenvector of , That is to say β 2 \beta_2 β2 be not in λ 2 \lambda_2 λ2 In the characteristic subspace of , therefore β 2 \beta_2 β2 The corresponding characteristic value is not λ 2 \lambda_2 λ2. In this way, orthogonalization is meaningless .

nature : if α 1 , α 2 , . . . , α t \alpha_1,\alpha_2,...,\alpha_t α1,α2,...,αt All are matrices. A A A Of are characteristic values λ \lambda λ Eigenvector of , Then when k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt When it's not zero , k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt Still matrix A A A Belong to characteristic value λ \lambda λ Eigenvector of .


To show more clearly , We expand into three-dimensional space .

Hypothetical matrix A = ( 3 2 0 1 2 0 2 0 1 2 0 3 2 ) A=\left(\begin{matrix} \frac{3}{2} & 0 & \frac{1}{2} \\ 0 & 2 & 0 \\ \frac{1}{2} & 0 & \frac{3}{2} \end{matrix}\right) A=2302102021023 Is a third-order real symmetric matrix , The three eigenvalues are λ 1 = 1 、 λ 2 = 2 、 λ 3 = 2 \lambda_1=1、\lambda_2=2、\lambda_3=2 λ1=1λ2=2λ3=2, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Because the number of linearly independent eigenvectors of a real symmetric matrix must be the same as the order , therefore λ 1 \lambda_1 λ1 The corresponding eigenvector is α 1 = ( − 1 , 0 , 1 ) T \alpha_1=(-1,0,1)^T α1=(1,0,1)T, λ 2 \lambda_2 λ2 and λ 3 \lambda_3 λ3 The corresponding eigenvectors are α 2 = ( 0 , 1 , 0 ) T \alpha_2=(0,1,0)^T α2=(0,1,0)T and α 3 = ( 1 , 2 , 1 ) T \alpha_3=(1,2,1)^T α3=(1,2,1)T, And the eigenvectors corresponding to different eigenvalues are orthogonal , namely ( α 1 , α 2 ) = 0 (\alpha_1,\alpha_2)=0 (α1,α2)=0, ( α 1 , α 3 ) = 0 (\alpha_1,\alpha_3)=0 (α1,α3)=0, α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Not orthogonal .

hypothesis α 1 \alpha_1 α1 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In 3D space, the following :

 Insert picture description here
Yes α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization :

 Insert picture description here
It can be found that the three vectors are orthogonal , And β 2 \beta_2 β2 and β 3 \beta_3 β3 The corresponding characteristic value is still 2 2 2.


If we deal with the matrix A A A Not a real symmetric matrix , Just a general matrix , That is to say, even the eigenvectors corresponding to different eigenvalues are not necessarily orthogonal , What happens if we orthogonalize two eigenvectors corresponding to different eigenvalues ?

First of all, we need to be clear about , If λ \lambda λ The characteristic subspace of is determined by two linearly independent vectors , Then the eigensubspace of the eigenvalue should be composed of all the vectors in the plane determined by the two linearly independent vectors , That is to say, the eigenvector satisfying the eigenvalue should be in the plane , The vectors that do not satisfy the eigenvalue are all out of plane .

Suppose for a non real symmetric third-order matrix A A A for , Its characteristic value is recorded as λ 1 、 λ 2 、 λ 3 \lambda_1、\lambda_2、\lambda_3 λ1λ2λ3, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Suppose the eigenvectors corresponding to the three eigenvalues α 1 \alpha_1 α1 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 as follows :

 Insert picture description here
Obviously, none of the three is orthogonal .

If the α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Schmidt orthogonalization , Because orthogonalization is only the addition and multiplication of two vectors , So it will not affect the plane formed by the two , in other words α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization β 2 \beta_2 β2 and β 3 \beta_3 β3 Still in α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane formed by , namely β 2 \beta_2 β2 and β 3 \beta_3 β3 In the eigensubspace of the same eigenvalue , That is to say, the eigenvalues of Schmidt orthogonality have not changed .

If the α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization , take α 2 \alpha_2 α2 As β 1 \beta_1 β1, take α 1 \alpha_1 α1 Convert to β 2 \beta_2 β2, You can get β 1 = α 2 = ( 1 , 1 , 0 ) T \beta_1 = \alpha_2 = (1, 1, 0)^T β1=α2=(1,1,0)T, β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T, The drawing is as follows :

 Insert picture description here
The light blue dotted line indicates λ 1 \lambda_1 λ1 The characteristic subspace of , One dimensional space . After Schmidt orthogonalization λ 1 \lambda_1 λ1 The corresponding eigenvectors α 1 \alpha_1 α1 Turned into ( 1 2 , 1 , 1 2 ) T (\frac{1}{2}, 1, \frac{1}{2})^T (21,1,21)T, Obviously the vector is no longer λ 1 \lambda_1 λ1 In the characteristic subspace of , This explanation β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T The corresponding characteristic value is not λ 1 \lambda_1 λ1;

Similarly , Can also be α 1 \alpha_1 α1 As β 1 \beta_1 β1, take α 2 \alpha_2 α2 Convert to β 2 \beta_2 β2, After Schmidt orthogonalization, we will find that β 2 \beta_2 β2 No longer by α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane of , explain β 2 \beta_2 β2 The characteristic value of is changed .


Through the example above , It can be concluded that the Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will lead to the change of eigenvalues corresponding to new vectors , Because the essence of Schmidt orthogonalization is the addition and multiplication of vectors , After Schmidt orthogonalization, the vector in the same feature subspace is still in the original feature subspace , But it is hard to say that the vectors of different feature subspaces are orthogonalized .

Tang Jiafeng's teaching plan mentioned a property :
set up A A A by n n n Order matrix , λ 1 \lambda_1 λ1, λ 2 \lambda_2 λ2 by A A A Two different eigenvalues of , also A α = λ 1 α A\alpha=\lambda_1\alpha Aα=λ1α, A β = λ 2 β A\beta=\lambda_2\beta Aβ=λ2β α \alpha α and β \beta β It's a non-zero vector ), For any a ≠ 0 a≠0 a=0, b ≠ 0 b ≠ 0 b=0, vector a α + b β a\alpha+b\beta aα+bβ It must not be an eigenvector .
  \space  
Schmidt orthogonalization is a process of linear combination , According to this property, it can also be explained that the eigenvectors corresponding to different eigenvalues are not eigenvectors obtained by orthogonalization .

Schmidt orthogonalization when doing problems

If you want to compute an orthogonal matrix Q Q Q, bring Q T A Q = Λ Q^TAQ=\Lambda QTAQ=Λ, be A A A It must be a real symmetric matrix , And it is necessary to Schmidt orthogonalize the eigenvector corresponding to the same eigenvalue , Then normalize all eigenvectors , The normalized eigenvectors are arranged to obtain Q Q Q;

If you want to calculate a reversible matrix P P P, bring P − 1 A P = Λ P^{-1}AP=\Lambda P1AP=Λ, be A A A The requirements are not so strict . If A A A Is a non real symmetric matrix , that P P P It must not be transformed into an orthogonal matrix .

原网站

版权声明
本文为[No card, no change]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/02/202202161719367904.html