当前位置:网站首页>Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"
Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"
2022-06-26 23:40:00 【No card, no change】
The following mainly focuses on the following issues :
- The physical meaning of Schmidt orthogonalization
- The meaning of characteristic subspace
- How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”( a key !)
The physical meaning of Schmidt orthogonalization
The process of Schmidt orthogonalization :
Yes α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization ( Hereinafter referred to as ” Orthogonalization “)
- β 1 = α 1 \beta_1 = \alpha_1 β1=α1
- β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2−(β1,β1)(α2,β1)β1
Understand it as a whole , ( α 2 , β 1 ) (\alpha_2, \beta_1) (α2,β1) Express α 2 \alpha_2 α2 and β 1 \beta_1 β1 Inner product , Is a number ; similarly , ( β 1 , β 1 ) (\beta_1,\beta_1) (β1,β1) It's also a number ; β 1 \beta_1 β1 Namely α 1 \alpha_1 α1.
therefore β 2 \beta_2 β2 It's essentially α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The linear combination of .
Simplify β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2−(β1,β1)(α2,β1)β1 :
β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 = α 2 − ∣ α 2 ∣ ∣ β 1 ∣ c o s θ ∣ β 1 ∣ 2 β 1 = α 2 − ∣ α 2 ∣ c o s θ ∣ β 1 ∣ β 1 = α 2 − ∣ α 2 ∣ β 1 ∣ β 1 ∣ c o s θ \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 \\ =\alpha_2 - \frac{|\alpha_2||\beta_1|cos\theta}{|\beta_1|^2}\beta_1 \\ = \alpha_2 - \frac{|\alpha_2|cos\theta}{|\beta_1|}\beta_1 \\ = \alpha_2 - |\alpha_2|\frac{\beta_1}{|\beta_1|}cos\theta β2=α2−(β1,β1)(α2,β1)β1=α2−∣β1∣2∣α2∣∣β1∣cosθβ1=α2−∣β1∣∣α2∣cosθβ1=α2−∣α2∣∣β1∣β1cosθ
among θ \theta θ by α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The angle between .
After Visualization :

The meaning of characteristic subspace
Simple understanding : For a matrix A A A A characteristic value of λ \lambda λ , The linear independent eigenvector corresponding to the eigenvalue is α 1 \alpha_1 α1 、 α 2 \alpha_2 α2 、 …… 、 α s \alpha_s αs, this s s s The set of all vectors and zero vectors that can be formed by a linear combination of linearly independent eigenvectors is called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of .
Well defined :
Given n n n Order matrix A A A , Regulations V λ = { α ∈ C n ∣ A α = λ α } V_\lambda = \{\alpha∈C^n|A\alpha=\lambda \alpha\} Vλ={ α∈Cn∣Aα=λα}
(1) V λ V_\lambda Vλ Non empty : 0 ∈ V λ 0 ∈ V_\lambda 0∈Vλ
(2) V λ V_\lambda Vλ Close to addition :
α , β ∈ V λ ⇒ A α = λ α , A β = λ β ⇒ A ( α + β ) = λ ( α + β ) ⇒ α + β ∈ V λ \alpha,\beta∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha,A\beta=\lambda\beta \\ \Rightarrow A(\alpha+\beta)=\lambda(\alpha+\beta) \\ \Rightarrow \alpha+\beta∈V_\lambda α,β∈Vλ⇒Aα=λα,Aβ=λβ⇒A(α+β)=λ(α+β)⇒α+β∈Vλ
(3) V λ V_\lambda Vλ Logarithmic multiplication closed :
α ∈ V λ ⇒ A α = λ α ⇒ A ( k α ) = k ( A α ) = k ( λ α ) = λ ( k α ) ⇒ k α ∈ V λ \alpha∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha\\ \Rightarrow A(k\alpha)=k(A\alpha)=k(\lambda\alpha) = \lambda(k\alpha) \\ \Rightarrow k\alpha∈ V_\lambda α∈Vλ⇒Aα=λα⇒A(kα)=k(Aα)=k(λα)=λ(kα)⇒kα∈Vλ
(4) α 1 , α 2 , . . . , α s ∈ V λ ⇒ k 1 α 1 + k 2 α 2 + . . . + k s α s ∈ V λ , k i ∈ C \alpha_1,\alpha_2,...,\alpha_s∈ V_\lambda \Rightarrow k_1\alpha_1+k_2\alpha_2 + ... + k_s\alpha_s∈V_\lambda,k_i∈ C α1,α2,...,αs∈Vλ⇒k1α1+k2α2+...+ksαs∈Vλ,ki∈C
(5) V λ V_\lambda Vλ yes n n n Subspace of dimensional complex vector space
λ \lambda λ yes A A A The eigenvalues of the ⇔ \Leftrightarrow ⇔ V λ ≠ 0 V_\lambda ≠{0} Vλ=0
V λ V_\lambda Vλ be called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of
How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”
The reason for this thinking comes from a class of linear algebra problems :
Known matrix A = ? A=? A=?( Real symmetric matrix )
(1) Find invertible matrix P P P, bring P − 1 A P P^{-1}AP P−1AP It's a diagonal matrix
(2) Find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix
The students who have reviewed everything must be able to calculate it skillfully .
First question : All linearly independent eigenvectors corresponding to each eigenvalue are arranged together ;
Second questions : take The same eigenvalue After Schmidt orthogonalization of all corresponding linear independent eigenvectors , Normalize all eigenvectors , Then put all the linearly independent eigenvectors together .
\space
We all know that the eigenvectors corresponding to different eigenvalues of a real symmetric matrix are not only linearly independent , But also orthogonal ( This is easier to prove ); So we only need to orthogonalize all eigenvectors corresponding to the same eigenvalue .
But you may have never seen a class of problems ,“ Prescribed matrix A A A Not a real symmetric matrix , It is required to find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix ”, because In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized .
\space
Why is that ? The reason will be explained below .
“ In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized ” The fundamental reason is that Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector .
For example, two-dimensional vector , hypothesis A = ( 1 − 1 0 2 ) A=\left(\begin{matrix} 1 & -1 \\ 0 & 2 \end{matrix}\right) A=(10−12), Its eigenvalue λ 1 = 1 、 λ 2 = 2 \lambda_1=1、\lambda_2 = 2 λ1=1、λ2=2, The corresponding eigenvectors are α 1 = ( 1 0 ) 、 α 2 = ( 1 − 1 ) \alpha_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\alpha_2=\left(\begin{matrix} 1 \\ -1 \end{matrix}\right) α1=(10)、α2=(1−1), By orthogonalizing them, we get β 1 = ( 1 0 ) 、 β 2 = ( 0 1 ) \beta_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\beta_2=\left(\begin{matrix} 0 \\ 1 \end{matrix}\right) β1=(10)、β2=(01). But clearly , A β 2 ≠ λ 2 β 2 A\beta_2≠\lambda_2\beta_2 Aβ2=λ2β2, That is to say, after orthogonalization β 2 \beta_2 β2 The corresponding characteristic value is no longer λ 2 \lambda_2 λ2 了 .
Why? “ Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector ”
Visualize the above example :
After Schmidt orthogonalization, the two vectors are indeed orthogonal , But according to the following properties of eigenvalues and eigenvectors , because β 2 \beta_2 β2 Not by α 2 \alpha_2 α2 Linear representation , therefore β 2 \beta_2 β2 Not eigenvalues λ 2 \lambda_2 λ2 Eigenvector of , That is to say β 2 \beta_2 β2 be not in λ 2 \lambda_2 λ2 In the characteristic subspace of , therefore β 2 \beta_2 β2 The corresponding characteristic value is not λ 2 \lambda_2 λ2. In this way, orthogonalization is meaningless .
nature : if α 1 , α 2 , . . . , α t \alpha_1,\alpha_2,...,\alpha_t α1,α2,...,αt All are matrices. A A A Of are characteristic values λ \lambda λ Eigenvector of , Then when k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt When it's not zero , k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt Still matrix A A A Belong to characteristic value λ \lambda λ Eigenvector of .
To show more clearly , We expand into three-dimensional space .
Hypothetical matrix A = ( 3 2 0 1 2 0 2 0 1 2 0 3 2 ) A=\left(\begin{matrix} \frac{3}{2} & 0 & \frac{1}{2} \\ 0 & 2 & 0 \\ \frac{1}{2} & 0 & \frac{3}{2} \end{matrix}\right) A=⎝⎛2302102021023⎠⎞ Is a third-order real symmetric matrix , The three eigenvalues are λ 1 = 1 、 λ 2 = 2 、 λ 3 = 2 \lambda_1=1、\lambda_2=2、\lambda_3=2 λ1=1、λ2=2、λ3=2, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Because the number of linearly independent eigenvectors of a real symmetric matrix must be the same as the order , therefore λ 1 \lambda_1 λ1 The corresponding eigenvector is α 1 = ( − 1 , 0 , 1 ) T \alpha_1=(-1,0,1)^T α1=(−1,0,1)T, λ 2 \lambda_2 λ2 and λ 3 \lambda_3 λ3 The corresponding eigenvectors are α 2 = ( 0 , 1 , 0 ) T \alpha_2=(0,1,0)^T α2=(0,1,0)T and α 3 = ( 1 , 2 , 1 ) T \alpha_3=(1,2,1)^T α3=(1,2,1)T, And the eigenvectors corresponding to different eigenvalues are orthogonal , namely ( α 1 , α 2 ) = 0 (\alpha_1,\alpha_2)=0 (α1,α2)=0, ( α 1 , α 3 ) = 0 (\alpha_1,\alpha_3)=0 (α1,α3)=0, α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Not orthogonal .
hypothesis α 1 \alpha_1 α1、 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In 3D space, the following :

Yes α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization :

It can be found that the three vectors are orthogonal , And β 2 \beta_2 β2 and β 3 \beta_3 β3 The corresponding characteristic value is still 2 2 2.
If we deal with the matrix A A A Not a real symmetric matrix , Just a general matrix , That is to say, even the eigenvectors corresponding to different eigenvalues are not necessarily orthogonal , What happens if we orthogonalize two eigenvectors corresponding to different eigenvalues ?
First of all, we need to be clear about , If λ \lambda λ The characteristic subspace of is determined by two linearly independent vectors , Then the eigensubspace of the eigenvalue should be composed of all the vectors in the plane determined by the two linearly independent vectors , That is to say, the eigenvector satisfying the eigenvalue should be in the plane , The vectors that do not satisfy the eigenvalue are all out of plane .
Suppose for a non real symmetric third-order matrix A A A for , Its characteristic value is recorded as λ 1 、 λ 2 、 λ 3 \lambda_1、\lambda_2、\lambda_3 λ1、λ2、λ3, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Suppose the eigenvectors corresponding to the three eigenvalues α 1 \alpha_1 α1、 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 as follows :

Obviously, none of the three is orthogonal .
If the α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Schmidt orthogonalization , Because orthogonalization is only the addition and multiplication of two vectors , So it will not affect the plane formed by the two , in other words α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization β 2 \beta_2 β2 and β 3 \beta_3 β3 Still in α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane formed by , namely β 2 \beta_2 β2 and β 3 \beta_3 β3 In the eigensubspace of the same eigenvalue , That is to say, the eigenvalues of Schmidt orthogonality have not changed .
If the α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization , take α 2 \alpha_2 α2 As β 1 \beta_1 β1, take α 1 \alpha_1 α1 Convert to β 2 \beta_2 β2, You can get β 1 = α 2 = ( 1 , 1 , 0 ) T \beta_1 = \alpha_2 = (1, 1, 0)^T β1=α2=(1,1,0)T, β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T, The drawing is as follows :

The light blue dotted line indicates λ 1 \lambda_1 λ1 The characteristic subspace of , One dimensional space . After Schmidt orthogonalization λ 1 \lambda_1 λ1 The corresponding eigenvectors α 1 \alpha_1 α1 Turned into ( 1 2 , 1 , 1 2 ) T (\frac{1}{2}, 1, \frac{1}{2})^T (21,1,21)T, Obviously the vector is no longer λ 1 \lambda_1 λ1 In the characteristic subspace of , This explanation β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T The corresponding characteristic value is not λ 1 \lambda_1 λ1;
Similarly , Can also be α 1 \alpha_1 α1 As β 1 \beta_1 β1, take α 2 \alpha_2 α2 Convert to β 2 \beta_2 β2, After Schmidt orthogonalization, we will find that β 2 \beta_2 β2 No longer by α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane of , explain β 2 \beta_2 β2 The characteristic value of is changed .
Through the example above , It can be concluded that the Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will lead to the change of eigenvalues corresponding to new vectors , Because the essence of Schmidt orthogonalization is the addition and multiplication of vectors , After Schmidt orthogonalization, the vector in the same feature subspace is still in the original feature subspace , But it is hard to say that the vectors of different feature subspaces are orthogonalized .
Tang Jiafeng's teaching plan mentioned a property :
set up A A A by n n n Order matrix , λ 1 \lambda_1 λ1, λ 2 \lambda_2 λ2 by A A A Two different eigenvalues of , also A α = λ 1 α A\alpha=\lambda_1\alpha Aα=λ1α, A β = λ 2 β A\beta=\lambda_2\beta Aβ=λ2β( α \alpha α and β \beta β It's a non-zero vector ), For any a ≠ 0 a≠0 a=0, b ≠ 0 b ≠ 0 b=0, vector a α + b β a\alpha+b\beta aα+bβ It must not be an eigenvector .
\space
Schmidt orthogonalization is a process of linear combination , According to this property, it can also be explained that the eigenvectors corresponding to different eigenvalues are not eigenvectors obtained by orthogonalization .
Schmidt orthogonalization when doing problems
If you want to compute an orthogonal matrix Q Q Q, bring Q T A Q = Λ Q^TAQ=\Lambda QTAQ=Λ, be A A A It must be a real symmetric matrix , And it is necessary to Schmidt orthogonalize the eigenvector corresponding to the same eigenvalue , Then normalize all eigenvectors , The normalized eigenvectors are arranged to obtain Q Q Q;
If you want to calculate a reversible matrix P P P, bring P − 1 A P = Λ P^{-1}AP=\Lambda P−1AP=Λ, be A A A The requirements are not so strict . If A A A Is a non real symmetric matrix , that P P P It must not be transformed into an orthogonal matrix .
边栏推荐
- Unityeditor Editor Extension - table function
- DAST black box vulnerability scanner part 5: vulnerability scanning engine and service capability
- Installation of xshell and xftp
- 简单测试轻量级表达式计算器Flee
- CVE-2022-30190 Follina Office RCE分析【附自定义word模板POC】
- 代码之外:写作是倒逼成长的最佳方式
- 【老卫搞机】090期:键盘?主机?全功能键盘主机!
- Simple test lightweight expression calculator fly
- Leetcode 718. Longest repeating subarray (violence enumeration, to be solved)
- 软件工程导论——第四章——形式化说明技术
猜你喜欢
![[interface] pyqt5 and swing transformer for face recognition](/img/37/b259627a8ffd82afe8e8f3029bf290.png)
[interface] pyqt5 and swing transformer for face recognition

【界面】pyqt5和Swin Transformer对人脸进行识别

A simple and crude method for exporting R language list to local
![[microservices] understanding microservices](/img/62/e826e692e7fd6e6e8dab2baa4dd170.png)
[microservices] understanding microservices

The user adds a timer function in the handler () goroutine. If it times out, it will be kicked out

不会写免杀也能轻松过defender上线CS

An article takes you to learn container escape

go中的微服务和容器编排

UnityEditor編輯器擴展-錶格功能

Reading graph augmentations to learn graph representations (lg2ar)
随机推荐
Is the low commission free account opening channel safe?
入侵痕迹清理
Smartbi gives you a piece to play with Boston matrix
【LeetCode】1984. Minimum difference between highest and lowest of K scores
300 questions lesson 3 vector group
【界面】pyqt5和Swin Transformer对人脸进行识别
[kotlin] keyword suspend learning of thread operation and async understanding
微信小程序自动生成打卡海报
为什么我不推荐去SAP培训机构参加培训?
股票怎样在手机上开户安全吗 网上开户炒股安全吗
NPM command prompt error: eacces: permission denied
[strong foundation program] video of calculus in mathematics and Physics Competition
Nacos安装指南
代码之外:写作是倒逼成长的最佳方式
Is it reliable to open an account for stock trading on the mobile phone? Is it safe to open an account for stock trading on the Internet
不同的子序列问题I
让敏捷回归本源——读《敏捷整洁之道》有感
[microservices] Understanding microservices
客户端实现client.go客户端类型定义连接
My advanced learning notes of C language ----- keywords