当前位置:网站首页>Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"
Understanding of "the eigenvectors corresponding to different eigenvalues cannot be orthogonalized"
2022-06-26 23:40:00 【No card, no change】
The following mainly focuses on the following issues :
- The physical meaning of Schmidt orthogonalization
- The meaning of characteristic subspace
- How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”( a key !)
The physical meaning of Schmidt orthogonalization
The process of Schmidt orthogonalization :
Yes α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization ( Hereinafter referred to as ” Orthogonalization “)
- β 1 = α 1 \beta_1 = \alpha_1 β1=α1
- β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2−(β1,β1)(α2,β1)β1
Understand it as a whole , ( α 2 , β 1 ) (\alpha_2, \beta_1) (α2,β1) Express α 2 \alpha_2 α2 and β 1 \beta_1 β1 Inner product , Is a number ; similarly , ( β 1 , β 1 ) (\beta_1,\beta_1) (β1,β1) It's also a number ; β 1 \beta_1 β1 Namely α 1 \alpha_1 α1.
therefore β 2 \beta_2 β2 It's essentially α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The linear combination of .
Simplify β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 β2=α2−(β1,β1)(α2,β1)β1 :
β 2 = α 2 − ( α 2 , β 1 ) ( β 1 , β 1 ) β 1 = α 2 − ∣ α 2 ∣ ∣ β 1 ∣ c o s θ ∣ β 1 ∣ 2 β 1 = α 2 − ∣ α 2 ∣ c o s θ ∣ β 1 ∣ β 1 = α 2 − ∣ α 2 ∣ β 1 ∣ β 1 ∣ c o s θ \beta_2 = \alpha_2 - \frac{(\alpha_2, \beta_1)}{(\beta_1,\beta_1)}\beta_1 \\ =\alpha_2 - \frac{|\alpha_2||\beta_1|cos\theta}{|\beta_1|^2}\beta_1 \\ = \alpha_2 - \frac{|\alpha_2|cos\theta}{|\beta_1|}\beta_1 \\ = \alpha_2 - |\alpha_2|\frac{\beta_1}{|\beta_1|}cos\theta β2=α2−(β1,β1)(α2,β1)β1=α2−∣β1∣2∣α2∣∣β1∣cosθβ1=α2−∣β1∣∣α2∣cosθβ1=α2−∣α2∣∣β1∣β1cosθ
among θ \theta θ by α 1 \alpha_1 α1 and α 2 \alpha_2 α2 The angle between .
After Visualization :

The meaning of characteristic subspace
Simple understanding : For a matrix A A A A characteristic value of λ \lambda λ , The linear independent eigenvector corresponding to the eigenvalue is α 1 \alpha_1 α1 、 α 2 \alpha_2 α2 、 …… 、 α s \alpha_s αs, this s s s The set of all vectors and zero vectors that can be formed by a linear combination of linearly independent eigenvectors is called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of .
Well defined :
Given n n n Order matrix A A A , Regulations V λ = { α ∈ C n ∣ A α = λ α } V_\lambda = \{\alpha∈C^n|A\alpha=\lambda \alpha\} Vλ={ α∈Cn∣Aα=λα}
(1) V λ V_\lambda Vλ Non empty : 0 ∈ V λ 0 ∈ V_\lambda 0∈Vλ
(2) V λ V_\lambda Vλ Close to addition :
α , β ∈ V λ ⇒ A α = λ α , A β = λ β ⇒ A ( α + β ) = λ ( α + β ) ⇒ α + β ∈ V λ \alpha,\beta∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha,A\beta=\lambda\beta \\ \Rightarrow A(\alpha+\beta)=\lambda(\alpha+\beta) \\ \Rightarrow \alpha+\beta∈V_\lambda α,β∈Vλ⇒Aα=λα,Aβ=λβ⇒A(α+β)=λ(α+β)⇒α+β∈Vλ
(3) V λ V_\lambda Vλ Logarithmic multiplication closed :
α ∈ V λ ⇒ A α = λ α ⇒ A ( k α ) = k ( A α ) = k ( λ α ) = λ ( k α ) ⇒ k α ∈ V λ \alpha∈V_\lambda \\ \Rightarrow A\alpha=\lambda\alpha\\ \Rightarrow A(k\alpha)=k(A\alpha)=k(\lambda\alpha) = \lambda(k\alpha) \\ \Rightarrow k\alpha∈ V_\lambda α∈Vλ⇒Aα=λα⇒A(kα)=k(Aα)=k(λα)=λ(kα)⇒kα∈Vλ
(4) α 1 , α 2 , . . . , α s ∈ V λ ⇒ k 1 α 1 + k 2 α 2 + . . . + k s α s ∈ V λ , k i ∈ C \alpha_1,\alpha_2,...,\alpha_s∈ V_\lambda \Rightarrow k_1\alpha_1+k_2\alpha_2 + ... + k_s\alpha_s∈V_\lambda,k_i∈ C α1,α2,...,αs∈Vλ⇒k1α1+k2α2+...+ksαs∈Vλ,ki∈C
(5) V λ V_\lambda Vλ yes n n n Subspace of dimensional complex vector space
λ \lambda λ yes A A A The eigenvalues of the ⇔ \Leftrightarrow ⇔ V λ ≠ 0 V_\lambda ≠{0} Vλ=0
V λ V_\lambda Vλ be called A A A The eigenvalues of the λ \lambda λ The characteristic subspace of
How to understand “ The real symmetric matrix can only orthogonalize the eigenvectors corresponding to the same eigenvalue ”
The reason for this thinking comes from a class of linear algebra problems :
Known matrix A = ? A=? A=?( Real symmetric matrix )
(1) Find invertible matrix P P P, bring P − 1 A P P^{-1}AP P−1AP It's a diagonal matrix
(2) Find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix
The students who have reviewed everything must be able to calculate it skillfully .
First question : All linearly independent eigenvectors corresponding to each eigenvalue are arranged together ;
Second questions : take The same eigenvalue After Schmidt orthogonalization of all corresponding linear independent eigenvectors , Normalize all eigenvectors , Then put all the linearly independent eigenvectors together .
\space
We all know that the eigenvectors corresponding to different eigenvalues of a real symmetric matrix are not only linearly independent , But also orthogonal ( This is easier to prove ); So we only need to orthogonalize all eigenvectors corresponding to the same eigenvalue .
But you may have never seen a class of problems ,“ Prescribed matrix A A A Not a real symmetric matrix , It is required to find the orthogonal matrix Q Q Q, bring Q T A Q Q^{T}AQ QTAQ It's a diagonal matrix ”, because In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized .
\space
Why is that ? The reason will be explained below .
“ In the process of similar diagonalization of non real symmetric matrices, the eigenvectors cannot be orthogonally normalized ” The fundamental reason is that Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector .
For example, two-dimensional vector , hypothesis A = ( 1 − 1 0 2 ) A=\left(\begin{matrix} 1 & -1 \\ 0 & 2 \end{matrix}\right) A=(10−12), Its eigenvalue λ 1 = 1 、 λ 2 = 2 \lambda_1=1、\lambda_2 = 2 λ1=1、λ2=2, The corresponding eigenvectors are α 1 = ( 1 0 ) 、 α 2 = ( 1 − 1 ) \alpha_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\alpha_2=\left(\begin{matrix} 1 \\ -1 \end{matrix}\right) α1=(10)、α2=(1−1), By orthogonalizing them, we get β 1 = ( 1 0 ) 、 β 2 = ( 0 1 ) \beta_1=\left(\begin{matrix} 1 \\ 0 \end{matrix}\right)、\beta_2=\left(\begin{matrix} 0 \\ 1 \end{matrix}\right) β1=(10)、β2=(01). But clearly , A β 2 ≠ λ 2 β 2 A\beta_2≠\lambda_2\beta_2 Aβ2=λ2β2, That is to say, after orthogonalization β 2 \beta_2 β2 The corresponding characteristic value is no longer λ 2 \lambda_2 λ2 了 .
Why? “ Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will change the eigenvalues corresponding to the new vector ”
Visualize the above example :
After Schmidt orthogonalization, the two vectors are indeed orthogonal , But according to the following properties of eigenvalues and eigenvectors , because β 2 \beta_2 β2 Not by α 2 \alpha_2 α2 Linear representation , therefore β 2 \beta_2 β2 Not eigenvalues λ 2 \lambda_2 λ2 Eigenvector of , That is to say β 2 \beta_2 β2 be not in λ 2 \lambda_2 λ2 In the characteristic subspace of , therefore β 2 \beta_2 β2 The corresponding characteristic value is not λ 2 \lambda_2 λ2. In this way, orthogonalization is meaningless .
nature : if α 1 , α 2 , . . . , α t \alpha_1,\alpha_2,...,\alpha_t α1,α2,...,αt All are matrices. A A A Of are characteristic values λ \lambda λ Eigenvector of , Then when k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt When it's not zero , k 1 α + k 2 α 2 + . . . + k t α t k_1\alpha_+k_2\alpha_2+...+k_t\alpha_t k1α+k2α2+...+ktαt Still matrix A A A Belong to characteristic value λ \lambda λ Eigenvector of .
To show more clearly , We expand into three-dimensional space .
Hypothetical matrix A = ( 3 2 0 1 2 0 2 0 1 2 0 3 2 ) A=\left(\begin{matrix} \frac{3}{2} & 0 & \frac{1}{2} \\ 0 & 2 & 0 \\ \frac{1}{2} & 0 & \frac{3}{2} \end{matrix}\right) A=⎝⎛2302102021023⎠⎞ Is a third-order real symmetric matrix , The three eigenvalues are λ 1 = 1 、 λ 2 = 2 、 λ 3 = 2 \lambda_1=1、\lambda_2=2、\lambda_3=2 λ1=1、λ2=2、λ3=2, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Because the number of linearly independent eigenvectors of a real symmetric matrix must be the same as the order , therefore λ 1 \lambda_1 λ1 The corresponding eigenvector is α 1 = ( − 1 , 0 , 1 ) T \alpha_1=(-1,0,1)^T α1=(−1,0,1)T, λ 2 \lambda_2 λ2 and λ 3 \lambda_3 λ3 The corresponding eigenvectors are α 2 = ( 0 , 1 , 0 ) T \alpha_2=(0,1,0)^T α2=(0,1,0)T and α 3 = ( 1 , 2 , 1 ) T \alpha_3=(1,2,1)^T α3=(1,2,1)T, And the eigenvectors corresponding to different eigenvalues are orthogonal , namely ( α 1 , α 2 ) = 0 (\alpha_1,\alpha_2)=0 (α1,α2)=0, ( α 1 , α 3 ) = 0 (\alpha_1,\alpha_3)=0 (α1,α3)=0, α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Not orthogonal .
hypothesis α 1 \alpha_1 α1、 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In 3D space, the following :

Yes α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization :

It can be found that the three vectors are orthogonal , And β 2 \beta_2 β2 and β 3 \beta_3 β3 The corresponding characteristic value is still 2 2 2.
If we deal with the matrix A A A Not a real symmetric matrix , Just a general matrix , That is to say, even the eigenvectors corresponding to different eigenvalues are not necessarily orthogonal , What happens if we orthogonalize two eigenvectors corresponding to different eigenvalues ?
First of all, we need to be clear about , If λ \lambda λ The characteristic subspace of is determined by two linearly independent vectors , Then the eigensubspace of the eigenvalue should be composed of all the vectors in the plane determined by the two linearly independent vectors , That is to say, the eigenvector satisfying the eigenvalue should be in the plane , The vectors that do not satisfy the eigenvalue are all out of plane .
Suppose for a non real symmetric third-order matrix A A A for , Its characteristic value is recorded as λ 1 、 λ 2 、 λ 3 \lambda_1、\lambda_2、\lambda_3 λ1、λ2、λ3, among λ 2 = λ 3 \lambda_2 = \lambda_3 λ2=λ3, Suppose the eigenvectors corresponding to the three eigenvalues α 1 \alpha_1 α1、 α 2 \alpha_2 α2 and α 3 \alpha_3 α3 as follows :

Obviously, none of the three is orthogonal .
If the α 2 \alpha_2 α2 and α 3 \alpha_3 α3 Schmidt orthogonalization , Because orthogonalization is only the addition and multiplication of two vectors , So it will not affect the plane formed by the two , in other words α 2 \alpha_2 α2 and α 3 \alpha_3 α3 After Schmidt orthogonalization β 2 \beta_2 β2 and β 3 \beta_3 β3 Still in α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane formed by , namely β 2 \beta_2 β2 and β 3 \beta_3 β3 In the eigensubspace of the same eigenvalue , That is to say, the eigenvalues of Schmidt orthogonality have not changed .
If the α 1 \alpha_1 α1 and α 2 \alpha_2 α2 Schmidt orthogonalization , take α 2 \alpha_2 α2 As β 1 \beta_1 β1, take α 1 \alpha_1 α1 Convert to β 2 \beta_2 β2, You can get β 1 = α 2 = ( 1 , 1 , 0 ) T \beta_1 = \alpha_2 = (1, 1, 0)^T β1=α2=(1,1,0)T, β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T, The drawing is as follows :

The light blue dotted line indicates λ 1 \lambda_1 λ1 The characteristic subspace of , One dimensional space . After Schmidt orthogonalization λ 1 \lambda_1 λ1 The corresponding eigenvectors α 1 \alpha_1 α1 Turned into ( 1 2 , 1 , 1 2 ) T (\frac{1}{2}, 1, \frac{1}{2})^T (21,1,21)T, Obviously the vector is no longer λ 1 \lambda_1 λ1 In the characteristic subspace of , This explanation β 2 = ( 1 2 , 1 , 1 2 ) T \beta_2 = (\frac{1}{2}, 1, \frac{1}{2})^T β2=(21,1,21)T The corresponding characteristic value is not λ 1 \lambda_1 λ1;
Similarly , Can also be α 1 \alpha_1 α1 As β 1 \beta_1 β1, take α 2 \alpha_2 α2 Convert to β 2 \beta_2 β2, After Schmidt orthogonalization, we will find that β 2 \beta_2 β2 No longer by α 2 \alpha_2 α2 and α 3 \alpha_3 α3 In the plane of , explain β 2 \beta_2 β2 The characteristic value of is changed .
Through the example above , It can be concluded that the Schmidt orthogonalization of eigenvectors corresponding to different eigenvalues will lead to the change of eigenvalues corresponding to new vectors , Because the essence of Schmidt orthogonalization is the addition and multiplication of vectors , After Schmidt orthogonalization, the vector in the same feature subspace is still in the original feature subspace , But it is hard to say that the vectors of different feature subspaces are orthogonalized .
Tang Jiafeng's teaching plan mentioned a property :
set up A A A by n n n Order matrix , λ 1 \lambda_1 λ1, λ 2 \lambda_2 λ2 by A A A Two different eigenvalues of , also A α = λ 1 α A\alpha=\lambda_1\alpha Aα=λ1α, A β = λ 2 β A\beta=\lambda_2\beta Aβ=λ2β( α \alpha α and β \beta β It's a non-zero vector ), For any a ≠ 0 a≠0 a=0, b ≠ 0 b ≠ 0 b=0, vector a α + b β a\alpha+b\beta aα+bβ It must not be an eigenvector .
\space
Schmidt orthogonalization is a process of linear combination , According to this property, it can also be explained that the eigenvectors corresponding to different eigenvalues are not eigenvectors obtained by orthogonalization .
Schmidt orthogonalization when doing problems
If you want to compute an orthogonal matrix Q Q Q, bring Q T A Q = Λ Q^TAQ=\Lambda QTAQ=Λ, be A A A It must be a real symmetric matrix , And it is necessary to Schmidt orthogonalize the eigenvector corresponding to the same eigenvalue , Then normalize all eigenvectors , The normalized eigenvectors are arranged to obtain Q Q Q;
If you want to calculate a reversible matrix P P P, bring P − 1 A P = Λ P^{-1}AP=\Lambda P−1AP=Λ, be A A A The requirements are not so strict . If A A A Is a non real symmetric matrix , that P P P It must not be transformed into an orthogonal matrix .
边栏推荐
- Reading graph augmentations to learn graph representations (lg2ar)
- UnityEditor编辑器扩展-表格功能
- 泰国安全又划算的支付方式
- Share three methods of automatic summation in Excel
- BS-GX-016基于SSM实现教材管理系统
- 【测试】最火的测试开发学习路线内容再次大更新,助力通关大厂测开
- From bitmap to bloom filter, C # implementation
- ASP. Net core create MVC project upload file (buffer mode)
- DAST black box vulnerability scanner part 5: vulnerability scanning engine and service capability
- PHP代码审计系列(一) 基础:方法、思路、流程
猜你喜欢

Why does EDR need defense in depth to combat ransomware?

300题 第三讲 向量组

论文学习——降雨场次划分方法对降雨控制率的影响分析
![Selenium电脑上怎么下载-Selenium下载和安装图文教程[超详细]](/img/ec/1c324dcf38d07742a139aac2bab02e.png)
Selenium电脑上怎么下载-Selenium下载和安装图文教程[超详细]

【界面】pyqt5和Swin Transformer对人脸进行识别

Common techniques of email attachment phishing

Introduction to message queuing
![[微服务]Nacos](/img/69/6641e943c4366d5591acdf9e12389c.png)
[微服务]Nacos
![[microservices] understanding microservices](/img/62/e826e692e7fd6e6e8dab2baa4dd170.png)
[microservices] understanding microservices

UnityEditor編輯器擴展-錶格功能
随机推荐
Is it safe to open an account on the mobile phone to buy stocks? Is it safe to open an account on the Internet to speculate in stocks
Unity初学者肯定能用得上的50个小技巧
Why does EDR need defense in depth to combat ransomware?
Can I open an account for stock trading on my mobile phone? Is it safe to open an account for stock trading on the Internet
Tensorrt notes (VII) sorting out tensorrt use problems
Open world mecha games phantom Galaxy
Share three methods of automatic summation in Excel
Is it safe to open an account and speculate in stocks on the mobile phone? Is it safe to open an account and speculate in stocks on the Internet
[try to hack] forward shell and reverse shell
Which securities dealers recommend? Is it safe to open an account online now?
DAST black box vulnerability scanner part 5: vulnerability scanning engine and service capability
阿里云服务器的购买、基本配置、(xshell)远程连接、搭建环境
BS-GX-016基于SSM实现教材管理系统
Reading graph augmentations to learn graph representations (lg2ar)
Can't write to avoid killing and can easily go online CS through defender
有哪些劵商推荐?现在在线开户安全么?
Alibaba cloud server purchase, basic configuration, (xshell) remote connection and environment building
【老卫搞机】090期:键盘?主机?全功能键盘主机!
DAST 黑盒漏洞扫描器 第五篇:漏洞扫描引擎与服务能力
Tensorrt笔记(七)Tensorrt使用问题整理