当前位置:网站首页>The most commonly used objective weighting method -- entropy weight method
The most commonly used objective weighting method -- entropy weight method
2022-06-12 05:16:00 【Breeding apes】
Our journey is the sea of stars , Not smoke and dust .

List of articles
One 、 Principle of entropy weight method
Two 、 The main steps of entropy weight method
2.2 Calculate the ratio of each index under each scheme
2.3 Find the information entropy of each index .
2.4 Determine the weight of each indicator
2.4.1 Calculate the weight of each index through information entropy :
2.4.2 Calculate the weight by calculating the information redundancy :
2.5 Finally, calculate the comprehensive score of each scheme
4.1 The use of entropy weight method
4.2 The advantages of entropy weight method
4.3 Disadvantages of entropy weight method
One 、 Principle of entropy weight method
1.1 Information entropy
Entropy is a physical concept of thermodynamics , It is a measure of the degree of chaos or disorder of the system , The larger the entropy, the more chaotic the system is ( That is, the less information you carry ), The smaller the entropy, the more orderly the system is ( That is, the more information you carry ). Information entropy refers to the concept of entropy in thermodynamics , Shannon called the amount of information contained in the source information entropy , Used to describe the average amount of event information , So in Mathematics , Information entropy is the expectation of the amount of information contained in an event (mean, Or mean , Or expectation , Is the probability of each possible result in the test multiplied by the sum of its results ), According to the definition of expectation , It is conceivable that the formula of information entropy is :

among H It's information entropy ,q Is the number of source messages ,
It's the news
Probability of occurrence .
1.2 Entropy weight method
Information is a measure of the order of a system , Entropy is a measure of the disorder of a system ; According to the definition of information entropy , For an indicator , Entropy can be used to judge the dispersion degree of an index , The smaller the information entropy , The greater the dispersion of the index , The impact of this index on Comprehensive Evaluation ( That's the weight ) The greater the , If the values of an indicator are all equal , Then the index does not play a role in the comprehensive evaluation . therefore , Information entropy can be used as a tool , Calculate the weight of each index , Provide basis for multi index comprehensive evaluation .
Literal paraphrase : When we analyze a factor , Subjectively list some influential indicators , These indicators have some correlation with dependent variables , Whether it is positive or negative correlation , At the same amount of change , The value of each indicator has also changed , At some stage ,X1 Changed 1,X2 Changed 0.5, The dependent variable changes 1, From the perspective of relevance X1 The correlation ratio X2 Be big ,X1 Yes Y The impact of , Then come to see X1 and X2 The index value of , If we say that the proportion is the same ( One value divided by the sum of all values ) After normalization ,X1 It seems that the amount of change is relatively large , The greater the degree of data dispersion after normalization , such as [1 1.5 2] and [1.5 1.75 2] After normalization, it is [0.22 0.33 0.44] and [0.28 0.33 0.38], The dispersion of the former is greater than that of the latter . Although this method is simple , But it is very useful in the practical problems of enterprises , For example, the ability level of business personnel should be evaluated according to their benefit and output .
Two 、 The main steps of entropy weight method
2.1 Data standardization
First, each index is de dimensioned . Suppose given m Indicators :
![]()
among
![]()
It is assumed that the standardized value of each indicator data is
![]()
be


2.2 Calculate the ratio of each index under each scheme
Set a certain level of indicators to have m Two level index , And has obtained n Annual data , Write it down as a matrix
. Under the same index , Calculate the proportion of each year's value in the total value , The formula is as follows :

2.3 Find the information entropy of each index .
According to the definition of information entropy in information theory , The information entropy of a set of data is :

2.4 Determine the weight of each indicator
According to the calculation formula of information entropy , The information entropy of each index is calculated as E1,E2,…,Em.
2.4.1 Calculate the weight of each index through information entropy :

here k It refers to the number of indicators , namely k=m.
2.4.2 Calculate the weight by calculating the information redundancy :
![]()
Then calculate the index weight :

2.5 Finally, calculate the comprehensive score of each scheme

3、 ... and 、 Program (MATLAB)
X=[124.3000 2.4200 25.9800 19.0000 3.1000 79.0000 54.1000 6.1400 3.5700 64.0000
134.7000 2.5000 21.0000 19.2500 3.3400 84.0000 53.7000 6.7400 3.5500 64.9600
193.3000 2.5600 29.2600 19.3100 3.4500 92.0000 54.0000 7.1800 3.4000 65.6500
118.6000 2.5200 31.1900 19.3600 3.5700 105.0000 53.9000 8.2000 3.2700 67.0100
94.9000 2.6000 28.5600 19.4500 3.3900 108.0000 53.6000 8.3400 3.2000 65.3400
123.8000 2.6500 28.1200 19.6200 3.5800 108.0000 53.3000 8.5100 3.1000 66.9900];
[n,m]=size(X);
for i=1:n
for j=1:m
p(i,j)=X(i,j)/sum(X(:,j));
end
end
%% Computation first j The entropy of each index e(j)
k=1/log(n);
for j=1:m
e(j)=-k*sum(p(:,j).*log(p(:,j)));
end
d=ones(1,m)-e; % Computing information entropy redundancy
w=d./sum(d) % Find the weight wFour 、 summary
4.1 The use of entropy weight method
Entropy weight method is one of the most basic models in modeling competition , It is mainly used to solve evaluation problems ( for example : Choose which option is best 、 Which athlete or employee performed better ), Used to determine the weight of each indicator , Weights are used to calculate the final score .
4.2 The advantages of entropy weight method
Entropy method is to determine the index weight according to the variation degree of each index value , This is an objective weighting method , The deviation caused by human factors is avoided . Compared with those subjective assignment methods , Higher accuracy and objectivity , Can better explain the results obtained .
Subjective assignment : Analytic hierarchy process , Efficacy coefficient method , Fuzzy comprehensive evaluation method , Composite index method
4.3 Disadvantages of entropy weight method
(1) The importance of the indicator itself is ignored , Sometimes the index weights determined will be far from the expected results , At the same time, entropy method can not reduce the dimension of evaluation index , That is, the entropy weight method conforms to the mathematical law and has strict mathematical significance , But the subjective intention of decision makers is often ignored .
(2) If the change of index value is very small or suddenly becomes larger and smaller , The entropy weight method has its limitations
Succeed in learning 、 Progress in learning ! We study together and don't give up ~
Remember Sanlian ~ Your support is my biggest motivation !! Welcome to read previous articles ~
Xiao Bian's contact information is as follows , Welcome to communicate with you .
int[] arr=new int[]{4,8,3,2,6,5,1};
int[] index= new int[]{6,4,5,0,3,0,2,6,3,1};
String QQ = "";
for (int i : index){
QQ +=arr[i];
}
System.out.println(" Small make up the QQ:" + QQ);
边栏推荐
- 2022 "college entrance examination memory" has been packaged, please check!
- Interview must ask: summary of ten classic sorting algorithms
- 2022-02-28 WPF upper computer 126 understand mqtt
- asp. Net core theme Middleware
- New year news of osdu open underground data space Forum
- How to deploy dolphin scheduler Apache dolphin scheduler 1.2.0 in cdh5.16.2
- The master programmer "plays" a C program that is not like C
- JS to determine whether the tags of multiple classes are empty
- Harris corner detection principle-
- Main business objects of pupanvr record (5)
猜你喜欢

New year news of osdu open underground data space Forum

cellular automaton

National land use data of 30m precision secondary classification

Drive safety coding & troubleshooting guide

How Bi makes SaaS products have a "sense of security" and "sensitivity" (Part I)

SQL injection upload one sentence Trojan horse (turn)

Soil type, soil texture, soil nutrient and change data, soil organic matter, soil pH, soil nitrogen, phosphorus and potassium

InnoDB data storage structure – MySQL

Interview must ask: summary of ten classic sorting algorithms

2022“高考记忆” 已打包完成,请查收!
随机推荐
Understanding of day16 array create query static and dynamic array array array performance in memory
1007- stair climbing
4.3 simulate browser operation and page waiting (display waiting and implicit waiting, handle)
asp. Net core theme Middleware
The master programmer "plays" a C program that is not like C
Nbiot module me3616 at command mqtt connecting thingsboard
Applet pull-down load refresh onreachbottom
Thingsboard view telemetry data through database
How to deploy dolphin scheduler 1.3.1 on cdh5
one billion one hundred and eleven million one hundred and eleven thousand one hundred and eleven
Thingsboard create RCP widget
Calculation method notes for personal use
Can‘t find a suitable configuration file in this directory or any parent. Error reporting and resolution
[backtracking based on bit operation] queen n problem 2
Shandong University network security range experimental platform -- team and project introduction
12.26 exercise summary
12.24 day exercise -- Programming summation, 99 multiplication table, while loop and for loop exercises
1008 color classification
Overview of common classes
22-2-28 there are many things to do at work today, ETH analysis