当前位置:网站首页>Reading papers on false news detection (4): a novel self-learning semi supervised deep learning network to detect fake news on
Reading papers on false news detection (4): a novel self-learning semi supervised deep learning network to detect fake news on
2022-07-29 08:14:00 【Quinn-ntmy】
Paper title :A novel self-learning semi-supervised deep learning network to detect fake news on social media
date :2021
# Based on news text 、# Semi supervision 、# Self learning 、# Pseudo label
One 、 The basic content
The model trained with labeled data is unlabeled data Fake labels , The innovation of this work is to use a confidence function The way is to evaluate the false label , Select high-quality pseudo tags Put the sample into the labeled data , This improves the quality of pseudo tags , So as to better carry out semi supervised false news detection .
Two 、 The motivation of the article
(1) In the actual , Annotated datasets are difficult to obtain ( Because fake news spreads through websites );
(2) Supervised learning model cannot realize self-study , Because it ignores the correlation between real and false data .
3、 ... and 、 Main work
This paper studies a self-learning semi supervised deep learning network , Added one Confidence network layer , It can automatically return and add the correct results , Help neural networks accumulate positive sample cases , So as to improve the accuracy of neural network .
The network trains both supervised and unsupervised tasks to detect fake news , Specifically :
(1) Design a semi supervised deep learning network , Use the modified deep learning machine to train both supervised and unsupervised tasks ;
(2) Very accurate unlabeled data can be automatically added to the training set , And gradually expand the training set in the multiple iterative training process to realize self-learning (self-learning).
Four 、 Model framework

Data processing stage : Data cleaning 、 Divide the data into marked data and unmarked data .
Semi supervised self-learning stage :
The model uses an improved deep learning machine L L L Training at the same time Supervised and Unsupervised Mission . There are supervision tasks Only a small part of the marked data is required in , and Unsupervised task Predict the remaining unlabeled data , And return unmarked data Highly reliable fake tags To enrich the marked data set , So as to achieve the effect of self-learning .
1、 Model training process
D l D_l Dl—— Examples of tags in the training dataset , The size is ∣ L ∣ |L| ∣L∣, D l 0 = ( X 1 , y 1 ) , ( X 2 , y 2 ) , … , ( X l , y l ) D_l^0={(X1,y1),(X2,y2),…,(Xl,yl)} Dl0=(X1,y1),(X2,y2),…,(Xl,yl);
D u D_u Du—— Test unmarked examples in the training set , The size is ∣ U ∣ |U| ∣U∣, D u = X l + 1 , X l + 2 , … , X l + u D_u={Xl+1,Xl+2,…,Xl+u} Du=Xl+1,Xl+2,…,Xl+u.
Workflow :
(1) initialization : In the supervised deep learning module , Use D l 0 D_l^0 Dl0 Train deep learning machines as training sets L L L. then , In the unsupervised deep learning module , D u ′ = ( X l + 1 , y ^ l + 1 ) , ( X l + 2 , y ^ l + 2 ) , … , ( X l + u , y ^ l + u ) D_u^{'}={(X_{l+1},~\hat{y}_{l+1}),(X_{l+2},~\hat{y}_{l+2}),…,(X_{l+u},~\hat{y}_{l+u})} Du′=(Xl+1, y^l+1),(Xl+2, y^l+2),…,(Xl+u, y^l+u) The pseudo tags are trained by the deep learning machine L And their confidence values σ Generate , If σ 0 σ_0 σ0 It's filtering D u ′ D_u^{'} Du′ The threshold value of the false label of self-confidence , be D u ′ D_u^{'} Du′ The set of confidence pseudo tags can be expressed as D p s e u 0 = ( ( X l + i , y ^ l + i ) , ( X l + i + 1 , y ^ l + i + 1 ) , … , ( X l + p + i , y ^ l + p + 2 ) ) D_{pseu}^0={((X_{l+i},~\hat{y}_{l+i}),(X_{l+i+1},~\hat{y}_{l+i+1}),…,(X_{l+p+i},~\hat{y}_{l+p+2}))} Dpseu0=((Xl+i, y^l+i),(Xl+i+1, y^l+i+1),…,(Xl+p+i, y^l+p+2)), The size is ∣ P 0 ∣ |P_0 | ∣P0∣.
(2) repeat : New training set D l 1 = ∣ D l 0 ∪ D p s e u 0 ∣ = ( X 1 , y 1 ) , ( X 1 , y 1 ) , … , ( X l , y l ) , … , ( X l + p , y l + p ) D_l^1=|D_l^0∪D_{pseu}^0 |={(X_1,y_1 ),(X_1,y_1 ),…,(X_l,y_l ),…,(X_{l+p},y_{l+p})} Dl1=∣Dl0∪Dpseu0∣=(X1,y1),(X1,y1),…,(Xl,yl),…,(Xl+p,yl+p) For retraining deep learning machines L To generate a new set of confidence tags D p s e u 2 D_{pseu}^2 Dpseu2, The size is ∣ P 1 ∣ |P_1 | ∣P1∣ And a new training set D l 2 = ∣ D l 1 ∪ D p s e u 1 ∣ D_l^2={|{D_l^1}∪{D_{pseu}^1 }|} Dl2=∣Dl1∪Dpseu1∣. Repeat this step , until D p s e u t = D p s e u t + 1 D_{pseu}^t=D_{pseu}^{t+1} Dpseut=Dpseut+1.
2、 Deep learning machine L L L Basic framework
L L L Through existing neural networks ( for example RNN、CNN、LSTM and Bi-LSTM) Add confidence layer to build .
(1)Embedding layer
(2)Dropout layer( Set to 0.5)
(3)Bi-LSTM layer
(4)Softmax layer
(5)Confidence-function layer:
It is believed that the network layer will automatically return and add correction results , So as to help the neural network accumulate positive sample cases .
This layer is used to calculate D u D_u Du Confidence value of each element in σ σ σ, And in D u ′ D_u^{'} Du′ Generate pseudo tags in . For each input X i X_i Xi, σ ( X i ) = m a x ( 0 , p ( y = j ∣ X i ) ) σ_(X_i )=max(0,p(y=j|X_i)) σ(Xi)=max(0,p(y=j∣Xi));
hypothesis σ 0 σ_0 σ0 It's filtering D u ′ D_u^{'} Du′ The threshold value of the false label of self-confidence , be D u ′ D_u^{'} Du′ The elements in X i X_i Xi False label of confidence , If σ ( X i ) > σ 0 σ_{(X_i )}>σ_0 σ(Xi)>σ0, be y ^ = { 0 , o t h e r w i s e 1 , i f j = a r g m a x p ( y = j ∣ X i ) \hat{y}=\{^{1,~if~~j=argmax~p(y=j|X_i)}_{0,~otherwise} y^={ 0, otherwise1, if j=argmax p(y=j∣Xi).
And then we get D u ′ D_u^{'} Du′ The entire set of confidence pseudo tags D p s e u 0 D_{pseu}^0 Dpseu0, Its size is ∣ P 0 ∣ |P_0 | ∣P0∣.
The final new training set D l 1 = ∣ D l 0 ∪ D p s e u 0 ∣ = ( X 1 , y 1 ) , ( X 1 , y 1 ) , … , ( X l , y l ) , … , ( X l + p , y l + p ) D_l^1=|D_l^0∪D_pseu^0 |={(X_1,y_1 ),(X_1,y_1 ),…,(X_l,y_l ),…,(X_{l+p},y_{l+p})} Dl1=∣Dl0∪Dpseu0∣=(X1,y1),(X1,y1),…,(Xl,yl),…,(Xl+p,yl+p) For retraining deep learning machines L To generate a new set of confidence pseudo tags D p s e u 2 D_{pseu}^2 Dpseu2, The size is ∣ P 1 ∣ |P_1 | ∣P1∣ And a new training set D l 2 = ∣ D l 1 ∪ D p s e u 1 ∣ D_l^2=|D_l^1∪D_{pseu}^1 | Dl2=∣Dl1∪Dpseu1∣. Repeat this step , until D p s e u t = D p s e u t + 1 D_{pseu}^t=D_{pseu}^{t+1} Dpseut=Dpseut+1.
5、 ... and 、 Data sets
FakeNewsNet In the repository PolitiFact and GossipCop Data sets , Each data set contains news content 、 Social background and space-time information .
6、 ... and 、 experimental result

边栏推荐
- Qt/PyQt 窗口类型与窗口标志
- STM32 detection signal frequency
- Simple calculator wechat applet project source code
- Application of explosion-proof inclination sensor in safe operation of LNG
- Crawl notes
- Proteus simulation based on 51 MCU ADC0808
- Useful websites
- Temperature acquisition and control system based on WiFi
- Operator overloading
- MySQL中的时间函数
猜你喜欢

华为无线设备配置利用WDS技术部署WLAN业务

Proteus simulation based on 51 MCU ADC0808

Dynamic Thresholds Buffer Management in a Shared Buffer Packet Switch论文总结

Compatible with cc1101/cmt2300-dp4301 sub-1g wireless transceiver chip

pnpm install出现:ERR_PNPM_PEER_DEP_ISSUES Unmet peer dependencies

sql判断语句的编写

V-Ray 5 acescg workflow settings
![[beauty of software engineering - column notes] 28 | what is the core competitiveness of software engineers? (next)](/img/c1/49b4610c483933442fb96b8e3f5315.png)
[beauty of software engineering - column notes] 28 | what is the core competitiveness of software engineers? (next)
![[beauty of software engineering - column notes] 27 | what is the core competitiveness of software engineers? (top)](/img/23/288f6c946a44e36ab58eb0555f3650.png)
[beauty of software engineering - column notes] 27 | what is the core competitiveness of software engineers? (top)

Simplefoc parameter adjustment 2- speed and position control
随机推荐
Cv520 domestic replacement of ci521 13.56MHz contactless reader chip
In an SQL file, a test table and data are defined above, and you can select* from the test table below
Si12t and si14t low power capacitive touch chips
Simplefoc parameter adjustment 1-torque control
MySQL中的时间函数
[beauty of software engineering - column notes] 21 | architecture design: can ordinary programmers also implement complex systems?
Proteus simulation based on msp430f2491 (realize water lamp)
Implementation of simple matcap+fresnel shader in unity
UE4 highlight official reference value
[密码学实验] 0x00 安装NTL库
Network Security Learning chapter
Random lottery turntable wechat applet project source code
Dynamic thresholds buffer management in a shared buffer packet switch paper summary
File system I
分段分页以及段页结合
The difference between torch.tensor and torch.tensor
Stm8s003 domestic substitute for dp32g003 32-bit microcontroller chip
Tle5012b+stm32f103c8t6 (bluepill) reading angle data
Qt/pyqt window type and window flag
Vmstat memory consumption query