当前位置:网站首页>Complementary knowledge of auto encoder
Complementary knowledge of auto encoder
2022-07-04 12:24:00 【hello_ JeremyWang】
1. Yes Auto-Encoder Ask for more
stay Pytorch actual combat _ Image dimensionality reduction and clustering in , I have briefly introduced Auto-Encoder Principle . For the simplest Auto-Encoder, Our requirement is to minimize reconstruction loss, That is, the restored image or article should be as close as possible to the original image or article .
But beyond that , Can we talk about Auto-Encoder Put forward more requirements ? The answer is yes , Let's take a look :
- Not just reduce reconstruction loss
- Get more acceptable embedding
1.1 Demand one
Request a request that we not only reduce reconstruction loss, And ask us to get embedding Can represent our original pictures or words ( It's like writing wheel eyes represents the yuzhibo family ). How can we make the machine do this ?
From the bottom PPT It can be seen that , We need to build another classifier Discriminator To measure embedding How well it fits the original picture . The specific process is , We set the parameter to θ \theta θ Of Encoder Compress the picture , And compress the obtained embedding and Put pictures together Discriminator To classify , from Discriminator To determine whether the two fit . For each θ \theta θ Come on , We all adjust Discriminator Parameters of ϕ \phi ϕ To make Discriminator The training error is as small as possible , We define this error as L D ∗ L_D^{*} LD∗. Finally, we should adjust the parameters θ \theta θ bring L D ∗ L_D^{*} LD∗ As small as possible .
1.2 Requirement 2
Claim 2 requires us to get embedding More explanatory . Usually we get embedding It looks like a mess , Just like below PPT The picture in the upper right corner is the same . We want to know embedding What information does each part represent . As shown in the figure below , In speech training , What we got embedding It may contain the information of the speaker ( Such as : Pronunciation habits and so on ) And the information in the discourse itself , We want to separate them .
How to do it specifically ? A simple and natural idea is , We train two Encoder, One of them is specially used to extract the information of the speech itself , The other is used to extract the information of the speaker . What's the use ? For example, we can combine the information of another speaker with the information of the discourse itself , Realize the effect of changing sound .
How to train? There are Encoder Well ? One way is reverse training . Similarly, we create a binary Discriminator, This Discriminator The function of is to eat the part that represents the information of the discourse itself embedding, And decide who said it . If our Encoder Be able to cheat Discriminator, He couldn't tell who said it , That explains this part embedding The information of the speaker is no longer contained in .
边栏推荐
- First knowledge of spark - 7000 words +15 diagrams, and learn the basic knowledge of spark
- IPv6 experiment
- Anti clockwise rotation method of event arrangement -- PHP implementation
- Local MySQL forget password modification method (Windows) [easy to understand]
- CSDN documentation specification
- Dos and path
- QQ one click cookie acquisition
- Xiaobing · beauty appraisal
- LxC shared directory addition and deletion
- Memory computing integration: AI chip architecture in the post Moorish Era
猜你喜欢
![Entitas learning [3] multi context system](/img/f9/a3ce86ff2121dd1043305b7e834cc5.jpg)
Entitas learning [3] multi context system

Azure solution: how can third-party tools call azure blob storage to store data?

ASP. Net razor – introduction to VB loops and arrays

Login operation (for user name and password)

MySQL advanced review

How to judge the advantages and disadvantages of low code products in the market?

How to use "bottom logic" to see the cards in the world?
![[Yunju entrepreneurial foundation notes] Chapter II entrepreneur test 11](/img/6a/398d9cceecdd9d7c9c4613d8b5ca27.jpg)
[Yunju entrepreneurial foundation notes] Chapter II entrepreneur test 11

Tableau makes data summary after linking the database, and summary exceptions occasionally occur.

QQ group administrators
随机推荐
Data communication and network: ch13 Ethernet
'using an alias column in the where clause in PostgreSQL' - using an alias column in the where clause in PostgreSQL
Reptile learning winter vacation series (2)
OSI model notes
Clion configuration of opencv
[Chongqing Guangdong education] National Open University spring 2019 2727 tax basis reference questions
Simple understanding of seesion, cookies, tokens
How to judge the advantages and disadvantages of low code products in the market?
Configure SSH key to realize login free
Googgle guava ImmutableCollections
Map container
[Yunju entrepreneurial foundation notes] Chapter II entrepreneur test 24
[directory] search
template<typename MAP, typename LIST, typename First, typename ... Keytypes > recursive call with indefinite parameters - beauty of Pan China
CSDN documentation specification
A few words explain redis cache penetration, breakdown, avalanche, and redis sentinel
Day01 preliminary packet capture
DVC use case (VI): Data Registry
Local MySQL forget password modification method (Windows) [easy to understand]
QQ group administrators