当前位置:网站首页>Pytorch parameter initialization

Pytorch parameter initialization

2022-07-07 07:43:00 Melody2050

feasible initializer Yes kaiming_normal、xavier_normal.

Reference blog weight-initialization-in-neural-networks-a-journey-from-the-basics-to-kaiming

Gradient explosion and gradient disappearance

Back propagation

Chain according to gradient Back propagation , When the deep gradient is going to spread to the shallow , Will be multiplied by the gradient of this layer . We refer to Deep feedforward network and Xavier Initialization principle Give examples . Suppose there is a linear connection layer followed by an activation layer , Here's the picture :

  • Linear connection layer f2 The input is x, Output is z. namely z = f 2
原网站

版权声明
本文为[Melody2050]所创,转载请带上原文链接,感谢
https://yzsam.com/2022/188/202207070411402547.html