Deep learning II - I Practical aspects of deep learning - Regularizing your neural network 神经网络范数正则化

Regularizing your neural network 神经网络正则化


Logistic regression regularization

先用简单的逻辑回归正则化作为例子,因为神经网络的参数 W 是2维的。

  1. 无正则

    J ( w , b ) = 1 m i = 1 m L ( y ^ ( i ) y ( i ) )

  2. L 2 正则

    J ( w , b ) = 1 m i = 1 m L ( y ^ ( i ) y ( i ) ) + λ 2 m | | w | | 2 2

    | | w | | 2 2 = j = 1 n x w j 2 = w T w

  3. L 1 正则
    J ( w , b ) = 1 m i = 1 m L ( y ^ ( i ) y ( i ) ) + λ m | | w | | 1

| | w | | 1 = j = 1 n x | w | j

Neural network regularization

  1. Frobenius正则(类似 L 2 正则)
    J ( w [ 1 ] , b [ 1 ] , , w [ l ] , b [ l ] ) = 1 m i = 1 m L ( y ^ ( i ) , y ( i ) ) + 1 2 m l = 1 L | | w [ l ] | | F 2

    | | w [ l ] | | F 2 = i = 1 n [ l ] j = 1 n [ l 1 ] ( w i j [ l ] ) 2

相较于无正则化的反向传播,正则化的反向传播在更新 W 时,会对其进行权重衰减(weight decay),并下降。

d w [ l ] = ( f r o m   b a c k p r o p a g a t i o n ) + λ m w [ l ]

w [ l ] : = w [ l ] α d w [ l ] = w [ l ] α λ m w [ l ] α ( f r o m   b a c k p r o p a g a t i o n ) = ( 1 α λ m ) w [ l ] α ( f r o m   b a c k p r o p a g a t i o n )

这里写图片描述
这里写图片描述

猜你喜欢

转载自blog.csdn.net/zfcjhdq/article/details/80741685