嘿,我是新来tensorflow,甚至经过大量的努力可以在此之后 L1则项不会增加误差项如何准确地添加L1正规化tensorflow误差函数
x = tf.placeholder("float", [None, n_input])
# Weights and biases to hidden layer
ae_Wh1 = tf.Variable(tf.random_uniform((n_input, n_hidden1), -1.0/math.sqrt(n_input), 1.0/math.sqrt(n_input)))
ae_bh1 = tf.Variable(tf.zeros([n_hidden1]))
ae_h1 = tf.nn.tanh(tf.matmul(x,ae_Wh1) + ae_bh1)
ae_Wh2 = tf.Variable(tf.random_uniform((n_hidden1, n_hidden2), -1.0/math.sqrt(n_hidden1), 1.0/math.sqrt(n_hidden1)))
ae_bh2 = tf.Variable(tf.zeros([n_hidden2]))
ae_h2 = tf.nn.tanh(tf.matmul(ae_h1,ae_Wh2) + ae_bh2)
ae_Wh3 = tf.transpose(ae_Wh2)
ae_bh3 = tf.Variable(tf.zeros([n_hidden1]))
ae_h1_O = tf.nn.tanh(tf.matmul(ae_h2,ae_Wh3) + ae_bh3)
ae_Wh4 = tf.transpose(ae_Wh1)
ae_bh4 = tf.Variable(tf.zeros([n_input]))
ae_y_pred = tf.nn.tanh(tf.matmul(ae_h1_O,ae_Wh4) + ae_bh4)
ae_y_actual = tf.placeholder("float", [None,n_input])
meansq = tf.reduce_mean(tf.square(ae_y_actual - ae_y_pred))
train_step = tf.train.GradientDescentOptimizer(0.05).minimize(meansq)
我运行上面使用
init = tf.initialize_all_variables()
sess = tf.Session()
sess.run(init)
n_rounds = 100
batch_size = min(500, n_samp)
for i in range(100):
sample = np.random.randint(n_samp, size=batch_size)
batch_xs = input_data[sample][:]
batch_ys = output_data_ae[sample][:]
sess.run(train_step, feed_dict={x: batch_xs, ae_y_actual:batch_ys})
以上为4层自动编码器的代码图表,“meansq”是我的平方损失本功能离子。我怎样才能为网络中的权重矩阵(张量)添加L1 reguarisation?
L1可以用sum和abs运算符来实现,这两个运算符都存在于tensorflow(包括它们的梯度) –
'0.001 * tf.reduce_sum(tf.abs(parameters))'给出了参数向量的L1范数在技术上可能是一个更高的排名张量),所以惩罚你的学习 –
非常感谢你+雅罗斯拉夫。因此,对于我的情况,它应该像(?) meansq = tf.reduce_mean(tf.square(ae_y_actual-ae_y_pred))+ 0.001 * tf.reduce_sum(tf.abs(ae_Wh1))+ 0.001 * tf。 reduce_sum(tf.abs(ae_Wh1)) 我是否正确? – Abhishek