c56dc1933bd601a4630e963d4f0d3142fdc9c0d9,tutorials/mnist.py,,loss,#,82
Before Change
cross_entropy = tf.nn.softmax_cross_entropy_with_logits(logits,
onehot_labels,
name="xentropy")
loss = tf.reduce_mean(cross_entropy, name="xentropy_mean")
return loss
def training(loss, learning_rate):
After Change
cross_entropy = tf.nn.softmax_cross_entropy_with_logits(
logits, onehot_labels, name="xentropy"
)
return tf.reduce_mean(cross_entropy, name="xentropy_mean")
def convert_labels_to_onehot(labels):
In pattern: SUPERPATTERN
Frequency: 3
Non-data size: 2
Instances
Project Name: lene/nn-wtf
Commit Name: c56dc1933bd601a4630e963d4f0d3142fdc9c0d9
Time: 2016-02-14
Author: lene.preuss@gmail.com
File Name: tutorials/mnist.py
Class Name:
Method Name: loss
Project Name: tensorlayer/tensorlayer
Commit Name: fee9a89624746a09024ddaf8582a3a0fd86bffa4
Time: 2019-01-16
Author: dhsig552@163.com
File Name: tensorlayer/cost.py
Class Name:
Method Name: mean_squared_error
Project Name: tensorlayer/tensorlayer
Commit Name: fee9a89624746a09024ddaf8582a3a0fd86bffa4
Time: 2019-01-16
Author: dhsig552@163.com
File Name: tensorlayer/cost.py
Class Name:
Method Name: absolute_difference_error