diff --git a/python/paddle/fluid/layers/nn.py b/python/paddle/fluid/layers/nn.py index e1387cec1daaeb09d372606fd320928a26838935..16514fc214a57bf8730179b88a354b4cd43cb57d 100644 --- a/python/paddle/fluid/layers/nn.py +++ b/python/paddle/fluid/layers/nn.py @@ -5878,7 +5878,7 @@ def sampled_softmax_with_cross_entropy(logits, 'ignore_index': False, 'numeric_stable_mode': False }) - return outputs / num_true + return loss / num_true def smooth_l1(x, y, inside_weight=None, outside_weight=None, sigma=None): diff --git a/python/paddle/fluid/tests/unittests/test_layers.py b/python/paddle/fluid/tests/unittests/test_layers.py index b73a2fb86617ebcb6c3bde13ee7a7185bfc58874..30194f8cacfea2361ffe4afe537287a261cf470b 100644 --- a/python/paddle/fluid/tests/unittests/test_layers.py +++ b/python/paddle/fluid/tests/unittests/test_layers.py @@ -378,9 +378,10 @@ class TestBook(unittest.TestCase): program = Program() with program_guard(program): logits = layers.data(name='Logits', shape=[256], dtype='float64') - label = layers.data(name='Label', shape=[5], dtype='int64') + label = layers.data(name='Label', shape=[1], dtype='int64') num_samples = 25 - output = layers.sample_logits(logits, label, num_samples) + output = layers.sampled_softmax_with_cross_entropy(logits, label, + num_samples) self.assertIsNotNone(output) print(str(program))