提交 810247de 编写于 作者: F Francois Chollet

Fix activations tests

上级 f8763efb
......@@ -164,19 +164,18 @@ def test_relu():
assert_allclose(result, test_values, rtol=1e-05)
# Test max_value
test_values = [0.5, 1.5]
test_values = np.array([[0.5, 1.5]], dtype=K.floatx())
f = K.function([x], [activations.relu(x, max_value=1.)])
result = f([test_values])[0]
assert np.max(result) <= 1.
# Test max_value == 6.
test_values = [0.5, 6.]
test_values = np.array([[0.5, 6.]], dtype=K.floatx())
f = K.function([x], [activations.relu(x, max_value=1.)])
result = f([test_values])[0]
assert np.max(result) <= 6.
def test_elu():
x = K.placeholder(ndim=2)
f = K.function([x], [activations.elu(x, 0.5)])
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册