From cf9247282293fe447ecd5c089d4019b69e196a1b Mon Sep 17 00:00:00 2001 From: dongzhihong Date: Tue, 8 Aug 2017 17:53:50 +0800 Subject: [PATCH] clang format --- paddle/operators/softmax_op.h | 20 +++++++++++--------- 1 file changed, 11 insertions(+), 9 deletions(-) diff --git a/paddle/operators/softmax_op.h b/paddle/operators/softmax_op.h index f90056e1e47..4fa6b595404 100644 --- a/paddle/operators/softmax_op.h +++ b/paddle/operators/softmax_op.h @@ -45,19 +45,21 @@ class SoftmaxKernel : public framework::OpKernel { Eigen::DSizes batch_by_one(batch_size, 1); Eigen::DSizes one_by_class(1, num_classes); - auto shifted_logits = (logits - logits.maximum(along_class) - .eval() - .reshape(batch_by_one) - .broadcast(one_by_class)); + auto shifted_logits = (logits - + logits.maximum(along_class) + .eval() + .reshape(batch_by_one) + .broadcast(one_by_class)); softmax.device(context.GetEigenDevice()) = shifted_logits.exp(); softmax.device(context.GetEigenDevice()) = - (softmax * softmax.sum(along_class) - .inverse() - .eval() - .reshape(batch_by_one) - .broadcast(one_by_class)); + (softmax * + softmax.sum(along_class) + .inverse() + .eval() + .reshape(batch_by_one) + .broadcast(one_by_class)); } }; -- GitLab