提交 ff9cf6f4 编写于 作者: D dingminghui 提交者: jackzhang235

refactor(act_op): replace npu with mlu

上级 c8816dde
......@@ -116,7 +116,7 @@ void test_act(std::vector<int64_t> x_shape, std::string op_type) {
opdesc.SetAttr("offset", 0.5f);
}
// create and convert op to NPU model, then run it on NPU
// create and convert op to MLU model, then run it on MLU
auto op = CreateOp<operators::ActivationOp>(opdesc, &scope);
// execute reference implementation and save to output tensor
act_ref(op);
......
......@@ -153,7 +153,7 @@ void test_elementwise_add(const std::vector<int64_t>& x_shape,
opdesc.SetOutput("Out", {out_var_name});
opdesc.SetAttr("axis", axis);
// create and convert op to NPU model, then run it on NPU
// create and convert op to MLU model, then run it on MLU
auto op = CreateOp<operators::ElementwiseOp>(opdesc, &scope);
// execute reference implementation and save to output tensor
......
......@@ -93,7 +93,7 @@ void test_softmax(const std::vector<int64_t>& input_shape, int axis) {
opdesc.SetOutput("Out", {out_var_name});
opdesc.SetAttr("axis", axis);
// create and convert op to NPU model, then run it on NPU
// create and convert op to MLU model, then run it on MLU
auto op = CreateOp<operators::SoftmaxOp>(opdesc, &scope);
// execute reference implementation and save to output tensor
softmax_ref<float>(op);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册