提交 22ec976e 编写于 作者: Z zhoufeng

fix compilation warnings

Signed-off-by: Nzhoufeng <zhoufeng54@huawei.com>
上级 0fe88f07
......@@ -23,7 +23,7 @@ endif()
set(CMAKE_CXX_FLAGS_DEBUG "$ENV{CXXFLAGS} -O0 -g2 -ggdb -fno-inline-functions -fno-omit-frame-pointer -Wl,--allow-shlib-undefined -D_LIBCPP_INLINE_VISIBILITY='' -D_LIBCPP_DISABLE_EXTERN_TEMPLATE=1 -DHALF_ENABLE_CPP11_USER_LITERALS=0 -D_FORTIFY_SOURCE=2 -Wno-cpp")
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -I/usr/local/include -std=c++17 -Werror -Wall -Wno-deprecated-declarations -fPIC")
set(CMAKE_CXX_FLAGS "${CMAKE_CXX_FLAGS} -I/usr/local/include -std=c++17 -Werror -Wall -fPIC")
set(CMAKE_EXPORT_COMPILE_COMMANDS ON)
set(PYBIND11_CPP_STANDARD -std=c++17)
......
......@@ -1818,7 +1818,7 @@ class IrParser {
py::object tensor_obj = LoadObject(lexer_.GetTokenText());
py::array tensor_data = py::cast<py::array>(tensor_obj);
if (tensor_data == nullptr) {
if (!tensor_data) {
return TOK_ERROR;
}
*val_ptr = TensorPy::MakeTensor(tensor_data, TypeIdToType(type));
......
......@@ -49,7 +49,8 @@ void MultitypeFuncGraph::Register(const TypePtrList &types, specialize_fn s_fn)
}
void MultitypeFuncGraph::Register(const TypePtrList &types, const py::function &py_fn) {
MS_LOG(DEBUG) << "Register type (" << ::mindspore::ToString(types) << ", " << std::string(py_fn.str()) << ").";
MS_LOG(DEBUG) << "Register type (" << ::mindspore::ToString(types) << ", " << py::str(py_fn.cast<py::object>())
<< ").";
auto fn = fn_cache_.find(types);
if (fn != fn_cache_.end()) {
MS_LOG(EXCEPTION) << "Cannot register as (" << ::mindspore::ToString(types) << ", already registered.";
......@@ -116,7 +117,7 @@ FuncGraphPtr MultitypeFuncGraph::GenerateFromTypes(const TypePtrList &types) {
auto py_fn = SignMatch(types);
std::ostringstream buffer;
buffer << types;
if (py_fn != py::none()) {
if (!py_fn.is_none()) {
FuncGraphPtr func_graph = parse::ParsePythonCode(py_fn);
if (func_graph == nullptr) {
MS_LOG(EXCEPTION) << "Fail to parse overload function " << buffer.str();
......
......@@ -50,7 +50,7 @@ FuncGraphPtr KPrim::GetBprop(const PrimitivePtr &prim) {
grad_op_child_scope_prefix + prim->name());
ScopeGuard scope_guard(scope);
py::function fn = prim->is_base() ? GetBpropFunction(prim->name()) : prim->cast<PrimitivePyPtr>()->GetBpropFunction();
if (fn == nullptr || py::isinstance<py::none>(fn)) {
if (!fn || py::isinstance<py::none>(fn)) {
MS_LOG(DEBUG) << "Fail to find bprop function for " << prim->name() << ".";
return nullptr;
}
......
......@@ -285,7 +285,7 @@ Status BatchOp::MapColumns(std::pair<std::unique_ptr<TensorQTable>, CBatchInfo>
Status BatchOp::GetBatchSize(int32_t *batch_size, CBatchInfo info) {
#ifdef ENABLE_PYTHON
if (batch_size_func_ != nullptr) {
if (batch_size_func_) {
RETURN_IF_NOT_OK(InvokeBatchSizeFunc(batch_size, info));
} else {
(*batch_size) = start_batch_size_;
......
......@@ -36,7 +36,7 @@ namespace parse {
FuncGraphPtr ParsePythonCode(const py::object &obj, const std::string &python_mod_get_parse_method) {
(void)python_adapter::set_python_scoped();
if (obj == nullptr || py::isinstance<py::none>(obj)) {
if (!obj || py::isinstance<py::none>(obj)) {
MS_LOG(ERROR) << "Parse the python code failed, obj is nullptr or none";
return nullptr;
}
......@@ -265,7 +265,7 @@ FunctionBlockPtr Parser::ParseFunction(const py::object &node, const FunctionBlo
}
bool set_flag = UpdateFuncGraphFlags(ast_->function(), current_fg);
if (ast_->obj() != ast_->function()) {
if (!ast_->obj().is(ast_->function())) {
set_flag = set_flag && UpdateFuncGraphFlags(ast_->obj(), current_fg);
}
......
......@@ -190,7 +190,7 @@ bool AddDFGraph(const std::map<std::string, ExecutorInfoPtr> &info, const py::di
ConvertObjectToTensors(init_params, &init_tensors);
(void)convertor.ConvertAllNode().InitParam(init_tensors).BuildGraph();
if (broadcast_params != py::none()) {
if (!broadcast_params.is_none()) {
if (!py::isinstance<py::dict>(broadcast_params)) {
MS_LOG(ERROR) << "Invalid broadcast params, it must be py::dict type";
return false;
......
......@@ -103,7 +103,7 @@ py::tuple check_bprop_out(const py::object &grads_obj, const py::tuple &py_args)
py::object grad_dtype = grads[i].attr("dtype");
py::tuple arg_shape = py_args[i].attr("shape");
py::object arg_dtype = py_args[i].attr("dtype");
if (!grad_shape.equal(arg_shape) || grad_dtype != arg_dtype) {
if (!grad_shape.equal(arg_shape) || !grad_dtype.is(arg_dtype)) {
MS_EXCEPTION(ValueError) << "For user define net bprop, the gradient of the " << i
<< "th arg should have the same shape and dtype as the " << i << "th arg, but the "
<< i << "th arg shape: " << py::cast<py::str>(arg_shape)
......
......@@ -63,7 +63,7 @@ class PrimitivePy : public Primitive {
const py::object &GetPyObj() const { return python_obj_; }
py::dict RunInfer(const py::tuple &args);
bool ObjHasAttr(const char *attr_name) { return py::hasattr(python_obj_, attr_name); }
bool HasPyObj() { return python_obj_ != nullptr; }
bool HasPyObj() { return python_obj_.operator bool(); }
PrimitivePtr Clone() override;
bool is_tuple_input_ = false;
......
......@@ -24,5 +24,5 @@ bool PyObjectRef::operator==(const BaseRef &other) const {
return *this == utils::cast<PyObjectRef>(other);
}
bool PyObjectRef::operator==(const PyObjectRef &other) const { return object_ == other.object_; }
bool PyObjectRef::operator==(const PyObjectRef &other) const { return object_.is(other.object_); }
} // namespace mindspore
......@@ -61,7 +61,7 @@ void PrintMeTensor(MeTensor* tensor) {
py::array tensor_data = TensorPy::AsNumpy(*tensor);
std::cout << std::string(py::str(tensor_data)) << std::endl;
std::cout << "tensor dtype is: " << std::string(tensor_data.dtype().str()) << std::endl;
std::cout << "tensor dtype is: " << py::str(tensor_data.dtype()) << std::endl;
}
FuncGraphPtr MakeFuncGraph(const PrimitivePtr prim, unsigned int nparam) {
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册