// Copyright (c) 2020 PaddlePaddle Authors. All Rights Reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. #pragma once #include #include #include #include "paddle/fluid/framework/variable.h" #include "paddle/fluid/imperative/hooks.h" namespace paddle { namespace imperative { class InteriorVarHookPipeline; class LeafVarHookPipeline; class VarBase; class GradOpNode; class VariableWrapper { public: friend class VarBase; explicit VariableWrapper(const std::string& name) : name_(name) {} ~VariableWrapper() { VLOG(10) << "Destruct VariableWrapper: " << Name(); } const framework::Variable& Var() const { return var_; } framework::Variable* MutableVar() { return &var_; } // This is used for python api void SetOverridedStopGradient(bool stop_gradient) { overrided_stop_gradient_ = static_cast(stop_gradient); if (auto grad_var = grad_var_.lock()) { grad_var->SetOverridedStopGradient(stop_gradient); } } // This is used for python api bool OverridedStopGradient() const { return overrided_stop_gradient_ != 0; } // This is used inside C++ int InnerOverridedStopGradient() const { return overrided_stop_gradient_; } // This is used inside C++ void InnerSetOverridedStopGradient(bool stop_gradient) { if (overrided_stop_gradient_ == -1) { overrided_stop_gradient_ = static_cast(stop_gradient); } else { VLOG(6) << "Ignore Stop gradient conversion for Var: " << Name() << "Set value is: " << overrided_stop_gradient_; } if (auto grad_var = grad_var_.lock()) { grad_var->InnerSetOverridedStopGradient(stop_gradient); } } bool IsLeaf() const { if (OverridedStopGradient()) { return true; } if (HasGradVar() && !GetGradVar()->HasGradNode()) { return true; } return false; } bool IsLeafGrad() const { if (!HasGradNode() && !OverridedStopGradient()) { return true; } return false; } void SetPersistable(bool persistable) { persistable_ = persistable; } bool Persistable() const { return persistable_; } bool IsEmpty() const { bool is_empty = true; if (var_.IsInitialized()) { const framework::Tensor* tensor = nullptr; if (var_.IsType()) { tensor = &(var_.Get()); } else if (var_.IsType()) { tensor = &(var_.Get().value()); } else { PADDLE_THROW(platform::errors::PermissionDenied( "Only support LoDTensor and SelectedRows for gradient var")); } if (tensor && tensor->IsInitialized()) { is_empty = false; } } return is_empty || is_empty_; } // TODO(zhouwei): fix Tensor.clear_gradient() bug, function SetIsEmpty() isn't // need void SetIsEmpty(bool is_empty) { is_empty_ = is_empty; } const std::string& Name() const { return name_; } void SetName(const std::string& name) { name_ = name; } void SetType(framework::proto::VarType::Type type) { type_ = type; } framework::proto::VarType::Type Type() const { return type_; } std::shared_ptr GetGradVar() const { return grad_var_.lock(); } const std::weak_ptr& GetWeakGradVar() const { return grad_var_; } std::shared_ptr GetGradNode() const { return grad_node_.lock(); } bool HasGradNode() const { return !grad_node_.expired(); } bool HasGradVar() const { return !grad_var_.expired(); } void SetDataType(framework::proto::VarType::Type data_type) { data_type_ = data_type; } framework::proto::VarType::Type DataType() const { const framework::Tensor* tensor = nullptr; if (var_.IsInitialized()) { if (type_ == framework::proto::VarType::LOD_TENSOR) { tensor = &(var_.Get()); } else if (type_ == framework::proto::VarType::SELECTED_ROWS) { tensor = &(var_.Get().value()); } else { VLOG(6) << "Variable " << name_ << " is not initialized"; return data_type_; } } if (tensor && tensor->IsInitialized()) { return tensor->type(); } else { VLOG(6) << "The tensor of variable " << name_ << " is not initialized"; return data_type_; } } void SetForwardDataType(framework::proto::VarType::Type data_type) { fwd_data_type_ = data_type; } framework::proto::VarType::Type ForwardDataType() const { return fwd_data_type_; } const platform::Place Place() const { const framework::Tensor* tensor = nullptr; auto place = platform::CPUPlace(); // Default place for var not initialized. if (var_.IsInitialized()) { if (type_ == framework::proto::VarType::LOD_TENSOR) { tensor = &(var_.Get()); } else if (type_ == framework::proto::VarType::SELECTED_ROWS) { tensor = &(var_.Get().value()); } else { VLOG(6) << "Variable " << name_ << " is not initialized"; return place; } } if (tensor && tensor->IsInitialized()) { return tensor->place(); } else { VLOG(6) << "The tensor of variable " << name_ << " is not initialized"; return place; } } /* Hook related method: only can be call by GradVarBase */ bool HasInteriorHooks() const { return interior_hooks_ != nullptr; } bool HasLeafHooks() const { return leaf_hooks_ != nullptr; } void AddGradVarInteriorHook(std::unique_ptr&& hook) { auto interior_hooks = GetGradVarInteriorHooksSafely(); interior_hooks->add_hook(std::move(hook)); } void AddGradVarLeafHook(std::unique_ptr&& hook) { auto leaf_hooks = GetGradVarLeafHooksSafely(); leaf_hooks->add_hook(std::move(hook)); } void AddGradVarLeafBackwardHook( std::unique_ptr&& hook) { auto leaf_hooks = GetGradVarLeafHooksSafely(); leaf_hooks->add_backward_hook(std::move(hook)); } const std::shared_ptr& GetInteriorHooks() const { return interior_hooks_; } std::shared_ptr& GetInteriorHooks() { return interior_hooks_; } const std::shared_ptr& GetLeafHooks() const { return leaf_hooks_; } std::shared_ptr& GetLeafHooks() { return leaf_hooks_; } uint32_t InplaceVersionSnapshot() const { return inplace_version_snapshot_; } void ResetInplaceVersion() { auto new_version = var_.CurrentInplaceVersion(); VLOG(6) << "The wrapper version of VariableWrapper '" << name_ << "' will be updated from " << inplace_version_snapshot_ << "to " << new_version; inplace_version_snapshot_ = new_version; } private: void SetGradVar(const std::shared_ptr& var) { auto shared_var = grad_var_.lock(); if (shared_var != var) { PADDLE_ENFORCE_EQ(shared_var, nullptr, platform::errors::PermissionDenied( "Cannot set gradient var wrapper twice")); grad_var_ = var; } } void SetGradNode(const std::shared_ptr& grad_node) { if (!grad_node) { grad_node_.reset(); return; } auto shared_node = grad_node_.lock(); if (shared_node != grad_node) { PADDLE_ENFORCE_EQ( shared_node, nullptr, platform::errors::PermissionDenied("Cannot set gradient op twice")); grad_node_ = grad_node; } } /* Hook related private methods */ std::shared_ptr GetGradVarSafely() const { auto shared_grad_var = grad_var_.lock(); PADDLE_ENFORCE_NOT_NULL( shared_grad_var, platform::errors::PermissionDenied( "Cannot add gradient hook on Tensor without gradient.")); return shared_grad_var; } std::shared_ptr& GetGradVarInteriorHooksSafely() { auto shared_grad_var = GetGradVarSafely(); PADDLE_ENFORCE_EQ(HasGradNode(), true, platform::errors::PermissionDenied( "Only interior Tensor in backward can register " "interior gradient hook.")); if (shared_grad_var->interior_hooks_ == nullptr) { shared_grad_var->interior_hooks_ = std::make_shared(); } return shared_grad_var->interior_hooks_; } std::shared_ptr& GetGradVarLeafHooksSafely() { auto shared_grad_var = GetGradVarSafely(); PADDLE_ENFORCE_EQ( HasGradNode(), false, platform::errors::PermissionDenied( "Only leaf Tensor in backward can register leaf gradient hook.")); if (shared_grad_var->leaf_hooks_ == nullptr) { shared_grad_var->leaf_hooks_ = std::make_shared(); } return shared_grad_var->leaf_hooks_; } private: framework::Variable var_; std::string name_; // add this property for users may set stop_gradient themselves and this // should override the frameworks setting (-1) unset, (1) true, (0) false int overrided_stop_gradient_{-1}; bool persistable_{false}; // Used for checking whether there is any inplace operation affecting gradient // calculation. uint32_t inplace_version_snapshot_{0}; framework::proto::VarType::Type type_{framework::proto::VarType::LOD_TENSOR}; framework::proto::VarType::Type data_type_{framework::proto::VarType::FP32}; // See [ Why need handle complex gradient to real gradient? ] // Used for grad var to get the data type of its corresponding forward var, // if inconsistent, the data type of grad var needs to be casted to be // consistent with forward var framework::proto::VarType::Type fwd_data_type_{ static_cast(-1)}; std::weak_ptr grad_var_; std::weak_ptr grad_node_; // TODO(zhouwei): fix bug of Tensor.clear_gradient(), function SetIsEmpty() // isn't need bool is_empty_{false}; // NOTE: only grad var can hold hooks now // only interior var can hold interior hooks std::shared_ptr interior_hooks_; // only leaf var can hold leaf hooks std::shared_ptr leaf_hooks_; }; } // namespace imperative } // namespace paddle