提交 6c20e08b 编写于 作者: Y Yu Yang

Try using status to handle Paddle Error

上级 02480316
...@@ -69,8 +69,14 @@ static ClassRegistrar<ActivationFunction> gActivationRegistrar; ...@@ -69,8 +69,14 @@ static ClassRegistrar<ActivationFunction> gActivationRegistrar;
class IdentityActivation : public ActivationFunction { class IdentityActivation : public ActivationFunction {
public: public:
static const std::string name; static const std::string name;
void forward(Argument& act) { (void)act; } Status forward(Argument& act) {
void backward(Argument& act) { (void)act; } (void)act;
return Status();
}
Status backward(Argument& act) {
(void)act;
return Status();
}
const std::string& getName() const { return name; } const std::string& getName() const { return name; }
}; };
const std::string IdentityActivation::name = ""; const std::string IdentityActivation::name = "";
...@@ -86,8 +92,14 @@ static InitFunction __reg_activation__identity([] { ...@@ -86,8 +92,14 @@ static InitFunction __reg_activation__identity([] {
* \f] * \f]
*/ */
BEGIN_DEFINE_ACTIVATION(sigmoid) BEGIN_DEFINE_ACTIVATION(sigmoid)
void forward(Argument& act) { act.value->sigmoid(*act.value); } Status forward(Argument& act) {
void backward(Argument& act) { act.grad->sigmoidDerivative(*act.value); } act.value->sigmoid(*act.value);
return Status();
}
Status backward(Argument& act) {
act.grad->sigmoidDerivative(*act.value);
return Status();
}
END_DEFINE_ACTIVATION(sigmoid) END_DEFINE_ACTIVATION(sigmoid)
/** /**
...@@ -103,9 +115,12 @@ MatrixPtr sftMaxDot_; ...@@ -103,9 +115,12 @@ MatrixPtr sftMaxDot_;
MatrixPtr one_; MatrixPtr one_;
public: public:
void forward(Argument& act) { act.value->softmax(*act.value); } Status forward(Argument& act) {
act.value->softmax(*act.value);
return Status();
}
void backward(Argument& act) { Status backward(Argument& act) {
MatrixPtr outputV = act.value; MatrixPtr outputV = act.value;
MatrixPtr outputG = act.grad; MatrixPtr outputG = act.grad;
...@@ -137,6 +152,7 @@ void backward(Argument& act) { ...@@ -137,6 +152,7 @@ void backward(Argument& act) {
act.grad->softmaxDerivative(*act.value, *sftMaxSum_); act.grad->softmaxDerivative(*act.value, *sftMaxSum_);
} }
return Status();
} }
END_DEFINE_ACTIVATION(softmax) END_DEFINE_ACTIVATION(softmax)
...@@ -151,8 +167,11 @@ ACTIVATION_CLASS_NAME(softmax) softmax_; ...@@ -151,8 +167,11 @@ ACTIVATION_CLASS_NAME(softmax) softmax_;
Argument argument_; Argument argument_;
public: public:
void forward(Argument& act) { Status forward(Argument& act) {
CHECK_EQ(act.value->getWidth(), 1UL); if (act.value->getWidth() != 1UL) {
return Status(
"Input width for each timestep of sequence softmax should be 1");
}
if (!argument_.value) { if (!argument_.value) {
argument_.value = Matrix::create(nullptr, argument_.value = Matrix::create(nullptr,
...@@ -169,10 +188,14 @@ void forward(Argument& act) { ...@@ -169,10 +188,14 @@ void forward(Argument& act) {
auto starts = act.sequenceStartPositions->getVector(useGpu(act.deviceId)); auto starts = act.sequenceStartPositions->getVector(useGpu(act.deviceId));
act.value->sequenceSoftmax(*act.value, *starts); act.value->sequenceSoftmax(*act.value, *starts);
return Status();
} }
void backward(Argument& act) { Status backward(Argument& act) {
CHECK_EQ(act.grad->getWidth(), 1UL); if (act.value->getWidth() != 1UL) {
return Status(
"Input width for each timestep of sequence softmax should be 1");
}
size_t numSequences = act.getNumSequences(); size_t numSequences = act.getNumSequences();
const int* starts = act.sequenceStartPositions->getData(false); const int* starts = act.sequenceStartPositions->getData(false);
...@@ -186,6 +209,7 @@ void backward(Argument& act) { ...@@ -186,6 +209,7 @@ void backward(Argument& act) {
softmax_.backward(argument_); softmax_.backward(argument_);
} }
return Status();
} }
END_DEFINE_ACTIVATION(sequence_softmax) END_DEFINE_ACTIVATION(sequence_softmax)
...@@ -200,9 +224,15 @@ END_DEFINE_ACTIVATION(sequence_softmax) ...@@ -200,9 +224,15 @@ END_DEFINE_ACTIVATION(sequence_softmax)
* 0 otherwise. * 0 otherwise.
*/ */
BEGIN_DEFINE_ACTIVATION(relu) BEGIN_DEFINE_ACTIVATION(relu)
void forward(Argument& act) { act.value->relu(*act.value); } Status forward(Argument& act) {
act.value->relu(*act.value);
return Status();
}
void backward(Argument& act) { act.grad->reluDerivative(*act.value); } Status backward(Argument& act) {
act.grad->reluDerivative(*act.value);
return Status();
}
END_DEFINE_ACTIVATION(relu) END_DEFINE_ACTIVATION(relu)
/** /**
...@@ -219,9 +249,15 @@ END_DEFINE_ACTIVATION(relu) ...@@ -219,9 +249,15 @@ END_DEFINE_ACTIVATION(relu)
* TODO(yuyang18): Remove magic number 24 or make it configuable. * TODO(yuyang18): Remove magic number 24 or make it configuable.
*/ */
BEGIN_DEFINE_ACTIVATION(brelu) BEGIN_DEFINE_ACTIVATION(brelu)
void forward(Argument& act) { act.value->brelu(*act.value); } Status forward(Argument& act) {
act.value->brelu(*act.value);
return Status();
}
void backward(Argument& act) { act.grad->breluDerivative(*act.value); } Status backward(Argument& act) {
act.grad->breluDerivative(*act.value);
return Status();
}
END_DEFINE_ACTIVATION(brelu) END_DEFINE_ACTIVATION(brelu)
/** /**
...@@ -231,9 +267,15 @@ END_DEFINE_ACTIVATION(brelu) ...@@ -231,9 +267,15 @@ END_DEFINE_ACTIVATION(brelu)
* \f] * \f]
*/ */
BEGIN_DEFINE_ACTIVATION(tanh) BEGIN_DEFINE_ACTIVATION(tanh)
void forward(Argument& act) { act.value->tanh(*act.value); } Status forward(Argument& act) {
act.value->tanh(*act.value);
return Status();
}
void backward(Argument& act) { act.grad->tanhDerivative(*act.value); } Status backward(Argument& act) {
act.grad->tanhDerivative(*act.value);
return Status();
}
END_DEFINE_ACTIVATION(tanh) END_DEFINE_ACTIVATION(tanh)
/** /**
...@@ -248,10 +290,14 @@ real a, b; ...@@ -248,10 +290,14 @@ real a, b;
public: public:
ACTIVATION_CLASS_NAME(stanh)() : a(1.7159), b(2. / 3.) {} ACTIVATION_CLASS_NAME(stanh)() : a(1.7159), b(2. / 3.) {}
void forward(Argument& act) { act.value->scaledTanh(*act.value, a, b); } Status forward(Argument& act) {
act.value->scaledTanh(*act.value, a, b);
return Status();
}
void backward(Argument& act) { Status backward(Argument& act) {
act.grad->scaledTanhDerivative(*act.value, a, b); act.grad->scaledTanhDerivative(*act.value, a, b);
return Status();
} }
END_DEFINE_ACTIVATION(stanh) END_DEFINE_ACTIVATION(stanh)
...@@ -262,9 +308,15 @@ END_DEFINE_ACTIVATION(stanh) ...@@ -262,9 +308,15 @@ END_DEFINE_ACTIVATION(stanh)
* \f] * \f]
*/ */
BEGIN_DEFINE_ACTIVATION(softrelu) BEGIN_DEFINE_ACTIVATION(softrelu)
void forward(Argument& act) { act.value->softrelu(*act.value); } Status forward(Argument& act) {
act.value->softrelu(*act.value);
return Status();
}
void backward(Argument& act) { act.grad->softreluDerivative(*act.value); } Status backward(Argument& act) {
act.grad->softreluDerivative(*act.value);
return Status();
}
END_DEFINE_ACTIVATION(softrelu) END_DEFINE_ACTIVATION(softrelu)
/** /**
...@@ -280,7 +332,7 @@ END_DEFINE_ACTIVATION(softrelu) ...@@ -280,7 +332,7 @@ END_DEFINE_ACTIVATION(softrelu)
* 0 if z=0 * 0 if z=0
*/ */
BEGIN_DEFINE_ACTIVATION(abs) BEGIN_DEFINE_ACTIVATION(abs)
void forward(Argument& act) { Status forward(Argument& act) {
SetDevice device(act.deviceId); SetDevice device(act.deviceId);
Matrix::resizeOrCreate(act.in, Matrix::resizeOrCreate(act.in,
act.value->getHeight(), act.value->getHeight(),
...@@ -290,9 +342,13 @@ void forward(Argument& act) { ...@@ -290,9 +342,13 @@ void forward(Argument& act) {
act.in->copyFrom(*act.value); act.in->copyFrom(*act.value);
act.value->abs2(*act.value); act.value->abs2(*act.value);
return Status();
} }
void backward(Argument& act) { act.grad->absDerivative(*act.in); } Status backward(Argument& act) {
act.grad->absDerivative(*act.in);
return Status();
}
END_DEFINE_ACTIVATION(abs) END_DEFINE_ACTIVATION(abs)
/** /**
...@@ -302,7 +358,7 @@ END_DEFINE_ACTIVATION(abs) ...@@ -302,7 +358,7 @@ END_DEFINE_ACTIVATION(abs)
* \f] * \f]
*/ */
BEGIN_DEFINE_ACTIVATION(square) BEGIN_DEFINE_ACTIVATION(square)
void forward(Argument& act) { Status forward(Argument& act) {
SetDevice device(act.deviceId); SetDevice device(act.deviceId);
Matrix::resizeOrCreate(act.in, Matrix::resizeOrCreate(act.in,
act.value->getHeight(), act.value->getHeight(),
...@@ -312,9 +368,13 @@ void forward(Argument& act) { ...@@ -312,9 +368,13 @@ void forward(Argument& act) {
act.in->copyFrom(*act.value); act.in->copyFrom(*act.value);
act.value->square2(*act.value); act.value->square2(*act.value);
return Status();
} }
void backward(Argument& act) { act.grad->squareDerivative(*act.in); } Status backward(Argument& act) {
act.grad->squareDerivative(*act.in);
return Status();
}
END_DEFINE_ACTIVATION(square) END_DEFINE_ACTIVATION(square)
/** /**
...@@ -324,9 +384,15 @@ END_DEFINE_ACTIVATION(square) ...@@ -324,9 +384,15 @@ END_DEFINE_ACTIVATION(square)
* \f] * \f]
*/ */
BEGIN_DEFINE_ACTIVATION(exponential) BEGIN_DEFINE_ACTIVATION(exponential)
void forward(Argument& act) { act.value->exp2(*act.value); } Status forward(Argument& act) {
act.value->exp2(*act.value);
return Status();
}
void backward(Argument& act) { act.grad->expDerivative(*act.value); } Status backward(Argument& act) {
act.grad->expDerivative(*act.value);
return Status();
}
END_DEFINE_ACTIVATION(exponential) END_DEFINE_ACTIVATION(exponential)
/** /**
...@@ -336,7 +402,7 @@ END_DEFINE_ACTIVATION(exponential) ...@@ -336,7 +402,7 @@ END_DEFINE_ACTIVATION(exponential)
* \f] * \f]
*/ */
BEGIN_DEFINE_ACTIVATION(log) BEGIN_DEFINE_ACTIVATION(log)
void forward(Argument& act) { Status forward(Argument& act) {
SetDevice device(act.deviceId); SetDevice device(act.deviceId);
Matrix::resizeOrCreate(act.in, Matrix::resizeOrCreate(act.in,
act.value->getHeight(), act.value->getHeight(),
...@@ -346,9 +412,13 @@ void forward(Argument& act) { ...@@ -346,9 +412,13 @@ void forward(Argument& act) {
act.in->copyFrom(*act.value); act.in->copyFrom(*act.value);
act.value->log2(*act.value); act.value->log2(*act.value);
return Status();
} }
void backward(Argument& act) { act.grad->dotDiv(*act.grad, *act.in); } Status backward(Argument& act) {
act.grad->dotDiv(*act.grad, *act.in);
return Status();
}
END_DEFINE_ACTIVATION(log) END_DEFINE_ACTIVATION(log)
ActivationFunction* ActivationFunction::create(const std::string& type) { ActivationFunction* ActivationFunction::create(const std::string& type) {
......
...@@ -15,6 +15,7 @@ limitations under the License. */ ...@@ -15,6 +15,7 @@ limitations under the License. */
#pragma once #pragma once
#include <string> #include <string>
#include <vector> #include <vector>
#include "paddle/utils/Status.h"
namespace paddle { namespace paddle {
...@@ -48,7 +49,7 @@ public: ...@@ -48,7 +49,7 @@ public:
* *
* Usually, act is Layer::output_ * Usually, act is Layer::output_
*/ */
virtual void forward(Argument& act) = 0; virtual Status forward(Argument& act) = 0;
/** /**
* @brief Backward propagaion * @brief Backward propagaion
...@@ -57,7 +58,7 @@ public: ...@@ -57,7 +58,7 @@ public:
* - Before calling backward(), act.grad = dE / dy, where E is the error/cost * - Before calling backward(), act.grad = dE / dy, where E is the error/cost
* - After backward() returns, act.grad = dE / dx = (dE/dy) * (dy/dx) * - After backward() returns, act.grad = dE / dx = (dE/dy) * (dy/dx)
*/ */
virtual void backward(Argument& act) = 0; virtual Status backward(Argument& act) = 0;
virtual const std::string& getName() const = 0; virtual const std::string& getName() const = 0;
}; };
......
...@@ -16,6 +16,7 @@ limitations under the License. */ ...@@ -16,6 +16,7 @@ limitations under the License. */
#include "paddle/math/SparseMatrix.h" #include "paddle/math/SparseMatrix.h"
#include "paddle/utils/Logging.h" #include "paddle/utils/Logging.h"
#include "paddle/utils/Status.h"
#include "AddtoLayer.h" #include "AddtoLayer.h"
#include "CRFLayer.h" #include "CRFLayer.h"
...@@ -334,7 +335,8 @@ void Layer::showOutputStats() { ...@@ -334,7 +335,8 @@ void Layer::showOutputStats() {
void Layer::forwardActivation() { void Layer::forwardActivation() {
/* activation */ /* activation */
activation_->forward(output_); auto status = activation_->forward(output_);
CHECK(status.isOK()) << status.what();
/* dropout */ /* dropout */
if (config_.drop_rate() > 0) { if (config_.drop_rate() > 0) {
...@@ -372,7 +374,8 @@ void Layer::backwardActivation() { ...@@ -372,7 +374,8 @@ void Layer::backwardActivation() {
oGrad->dotMul(*oGrad, *dropOutMask_); oGrad->dotMul(*oGrad, *dropOutMask_);
} }
activation_->backward(output_); auto status = activation_->backward(output_);
CHECK(status.isOK()) << status.what();
} }
void Layer::forwardDropOut() { void Layer::forwardDropOut() {
......
...@@ -11,18 +11,44 @@ distributed under the License is distributed on an "AS IS" BASIS, ...@@ -11,18 +11,44 @@ distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and See the License for the specific language governing permissions and
limitations under the License. */ limitations under the License. */
#pragma once
#include <memory> #include <memory>
#include <string> #include <string>
namespace paddle { namespace paddle {
/**
* Status is Paddle error code. It only contain a std::string as error message.
* Although Status inherits the std::exception, but do not throw it except you
* know what you are doing.
*/
class Status final : public std::exception { class Status final : public std::exception {
public: public:
/**
* Default Status. OK
*/
Status() noexcept {} Status() noexcept {}
Status(const std::string& msg) : errMsg_(new std::string(msg)) {} /**
* @brief Create Status with error message
* @param msg
*/
explicit Status(const std::string& msg) : errMsg_(new std::string(msg)) {}
/**
* @brief set a error message for status.
* @param msg
*/
inline void set(const std::string& msg) noexcept {
errMsg_.reset(new std::string(msg));
}
virtual const char* what() const noexcept override { /**
* @brief what will return the error message. If status is OK, return nullptr.
*/
const char* what() const noexcept override {
if (errMsg_) { if (errMsg_) {
return errMsg_->data(); return errMsg_->data();
} else { } else {
...@@ -30,10 +56,14 @@ public: ...@@ -30,10 +56,14 @@ public:
} }
} }
/**
* @brief isOK
* @return true if OK.
*/
inline bool isOK() const noexcept { return errMsg_ == nullptr; } inline bool isOK() const noexcept { return errMsg_ == nullptr; }
private: private:
std::unique_ptr<std::string> errMsg_; std::shared_ptr<std::string> errMsg_;
}; };
} // namespace paddle } // namespace paddle
...@@ -4,6 +4,7 @@ add_simple_unittest(test_CustomStackTrace) ...@@ -4,6 +4,7 @@ add_simple_unittest(test_CustomStackTrace)
add_simple_unittest(test_ThreadBarrier) add_simple_unittest(test_ThreadBarrier)
add_simple_unittest(test_SpinLock) add_simple_unittest(test_SpinLock)
add_simple_unittest(test_SIMDFlags) add_simple_unittest(test_SIMDFlags)
add_simple_unittest(test_Status)
add_executable( add_executable(
test_CustomStackTracePrint test_CustomStackTracePrint
......
/* Copyright (c) 2016 PaddlePaddle Authors. All Rights Reserve.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License. */
#include "paddle/utils/Status.h"
#include <gtest/gtest.h>
TEST(Status, testAll) {
paddle::Status status;
ASSERT_TRUE(status.isOK());
status.set("I'm the error");
ASSERT_FALSE(status.isOK());
ASSERT_STREQ("I'm the error", status.what());
paddle::Status status2("error2");
ASSERT_FALSE(status2.isOK());
ASSERT_STREQ("error2", status2.what());
}
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册