Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
BaiXuePrincess
Paddle
提交
70c4f7b5
P
Paddle
项目概览
BaiXuePrincess
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
70c4f7b5
编写于
2月 27, 2017
作者:
L
Luo Tao
浏览文件
操作
浏览文件
下载
差异文件
Merge branch 'develop' into layer
上级
5d7e7bc0
111e7710
变更
18
隐藏空白更改
内联
并排
Showing
18 changed file
with
101 addition
and
27 deletion
+101
-27
demo/image_classification/prediction.py
demo/image_classification/prediction.py
+1
-1
demo/model_zoo/resnet/classify.py
demo/model_zoo/resnet/classify.py
+1
-1
paddle/api/Arguments.cpp
paddle/api/Arguments.cpp
+7
-0
paddle/api/GradientMachine.cpp
paddle/api/GradientMachine.cpp
+4
-4
paddle/api/PaddleAPI.h
paddle/api/PaddleAPI.h
+3
-2
paddle/api/Trainer.cpp
paddle/api/Trainer.cpp
+4
-5
paddle/gserver/gradientmachines/GradientMachine.h
paddle/gserver/gradientmachines/GradientMachine.h
+4
-0
paddle/gserver/gradientmachines/MultiGradientMachine.cpp
paddle/gserver/gradientmachines/MultiGradientMachine.cpp
+12
-0
paddle/gserver/gradientmachines/MultiGradientMachine.h
paddle/gserver/gradientmachines/MultiGradientMachine.h
+4
-0
paddle/gserver/gradientmachines/NeuralNetwork.cpp
paddle/gserver/gradientmachines/NeuralNetwork.cpp
+3
-4
paddle/gserver/gradientmachines/NeuralNetwork.h
paddle/gserver/gradientmachines/NeuralNetwork.h
+2
-1
paddle/gserver/layers/CosSimLayer.cpp
paddle/gserver/layers/CosSimLayer.cpp
+2
-2
paddle/gserver/layers/CosSimVecMatLayer.cpp
paddle/gserver/layers/CosSimVecMatLayer.cpp
+2
-2
paddle/math/tests/test_RowBuffer.cpp
paddle/math/tests/test_RowBuffer.cpp
+4
-4
paddle/py_paddle/util.py
paddle/py_paddle/util.py
+1
-1
python/paddle/v2/dataset/__init__.py
python/paddle/v2/dataset/__init__.py
+0
-0
python/paddle/v2/dataset/config.py
python/paddle/v2/dataset/config.py
+8
-0
python/paddle/v2/dataset/mnist.py
python/paddle/v2/dataset/mnist.py
+39
-0
未找到文件。
demo/image_classification/prediction.py
浏览文件 @
70c4f7b5
...
...
@@ -126,7 +126,7 @@ class ImageClassifier():
# For oversampling, average predictions across crops.
# If not, the shape of output[name]: (1, class_number),
# the mean is also applicable.
return
output
[
output_layer
].
mean
(
0
)
return
output
[
output_layer
]
[
'value'
]
.
mean
(
0
)
def
predict
(
self
,
image
=
None
,
output_layer
=
None
):
assert
isinstance
(
image
,
basestring
)
...
...
demo/model_zoo/resnet/classify.py
浏览文件 @
70c4f7b5
...
...
@@ -156,7 +156,7 @@ class ImageClassifier():
# For oversampling, average predictions across crops.
# If not, the shape of output[name]: (1, class_number),
# the mean is also applicable.
res
[
name
]
=
output
[
name
].
mean
(
0
)
res
[
name
]
=
output
[
name
]
[
'value'
]
.
mean
(
0
)
return
res
...
...
paddle/api/Arguments.cpp
浏览文件 @
70c4f7b5
...
...
@@ -38,6 +38,13 @@ Arguments* Arguments::createByPaddleArgumentVector(void* ptr) {
return
args
;
}
Arguments
*
Arguments
::
createByPaddleArgument
(
const
void
*
ptr
)
{
auto
p
=
(
paddle
::
Argument
*
)(
ptr
);
auto
args
=
new
Arguments
();
args
->
m
->
outputs
.
push_back
(
*
p
);
return
args
;
}
Matrix
*
Arguments
::
getSlotValue
(
size_t
idx
)
const
throw
(
RangeError
)
{
auto
&
a
=
m
->
getArg
(
idx
);
return
Matrix
::
createByPaddleMatrixPtr
(
&
a
.
value
);
...
...
paddle/api/GradientMachine.cpp
浏览文件 @
70c4f7b5
...
...
@@ -144,12 +144,12 @@ Parameter* GradientMachine::getParameter(size_t i) throw(RangeError) {
void
GradientMachine
::
randParameters
()
{
m
->
machine
->
randParameters
();
}
Matrix
*
GradientMachine
::
getLayerOutput
(
const
std
::
string
&
layerName
)
const
Arguments
*
GradientMachine
::
getLayerOutput
(
const
std
::
string
&
layerName
)
const
throw
(
UnsupportError
)
{
auto
nn
=
std
::
dynamic_pointer_cast
<
paddle
::
NeuralNetwork
>
(
m
->
machine
)
;
auto
nn
=
m
->
machine
;
if
(
nn
)
{
auto
mat
=
nn
->
getLayerOutput
(
layerName
);
return
Matrix
::
createByPaddleMatrixPtr
(
&
mat
);
auto
arg
=
nn
->
getLayerOutput
(
layerName
);
return
Arguments
::
createByPaddleArgument
(
&
arg
);
}
else
{
throw
UnsupportError
();
}
...
...
paddle/api/PaddleAPI.h
浏览文件 @
70c4f7b5
...
...
@@ -454,6 +454,7 @@ public:
private:
static
Arguments
*
createByPaddleArgumentVector
(
void
*
ptr
);
static
Arguments
*
createByPaddleArgument
(
const
void
*
ptr
);
void
*
getInternalArgumentsPtr
()
const
;
private:
...
...
@@ -769,7 +770,7 @@ public:
void
randParameters
();
Matrix
*
getLayerOutput
(
const
std
::
string
&
layerName
)
const
Arguments
*
getLayerOutput
(
const
std
::
string
&
layerName
)
const
throw
(
UnsupportError
);
/**
...
...
@@ -956,7 +957,7 @@ public:
Arguments
*
getForwardOutput
();
Matrix
*
getLayerOutput
(
const
std
::
string
&
layerName
)
;
Arguments
*
getLayerOutput
(
const
std
::
string
&
layerName
)
const
;
};
/// the N-Best results generated from one input sequence.
...
...
paddle/api/Trainer.cpp
浏览文件 @
70c4f7b5
...
...
@@ -131,12 +131,11 @@ void Trainer::testOneDataBatch(size_t batchSize, const Arguments& args) {
void
TrainerPrivate
::
finishTestPeriod
()
{
tester_
->
finishTestPeriod
();
}
void
Trainer
::
finishTestPeriod
()
{
m
->
finishTestPeriod
();
}
Matrix
*
Trainer
::
getLayerOutput
(
const
std
::
string
&
layerName
)
{
auto
nn
=
std
::
dynamic_pointer_cast
<
paddle
::
NeuralNetwork
>
(
this
->
m
->
getGradientMachine
());
Arguments
*
Trainer
::
getLayerOutput
(
const
std
::
string
&
layerName
)
const
{
auto
nn
=
this
->
m
->
getGradientMachine
();
CHECK
(
nn
)
<<
"trainerInternal_.getGradientMachine() is not NeuralNetwork"
;
auto
m
=
nn
->
getLayerOutput
(
layerName
);
return
Matrix
::
createByPaddleMatrixPtr
(
&
m
);
auto
arg
=
nn
->
getLayerOutput
(
layerName
);
return
Arguments
::
createByPaddleArgument
(
&
arg
);
}
void
Trainer
::
forwardOneBatch
(
size_t
batchSize
)
{
...
...
paddle/gserver/gradientmachines/GradientMachine.h
浏览文件 @
70c4f7b5
...
...
@@ -134,6 +134,10 @@ public:
backward
(
callback
);
}
virtual
Argument
getLayerOutput
(
const
std
::
string
&
layerName
)
{
return
*
((
Argument
*
)
nullptr
);
}
// see comment in Layer.h for the function with the same name
virtual
void
resetState
()
{}
...
...
paddle/gserver/gradientmachines/MultiGradientMachine.cpp
浏览文件 @
70c4f7b5
...
...
@@ -282,6 +282,18 @@ void MultiGradientMachine::forwardBackward(const std::vector<Argument>& inArgs,
backwardImp
(
callback
);
}
Argument
MultiGradientMachine
::
getLayerOutput
(
const
std
::
string
&
layerName
)
{
std
::
vector
<
Argument
>
args
;
args
.
reserve
(
threads_
.
size
());
for
(
auto
&
thread
:
threads_
)
{
args
.
push_back
(
thread
->
getGradientMachine
()
->
getLayerOutput
(
layerName
));
}
outLayerArgs_
.
concat
(
args
,
false
/* use_gpu */
,
outArgStream_
,
passType_
);
return
outLayerArgs_
;
}
void
MultiGradientMachine
::
backwardImp
(
const
UpdateCallback
&
callback
)
{
for
(
size_t
i
=
0
;
i
<
parameters_
.
size
();
i
++
)
{
if
(
!
parameters_
[
i
]
->
useGpu
()
||
parameters_
[
i
]
->
isStatic
())
continue
;
...
...
paddle/gserver/gradientmachines/MultiGradientMachine.h
浏览文件 @
70c4f7b5
...
...
@@ -189,6 +189,8 @@ public:
PassType
passType
,
const
UpdateCallback
&
callback
);
virtual
Argument
getLayerOutput
(
const
std
::
string
&
layerName
);
virtual
void
onPassEnd
();
virtual
void
finish
();
...
...
@@ -314,6 +316,8 @@ protected:
std
::
vector
<
Argument
>
outArgs_
;
hl_stream_t
outArgStream_
;
Argument
outLayerArgs_
;
/// ParameterType which needs to be merged from each GPU
std
::
vector
<
ParameterType
>
mergeTypes_
;
int
numDevices_
;
/* number of gpu devices */
...
...
paddle/gserver/gradientmachines/NeuralNetwork.cpp
浏览文件 @
70c4f7b5
...
...
@@ -293,11 +293,10 @@ void NeuralNetwork::backward(const UpdateCallback& callback) {
}
}
MatrixPtr
NeuralNetwork
::
getLayerOutput
(
const
std
::
string
&
layerName
)
{
auto
it
=
layerMap_
.
find
(
layerName
);
CHECK
(
it
!=
layerMap_
.
end
())
<<
"Cannot find layer: "
<<
layerName
;
return
it
->
second
->
getOutputValue
();
Argument
NeuralNetwork
::
getLayerOutput
(
const
std
::
string
&
layerName
)
{
return
getLayer
(
layerName
)
->
getOutput
();
}
void
NeuralNetwork
::
onPassEnd
()
{
for
(
auto
&
layer
:
layers_
)
{
layer
->
onPassEnd
();
...
...
paddle/gserver/gradientmachines/NeuralNetwork.h
浏览文件 @
70c4f7b5
...
...
@@ -87,7 +87,8 @@ public:
virtual
void
backward
(
const
UpdateCallback
&
callback
=
nullptr
);
MatrixPtr
getLayerOutput
(
const
std
::
string
&
layerName
);
virtual
Argument
getLayerOutput
(
const
std
::
string
&
layerName
);
const
LayerPtr
&
getLayer
(
const
std
::
string
&
layerName
)
const
{
auto
it
=
layerMap_
.
find
(
layerName
);
CHECK
(
it
!=
layerMap_
.
end
())
<<
"Unknown layer "
<<
layerName
;
...
...
paddle/gserver/layers/CosSimLayer.cpp
浏览文件 @
70c4f7b5
...
...
@@ -42,7 +42,7 @@ void CosSimLayer::forward(PassType passType) {
/* malloc memory for the output_ if necessary */
int
batchSize
=
getInputValue
(
0
)
->
getHeight
();
int
size
=
getSize
();
CHECK_EQ
(
forward_
.
size
(),
1
)
<<
"Only one forward function needed"
;
CHECK_EQ
(
forward_
.
size
(),
1
UL
)
<<
"Only one forward function needed"
;
{
REGISTER_TIMER_INFO
(
"CosFwResetTimer"
,
getName
().
c_str
());
...
...
@@ -68,7 +68,7 @@ void CosSimLayer::forward(PassType passType) {
void
CosSimLayer
::
backward
(
const
UpdateCallback
&
callback
)
{
/* activation */
{
REGISTER_TIMER_INFO
(
"CosBpAtvTimer"
,
getName
().
c_str
());
CHECK_EQ
(
backward_
.
size
(),
1
)
<<
"Only one backward function needed"
;
CHECK_EQ
(
backward_
.
size
(),
1
UL
)
<<
"Only one backward function needed"
;
const
auto
outG
=
this
->
getOutputGrad
();
const
auto
outV
=
this
->
getOutputValue
();
...
...
paddle/gserver/layers/CosSimVecMatLayer.cpp
浏览文件 @
70c4f7b5
...
...
@@ -112,7 +112,7 @@ bool CosSimVecMatLayer::init(const LayerMap& layerMap,
void
CosSimVecMatLayer
::
forward
(
PassType
passType
)
{
Layer
::
forward
(
passType
);
CHECK_EQ
(
forward_
.
size
(),
1
)
<<
"Only one forward function needed"
;
CHECK_EQ
(
forward_
.
size
(),
1
UL
)
<<
"Only one forward function needed"
;
MatrixPtr
inV0
=
getInputValue
(
0
);
MatrixPtr
inV1
=
getInputValue
(
1
);
...
...
@@ -145,7 +145,7 @@ void CosSimVecMatLayer::forward(PassType passType) {
}
void
CosSimVecMatLayer
::
backward
(
const
UpdateCallback
&
callback
)
{
CHECK_EQ
(
backward_
.
size
(),
1
)
<<
"Only one forward function needed"
;
CHECK_EQ
(
backward_
.
size
(),
1
UL
)
<<
"Only one forward function needed"
;
MatrixPtr
inV0
=
getInputValue
(
0
);
MatrixPtr
inV1
=
getInputValue
(
1
);
...
...
paddle/math/tests/test_RowBuffer.cpp
浏览文件 @
70c4f7b5
...
...
@@ -17,10 +17,10 @@ limitations under the License. */
TEST
(
RowBuffer
,
testAutoGrow
)
{
paddle
::
RowBuffer
buf
(
128
);
ASSERT_EQ
(
128
,
buf
.
getWidth
());
ASSERT_EQ
(
128
UL
,
buf
.
getWidth
());
ASSERT_TRUE
(
buf
.
isAutoGrowth
());
buf
.
resize
(
2
);
ASSERT_EQ
(
2
,
buf
.
getRowCount
());
ASSERT_EQ
(
2
UL
,
buf
.
getRowCount
());
for
(
size_t
i
=
0
;
i
<
buf
.
getWidth
()
*
2
;
++
i
)
{
buf
.
data
()[
i
]
=
i
;
}
...
...
@@ -35,7 +35,7 @@ TEST(RowBuffer, testAutoGrow) {
data
[
i
]
=
i
;
}
ASSERT_EQ
(
3
,
buf
.
getRowCount
());
ASSERT_EQ
(
3
UL
,
buf
.
getRowCount
());
for
(
size_t
i
=
0
;
i
<
buf
.
getRowCount
()
-
1
;
++
i
)
{
for
(
size_t
j
=
0
;
j
<
buf
.
getWidth
();
++
j
)
{
ASSERT_NEAR
(
i
*
buf
.
getWidth
()
+
j
,
buf
.
get
(
i
)[
j
],
1e-5
);
...
...
@@ -51,7 +51,7 @@ TEST(RowBuffer, testWithMemBuf) {
std
::
make_shared
<
paddle
::
CpuMemoryHandle
>
(
128
*
2
*
sizeof
(
real
));
paddle
::
RowBuffer
buf
(
mem
,
128
);
ASSERT_TRUE
(
!
buf
.
isAutoGrowth
());
ASSERT_EQ
(
2
,
buf
.
getRowCount
());
ASSERT_EQ
(
2
UL
,
buf
.
getRowCount
());
for
(
size_t
i
=
0
;
i
<
buf
.
getWidth
()
*
2
;
++
i
)
{
buf
.
data
()[
i
]
=
i
;
}
...
...
paddle/py_paddle/util.py
浏览文件 @
70c4f7b5
...
...
@@ -208,7 +208,7 @@ def __monkeypatch_gradient_machine__():
output
=
dict
()
for
name
in
layerNames
:
output
[
name
]
=
__
matrix_to_numpy__
(
self
.
getLayerOutput
(
name
))
output
[
name
]
=
__
arguments_to_numpy__
(
0
,
self
.
getLayerOutput
(
name
))
return
output
swig_paddle
.
GradientMachine
.
getLayerOutputs
=
getLayerOutputs
...
...
python/paddle/v2/dataset/__init__.py
0 → 100644
浏览文件 @
70c4f7b5
python/paddle/v2/dataset/config.py
0 → 100644
浏览文件 @
70c4f7b5
import
os
__all__
=
[
'DATA_HOME'
]
DATA_HOME
=
os
.
path
.
expanduser
(
'~/.cache/paddle_data_set'
)
if
not
os
.
path
.
exists
(
DATA_HOME
):
os
.
makedirs
(
DATA_HOME
)
python/paddle/v2/dataset/mnist.py
0 → 100644
浏览文件 @
70c4f7b5
import
sklearn.datasets.mldata
import
sklearn.model_selection
import
numpy
from
config
import
DATA_HOME
__all__
=
[
'train_creator'
,
'test_creator'
]
def
__mnist_reader_creator__
(
data
,
target
):
def
reader
():
n_samples
=
data
.
shape
[
0
]
for
i
in
xrange
(
n_samples
):
yield
(
data
[
i
]
/
255.0
).
astype
(
numpy
.
float32
),
int
(
target
[
i
])
return
reader
TEST_SIZE
=
10000
data
=
sklearn
.
datasets
.
mldata
.
fetch_mldata
(
"MNIST original"
,
data_home
=
DATA_HOME
)
X_train
,
X_test
,
y_train
,
y_test
=
sklearn
.
model_selection
.
train_test_split
(
data
.
data
,
data
.
target
,
test_size
=
TEST_SIZE
,
random_state
=
0
)
def
train_creator
():
return
__mnist_reader_creator__
(
X_train
,
y_train
)
def
test_creator
():
return
__mnist_reader_creator__
(
X_test
,
y_test
)
def
unittest
():
assert
len
(
list
(
test_creator
()()))
==
TEST_SIZE
if
__name__
==
'__main__'
:
unittest
()
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录