Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
Crayon鑫
Paddle
提交
14eb5b8e
P
Paddle
项目概览
Crayon鑫
/
Paddle
与 Fork 源项目一致
Fork自
PaddlePaddle / Paddle
通知
1
Star
1
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
1
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
Paddle
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
1
Issue
1
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
14eb5b8e
编写于
3月 24, 2017
作者:
Y
Yancey1989
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
rename fetch_all to fetch; add fetch_all function
上级
7b72c792
变更
10
隐藏空白更改
内联
并排
Showing
10 changed file
with
42 addition
and
35 deletion
+42
-35
python/paddle/v2/dataset/cifar.py
python/paddle/v2/dataset/cifar.py
+8
-13
python/paddle/v2/dataset/common.py
python/paddle/v2/dataset/common.py
+12
-0
python/paddle/v2/dataset/conll05.py
python/paddle/v2/dataset/conll05.py
+6
-6
python/paddle/v2/dataset/imdb.py
python/paddle/v2/dataset/imdb.py
+1
-1
python/paddle/v2/dataset/imikolov.py
python/paddle/v2/dataset/imikolov.py
+1
-1
python/paddle/v2/dataset/mnist.py
python/paddle/v2/dataset/mnist.py
+3
-1
python/paddle/v2/dataset/movielens.py
python/paddle/v2/dataset/movielens.py
+2
-2
python/paddle/v2/dataset/sentiment.py
python/paddle/v2/dataset/sentiment.py
+2
-2
python/paddle/v2/dataset/uci_housing.py
python/paddle/v2/dataset/uci_housing.py
+2
-2
python/paddle/v2/dataset/wmt14.py
python/paddle/v2/dataset/wmt14.py
+5
-7
未找到文件。
python/paddle/v2/dataset/cifar.py
浏览文件 @
14eb5b8e
...
...
@@ -20,7 +20,7 @@ TODO(yuyang18): Complete the comments.
import
cPickle
import
itertools
import
numpy
import
paddle.v2.dataset.common
from
common
import
download
import
tarfile
__all__
=
[
'train100'
,
'test100'
,
'train10'
,
'test10'
]
...
...
@@ -55,28 +55,23 @@ def reader_creator(filename, sub_name):
def
train100
():
return
reader_creator
(
paddle
.
v2
.
dataset
.
common
.
download
(
CIFAR100_URL
,
'cifar'
,
CIFAR100_MD5
),
'train'
)
download
(
CIFAR100_URL
,
'cifar'
,
CIFAR100_MD5
),
'train'
)
def
test100
():
return
reader_creator
(
paddle
.
v2
.
dataset
.
common
.
download
(
CIFAR100_URL
,
'cifar'
,
CIFAR100_MD5
),
'test'
)
return
reader_creator
(
download
(
CIFAR100_URL
,
'cifar'
,
CIFAR100_MD5
),
'test'
)
def
train10
():
return
reader_creator
(
paddle
.
v2
.
dataset
.
common
.
download
(
CIFAR10_URL
,
'cifar'
,
CIFAR10_MD5
),
'data_batch'
)
download
(
CIFAR10_URL
,
'cifar'
,
CIFAR10_MD5
),
'data_batch'
)
def
test10
():
return
reader_creator
(
paddle
.
v2
.
dataset
.
common
.
download
(
CIFAR10_URL
,
'cifar'
,
CIFAR10_MD5
),
'test_batch'
)
download
(
CIFAR10_URL
,
'cifar'
,
CIFAR10_MD5
),
'test_batch'
)
def
fetch
_data
():
paddle
.
v2
.
dataset
.
common
.
download
(
CIFAR10_URL
,
'cifar'
,
CIFAR10_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
CIFAR100_URL
,
'cifar'
,
CIFAR100_MD5
)
def
fetch
():
download
(
CIFAR10_URL
,
'cifar'
,
CIFAR10_MD5
)
download
(
CIFAR100_URL
,
'cifar'
,
CIFAR100_MD5
)
python/paddle/v2/dataset/common.py
浏览文件 @
14eb5b8e
...
...
@@ -17,6 +17,8 @@ import hashlib
import
os
import
shutil
import
sys
import
importlib
import
paddle.v2.dataset
__all__
=
[
'DATA_HOME'
,
'download'
,
'md5file'
]
...
...
@@ -69,3 +71,13 @@ def dict_add(a_dict, ele):
a_dict
[
ele
]
+=
1
else
:
a_dict
[
ele
]
=
1
def
fetch_all
():
for
module_name
in
filter
(
lambda
x
:
not
x
.
startswith
(
"__"
),
dir
(
paddle
.
v2
.
dataset
)):
if
"fetch"
in
dir
(
importlib
.
import_module
(
"paddle.v2.dataset.%s"
%
module_name
)):
getattr
(
importlib
.
import_module
(
"paddle.v2.dataset.%s"
%
module_name
),
"fetch"
)()
python/paddle/v2/dataset/conll05.py
浏览文件 @
14eb5b8e
...
...
@@ -198,9 +198,9 @@ def test():
return
reader_creator
(
reader
,
word_dict
,
verb_dict
,
label_dict
)
def
fetch
_data
():
paddle
.
v2
.
dataset
.
common
.
download
(
WORDDICT_URL
,
'conll05st'
,
WORDDICT_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
VERBDICT_URL
,
'conll05st'
,
VERBDICT_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
TRGDICT_URL
,
'conll05st'
,
TRGDICT_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
EMB_URL
,
'conll05st'
,
EMB_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
DATA_URL
,
'conll05st'
,
DATA_MD5
)
def
fetch
():
download
(
WORDDICT_URL
,
'conll05st'
,
WORDDICT_MD5
)
download
(
VERBDICT_URL
,
'conll05st'
,
VERBDICT_MD5
)
download
(
TRGDICT_URL
,
'conll05st'
,
TRGDICT_MD5
)
download
(
EMB_URL
,
'conll05st'
,
EMB_MD5
)
download
(
DATA_URL
,
'conll05st'
,
DATA_MD5
)
python/paddle/v2/dataset/imdb.py
浏览文件 @
14eb5b8e
...
...
@@ -125,5 +125,5 @@ def word_dict():
re
.
compile
(
"aclImdb/((train)|(test))/((pos)|(neg))/.*\.txt$"
),
150
)
def
fetch
_data
():
def
fetch
():
paddle
.
v2
.
dataset
.
common
.
download
(
URL
,
'imdb'
,
MD5
)
python/paddle/v2/dataset/imikolov.py
浏览文件 @
14eb5b8e
...
...
@@ -91,5 +91,5 @@ def test(word_idx, n):
return
reader_creator
(
'./simple-examples/data/ptb.valid.txt'
,
word_idx
,
n
)
def
fetch
_data
():
def
fetch
():
paddle
.
v2
.
dataset
.
common
.
download
(
URL
,
"imikolov"
,
MD5
)
python/paddle/v2/dataset/mnist.py
浏览文件 @
14eb5b8e
...
...
@@ -108,6 +108,8 @@ def test():
TEST_LABEL_MD5
),
100
)
def
fetch
_data
():
def
fetch
():
paddle
.
v2
.
dataset
.
common
.
download
(
TRAIN_IMAGE_URL
,
'mnist'
,
TRAIN_IMAGE_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
TRAIN_LABEL_URL
,
'mnist'
,
TRAIN_LABEL_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
TEST_IMAGE_URL
,
'mnist'
,
TEST_IMAGE_MD5
)
paddle
.
v2
.
dataset
.
common
.
download
(
TEST_LABEL_URL
,
'mnist'
,
TRAIN_LABEL_MD5
)
python/paddle/v2/dataset/movielens.py
浏览文件 @
14eb5b8e
...
...
@@ -205,8 +205,8 @@ def unittest():
print
train_count
,
test_count
def
fetch
_data
():
paddle
.
v2
.
dataset
.
common
.
download
(
URL
,
"movielens"
,
MD5
)
def
fetch
():
download
(
URL
,
"movielens"
,
MD5
)
if
__name__
==
'__main__'
:
...
...
python/paddle/v2/dataset/sentiment.py
浏览文件 @
14eb5b8e
...
...
@@ -26,7 +26,7 @@ from itertools import chain
import
nltk
from
nltk.corpus
import
movie_reviews
import
paddle.v2.dataset.
common
import
common
__all__
=
[
'train'
,
'test'
,
'get_word_dict'
]
NUM_TRAINING_INSTANCES
=
1600
...
...
@@ -127,5 +127,5 @@ def test():
return
reader_creator
(
data_set
[
NUM_TRAINING_INSTANCES
:])
def
fetch
_data
():
def
fetch
():
nltk
.
download
(
'movie_reviews'
,
download_dir
=
common
.
DATA_HOME
)
python/paddle/v2/dataset/uci_housing.py
浏览文件 @
14eb5b8e
...
...
@@ -91,5 +91,5 @@ def test():
return
reader
def
fetch
_data
():
paddle
.
v2
.
dataset
.
common
.
download
(
URL
,
'uci_housing'
,
MD5
)
def
fetch
():
download
(
URL
,
'uci_housing'
,
MD5
)
python/paddle/v2/dataset/wmt14.py
浏览文件 @
14eb5b8e
...
...
@@ -16,7 +16,7 @@ wmt14 dataset
"""
import
tarfile
import
paddle.v2.dataset.common
from
paddle.v2.dataset.common
import
download
__all__
=
[
'train'
,
'test'
,
'build_dict'
]
...
...
@@ -95,15 +95,13 @@ def reader_creator(tar_file, file_name, dict_size):
def
train
(
dict_size
):
return
reader_creator
(
paddle
.
v2
.
dataset
.
common
.
download
(
URL_TRAIN
,
'wmt14'
,
MD5_TRAIN
),
'train/train'
,
dict_size
)
download
(
URL_TRAIN
,
'wmt14'
,
MD5_TRAIN
),
'train/train'
,
dict_size
)
def
test
(
dict_size
):
return
reader_creator
(
paddle
.
v2
.
dataset
.
common
.
download
(
URL_TRAIN
,
'wmt14'
,
MD5_TRAIN
),
'test/test'
,
dict_size
)
download
(
URL_TRAIN
,
'wmt14'
,
MD5_TRAIN
),
'test/test'
,
dict_size
)
def
fetch
_data
():
paddle
.
v2
.
dataset
.
common
.
download
(
URL_TRAIN
,
'wmt14'
,
MD5_TRAIN
)
def
fetch
():
download
(
URL_TRAIN
,
'wmt14'
,
MD5_TRAIN
)
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录