Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
s920243400
PaddleDetection
提交
a1ae96b2
P
PaddleDetection
项目概览
s920243400
/
PaddleDetection
与 Fork 源项目一致
Fork自
PaddlePaddle / PaddleDetection
通知
2
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleDetection
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
未验证
提交
a1ae96b2
编写于
5月 07, 2020
作者:
Y
Yang Zhang
提交者:
GitHub
5月 07, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Make logger less noisy (#583)
* Make logger less noisy * Fix indent
上级
c62e6145
变更
7
隐藏空白更改
内联
并排
Showing
7 changed file
with
26 addition
and
26 deletion
+26
-26
ppdet/data/parallel_map.py
ppdet/data/parallel_map.py
+2
-2
ppdet/data/reader.py
ppdet/data/reader.py
+2
-2
ppdet/data/source/coco.py
ppdet/data/source/coco.py
+1
-1
ppdet/data/source/voc.py
ppdet/data/source/voc.py
+1
-1
ppdet/data/source/widerface.py
ppdet/data/source/widerface.py
+1
-1
ppdet/utils/checkpoint.py
ppdet/utils/checkpoint.py
+2
-2
ppdet/utils/download.py
ppdet/utils/download.py
+17
-17
未找到文件。
ppdet/data/parallel_map.py
浏览文件 @
a1ae96b2
...
...
@@ -70,8 +70,8 @@ class ParallelMap(object):
self
.
_bufsize
=
bufsize
self
.
_use_process
=
use_process
if
self
.
_use_process
and
sys
.
platform
==
"win32"
:
logger
.
info
(
"Use multi-thread reader instead of "
"multi-process reader on Windows."
)
logger
.
debug
(
"Use multi-thread reader instead of "
"multi-process reader on Windows."
)
self
.
_use_process
=
False
if
self
.
_use_process
and
type
(
memsize
)
is
str
:
assert
memsize
[
-
1
].
lower
()
in
[
'g'
,
'm'
],
\
...
...
ppdet/data/reader.py
浏览文件 @
a1ae96b2
...
...
@@ -286,8 +286,8 @@ class Reader(object):
np
.
random
.
shuffle
(
self
.
indexes
)
if
self
.
_mixup_epoch
>
0
and
len
(
self
.
indexes
)
<
2
:
logger
.
info
(
"Disable mixup for dataset samples "
"less than 2 samples"
)
logger
.
debug
(
"Disable mixup for dataset samples "
"less than 2 samples"
)
self
.
_mixup_epoch
=
-
1
if
self
.
_epoch
<
0
:
...
...
ppdet/data/source/coco.py
浏览文件 @
a1ae96b2
...
...
@@ -154,5 +154,5 @@ class COCODataSet(DataSet):
if
self
.
sample_num
>
0
and
ct
>=
self
.
sample_num
:
break
assert
len
(
records
)
>
0
,
'not found any coco record in %s'
%
(
anno_path
)
logger
.
info
(
'{} samples in file {}'
.
format
(
ct
,
anno_path
))
logger
.
debug
(
'{} samples in file {}'
.
format
(
ct
,
anno_path
))
self
.
roidbs
,
self
.
cname2cid
=
records
,
cname2cid
ppdet/data/source/voc.py
浏览文件 @
a1ae96b2
...
...
@@ -160,7 +160,7 @@ class VOCDataSet(DataSet):
break
assert
len
(
records
)
>
0
,
'not found any voc record in %s'
%
(
self
.
anno_path
)
logger
.
info
(
'{} samples in file {}'
.
format
(
ct
,
anno_path
))
logger
.
debug
(
'{} samples in file {}'
.
format
(
ct
,
anno_path
))
self
.
roidbs
,
self
.
cname2cid
=
records
,
cname2cid
...
...
ppdet/data/source/widerface.py
浏览文件 @
a1ae96b2
...
...
@@ -102,7 +102,7 @@ class WIDERFaceDataSet(DataSet):
if
self
.
sample_num
>
0
and
ct
>=
self
.
sample_num
:
break
assert
len
(
records
)
>
0
,
'not found any widerface in %s'
%
(
anno_path
)
logger
.
info
(
'{} samples in file {}'
.
format
(
ct
,
anno_path
))
logger
.
debug
(
'{} samples in file {}'
.
format
(
ct
,
anno_path
))
self
.
roidbs
,
self
.
cname2cid
=
records
,
cname2cid
...
...
ppdet/utils/checkpoint.py
浏览文件 @
a1ae96b2
...
...
@@ -119,7 +119,7 @@ def load_params(exe, prog, path, ignore_params=[]):
raise
ValueError
(
"Model pretrain path {} does not "
"exists."
.
format
(
path
))
logger
.
info
(
'Loading parameters from {}...'
.
format
(
path
))
logger
.
debug
(
'Loading parameters from {}...'
.
format
(
path
))
ignore_set
=
set
()
state
=
_load_state
(
path
)
...
...
@@ -208,7 +208,7 @@ def load_and_fusebn(exe, prog, path):
prog (fluid.Program): save weight from which Program object.
path (string): the path to save model.
"""
logger
.
info
(
'Load model and fuse batch norm if have from {}...'
.
format
(
logger
.
debug
(
'Load model and fuse batch norm if have from {}...'
.
format
(
path
))
if
is_url
(
path
):
...
...
ppdet/utils/download.py
浏览文件 @
a1ae96b2
...
...
@@ -108,8 +108,8 @@ def get_dataset_path(path, annotation, image_dir):
data_name
=
os
.
path
.
split
(
path
.
strip
().
lower
())[
-
1
]
for
name
,
dataset
in
DATASETS
.
items
():
if
data_name
==
name
:
logger
.
info
(
"Parse dataset_dir {} as dataset "
"{}"
.
format
(
path
,
name
))
logger
.
debug
(
"Parse dataset_dir {} as dataset "
"{}"
.
format
(
path
,
name
))
if
name
==
'objects365'
:
raise
NotImplementedError
(
"Dataset {} is not valid for download automatically. "
...
...
@@ -146,7 +146,7 @@ def get_dataset_path(path, annotation, image_dir):
def
create_voc_list
(
data_dir
,
devkit_subdir
=
'VOCdevkit'
):
logger
.
info
(
"Create voc file list..."
)
logger
.
debug
(
"Create voc file list..."
)
devkit_dir
=
osp
.
join
(
data_dir
,
devkit_subdir
)
years
=
[
'2007'
,
'2012'
]
...
...
@@ -155,7 +155,7 @@ def create_voc_list(data_dir, devkit_subdir='VOCdevkit'):
# do not generate label_list.txt here. For default
# label, see ../data/source/voc.py
create_list
(
devkit_dir
,
years
,
data_dir
)
logger
.
info
(
"Create voc file list finished"
)
logger
.
debug
(
"Create voc file list finished"
)
def
map_path
(
url
,
root_dir
):
...
...
@@ -197,7 +197,7 @@ def get_path(url, root_dir, md5sum=None, check_exist=True):
exist_flag
=
False
if
osp
.
exists
(
fullpath
)
and
check_exist
:
exist_flag
=
True
logger
.
info
(
"Found {}"
.
format
(
fullpath
))
logger
.
debug
(
"Found {}"
.
format
(
fullpath
))
else
:
exist_flag
=
False
fullname
=
_download
(
url
,
root_dir
,
md5sum
)
...
...
@@ -218,7 +218,7 @@ def download_dataset(path, dataset=None):
dataset_info
=
DATASETS
[
dataset
][
0
]
for
info
in
dataset_info
:
get_path
(
info
[
0
],
path
,
info
[
1
],
False
)
logger
.
info
(
"Download dataset {} finished."
.
format
(
dataset
))
logger
.
debug
(
"Download dataset {} finished."
.
format
(
dataset
))
def
_dataset_exists
(
path
,
annotation
,
image_dir
):
...
...
@@ -226,23 +226,23 @@ def _dataset_exists(path, annotation, image_dir):
Check if user define dataset exists
"""
if
not
osp
.
exists
(
path
):
logger
.
info
(
"Config dataset_dir {} is not exits, "
"dataset config is not valid"
.
format
(
path
))
logger
.
debug
(
"Config dataset_dir {} is not exits, "
"dataset config is not valid"
.
format
(
path
))
return
False
if
annotation
:
annotation_path
=
osp
.
join
(
path
,
annotation
)
if
not
osp
.
isfile
(
annotation_path
):
logger
.
info
(
"Config annotation {} is not a "
"file, dataset config is not "
"valid"
.
format
(
annotation_path
))
logger
.
debug
(
"Config annotation {} is not a "
"file, dataset config is not "
"valid"
.
format
(
annotation_path
))
return
False
if
image_dir
:
image_path
=
osp
.
join
(
path
,
image_dir
)
if
not
osp
.
isdir
(
image_path
):
logger
.
info
(
"Config image_dir {} is not a "
"directory, dataset config is not "
"valid"
.
format
(
image_path
))
logger
.
warning
(
"Config image_dir {} is not a "
"directory, dataset config is not "
"valid"
.
format
(
image_path
))
return
False
return
True
...
...
@@ -300,7 +300,7 @@ def _md5check(fullname, md5sum=None):
if
md5sum
is
None
:
return
True
logger
.
info
(
"File {} md5 checking..."
.
format
(
fullname
))
logger
.
debug
(
"File {} md5 checking..."
.
format
(
fullname
))
md5
=
hashlib
.
md5
()
with
open
(
fullname
,
'rb'
)
as
f
:
for
chunk
in
iter
(
lambda
:
f
.
read
(
4096
),
b
""
):
...
...
@@ -308,8 +308,8 @@ def _md5check(fullname, md5sum=None):
calc_md5sum
=
md5
.
hexdigest
()
if
calc_md5sum
!=
md5sum
:
logger
.
info
(
"File {} md5 check failed, {}(calc) != "
"{}(base)"
.
format
(
fullname
,
calc_md5sum
,
md5sum
))
logger
.
warning
(
"File {} md5 check failed, {}(calc) != "
"{}(base)"
.
format
(
fullname
,
calc_md5sum
,
md5sum
))
return
False
return
True
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录