Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
weixin_45480109
yolov3
提交
4a072808
Y
yolov3
项目概览
weixin_45480109
/
yolov3
与 Fork 源项目一致
从无法访问的项目Fork
通知
2
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
Y
yolov3
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
未验证
提交
4a072808
编写于
12月 06, 2020
作者:
G
Glenn Jocher
提交者:
GitHub
12月 06, 2020
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Pycocotools best.pt after COCO train (#1593)
上级
adc49abc
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
25 addition
and
15 deletion
+25
-15
test.py
test.py
+2
-3
train.py
train.py
+22
-11
utils/google_utils.py
utils/google_utils.py
+1
-1
未找到文件。
test.py
浏览文件 @
4a072808
import
argparse
import
glob
import
json
import
os
from
pathlib
import
Path
...
...
@@ -246,7 +245,7 @@ def test(data,
# Save JSON
if
save_json
and
len
(
jdict
):
w
=
Path
(
weights
[
0
]
if
isinstance
(
weights
,
list
)
else
weights
).
stem
if
weights
is
not
None
else
''
# weights
anno_json
=
glob
.
glob
(
'../coco/annotations/instances_val*.json'
)[
0
]
# annotations json
anno_json
=
'../coco/annotations/instances_val2017.json'
# annotations json
pred_json
=
str
(
save_dir
/
f
"
{
w
}
_predictions.json"
)
# predictions json
print
(
'
\n
Evaluating pycocotools mAP... saving %s...'
%
pred_json
)
with
open
(
pred_json
,
'w'
)
as
f
:
...
...
@@ -266,7 +265,7 @@ def test(data,
eval
.
summarize
()
map
,
map50
=
eval
.
stats
[:
2
]
# update results (mAP@0.5:0.95, mAP@0.5)
except
Exception
as
e
:
print
(
'ERROR: pycocotools unable to run: %s'
%
e
)
print
(
f
'pycocotools unable to run:
{
e
}
'
)
# Return results
if
not
training
:
...
...
train.py
浏览文件 @
4a072808
...
...
@@ -22,6 +22,7 @@ from torch.utils.tensorboard import SummaryWriter
from
tqdm
import
tqdm
import
test
# import test.py to get mAP after each epoch
from
models.experimental
import
attempt_load
from
models.yolo
import
Model
from
utils.autoanchor
import
check_anchors
from
utils.datasets
import
create_dataloader
...
...
@@ -193,9 +194,9 @@ def train(hyp, opt, device, tb_writer=None, wandb=None):
# Process 0
if
rank
in
[
-
1
,
0
]:
ema
.
updates
=
start_epoch
*
nb
//
accumulate
# set EMA updates
testloader
=
create_dataloader
(
test_path
,
imgsz_test
,
total_batch_size
,
gs
,
opt
,
testloader
=
create_dataloader
(
test_path
,
imgsz_test
,
total_batch_size
,
gs
,
opt
,
# testloader
hyp
=
hyp
,
cache
=
opt
.
cache_images
and
not
opt
.
notest
,
rect
=
True
,
rank
=-
1
,
world_size
=
opt
.
world_size
,
workers
=
opt
.
workers
)[
0
]
# testloader
rank
=-
1
,
world_size
=
opt
.
world_size
,
workers
=
opt
.
workers
,
pad
=
0.5
)[
0
]
if
not
opt
.
resume
:
labels
=
np
.
concatenate
(
dataset
.
labels
,
0
)
...
...
@@ -385,15 +386,12 @@ def train(hyp, opt, device, tb_writer=None, wandb=None):
if
rank
in
[
-
1
,
0
]:
# Strip optimizers
n
=
opt
.
name
if
opt
.
name
.
isnumeric
()
else
''
fresults
,
flast
,
fbest
=
save_dir
/
f
'results
{
n
}
.txt'
,
wdir
/
f
'last
{
n
}
.pt'
,
wdir
/
f
'best
{
n
}
.pt'
for
f1
,
f2
in
zip
([
wdir
/
'last.pt'
,
wdir
/
'best.pt'
,
results_file
],
[
flast
,
fbest
,
fresults
]):
if
f1
.
exists
():
os
.
rename
(
f1
,
f2
)
# rename
if
str
(
f2
).
endswith
(
'.pt'
):
# is *.pt
strip_optimizer
(
f2
)
# strip optimizer
os
.
system
(
'gsutil cp %s gs://%s/weights'
%
(
f2
,
opt
.
bucket
))
if
opt
.
bucket
else
None
# upload
# Finish
for
f
in
[
last
,
best
]:
if
f
.
exists
():
# is *.pt
strip_optimizer
(
f
)
# strip optimizer
os
.
system
(
'gsutil cp %s gs://%s/weights'
%
(
f
,
opt
.
bucket
))
if
opt
.
bucket
else
None
# upload
# Plots
if
plots
:
plot_results
(
save_dir
=
save_dir
)
# save as results.png
if
wandb
:
...
...
@@ -401,6 +399,19 @@ def train(hyp, opt, device, tb_writer=None, wandb=None):
wandb
.
log
({
"Results"
:
[
wandb
.
Image
(
str
(
save_dir
/
f
),
caption
=
f
)
for
f
in
files
if
(
save_dir
/
f
).
exists
()]})
logger
.
info
(
'%g epochs completed in %.3f hours.
\n
'
%
(
epoch
-
start_epoch
+
1
,
(
time
.
time
()
-
t0
)
/
3600
))
# Test best.pt
if
opt
.
data
.
endswith
(
'coco.yaml'
)
and
nc
==
80
:
# if COCO
results
,
_
,
_
=
test
.
test
(
opt
.
data
,
batch_size
=
total_batch_size
,
imgsz
=
imgsz_test
,
model
=
attempt_load
(
best
if
best
.
exists
()
else
last
,
device
).
half
(),
single_cls
=
opt
.
single_cls
,
dataloader
=
testloader
,
save_dir
=
save_dir
,
save_json
=
True
,
# use pycocotools
plots
=
False
)
else
:
dist
.
destroy_process_group
()
...
...
utils/google_utils.py
浏览文件 @
4a072808
...
...
@@ -17,7 +17,7 @@ def gsutil_getsize(url=''):
def
attempt_download
(
weights
):
# Attempt to download pretrained weights if not found locally
weights
=
weights
.
strip
().
replace
(
"'"
,
''
)
weights
=
str
(
weights
)
.
strip
().
replace
(
"'"
,
''
)
file
=
Path
(
weights
).
name
.
lower
()
msg
=
weights
+
' missing, try downloading from https://github.com/ultralytics/yolov3/releases/'
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录