Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
曾经的那一瞬间
Models
提交
595dbabe
M
Models
项目概览
曾经的那一瞬间
/
Models
10 个月 前同步成功
通知
1
Star
0
Fork
0
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
0
列表
看板
标记
里程碑
合并请求
0
DevOps
流水线
流水线任务
计划
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
M
Models
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
0
Issue
0
列表
看板
标记
里程碑
合并请求
0
合并请求
0
Pages
DevOps
DevOps
流水线
流水线任务
计划
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
流水线任务
提交
Issue看板
前往新版Gitcode,体验更适合开发者的 AI 搜索 >>
提交
595dbabe
编写于
7月 11, 2023
作者:
Y
Yilei Yang
提交者:
A. Unique TensorFlower
7月 11, 2023
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
Internal change
PiperOrigin-RevId: 547338367
上级
843bdf24
变更
10
隐藏空白更改
内联
并排
Showing
10 changed file
with
128 addition
and
58 deletion
+128
-58
official/nlp/tasks/masked_lm.py
official/nlp/tasks/masked_lm.py
+16
-6
official/projects/mosaic/configs/mosaic_config.py
official/projects/mosaic/configs/mosaic_config.py
+28
-12
official/projects/pix2seq/configs/pix2seq.py
official/projects/pix2seq/configs/pix2seq.py
+17
-8
official/projects/s3d/configs/s3d.py
official/projects/s3d/configs/s3d.py
+2
-2
official/projects/simclr/configs/multitask_config.py
official/projects/simclr/configs/multitask_config.py
+18
-8
official/projects/teams/teams.py
official/projects/teams/teams.py
+6
-2
official/projects/teams/teams_task.py
official/projects/teams/teams_task.py
+7
-3
official/projects/text_classification_example/classification_example.py
...cts/text_classification_example/classification_example.py
+7
-3
official/projects/unified_detector/configs/ocr_config.py
official/projects/unified_detector/configs/ocr_config.py
+1
-1
official/projects/video_ssl/configs/video_ssl.py
official/projects/video_ssl/configs/video_ssl.py
+26
-13
未找到文件。
official/nlp/tasks/masked_lm.py
浏览文件 @
595dbabe
...
...
@@ -31,15 +31,25 @@ from official.nlp.modeling import models
@
dataclasses
.
dataclass
class
MaskedLMConfig
(
cfg
.
TaskConfig
):
"""The model config."""
model
:
bert
.
PretrainerConfig
=
bert
.
PretrainerConfig
(
cls_heads
=
[
bert
.
ClsHeadConfig
(
inner_dim
=
768
,
num_classes
=
2
,
dropout_rate
=
0.1
,
name
=
'next_sentence'
)
])
model
:
bert
.
PretrainerConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
bert
.
PretrainerConfig
(
# pylint: disable=g-long-lambda
cls_heads
=
[
bert
.
ClsHeadConfig
(
inner_dim
=
768
,
num_classes
=
2
,
dropout_rate
=
0.1
,
name
=
'next_sentence'
,
)
]
)
)
# TODO(b/154564893): Mathematically, scale_loss should be True.
# However, it works better with scale_loss being False.
scale_loss
:
bool
=
False
train_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
()
validation_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
()
train_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
validation_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
@
task_factory
.
register_task_cls
(
MaskedLMConfig
)
...
...
official/projects/mosaic/configs/mosaic_config.py
浏览文件 @
595dbabe
...
...
@@ -64,23 +64,37 @@ class MosaicSemanticSegmentationModel(hyperparams.Config):
"""MOSAIC semantic segmentation model config."""
num_classes
:
int
=
19
input_size
:
List
[
int
]
=
dataclasses
.
field
(
default_factory
=
list
)
head
:
MosaicDecoderHead
=
MosaicDecoderHead
()
backbone
:
backbones
.
Backbone
=
backbones
.
Backbone
(
type
=
'mobilenet'
,
mobilenet
=
backbones
.
MobileNet
())
neck
:
MosaicEncoderNeck
=
MosaicEncoderNeck
()
head
:
MosaicDecoderHead
=
dataclasses
.
field
(
default_factory
=
MosaicDecoderHead
)
backbone
:
backbones
.
Backbone
=
dataclasses
.
field
(
default_factory
=
lambda
:
backbones
.
Backbone
(
# pylint: disable=g-long-lambda
type
=
'mobilenet'
,
mobilenet
=
backbones
.
MobileNet
()
)
)
neck
:
MosaicEncoderNeck
=
dataclasses
.
field
(
default_factory
=
MosaicEncoderNeck
)
mask_scoring_head
:
Optional
[
seg_cfg
.
MaskScoringHead
]
=
None
norm_activation
:
common
.
NormActivation
=
common
.
NormActivation
(
use_sync_bn
=
True
,
norm_momentum
=
0.99
,
norm_epsilon
=
0.001
)
norm_activation
:
common
.
NormActivation
=
dataclasses
.
field
(
default_factory
=
lambda
:
common
.
NormActivation
(
# pylint: disable=g-long-lambda
use_sync_bn
=
True
,
norm_momentum
=
0.99
,
norm_epsilon
=
0.001
)
)
@
dataclasses
.
dataclass
class
MosaicSemanticSegmentationTask
(
seg_cfg
.
SemanticSegmentationTask
):
"""The config for MOSAIC segmentation task."""
model
:
MosaicSemanticSegmentationModel
=
MosaicSemanticSegmentationModel
()
train_data
:
seg_cfg
.
DataConfig
=
seg_cfg
.
DataConfig
(
is_training
=
True
)
validation_data
:
seg_cfg
.
DataConfig
=
seg_cfg
.
DataConfig
(
is_training
=
False
)
losses
:
seg_cfg
.
Losses
=
seg_cfg
.
Losses
()
evaluation
:
seg_cfg
.
Evaluation
=
seg_cfg
.
Evaluation
()
model
:
MosaicSemanticSegmentationModel
=
dataclasses
.
field
(
default_factory
=
MosaicSemanticSegmentationModel
)
train_data
:
seg_cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
seg_cfg
.
DataConfig
(
is_training
=
True
)
)
validation_data
:
seg_cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
seg_cfg
.
DataConfig
(
is_training
=
False
)
)
losses
:
seg_cfg
.
Losses
=
dataclasses
.
field
(
default_factory
=
seg_cfg
.
Losses
)
evaluation
:
seg_cfg
.
Evaluation
=
dataclasses
.
field
(
default_factory
=
seg_cfg
.
Evaluation
)
train_input_partition_dims
:
List
[
int
]
=
dataclasses
.
field
(
default_factory
=
list
)
eval_input_partition_dims
:
List
[
int
]
=
dataclasses
.
field
(
...
...
@@ -88,7 +102,9 @@ class MosaicSemanticSegmentationTask(seg_cfg.SemanticSegmentationTask):
init_checkpoint
:
Optional
[
str
]
=
None
init_checkpoint_modules
:
Union
[
str
,
List
[
str
]]
=
'all'
# all, backbone, and/or neck.
export_config
:
seg_cfg
.
ExportConfig
=
seg_cfg
.
ExportConfig
()
export_config
:
seg_cfg
.
ExportConfig
=
dataclasses
.
field
(
default_factory
=
seg_cfg
.
ExportConfig
)
# Cityscapes Dataset (Download and process the dataset yourself)
...
...
official/projects/pix2seq/configs/pix2seq.py
浏览文件 @
595dbabe
...
...
@@ -67,7 +67,9 @@ class DataConfig(cfg.DataConfig):
global_batch_size
:
int
=
0
is_training
:
bool
=
False
dtype
:
str
=
'float32'
decoder
:
common
.
DataDecoder
=
common
.
DataDecoder
()
decoder
:
common
.
DataDecoder
=
dataclasses
.
field
(
default_factory
=
common
.
DataDecoder
)
shuffle_buffer_size
:
int
=
10000
file_type
:
str
=
'tfrecord'
drop_remainder
:
bool
=
True
...
...
@@ -97,10 +99,15 @@ class Pix2Seq(hyperparams.Config):
shared_decoder_embedding
:
bool
=
True
decoder_output_bias
:
bool
=
True
input_size
:
List
[
int
]
=
dataclasses
.
field
(
default_factory
=
list
)
backbone
:
backbones
.
Backbone
=
backbones
.
Backbone
(
type
=
'resnet'
,
resnet
=
backbones
.
ResNet
(
model_id
=
50
,
bn_trainable
=
False
)
backbone
:
backbones
.
Backbone
=
dataclasses
.
field
(
default_factory
=
lambda
:
backbones
.
Backbone
(
# pylint: disable=g-long-lambda
type
=
'resnet'
,
resnet
=
backbones
.
ResNet
(
model_id
=
50
,
bn_trainable
=
False
),
)
)
norm_activation
:
common
.
NormActivation
=
dataclasses
.
field
(
default_factory
=
common
.
NormActivation
)
norm_activation
:
common
.
NormActivation
=
common
.
NormActivation
()
backbone_endpoint_name
:
str
=
'5'
drop_path
:
float
=
0.1
drop_units
:
float
=
0.1
...
...
@@ -110,10 +117,12 @@ class Pix2Seq(hyperparams.Config):
@
dataclasses
.
dataclass
class
Pix2SeqTask
(
cfg
.
TaskConfig
):
model
:
Pix2Seq
=
Pix2Seq
()
train_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
()
validation_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
()
losses
:
Losses
=
Losses
()
model
:
Pix2Seq
=
dataclasses
.
field
(
default_factory
=
Pix2Seq
)
train_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
validation_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
losses
:
Losses
=
dataclasses
.
field
(
default_factory
=
Losses
)
init_checkpoint
:
Optional
[
str
]
=
None
init_checkpoint_modules
:
Union
[
str
,
List
[
str
]]
=
'all'
# all, backbone
annotation_file
:
Optional
[
str
]
=
None
...
...
official/projects/s3d/configs/s3d.py
浏览文件 @
595dbabe
...
...
@@ -83,7 +83,7 @@ class Backbone3D(backbones_3d.Backbone3D):
s3d: s3d backbone config.
"""
type
:
str
=
's3d'
s3d
:
S3D
=
S3D
(
)
s3d
:
S3D
=
dataclasses
.
field
(
default_factory
=
S3D
)
@
dataclasses
.
dataclass
...
...
@@ -95,4 +95,4 @@ class S3DModel(video_classification.VideoClassificationModel):
backbone: backbone config.
"""
model_type
:
str
=
's3d'
backbone
:
Backbone3D
=
Backbone3D
(
)
backbone
:
Backbone3D
=
dataclasses
.
field
(
default_factory
=
Backbone3D
)
official/projects/simclr/configs/multitask_config.py
浏览文件 @
595dbabe
...
...
@@ -31,8 +31,9 @@ class SimCLRMTHeadConfig(hyperparams.Config):
"""Per-task specific configs."""
task_name
:
str
=
'task_name'
# Supervised head is required for finetune, but optional for pretrain.
supervised_head
:
simclr_configs
.
SupervisedHead
=
simclr_configs
.
SupervisedHead
(
num_classes
=
1001
)
supervised_head
:
simclr_configs
.
SupervisedHead
=
dataclasses
.
field
(
default_factory
=
lambda
:
simclr_configs
.
SupervisedHead
(
num_classes
=
1001
)
)
mode
:
str
=
simclr_model
.
PRETRAIN
...
...
@@ -40,13 +41,22 @@ class SimCLRMTHeadConfig(hyperparams.Config):
class
SimCLRMTModelConfig
(
hyperparams
.
Config
):
"""Model config for multi-task SimCLR model."""
input_size
:
List
[
int
]
=
dataclasses
.
field
(
default_factory
=
list
)
backbone
:
backbones
.
Backbone
=
backbones
.
Backbone
(
type
=
'resnet'
,
resnet
=
backbones
.
ResNet
())
backbone
:
backbones
.
Backbone
=
dataclasses
.
field
(
default_factory
=
lambda
:
backbones
.
Backbone
(
# pylint: disable=g-long-lambda
type
=
'resnet'
,
resnet
=
backbones
.
ResNet
()
)
)
backbone_trainable
:
bool
=
True
projection_head
:
simclr_configs
.
ProjectionHead
=
simclr_configs
.
ProjectionHead
(
proj_output_dim
=
128
,
num_proj_layers
=
3
,
ft_proj_idx
=
1
)
norm_activation
:
common
.
NormActivation
=
common
.
NormActivation
(
norm_momentum
=
0.9
,
norm_epsilon
=
1e-5
,
use_sync_bn
=
False
)
projection_head
:
simclr_configs
.
ProjectionHead
=
dataclasses
.
field
(
default_factory
=
lambda
:
simclr_configs
.
ProjectionHead
(
# pylint: disable=g-long-lambda
proj_output_dim
=
128
,
num_proj_layers
=
3
,
ft_proj_idx
=
1
)
)
norm_activation
:
common
.
NormActivation
=
dataclasses
.
field
(
default_factory
=
lambda
:
common
.
NormActivation
(
# pylint: disable=g-long-lambda
norm_momentum
=
0.9
,
norm_epsilon
=
1e-5
,
use_sync_bn
=
False
)
)
heads
:
Tuple
[
SimCLRMTHeadConfig
,
...]
=
()
# L2 weight decay is used in the model, not in task.
# Note that this can not be used together with lars optimizer.
...
...
official/projects/teams/teams.py
浏览文件 @
595dbabe
...
...
@@ -43,8 +43,12 @@ class TeamsPretrainerConfig(base_config.Config):
num_shared_generator_hidden_layers
:
int
=
3
# Number of bottom layers shared between different discriminator tasks.
num_discriminator_task_agnostic_layers
:
int
=
11
generator
:
encoders
.
BertEncoderConfig
=
encoders
.
BertEncoderConfig
()
discriminator
:
encoders
.
BertEncoderConfig
=
encoders
.
BertEncoderConfig
()
generator
:
encoders
.
BertEncoderConfig
=
dataclasses
.
field
(
default_factory
=
encoders
.
BertEncoderConfig
)
discriminator
:
encoders
.
BertEncoderConfig
=
dataclasses
.
field
(
default_factory
=
encoders
.
BertEncoderConfig
)
class
TeamsEncoderConfig
(
encoders
.
BertEncoderConfig
):
...
...
official/projects/teams/teams_task.py
浏览文件 @
595dbabe
...
...
@@ -30,9 +30,13 @@ from official.projects.teams import teams_pretrainer
@
dataclasses
.
dataclass
class
TeamsPretrainTaskConfig
(
cfg
.
TaskConfig
):
"""The model config."""
model
:
teams
.
TeamsPretrainerConfig
=
teams
.
TeamsPretrainerConfig
()
train_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
()
validation_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
()
model
:
teams
.
TeamsPretrainerConfig
=
dataclasses
.
field
(
default_factory
=
teams
.
TeamsPretrainerConfig
)
train_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
validation_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
def
_get_generator_hidden_layers
(
discriminator_network
,
num_hidden_layers
,
...
...
official/projects/text_classification_example/classification_example.py
浏览文件 @
595dbabe
...
...
@@ -34,7 +34,9 @@ from official.projects.text_classification_example import classification_data_lo
@
dataclasses
.
dataclass
class
ModelConfig
(
base_config
.
Config
):
"""A base span labeler configuration."""
encoder
:
encoders
.
EncoderConfig
=
encoders
.
EncoderConfig
()
encoder
:
encoders
.
EncoderConfig
=
dataclasses
.
field
(
default_factory
=
encoders
.
EncoderConfig
)
head_dropout
:
float
=
0.1
head_initializer_range
:
float
=
0.02
...
...
@@ -49,9 +51,11 @@ class ClassificationExampleConfig(cfg.TaskConfig):
num_classes
=
2
class_names
=
[
'A'
,
'B'
]
train_data
:
cfg
.
DataConfig
=
classification_data_loader
.
ClassificationExampleDataConfig
(
train_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
classification_data_loader
.
ClassificationExampleDataConfig
)
validation_data
:
cfg
.
DataConfig
=
classification_data_loader
.
ClassificationExampleDataConfig
(
validation_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
classification_data_loader
.
ClassificationExampleDataConfig
)
...
...
official/projects/unified_detector/configs/ocr_config.py
浏览文件 @
595dbabe
...
...
@@ -22,7 +22,7 @@ from official.modeling import optimization
@
dataclasses
.
dataclass
class
OcrTaskConfig
(
cfg
.
TaskConfig
):
train_data
:
cfg
.
DataConfig
=
cfg
.
DataConfig
(
)
train_data
:
cfg
.
DataConfig
=
dataclasses
.
field
(
default_factory
=
cfg
.
DataConfig
)
model_call_needs_labels
:
bool
=
False
...
...
official/projects/video_ssl/configs/video_ssl.py
浏览文件 @
595dbabe
...
...
@@ -43,8 +43,11 @@ class VideoSSLModel(VideoClassificationModel):
hidden_dim
:
int
=
2048
hidden_layer_num
:
int
=
3
projection_dim
:
int
=
128
hidden_norm_activation
:
common
.
NormActivation
=
common
.
NormActivation
(
use_sync_bn
=
False
,
norm_momentum
=
0.997
,
norm_epsilon
=
1.0e-05
)
hidden_norm_activation
:
common
.
NormActivation
=
dataclasses
.
field
(
default_factory
=
lambda
:
common
.
NormActivation
(
use_sync_bn
=
False
,
norm_momentum
=
0.997
,
norm_epsilon
=
1.0e-05
)
)
@
dataclasses
.
dataclass
...
...
@@ -55,21 +58,31 @@ class SSLLosses(Losses):
@
dataclasses
.
dataclass
class
VideoSSLPretrainTask
(
VideoClassificationTask
):
model
:
VideoSSLModel
=
VideoSSLModel
()
losses
:
SSLLosses
=
SSLLosses
()
train_data
:
DataConfig
=
DataConfig
(
is_training
=
True
,
drop_remainder
=
True
)
validation_data
:
DataConfig
=
DataConfig
(
is_training
=
False
,
drop_remainder
=
False
)
losses
:
SSLLosses
=
SSLLosses
()
model
:
VideoSSLModel
=
dataclasses
.
field
(
default_factory
=
VideoSSLModel
)
losses
:
SSLLosses
=
dataclasses
.
field
(
default_factory
=
SSLLosses
)
train_data
:
DataConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
DataConfig
(
is_training
=
True
,
drop_remainder
=
True
)
)
validation_data
:
DataConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
DataConfig
(
# pylint: disable=g-long-lambda
is_training
=
False
,
drop_remainder
=
False
)
)
losses
:
SSLLosses
=
dataclasses
.
field
(
default_factory
=
SSLLosses
)
@
dataclasses
.
dataclass
class
VideoSSLEvalTask
(
VideoClassificationTask
):
model
:
VideoSSLModel
=
VideoSSLModel
()
train_data
:
DataConfig
=
DataConfig
(
is_training
=
True
,
drop_remainder
=
True
)
validation_data
:
DataConfig
=
DataConfig
(
is_training
=
False
,
drop_remainder
=
False
)
losses
:
SSLLosses
=
SSLLosses
()
model
:
VideoSSLModel
=
dataclasses
.
field
(
default_factory
=
VideoSSLModel
)
train_data
:
DataConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
DataConfig
(
is_training
=
True
,
drop_remainder
=
True
)
)
validation_data
:
DataConfig
=
dataclasses
.
field
(
default_factory
=
lambda
:
DataConfig
(
# pylint: disable=g-long-lambda
is_training
=
False
,
drop_remainder
=
False
)
)
losses
:
SSLLosses
=
dataclasses
.
field
(
default_factory
=
SSLLosses
)
@
exp_factory
.
register_config_factory
(
'video_ssl_pretrain_kinetics400'
)
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录