Skip to content
体验新版
项目
组织
正在加载...
登录
切换导航
打开侧边栏
PaddlePaddle
PaddleClas
提交
23f5af9f
P
PaddleClas
项目概览
PaddlePaddle
/
PaddleClas
大约 1 年 前同步成功
通知
115
Star
4999
Fork
1114
代码
文件
提交
分支
Tags
贡献者
分支图
Diff
Issue
19
列表
看板
标记
里程碑
合并请求
6
Wiki
0
Wiki
分析
仓库
DevOps
项目成员
Pages
P
PaddleClas
项目概览
项目概览
详情
发布
仓库
仓库
文件
提交
分支
标签
贡献者
分支图
比较
Issue
19
Issue
19
列表
看板
标记
里程碑
合并请求
6
合并请求
6
Pages
分析
分析
仓库分析
DevOps
Wiki
0
Wiki
成员
成员
收起侧边栏
关闭侧边栏
动态
分支图
创建新Issue
提交
Issue看板
提交
23f5af9f
编写于
2月 07, 2023
作者:
Z
zh-hike
提交者:
zengshao0622
2月 07, 2023
浏览文件
操作
浏览文件
下载
电子邮件补丁
差异文件
add field vit to foundationvit's name
上级
dec7b024
变更
3
隐藏空白更改
内联
并排
Showing
3 changed file
with
40 addition
and
40 deletion
+40
-40
ppcls/arch/backbone/__init__.py
ppcls/arch/backbone/__init__.py
+1
-1
ppcls/arch/backbone/model_zoo/foundation_vit.py
ppcls/arch/backbone/model_zoo/foundation_vit.py
+38
-38
ppcls/arch/backbone/variant_models/foundation_vit_variant.py
ppcls/arch/backbone/variant_models/foundation_vit_variant.py
+1
-1
未找到文件。
ppcls/arch/backbone/__init__.py
浏览文件 @
23f5af9f
...
@@ -71,7 +71,7 @@ from .model_zoo.mobilevit import MobileViT_XXS, MobileViT_XS, MobileViT_S
...
@@ -71,7 +71,7 @@ from .model_zoo.mobilevit import MobileViT_XXS, MobileViT_XS, MobileViT_S
from
.model_zoo.repvgg
import
RepVGG_A0
,
RepVGG_A1
,
RepVGG_A2
,
RepVGG_B0
,
RepVGG_B1
,
RepVGG_B2
,
RepVGG_B1g2
,
RepVGG_B1g4
,
RepVGG_B2g4
,
RepVGG_B3
,
RepVGG_B3g4
,
RepVGG_D2se
from
.model_zoo.repvgg
import
RepVGG_A0
,
RepVGG_A1
,
RepVGG_A2
,
RepVGG_B0
,
RepVGG_B1
,
RepVGG_B2
,
RepVGG_B1g2
,
RepVGG_B1g4
,
RepVGG_B2g4
,
RepVGG_B3
,
RepVGG_B3g4
,
RepVGG_D2se
from
.model_zoo.van
import
VAN_B0
,
VAN_B1
,
VAN_B2
,
VAN_B3
from
.model_zoo.van
import
VAN_B0
,
VAN_B1
,
VAN_B2
,
VAN_B3
from
.model_zoo.peleenet
import
PeleeNet
from
.model_zoo.peleenet
import
PeleeNet
from
.model_zoo.foundation_vit
import
CLIP_
base_patch32_224
,
CLIP_base_patch16_224
,
CLIP_large_patch14_336
,
CLIP_large_patch14_224
,
BEiTv2_base_patch16_224
,
BEiTv2_large_patch16_224
,
CAE_base_patch16_224
,
EVA_huge_patch14
,
MOCOV3_small
,
MOCOV3_base
,
MAE_huge_patch14
,
MAE_large_patch16
,
MAE
_base_patch16
from
.model_zoo.foundation_vit
import
CLIP_
vit_base_patch32_224
,
CLIP_vit_base_patch16_224
,
CLIP_vit_large_patch14_336
,
CLIP_vit_large_patch14_224
,
BEiTv2_vit_base_patch16_224
,
BEiTv2_vit_large_patch16_224
,
CAE_vit_base_patch16_224
,
EVA_vit_huge_patch14
,
MOCOV3_vit_small
,
MOCOV3_vit_base
,
MAE_vit_huge_patch14
,
MAE_vit_large_patch16
,
MAE_vit
_base_patch16
from
.model_zoo.convnext
import
ConvNeXt_tiny
,
ConvNeXt_small
,
ConvNeXt_base_224
,
ConvNeXt_base_384
,
ConvNeXt_large_224
,
ConvNeXt_large_384
from
.model_zoo.convnext
import
ConvNeXt_tiny
,
ConvNeXt_small
,
ConvNeXt_base_224
,
ConvNeXt_base_384
,
ConvNeXt_large_224
,
ConvNeXt_large_384
from
.model_zoo.nextvit
import
NextViT_small_224
,
NextViT_base_224
,
NextViT_large_224
,
NextViT_small_384
,
NextViT_base_384
,
NextViT_large_384
from
.model_zoo.nextvit
import
NextViT_small_224
,
NextViT_base_224
,
NextViT_large_224
,
NextViT_small_384
,
NextViT_base_384
,
NextViT_large_384
from
.model_zoo.cae
import
cae_base_patch16_224
,
cae_large_patch16_224
from
.model_zoo.cae
import
cae_base_patch16_224
,
cae_large_patch16_224
...
...
ppcls/arch/backbone/model_zoo/foundation_vit.py
浏览文件 @
23f5af9f
...
@@ -26,19 +26,19 @@ from paddle.nn.initializer import TruncatedNormal, Constant, Normal
...
@@ -26,19 +26,19 @@ from paddle.nn.initializer import TruncatedNormal, Constant, Normal
from
....utils.save_load
import
load_dygraph_pretrain
,
load_dygraph_pretrain_from_url
from
....utils.save_load
import
load_dygraph_pretrain
,
load_dygraph_pretrain_from_url
MODEL_URLS
=
{
MODEL_URLS
=
{
"CLIP_
base_patch32_224"
:
None
,
"CLIP_
vit_base_patch32_224"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/CLIP_vit_base_patch32_224.pdparams"
,
"CLIP_
base_patch16_224"
:
None
,
"CLIP_
vit_base_patch16_224"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/CLIP_vit_base_patch16_224.pdparams"
,
"CLIP_
large_patch14_336"
:
None
,
"CLIP_
vit_large_patch14_336"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/CLIP_vit_large_patch14_336.pdparams"
,
"CLIP_
large_patch14_224"
:
None
,
"CLIP_
vit_large_patch14_224"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/CLIP_vit_large_patch14_224.pdparams"
,
"BEiTv2_
base_patch16_224"
:
None
,
"BEiTv2_
vit_base_patch16_224"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/BEiTv2_vit_base_patch16_224.pdparams"
,
"BEiTv2_
large_patch16_224"
:
None
,
"BEiTv2_
vit_large_patch16_224"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/BEiTv2_vit_large_patch16_224.pdparams"
,
"CAE_
base_patch16_224"
:
None
,
"CAE_
vit_base_patch16_224"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/CAE_vit_base_patch16_224.pdparams"
,
'EVA_
huge_patch14'
:
None
,
'EVA_
vit_huge_patch14'
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/EVA_vit_huge_patch14.pdparams"
,
"MOCOV3_
small"
:
None
,
"MOCOV3_
vit_small"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/MOCOV3_vit_small.pdparams"
,
"MOCOV3_
base"
:
None
,
"MOCOV3_
vit_base"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/MOCOV3_vit_base.pdparams"
,
"MAE_
huge_patch14"
:
None
,
"MAE_
vit_huge_patch14"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/MAE_vit_huge_patch14.pdparams"
,
"MAE_
large_patch16"
:
None
,
"MAE_
vit_large_patch16"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/MAE_vit_large_patch16.pdparams"
,
"MAE_
base_patch16"
:
None
"MAE_
vit_base_patch16"
:
"https://paddle-imagenet-models-name.bj.bcebos.com/dygraph/foundation_models/MAE_vit_base_patch16.pdparams"
,
}
}
__all__
=
list
(
MODEL_URLS
.
keys
())
__all__
=
list
(
MODEL_URLS
.
keys
())
...
@@ -48,8 +48,8 @@ _model_diff = None
...
@@ -48,8 +48,8 @@ _model_diff = None
_CLIP_diff
=
{
_CLIP_diff
=
{
'add_layer_norm_before_encoder'
:
[
'add_layer_norm_before_encoder'
:
[
'base_
patch32_224'
,
'base_patch16_224'
,
'large
_patch14_336'
,
'base_
vit_patch32_224'
,
'base_vit_patch16_224'
,
'large_vit
_patch14_336'
,
'large_patch14_224'
'large_
vit_
patch14_224'
],
],
'add_relative_position_bias_in_msa'
:
[],
'add_relative_position_bias_in_msa'
:
[],
'add_shared_rel_pos_bias'
:
[],
'add_shared_rel_pos_bias'
:
[],
...
@@ -57,8 +57,8 @@ _CLIP_diff = {
...
@@ -57,8 +57,8 @@ _CLIP_diff = {
'remove_cls_token'
:
[],
'remove_cls_token'
:
[],
'remove_abs_pos_emb'
:
[],
'remove_abs_pos_emb'
:
[],
'replace_mlp_GELU'
:
[
'replace_mlp_GELU'
:
[
'base_
patch32_224'
,
'base_patch16_224'
,
'large
_patch14_336'
,
'base_
vit_patch32_224'
,
'base_vit_patch16_224'
,
'large_vit
_patch14_336'
,
'large_patch14_224'
'large_
vit_
patch14_224'
],
],
'head'
:
{
'head'
:
{
'fc_norm'
:
[],
'fc_norm'
:
[],
...
@@ -87,7 +87,7 @@ _CoCa_diff = {
...
@@ -87,7 +87,7 @@ _CoCa_diff = {
'add_relative_position_bias_in_msa'
:
[],
'add_relative_position_bias_in_msa'
:
[],
'add_shared_rel_pos_bias'
:
[],
'add_shared_rel_pos_bias'
:
[],
'add_mul_gamma_to_msa_mlp'
:
[],
'add_mul_gamma_to_msa_mlp'
:
[],
'remove_cls_token'
:
[
'small_patch16_224'
],
'remove_cls_token'
:
[
'small_
vit_
patch16_224'
],
'remove_abs_pos_emb'
:
[],
'remove_abs_pos_emb'
:
[],
'replace_mlp_GELU'
:
[],
'replace_mlp_GELU'
:
[],
'head'
:
{
'head'
:
{
...
@@ -100,11 +100,11 @@ _CoCa_diff = {
...
@@ -100,11 +100,11 @@ _CoCa_diff = {
_BEiTv2_diff
=
{
_BEiTv2_diff
=
{
'add_layer_norm_before_encoder'
:
[],
'add_layer_norm_before_encoder'
:
[],
'add_relative_position_bias_in_msa'
:
'add_relative_position_bias_in_msa'
:
[
'base_
patch16_224'
,
'large
_patch16_224'
],
[
'base_
vit_patch16_224'
,
'large_vit
_patch16_224'
],
'add_shared_rel_pos_bias'
:
[],
'add_shared_rel_pos_bias'
:
[],
'add_mul_gamma_to_msa_mlp'
:
[
'base_
patch16_224'
,
'large
_patch16_224'
],
'add_mul_gamma_to_msa_mlp'
:
[
'base_
vit_patch16_224'
,
'large_vit
_patch16_224'
],
'remove_cls_token'
:
[],
'remove_cls_token'
:
[],
'remove_abs_pos_emb'
:
[
'base_
patch16_224'
,
'large
_patch16_224'
],
'remove_abs_pos_emb'
:
[
'base_
vit_patch16_224'
,
'large_vit
_patch16_224'
],
'replace_mlp_GELU'
:
[],
'replace_mlp_GELU'
:
[],
'head'
:
{
'head'
:
{
'fc_norm'
:
[],
'fc_norm'
:
[],
...
@@ -115,9 +115,9 @@ _BEiTv2_diff = {
...
@@ -115,9 +115,9 @@ _BEiTv2_diff = {
_CAE_diff
=
{
_CAE_diff
=
{
'add_layer_norm_before_encoder'
:
[],
'add_layer_norm_before_encoder'
:
[],
'add_relative_position_bias_in_msa'
:
[
'base_patch16_224'
],
'add_relative_position_bias_in_msa'
:
[
'base_
vit_
patch16_224'
],
'add_shared_rel_pos_bias'
:
[],
'add_shared_rel_pos_bias'
:
[],
'add_mul_gamma_to_msa_mlp'
:
[
'base_patch16_224'
],
'add_mul_gamma_to_msa_mlp'
:
[
'base_
vit_
patch16_224'
],
'remove_cls_token'
:
[],
'remove_cls_token'
:
[],
'remove_abs_pos_emb'
:
[],
'remove_abs_pos_emb'
:
[],
'replace_mlp_GELU'
:
[],
'replace_mlp_GELU'
:
[],
...
@@ -137,7 +137,7 @@ _EVA_diff = {
...
@@ -137,7 +137,7 @@ _EVA_diff = {
'remove_abs_pos_emb'
:
[],
'remove_abs_pos_emb'
:
[],
'replace_mlp_GELU'
:
[],
'replace_mlp_GELU'
:
[],
'head'
:
{
'head'
:
{
'fc_norm'
:
[
'huge_patch14'
],
'fc_norm'
:
[
'huge_
vit_
patch14'
],
'return_all_tokens'
:
[],
'return_all_tokens'
:
[],
'return_patch_tokens'
:
[],
'return_patch_tokens'
:
[],
}
}
...
@@ -152,7 +152,7 @@ _MAE_diff = {
...
@@ -152,7 +152,7 @@ _MAE_diff = {
'remove_abs_pos_emb'
:
[],
'remove_abs_pos_emb'
:
[],
'replace_mlp_GELU'
:
[],
'replace_mlp_GELU'
:
[],
'head'
:
{
'head'
:
{
'fc_norm'
:
[
'huge_patch14'
],
'fc_norm'
:
[
'huge_
vit_
patch14'
],
'return_all_tokens'
:
[],
'return_all_tokens'
:
[],
'return_patch_tokens'
:
[],
'return_patch_tokens'
:
[],
}
}
...
@@ -659,7 +659,7 @@ def _load_pretrained(pretrained, model, model_url, use_ssld=False):
...
@@ -659,7 +659,7 @@ def _load_pretrained(pretrained, model, model_url, use_ssld=False):
)
)
def
CLIP_base_patch32_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
CLIP_
vit_
base_patch32_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -677,7 +677,7 @@ def CLIP_base_patch32_224(pretrained=False, use_ssld=False, **kwargs):
...
@@ -677,7 +677,7 @@ def CLIP_base_patch32_224(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
CLIP_base_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
CLIP_
vit_
base_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -695,7 +695,7 @@ def CLIP_base_patch16_224(pretrained=False, use_ssld=False, **kwargs):
...
@@ -695,7 +695,7 @@ def CLIP_base_patch16_224(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
CLIP_large_patch14_336
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
CLIP_
vit_
large_patch14_336
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -713,7 +713,7 @@ def CLIP_large_patch14_336(pretrained=False, use_ssld=False, **kwargs):
...
@@ -713,7 +713,7 @@ def CLIP_large_patch14_336(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
CLIP_large_patch14_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
CLIP_
vit_
large_patch14_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -731,7 +731,7 @@ def CLIP_large_patch14_224(pretrained=False, use_ssld=False, **kwargs):
...
@@ -731,7 +731,7 @@ def CLIP_large_patch14_224(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
BEiTv2_base_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
BEiTv2_
vit_
base_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -749,7 +749,7 @@ def BEiTv2_base_patch16_224(pretrained=False, use_ssld=False, **kwargs):
...
@@ -749,7 +749,7 @@ def BEiTv2_base_patch16_224(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
BEiTv2_large_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
BEiTv2_
vit_
large_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -767,7 +767,7 @@ def BEiTv2_large_patch16_224(pretrained=False, use_ssld=False, **kwargs):
...
@@ -767,7 +767,7 @@ def BEiTv2_large_patch16_224(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
MOCOV3_small
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
MOCOV3_
vit_
small
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
"""
"""
vit small in mocov3
vit small in mocov3
"""
"""
...
@@ -786,7 +786,7 @@ def MOCOV3_small(pretrained=False, use_ssld=False, **kwargs):
...
@@ -786,7 +786,7 @@ def MOCOV3_small(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
MOCOV3_base
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
MOCOV3_
vit_
base
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
"""
"""
vit base in mocov3
vit base in mocov3
"""
"""
...
@@ -805,7 +805,7 @@ def MOCOV3_base(pretrained=False, use_ssld=False, **kwargs):
...
@@ -805,7 +805,7 @@ def MOCOV3_base(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
MAE_base_patch16
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
MAE_
vit_
base_patch16
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -821,7 +821,7 @@ def MAE_base_patch16(pretrained=False, use_ssld=False, **kwargs):
...
@@ -821,7 +821,7 @@ def MAE_base_patch16(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
MAE_large_patch16
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
MAE_
vit_
large_patch16
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -837,7 +837,7 @@ def MAE_large_patch16(pretrained=False, use_ssld=False, **kwargs):
...
@@ -837,7 +837,7 @@ def MAE_large_patch16(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
MAE_huge_patch14
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
MAE_
vit_
huge_patch14
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -853,7 +853,7 @@ def MAE_huge_patch14(pretrained=False, use_ssld=False, **kwargs):
...
@@ -853,7 +853,7 @@ def MAE_huge_patch14(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
EVA_huge_patch14
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
EVA_
vit_
huge_patch14
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
@@ -871,7 +871,7 @@ def EVA_huge_patch14(pretrained=False, use_ssld=False, **kwargs):
...
@@ -871,7 +871,7 @@ def EVA_huge_patch14(pretrained=False, use_ssld=False, **kwargs):
return
model
return
model
def
CAE_base_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
def
CAE_
vit_
base_patch16_224
(
pretrained
=
False
,
use_ssld
=
False
,
**
kwargs
):
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model_name
=
sys
.
_getframe
().
f_code
.
co_name
model
=
VisionTransformer
(
model
=
VisionTransformer
(
model_name
=
model_name
,
model_name
=
model_name
,
...
...
ppcls/arch/backbone/variant_models/foundation_vit_variant.py
浏览文件 @
23f5af9f
import
paddle
import
paddle
import
paddle.nn
as
nn
import
paddle.nn
as
nn
import
paddle.nn.functional
as
F
import
paddle.nn.functional
as
F
from
..model_zoo.foundation_vit
import
CLIP_large_patch14_224
,
_load_pretrained
from
..model_zoo.foundation_vit
import
CLIP_
vit_
large_patch14_224
,
_load_pretrained
MODEL_URLS
=
{
MODEL_URLS
=
{
"CLIP_large_patch14_224_aesthetic"
:
"CLIP_large_patch14_224_aesthetic"
:
...
...
编辑
预览
Markdown
is supported
0%
请重试
或
添加新附件
.
添加附件
取消
You are about to add
0
people
to the discussion. Proceed with caution.
先完成此消息的编辑!
取消
想要评论请
注册
或
登录