提交 5025d09f 编写于 作者: G gaotingquan 提交者: Tingquan Gao

fix: fix a DeprecationWarning & style

上级 ba2dd01a
...@@ -12,7 +12,7 @@ ...@@ -12,7 +12,7 @@
# See the License for the specific language governing permissions and # See the License for the specific language governing permissions and
# limitations under the License. # limitations under the License.
from collections import Callable from collections.abc import Callable
import numpy as np import numpy as np
import paddle import paddle
...@@ -331,9 +331,7 @@ def _load_pretrained(pretrained, model, model_url, use_ssld=False): ...@@ -331,9 +331,7 @@ def _load_pretrained(pretrained, model, model_url, use_ssld=False):
) )
def ViT_small_patch16_224(pretrained=False, def ViT_small_patch16_224(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
patch_size=16, patch_size=16,
embed_dim=768, embed_dim=768,
...@@ -350,9 +348,7 @@ def ViT_small_patch16_224(pretrained=False, ...@@ -350,9 +348,7 @@ def ViT_small_patch16_224(pretrained=False,
return model return model
def ViT_base_patch16_224(pretrained=False, def ViT_base_patch16_224(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
patch_size=16, patch_size=16,
embed_dim=768, embed_dim=768,
...@@ -370,9 +366,7 @@ def ViT_base_patch16_224(pretrained=False, ...@@ -370,9 +366,7 @@ def ViT_base_patch16_224(pretrained=False,
return model return model
def ViT_base_patch16_384(pretrained=False, def ViT_base_patch16_384(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
img_size=384, img_size=384,
patch_size=16, patch_size=16,
...@@ -391,9 +385,7 @@ def ViT_base_patch16_384(pretrained=False, ...@@ -391,9 +385,7 @@ def ViT_base_patch16_384(pretrained=False,
return model return model
def ViT_base_patch32_384(pretrained=False, def ViT_base_patch32_384(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
img_size=384, img_size=384,
patch_size=32, patch_size=32,
...@@ -412,9 +404,7 @@ def ViT_base_patch32_384(pretrained=False, ...@@ -412,9 +404,7 @@ def ViT_base_patch32_384(pretrained=False,
return model return model
def ViT_large_patch16_224(pretrained=False, def ViT_large_patch16_224(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
patch_size=16, patch_size=16,
embed_dim=1024, embed_dim=1024,
...@@ -432,9 +422,7 @@ def ViT_large_patch16_224(pretrained=False, ...@@ -432,9 +422,7 @@ def ViT_large_patch16_224(pretrained=False,
return model return model
def ViT_large_patch16_384(pretrained=False, def ViT_large_patch16_384(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
img_size=384, img_size=384,
patch_size=16, patch_size=16,
...@@ -453,9 +441,7 @@ def ViT_large_patch16_384(pretrained=False, ...@@ -453,9 +441,7 @@ def ViT_large_patch16_384(pretrained=False,
return model return model
def ViT_large_patch32_384(pretrained=False, def ViT_large_patch32_384(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
img_size=384, img_size=384,
patch_size=32, patch_size=32,
...@@ -474,9 +460,7 @@ def ViT_large_patch32_384(pretrained=False, ...@@ -474,9 +460,7 @@ def ViT_large_patch32_384(pretrained=False,
return model return model
def ViT_huge_patch16_224(pretrained=False, def ViT_huge_patch16_224(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
patch_size=16, patch_size=16,
embed_dim=1280, embed_dim=1280,
...@@ -492,9 +476,7 @@ def ViT_huge_patch16_224(pretrained=False, ...@@ -492,9 +476,7 @@ def ViT_huge_patch16_224(pretrained=False,
return model return model
def ViT_huge_patch32_384(pretrained=False, def ViT_huge_patch32_384(pretrained=False, use_ssld=False, **kwargs):
use_ssld=False,
**kwargs):
model = VisionTransformer( model = VisionTransformer(
img_size=384, img_size=384,
patch_size=32, patch_size=32,
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册