提交 93d78299 编写于 作者: X xyzhou-puck

upload params and glue dataset

上级 da4de5f1
......@@ -157,7 +157,7 @@ def main():
labels,
device=device)
cls_model.bert_layer.load("./bert_small", reset_optimizer=True)
cls_model.bert_layer.load("./bert_uncased_L-12_H-768_A-12/bert", reset_optimizer=True)
# do train
cls_model.fit(train_data=train_dataloader.dataloader,
......
1. download data: wget https://paddle-hapi.bj.bcebos.com/data/bert_data.tar.gz
2. unzip data: tar -zvxf bert_data.tar.gz
3. download pretrained parameters: wget https://paddle-hapi.bj.bcebos.com/models/bert_uncased_L-12_H-768_A-12.tar.gz
4. unzip pretrained parameters: tar -zvxf bert_uncased_L-12_H-768_A-12.tar.gz
4. bash run_classifier_single_gpu.sh
#!/bin/bash
BERT_BASE_PATH="./bert_uncased_L-12_H-768_A-12/"
DATA_PATH="./data/glue_data/MNLI/"
CKPT_PATH="./data/saved_model/mnli_models"
# start fine-tuning
python3.7 -m paddle.distributed.launch --started_port 8899 --selected_gpus=1,2,3 bert_classifier.py\
--use_cuda true \
--do_train true \
--do_test true \
--batch_size 64 \
--data_dir ${DATA_PATH} \
--vocab_path ${BERT_BASE_PATH}/vocab.txt \
--checkpoints ${CKPT_PATH} \
--save_steps 1000 \
--weight_decay 0.01 \
--warmup_proportion 0.1 \
--validation_steps 100 \
--epoch 3 \
--max_seq_len 128 \
--bert_config_path ${BERT_BASE_PATH}/bert_config.json \
--learning_rate 5e-5 \
--skip_steps 10 \
--shuffle true
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册