run_predict.sh 4.7 KB
Newer Older
W
wenquan wu 已提交
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109
#!/bin/bash

# set gpu id to use
export CUDA_VISIBLE_DEVICES=1

# task_name can select from ["match", "match_kn", "match_kn_gene"]
# match task: do not use knowledge info (goal and knowledge) for retrieval model
# match_kn task: use knowledge info (goal and knowledge) for retrieval model
# match_kn_gene task: 1) use knowledge info (goal and knowledge) for retrieval model;
#                     2) generalizes target_a/target_b of goal, replaces them with slot mark
# more information about generalization in match_kn_gene,
# you can refer to ./tools/convert_conversation_corpus_to_model_text.py
TASK_NAME=$1

if [ "$TASK_NAME" = "match" ]
then
  DICT_NAME="./dict/char.dict"
  USE_KNOWLEDGE=0
  TOPIC_GENERALIZATION=0
elif [ "$TASK_NAME" = "match_kn" ]
then
  DICT_NAME="./dict/char.dict"
  USE_KNOWLEDGE=1
  TOPIC_GENERALIZATION=0
elif [ "$TASK_NAME" = "match_kn_gene" ]
then
  DICT_NAME="./dict/gene.dict"
  USE_KNOWLEDGE=1
  TOPIC_GENERALIZATION=1
else
  echo "task name error, should be match|match_kn|match_kn_gene"
fi

# in predict stage, FOR_PREDICT=1
FOR_PREDICT=1

# put all data set that used and generated for testing under this folder: INPUT_PATH
# for more details, please refer to the following data processing instructions
INPUT_PATH="./data"

# put the model files needed for testing under this folder: OUTPUT_PATH
OUTPUT_PATH="./models"

# set python path according to your actual environment
PYTHON_PATH="python"

# in test stage, you can eval dev.txt or test.txt
# the "dev.txt" and "test.txt" are the original data of DuConv and
# need to be placed in this folder: INPUT_PATH/resource/
# the following preprocessing will generate the actual data needed for model testing
# after testing, you can run eval.py to get the final eval score if the original data have answer
# DATA_TYPE = "dev" or "test"
DATA_TYPE="dev"

# candidate set, construct in train stage
candidate_set_file=${INPUT_PATH}/candidate_set.txt

# ensure that each file is in the correct path
#     1. put the data of DuConv under this folder: INPUT_PATH/resource/
#            - the data provided consists of three parts: train.txt dev.txt test.txt
#            - the train.txt and dev.txt are session data, the test.txt is sample data
#            - in test stage, we just use the dev.txt or test.txt
#     2. the sample data extracted from session data is in this folder: INPUT_PATH/resource/
#     3. the candidate data constructed from sample data is in this folder: INPUT_PATH/resource/
#     4. the text file required by the model is in this folder: INPUT_PATH
corpus_file=${INPUT_PATH}/resource/${DATA_TYPE}.txt
sample_file=${INPUT_PATH}/resource/sample.${DATA_TYPE}.txt
candidate_file=${INPUT_PATH}/resource/candidate.${DATA_TYPE}.txt
text_file=${INPUT_PATH}/test.txt
score_file=./output/score.txt
predict_file=./output/predict.txt

# step 1: if eval dev.txt, firstly have to convert session data to sample data
# if eval test.txt, we can use original test.txt of DuConv directly.
if [ "${DATA_TYPE}"x = "test"x ]; then
    sample_file=${corpus_file}
else
    ${PYTHON_PATH} ./tools/convert_session_to_sample.py ${corpus_file} ${sample_file}
fi

# step 2: construct candidate for sample data
${PYTHON_PATH} ./tools/construct_candidate.py ${sample_file} ${candidate_set_file} ${candidate_file} 10

# step 3: convert sample data with candidates to text data required by the model
${PYTHON_PATH} ./tools/convert_conversation_corpus_to_model_text.py ${candidate_file} ${text_file} ${USE_KNOWLEDGE} ${TOPIC_GENERALIZATION} ${FOR_PREDICT}

# inference_model can used for interact.py
inference_model="./models/inference_model"

# step 4: predict score by model
$PYTHON_PATH -u predict.py --task_name ${TASK_NAME} \
                   --use_cuda \
                   --batch_size 10 \
                   --init_checkpoint ${OUTPUT_PATH}/50 \
                   --data_dir ${INPUT_PATH} \
                   --vocab_path ${DICT_NAME} \
                   --save_inference_model_path ${inference_model} \
                   --max_seq_len 128 \
                   --output ${score_file}

# step 5: extract predict utterance by candidate_file and score_file
# if the original file has answers, the predict_file format is "predict \t gold \n predict \t gold \n ......"
# if the original file not has answers, the predict_file format is "predict \n predict \n predict \n predict \n ......"
${PYTHON_PATH} ./tools/extract_predict_utterance.py ${candidate_file} ${score_file} ${predict_file}

# step 6: if the original file has answers, you can run the following command to get result
# if the original file not has answers, you can upload the ./output/test.result.final 
# to the website(https://ai.baidu.com/broad/submission?dataset=duconv) to get the official automatic evaluation
${PYTHON_PATH} ./tools/eval.py ${predict_file}