- MODEL_TYPE="glm-large-generation"
- CHECKPOINT_PATH="/zhangpai21/checkpoints/glm-large-en-generation"
- MP_SIZE=1
- MODEL_ARGS="--model-parallel-size ${MP_SIZE} \
- --vocab 30592 \
- --num-layers 24 \
- --hidden-size 1024 \
- --num-attention-heads 16 \
- --max-sequence-length 513 \
- --tokenizer-type glm_BertWordPieceTokenizer \
- --tokenizer-model-type bert-large-uncased \
- --task-mask \
- --load ${CHECKPOINT_PATH}"
|