model_glm_6b.sh 564 B

1234567891011121314151617
  1. MODEL_TYPE="glm-6b"
  2. CHECKPOINT_PATH="<your checkpoint path>"
  3. MP_SIZE=1
  4. MODEL_ARGS="--model-parallel-size ${MP_SIZE} \
  5. --num-layers 28 \
  6. --hidden-size 4096 \
  7. --inner-hidden-size 16384 \
  8. --vocab-size 150528 \
  9. --num-attention-heads 32 \
  10. --max-sequence-length 2048 \
  11. --tokenizer-type icetk-glm-130B \
  12. --layernorm-order post \
  13. --position-encoding-2d \
  14. --no-glu \
  15. --load ${CHECKPOINT_PATH} \
  16. --skip-init \
  17. --fp16"