2
0

model_glm_130b_int8.sh 548 B

12345678910111213141516
  1. MODEL_TYPE="glm-130b"
  2. CHECKPOINT_PATH="<your checkpoint path>"
  3. MP_SIZE=8
  4. MODEL_ARGS="--model-parallel-size ${MP_SIZE} \
  5. --num-layers 70 \
  6. --hidden-size 12288 \
  7. --inner-hidden-size 32768 \
  8. --vocab-size 150528 \
  9. --num-attention-heads 96 \
  10. --max-sequence-length 2048 \
  11. --tokenizer-type icetk-glm-130B \
  12. --layernorm-order post \
  13. --quantization-bit-width 8 \
  14. --load ${CHECKPOINT_PATH} \
  15. --skip-init \
  16. --fp16"