Shaw před 3 roky
rodič
revize
b1fb318cf4
1 změnil soubory, kde provedl 2 přidání a 0 odebrání
  1. 2 0
      README.md

+ 2 - 0
README.md

@@ -16,6 +16,8 @@ GLM-130B is an open bilingual (English & Chinese) bidirectional dense model with
 - **Reproducibility:** all results (30+ tasks) can be easily reproduced with open-sourced code and model checkpoints.
 - **Cross-Platform:** supports training and inference on NVIDIA, Hygon DCU, Ascend 910, and Sunway (Will be released soon).
 
+If you find our work and our open-sourced model useful, starring our repo to encourage our following development! :)
+
 ## News
 
 - 🌟 **[2022.08.24]** We are proud to publish the quantized version for GLM-130B.  While preserving the activation precision as FP16, the model weights can be quantized to as low as **INT4 with almost no degradation of performance**, further reducing the hardware requirements of the GLM-130B to **a single server with 4 * RTX 3090 (24G)**! See [Quantization of GLM-130B](docs/quantization.md) for details.