|
@@ -16,6 +16,8 @@ GLM-130B is an open bilingual (English & Chinese) bidirectional dense model with
|
|
|
- **Reproducibility:** all results (30+ tasks) can be easily reproduced with open-sourced code and model checkpoints.
|
|
|
- **Cross-Platform:** supports training and inference on NVIDIA, Hygon DCU, Ascend 910, and Sunway (Will be released soon).
|
|
|
|
|
|
+If you find our work and our open-sourced model useful, starring our repo to encourage our following development! :)
|
|
|
+
|
|
|
## News
|
|
|
|
|
|
- 🌟 **[2022.08.24]** We are proud to publish the quantized version for GLM-130B. While preserving the activation precision as FP16, the model weights can be quantized to as low as **INT4 with almost no degradation of performance**, further reducing the hardware requirements of the GLM-130B to **a single server with 4 * RTX 3090 (24G)**! See [Quantization of GLM-130B](docs/quantization.md) for details.
|