Edit model card

Breakings!

We know what you want, and here you go!

  • Newly released lyraChatGLM model, suitable for Ampere (A100/A10) as well as Volta (V100)
  • lyraChatGLM has been further optimized, reaching 9000 tokens/s on A100 and 3900 tokens/s on V100, about 5.5x faster than the up-to-date official version (2023/6/1).
  • The memory usage was optimized too, now we can set batch_size up to 256 on A100!
  • INT8 weight only PTQ is supported

Note that the code was fully updated too, you need to use the new API, see Uses below

If you like our work and consider to join us, feel free to drop a line to benbinwu@tencent.com.

P.S. Recently we have received a lot of inquiries on accelerating customized models. Actually, we do not have plan to release the convertion tool at this moment, nor do we think it would be possible to apply your customized models based on our current release.


Model Card for lyraChatGLM

lyraChatGLM is currently the fastest ChatGLM-6B available. To the best of our knowledge, it is the first accelerated version of ChatGLM-6B.

The inference speed of lyraChatGLM has achieved 300x acceleration upon the early original version. We are still working hard to further improve the performance.

Among its main features are (updated on 2023-06-20):

  • weights: original ChatGLM-6B weights released by THUDM.
  • device: Nvidia GPU with Amperer architecture or Volta architecture (A100, A10, V100...).
  • batch_size: compiled with dynamic batch size, maximum depends on device.
  • We now support cuda version of both 11.X and 12.X
  • lyraChatGLM has been further optimized, with faster model load speed from few minutes to less than 10s for non-int8 mode, and around 1 min for int8 mode!

Speed

  • orginal version(fixed batch infer): commit id 1d240ba

test on A100 40G

  1. The maximum batch size and maximum speed table for each version of the model.
    version max_batch_size max_speed
    original 1 30 tokens/s
    original(fxied batch infer) 192 1638.52 tokens/s
    lyraChatGLM(current) 256 9082.60 tokens/s
  2. The speed table for the same batch size.
    version 1 batch_size 8 batch_size 64 batch_size 128 batch_size
    original 30 tokens/s - - -
    original(fxied batch infer) 34.48 tokens/s 356.29 tokens/s 1638.52 tokens/s 1338.45 tokens/s
    lyraChatGLM(current) 110.05 tokens/s 843.60 tokens/s 4926.92 tokens/s 7235.04 tokens/s

test on V100

  1. The maximum batch size and maximum speed table for each version of the model.
    version max_batch_size max_speed
    original 1 17.83 tokens/s
    original(fxied batch infer) 128 992.20 tokens/s
    lyraChatGLM(current) 192 3958.39 tokens/s
  2. The speed table for the same batch size.
    version 1 batch_size 8 batch_size 64 batch_size 128 batch_size
    original 17.83 tokens/s - - -
    original(fxied batch infer) 17.83 tokens/s 228.95 tokens/s 889.7 tokens/s 922.20 tokens/s
    lyraChatGLM(current) 59.33 tokens/s 514.15 tokens/s 2849.88 tokens/s 3958.39 tokens/s

Model Sources

Docker Environment Recommendation

  • For Cuda 11.X: we recommend nvcr.io/nvidia/pytorch:22.12-py3
  • For Cuda 12.0: we recommend nvcr.io/nvidia/pytorch:23.02-py3
docker pull nvcr.io/nvidia/pytorch:23.02-py3
docker run --rm -it --gpus all -v ./:/lyraChatGLM nvcr.io/nvidia/pytorch:23.02-py3

pip install -r requirements.txt
python demo.py

Uses

from lyraChatGLM import LyraChatGLM6B

model_path = "./models/1-gpu-fp16.bin"
tokenizer_path = "./models"
data_type = "fp16"
int8_mode = 0   # 1 for INT8 WEIGHT ONLY PTQ
max_output_length = 150
arch = "Ampere" # Ampere or Volta
cuda_version = 12

model = LyraChatGLM6B(model_path, tokenizer_path, data_type, int8_mode, arch, cuda_version)
prompt = "列出3个不同的机器学习算法,并说明它们的适用范围."
test_batch_size = 256

prompts = [prompt, ]

# If you want to get different output in same batch, you can set do_sample to True
output_texts = model.generate(prompts, output_length=max_output_length,top_k=30, top_p=0.85, temperature=0.35, repetition_penalty=1.2, do_sample=False)

print(output_texts)

Demo output

input

列出3个不同的机器学习算法,并说明它们的适用范围.

output

以下是三个常见的机器学习算法及其适用范围:

  1. 决策树(Decision Tree):决策树是一种基于分类和回归问题的朴素贝叶斯模型。它通过构建一系列逐步分裂的分支来预测结果。适用于那些具有简单特征、大量数据且数据集大小在可接受范围内的情况。

  2. 随机森林(Random Forest):随机森林是一种集成学习算法,由多个决策树组成。它的优点是能够处理大规模数据和高维度的特征。适用于需要对多个变量进行建模的场景,例如医疗诊断、金融风险评估等。

  3. 支持向量机(Support Vector Machine):支持向量机是一种监督学习方法,通常用于分类问题。它可以处理高维数据,并且具有较高的准确性。适用于需要对高维数据进行分类或回归的问题,例如图像识别、自然语言处理等。

INT8

Int8 usage:

Our current version supports INT8 weight only PTQ. To enable this mode, simply modify the int8_mode to 1 in the demo.py file.

In this mode, gpu memory can be further reduced by about half and the speed can be doubled.

This solves the issue mentioned in https://github.com/THUDM/ChatGLM-6B/issues/1042.

However, the speed gain is best achieved with a batch size of no more than 128. If you don't use A100 GPU, you can adjust the batch size to reduce it and get the benefits. We recommend a batch size of 64.This mode is very suitable for GPUs with limited VRAM or scenarios where it is difficult to use larger batch sizes in real-time services.

It should be noted that although we have aligned the accuracy in our test cases, there may be slight differences in accuracy in some untested scenarios with int8. Please be aware of this.

Citation

@Misc{lyraChatGLM2023,
  author =       {Kangjian Wu, Zhengtao Wang, Yibo Lu, Bin Wu},
  title =        {lyraChatGLM: Accelerating ChatGLM to 9000+ tokens/s},
  howpublished = {\url{https://huggingface.co/TMElyralab/lyraChatGLM}},
  year =         {2023}
}

Report bug

Downloads last month
0
Unable to determine this model's library. Check the docs .