ChatGLM-6B 是一个开源的、支持中英双语问答的对话语言模型,基于
General Language Model (GLM)
架构,具有 62 亿参数。结合模型量化技术,用户可以在消费级的显卡上进行本地部署(INT4 量化级别下最低只需 6GB 显存)。ChatGLM-6B 使用了和
ChatGLM
相同的技术,针对中文问答和对话进行了优化。经过约 1T 标识符的中英双语训练,辅以监督微调、反馈自助、人类反馈强化学习等技术的加持,62 亿参数的 ChatGLM-6B 已经能生成相当符合人类偏好的回答。 ChatGLM-6B 权重对学术研究
完全开放
,在填写
问卷
进行登记后
亦允许免费商业使用
。
ChatGLM-6B is an open bilingual language model based on
General Language Model (GLM)
framework, with 6.2 billion parameters. With the quantization technique, users can deploy locally on consumer-grade graphics cards (only 6GB of GPU memory is required at the INT4 quantization level). ChatGLM-6B uses technology similar to ChatGPT, optimized for Chinese QA and dialogue. The model is trained for about 1T tokens of Chinese and English corpus, supplemented by supervised fine-tuning, feedback bootstrap, and reinforcement learning with human feedback. With only about 6.2 billion parameters, the model is able to generate answers that are in line with human preference. ChatGLM-6B weights are
completely open
for academic research, and
free commercial use
is also allowed after completing the
questionnaire
.
@inproceedings{
zeng2023glm-130b,
title={{GLM}-130B: An Open Bilingual Pre-trained Model},
author={Aohan Zeng and Xiao Liu and Zhengxiao Du and Zihan Wang and Hanyu Lai and Ming Ding and Zhuoyi Yang and Yifan Xu and Wendi Zheng and Xiao Xia and Weng Lam Tam and Zixuan Ma and Yufei Xue and Jidong Zhai and Wenguang Chen and Zhiyuan Liu and Peng Zhang and Yuxiao Dong and Jie Tang},
booktitle={The Eleventh International Conference on Learning Representations (ICLR)},
year={2023},
url={https://openreview.net/forum?id=-Aw0rrrPUF}
}
@inproceedings{du2022glm,
title={GLM: General Language Model Pretraining with Autoregressive Blank Infilling},
author={Du, Zhengxiao and Qian, Yujie and Liu, Xiao and Ding, Ming and Qiu, Jiezhong and Yang, Zhilin and Tang, Jie},
booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers)},
pages={320--335},
year={2022}
}
Runs of THUDM chatglm-6b on huggingface.co
2.8K
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs
More Information About chatglm-6b huggingface.co Model
chatglm-6b huggingface.co
chatglm-6b huggingface.co is an AI model on huggingface.co that provides chatglm-6b's model effect (), which can be used instantly with this THUDM chatglm-6b model. huggingface.co supports a free trial of the chatglm-6b model, and also provides paid use of the chatglm-6b. Support call chatglm-6b model through api, including Node.js, Python, http.
chatglm-6b huggingface.co is an online trial and call api platform, which integrates chatglm-6b's modeling effects, including api services, and provides a free online trial of chatglm-6b, you can try chatglm-6b online for free by clicking the link below.
THUDM chatglm-6b online free url in huggingface.co:
chatglm-6b is an open source model from GitHub that offers a free installation service, and any user can find chatglm-6b on GitHub to install. At the same time, huggingface.co provides the effect of chatglm-6b install, users can directly use chatglm-6b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.