CodeGeeX2 is the second-generation model of the multilingual code generation model
CodeGeeX
(
KDD’23
), which is implemented based on the
ChatGLM2
architecture trained on more code data. Due to the advantage of ChatGLM2, CodeGeeX2 has been comprehensively improved in coding capability (+107% > CodeGeeX; with only 6B parameters, surpassing larger StarCoder-15B for some tasks). It has the following features:
More Powerful Coding Capabilities
: Based on the ChatGLM2-6B model, CodeGeeX2-6B has been further pre-trained on 600B code tokens, which has been comprehensively improved in coding capability compared to the first-generation. On the
HumanEval-X
benchmark, all six languages have been significantly improved (Python +57%, C++ +71%, Java +54%, JavaScript +83%, Go +56%, Rust +321%), and in Python it reached 35.9% of Pass@1 one-time pass rate, surpassing the larger StarCoder-15B.
More Useful Features
: Inheriting the ChatGLM2-6B model features, CodeGeeX2-6B better supports both Chinese and English prompts, maximum 8192 sequence length, and the inference speed is significantly improved compared to the first-generation. After quantization, it only needs 6GB of GPU memory for inference, thus supports lightweight local deployment.
Comprehensive AI Coding Assistant
: The backend of CodeGeeX plugin (
VS Code
,
Jetbrains
) is upgraded, supporting 100+ programming languages, and adding practical functions such as infilling and cross-file completion. Combined with the "Ask CodeGeeX" interactive AI coding assistant, it can be used to solve various programming problems via Chinese or English dialogue, including but not limited to code summarization, code translation, debugging, and comment generation, which helps increasing the efficiency of developpers.
Open Liscense
: CodeGeeX2-6B weights are fully open to academic research, and please apply for commercial use by filling in the
registration form
.
The code in this repository is open source under the
Apache-2.0
license. The model weights are licensed under the
Model License
.
引用 | Citation
如果觉得我们的工作有帮助,欢迎引用以下论文:
If you find our work helpful, please feel free to cite the following paper:
@inproceedings{zheng2023codegeex,
title={CodeGeeX: A Pre-Trained Model for Code Generation with Multilingual Evaluations on HumanEval-X},
author={Qinkai Zheng and Xiao Xia and Xu Zou and Yuxiao Dong and Shan Wang and Yufei Xue and Zihan Wang and Lei Shen and Andi Wang and Yang Li and Teng Su and Zhilin Yang and Jie Tang},
booktitle={KDD},
year={2023}
}
Runs of THUDM codegeex2-6b-int4 on huggingface.co
4
Total runs
0
24-hour runs
0
3-day runs
0
7-day runs
0
30-day runs
More Information About codegeex2-6b-int4 huggingface.co Model
codegeex2-6b-int4 huggingface.co
codegeex2-6b-int4 huggingface.co is an AI model on huggingface.co that provides codegeex2-6b-int4's model effect (), which can be used instantly with this THUDM codegeex2-6b-int4 model. huggingface.co supports a free trial of the codegeex2-6b-int4 model, and also provides paid use of the codegeex2-6b-int4. Support call codegeex2-6b-int4 model through api, including Node.js, Python, http.
codegeex2-6b-int4 huggingface.co is an online trial and call api platform, which integrates codegeex2-6b-int4's modeling effects, including api services, and provides a free online trial of codegeex2-6b-int4, you can try codegeex2-6b-int4 online for free by clicking the link below.
THUDM codegeex2-6b-int4 online free url in huggingface.co:
codegeex2-6b-int4 is an open source model from GitHub that offers a free installation service, and any user can find codegeex2-6b-int4 on GitHub to install. At the same time, huggingface.co provides the effect of codegeex2-6b-int4 install, users can directly use codegeex2-6b-int4 installed effect in huggingface.co for debugging and trial. It also supports api for free installation.