Welcome to our Code Model repository! Our model is specifically fine-tuned for code generation tasks. Bud Millenial Code Gen open-source models are currently the State of the Art (SOTA) for code generation, beating all the existing models of all sizes. We have achieved a HumanEval value of 80.48 @ Pass 1, beating proprietary models like Gemini Ultra, Claude, GPT-3.5 etc. by a large margin, and on par with GPT-4 (HumanEval ~ 82. Ref. WizardCoder). Our proprietary model (Bud Code Jr) beats GPT-4 as well with a HumanEval value of 88.2 & a context size of 168K, we will be releasing an API for Researchers, Enterprises, and potential Partners by January 2024 end. If interested, please reach out to
[email protected]
News 🔥🔥🔥
[2024/01/09] We released
Code Millenials 3B
, which achieves the
56.09 pass@1
on the
HumanEval Benchmarks
.
[2024/01/09] We released
Code Millenials 1B
, which achieves the
51.82 pass@1
on the
HumanEval Benchmarks
.
[2024/01/03] We released
Code Millenials 34B
, which achieves the
80.48 pass@1
on the
HumanEval Benchmarks
.
[2024/01/02] We released
Code Millenials 13B
, which achieves the
76.21 pass@1
on the
HumanEval Benchmarks
.
HumanEval
For the millenial models, the eval script in the github repo is used for the above result.
Inference code using the pre-trained model from the Hugging Face model hub
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("budecosystem/code-millenials-1b")
model = AutoModelForCausalLM.from_pretrained("budecosystem/code-millenials-1b")
template = """A chat between a curious user and an artificial intelligence assistant. The assistant gives helpful, detailed, and polite answers to the user's questions.### Instruction: {instruction} ### Response:"""
instruction = <Your code instruction here>
prompt = template.format(instruction=instruction)
inputs = tokenizer(prompt, return_tensors="pt")
sample = model.generate(**inputs, max_length=128)
print(tokenizer.decode(sample[0]))
Training details
The model is trained of 8 A100 80GB for approximately 6hrs.
Hyperparameters
Value
per_device_train_batch_size
6
gradient_accumulation_steps
1
epoch
3
steps
11502
learning_rate
2e-5
lr schedular type
cosine
warmup ratio
0.1
optimizer
adamw
fp16
True
GPU
8 A100 80GB
Important Note
Bias, Risks, and Limitations:
Model may sometimes make errors, produce misleading contents, or struggle to manage tasks that are not related to coding.
Runs of budecosystem code-millenials-1b on huggingface.co
25
Total runs
0
24-hour runs
2
3-day runs
1
7-day runs
5
30-day runs
More Information About code-millenials-1b huggingface.co Model
code-millenials-1b huggingface.co
code-millenials-1b huggingface.co is an AI model on huggingface.co that provides code-millenials-1b's model effect (), which can be used instantly with this budecosystem code-millenials-1b model. huggingface.co supports a free trial of the code-millenials-1b model, and also provides paid use of the code-millenials-1b. Support call code-millenials-1b model through api, including Node.js, Python, http.
code-millenials-1b huggingface.co is an online trial and call api platform, which integrates code-millenials-1b's modeling effects, including api services, and provides a free online trial of code-millenials-1b, you can try code-millenials-1b online for free by clicking the link below.
budecosystem code-millenials-1b online free url in huggingface.co:
code-millenials-1b is an open source model from GitHub that offers a free installation service, and any user can find code-millenials-1b on GitHub to install. At the same time, huggingface.co provides the effect of code-millenials-1b install, users can directly use code-millenials-1b installed effect in huggingface.co for debugging and trial. It also supports api for free installation.