internlm / AlchemistCoder-L-7B

huggingface.co
Total runs: 13
24-hour runs: 0
7-day runs: -1
30-day runs: 4
Model's Last Updated: May 30 2024
text-generation

Introduction of AlchemistCoder-L-7B

Model Details of AlchemistCoder-L-7B

AlchemistCoder: Harmonizing and Eliciting Code Capability by Hindsight Tuning on Multi-source Data

[ 🤗 HuggingFace ] [ 📃 Paper ] [ 🌐 Project Page ]

✨ Highlights

Abstract: Open-source Large Language Models (LLMs) and their specialized variants, particularly Code LLMs, have recently delivered impressive performance. However, previous Code LLMs are typically fine-tuned on single-source data with limited quality and diversity, which may insufficiently elicit the potential of pre-trained Code LLMs. In this paper, we present AlchemistCoder, a series of Code LLMs with enhanced code generation and generalization capabilities fine-tuned on multi-source data. To achieve this, we pioneer to unveil inherent conflicts among the various styles and qualities in multi-source code corpora and introduce data-specific prompts with hindsight relabeling, termed AlchemistPrompts, to harmonize different data sources and instruction-response pairs. Additionally, we propose incorporating the data construction process into the fine-tuning data as code comprehension tasks, including instruction evolution, data filtering, and code review. Extensive experiments demonstrate that AlchemistCoder holds a clear lead among all models of the same size (6.7B/7B) and rivals or even surpasses larger models (15B/33B/70B), showcasing the efficacy of our method in refining instruction-following capabilities and advancing the boundaries of code intelligence.

  • AlchemistPrompts : Designed as data-specific prompts for harmonizing inherent conflicts in multi-source data and mitigating the instruction/response misalignment at a fined-grained level.
  • Code Comprehenstion Tasks : Sourced from the process of data construction, consisting of instruction evolution, data filtering, and code review.
  • Harmonized Multi-source Data : Instruction tuned on 200M tokens, including 6 types of high-quality data.
  • Superior Model Performance : Surpassing all the open-source models of the same size (6.7/7B), and rivaling or even beating larger models (15B/33B/70B/ChatGPT) on 6 code benchmarks.
  • Advanced generic capabilities : Demonstrated by the significant improvements on MMLU, BBH, and GSM8K.
🚀 Quick Start
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("internlm/AlchemistCoder-L-7B", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("internlm/AlchemistCoder-L-7B", trust_remote_code=True, torch_dtype=torch.bfloat16).cuda()
model = model.eval()

input_text = "Implement the Dijkstra algorithm in Python"
inputs = tokenizer(input_text, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_length=128)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
🧪 Evaluation and Fine-tune

Please refer to AlchemistCoder and InternLM .

😃 Acknowledgments

AlchemistCoder is built with InternLM and OpenCompass . Thanks for their awesome work!

📧 Contact

If you have any questions, please create an issue on this repository or contact us at:

🌟 Citation

If you find our work useful, please consider citing:

@misc{song2024alchemistcoder,
      title={AlchemistCoder: Harmonizing and Eliciting Code Capability by Hindsight Tuning on Multi-source Data}, 
      author={Zifan Song and Yudong Wang and Wenwei Zhang and Kuikun Liu and Chengqi Lyu and Demin Song and Qipeng Guo and Hang Yan and Dahua Lin and Kai Chen and Cairong Zhao},
      year={2024},
      eprint={2405.19265},
      archivePrefix={arXiv},
      primaryClass={cs.CL}
}

Runs of internlm AlchemistCoder-L-7B on huggingface.co

13
Total runs
0
24-hour runs
0
3-day runs
-1
7-day runs
4
30-day runs

More Information About AlchemistCoder-L-7B huggingface.co Model

More AlchemistCoder-L-7B license Visit here:

https://choosealicense.com/licenses/apache-2.0

AlchemistCoder-L-7B huggingface.co

AlchemistCoder-L-7B huggingface.co is an AI model on huggingface.co that provides AlchemistCoder-L-7B's model effect (), which can be used instantly with this internlm AlchemistCoder-L-7B model. huggingface.co supports a free trial of the AlchemistCoder-L-7B model, and also provides paid use of the AlchemistCoder-L-7B. Support call AlchemistCoder-L-7B model through api, including Node.js, Python, http.

AlchemistCoder-L-7B huggingface.co Url

https://huggingface.co/internlm/AlchemistCoder-L-7B

internlm AlchemistCoder-L-7B online free

AlchemistCoder-L-7B huggingface.co is an online trial and call api platform, which integrates AlchemistCoder-L-7B's modeling effects, including api services, and provides a free online trial of AlchemistCoder-L-7B, you can try AlchemistCoder-L-7B online for free by clicking the link below.

internlm AlchemistCoder-L-7B online free url in huggingface.co:

https://huggingface.co/internlm/AlchemistCoder-L-7B

AlchemistCoder-L-7B install

AlchemistCoder-L-7B is an open source model from GitHub that offers a free installation service, and any user can find AlchemistCoder-L-7B on GitHub to install. At the same time, huggingface.co provides the effect of AlchemistCoder-L-7B install, users can directly use AlchemistCoder-L-7B installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

AlchemistCoder-L-7B install url in huggingface.co:

https://huggingface.co/internlm/AlchemistCoder-L-7B

Url of AlchemistCoder-L-7B

AlchemistCoder-L-7B huggingface.co Url

Provider of AlchemistCoder-L-7B huggingface.co

internlm
ORGANIZATIONS

Other API from internlm

huggingface.co

Total runs: 35.4K
Run Growth: -29.7K
Growth Rate: -83.81%
Updated:March 29 2026
huggingface.co

Total runs: 25.4K
Run Growth: -877
Growth Rate: -3.56%
Updated:March 13 2025
huggingface.co

Total runs: 2.0K
Run Growth: 1.2K
Growth Rate: 61.95%
Updated:July 03 2024
huggingface.co

Total runs: 1.2K
Run Growth: -395
Growth Rate: -32.04%
Updated:January 24 2024
huggingface.co

Total runs: 175
Run Growth: -20
Growth Rate: -11.43%
Updated:April 16 2026
huggingface.co

Total runs: 169
Run Growth: -36
Growth Rate: -21.30%
Updated:July 15 2025
huggingface.co

Total runs: 123
Run Growth: 91
Growth Rate: 73.98%
Updated:October 23 2025
huggingface.co

Total runs: 42
Run Growth: 1
Growth Rate: 2.38%
Updated:July 15 2025
huggingface.co

Total runs: 30
Run Growth: -65
Growth Rate: -216.67%
Updated:March 24 2026