m-a-p / OpenCodeInterpreter-CL-34B

huggingface.co
Total runs: 737
24-hour runs: 0
7-day runs: 106
30-day runs: 579
Model's Last Updated: March 03 2024
text-generation

Introduction of OpenCodeInterpreter-CL-34B

Model Details of OpenCodeInterpreter-CL-34B

OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement

OpenCodeInterpreter

[🏠Homepage] | [🛠️Code]


Introduction

OpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.

For further information and related work, refer to our paper: "OpenCodeInterpreter: A System for Enhanced Code Generation and Execution" available on arXiv.

Model Information

This model is based on CodeLlama-34b-Python-hf .

Benchmark Scores

The OpenCodeInterpreter Models series exemplifies the evolution of coding model performance, particularly highlighting the significant enhancements brought about by the integration of execution feedback. In an effort to quantify these improvements, we present a detailed comparison across two critical benchmarks: HumanEval and MBPP. This comparison not only showcases the individual performance metrics on each benchmark but also provides an aggregated view of the overall performance enhancement. The subsequent table succinctly encapsulates the performance data, offering a clear perspective on how execution feedback contributes to elevating the models' capabilities in code interpretation and execution tasks.

Benchmark HumanEval (+) MBPP (+) Average (+)
OpenCodeInterpreter-DS-1.3B 65.2 (61.0) 63.4 (52.4) 64.3 (56.7)
+ Execution Feedback 65.2 (62.2) 65.2 (55.6) 65.2 (58.9)
OpenCodeInterpreter-DS-6.7B 76.2 (72.0) 73.9 (63.7) 75.1 (67.9)
+ Execution Feedback 81.1 (78.7) 82.7 (72.4) 81.9 (75.6)
+ Synth. Human Feedback 87.2 (86.6) 86.2 (74.2) 86.7 (80.4)
+ Synth. Human Feedback (Oracle) 89.7 (86.6) 87.2 (75.2) 88.5 (80.9)
OpenCodeInterpreter-DS-33B 79.3 (74.3) 78.7 (66.4) 79.0 (70.4)
+ Execution Feedback 82.9 (80.5) 83.5 (72.2) 83.2 (76.4)
+ Synth. Human Feedback 88.4 (86.0) 87.5 (75.9) 88.0 (81.0)
+ Synth. Human Feedback (Oracle) 92.7 (89.7) 90.5 (79.5) 91.6 (84.6)
OpenCodeInterpreter-CL-7B 72.6 (67.7) 66.4 (55.4) 69.5 (61.6)
+ Execution Feedback 75.6 (70.1) 69.9 (60.7) 72.8 (65.4)
OpenCodeInterpreter-CL-13B 77.4 (73.8) 70.7 (59.2) 74.1 (66.5)
+ Execution Feedback 81.1 (76.8) 78.2 (67.2) 79.7 (72.0)
OpenCodeInterpreter-CL-34B 78.0 (72.6) 73.4 (61.4) 75.7 (67.0)
+ Execution Feedback 81.7 (78.7) 80.2 (67.9) 81.0 (73.3)
OpenCodeInterpreter-CL-70B 76.2 (70.7) 73.0 (61.9) 74.6 (66.3)
+ Execution Feedback 79.9 (77.4) 81.5 (69.9) 80.7 (73.7)
OpenCodeInterpreter-GM-7B 56.1 (50.0) 39.8 (34.6) 48.0 (42.3)
+ Execution Feedback 64.0 (54.3) 48.6 (40.9) 56.3 (47.6)
OpenCodeInterpreter-SC2-3B 65.2 (57.9) 62.7 (52.9) 64.0 (55.4)
+ Execution Feedback 67.1 (60.4) 63.4 (54.9) 65.3 (57.7)
OpenCodeInterpreter-SC2-7B 73.8 (68.9) 61.7 (51.1) 67.8 (60.0)
+ Execution Feedback 75.6 (69.5) 66.9 (55.4) 71.3 (62.5)
OpenCodeInterpreter-SC2-15B 75.6 (69.5) 71.2 (61.2) 73.4 (65.4)
+ Execution Feedback 77.4 (72.0) 74.2 (63.4) 75.8 (67.7)

Note: The "(+)" notation represents scores from extended versions of the HumanEval and MBPP benchmarks. To ensure a fair comparison, the results shown for adding execution feedback are based on outcomes after just one iteration of feedback, without unrestricted iterations. This approach highlights the immediate impact of execution feedback on performance improvements across benchmarks.

Model Usage
Inference
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_path="m-a-p/OpenCodeInterpreter-CL-34B"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)
model.eval()

prompt = "Write a function to find the shared elements from the given two lists."
inputs = tokenizer.apply_chat_template(
        [{'role': 'user', 'content': prompt }],
        return_tensors="pt"
    ).to(model.device)
outputs = model.generate(
    inputs, 
    max_new_tokens=1024,
    do_sample=False,
    pad_token_id=tokenizer.eos_token_id,
    eos_token_id=tokenizer.eos_token_id,
)
print(tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True))
Contact

If you have any inquiries, please feel free to raise an issue or reach out to us via email at: [email protected] , [email protected] . We're here to assist you!"

Runs of m-a-p OpenCodeInterpreter-CL-34B on huggingface.co

737
Total runs
0
24-hour runs
21
3-day runs
106
7-day runs
579
30-day runs

More Information About OpenCodeInterpreter-CL-34B huggingface.co Model

More OpenCodeInterpreter-CL-34B license Visit here:

https://choosealicense.com/licenses/apache-2.0

OpenCodeInterpreter-CL-34B huggingface.co

OpenCodeInterpreter-CL-34B huggingface.co is an AI model on huggingface.co that provides OpenCodeInterpreter-CL-34B's model effect (), which can be used instantly with this m-a-p OpenCodeInterpreter-CL-34B model. huggingface.co supports a free trial of the OpenCodeInterpreter-CL-34B model, and also provides paid use of the OpenCodeInterpreter-CL-34B. Support call OpenCodeInterpreter-CL-34B model through api, including Node.js, Python, http.

OpenCodeInterpreter-CL-34B huggingface.co Url

https://huggingface.co/m-a-p/OpenCodeInterpreter-CL-34B

m-a-p OpenCodeInterpreter-CL-34B online free

OpenCodeInterpreter-CL-34B huggingface.co is an online trial and call api platform, which integrates OpenCodeInterpreter-CL-34B's modeling effects, including api services, and provides a free online trial of OpenCodeInterpreter-CL-34B, you can try OpenCodeInterpreter-CL-34B online for free by clicking the link below.

m-a-p OpenCodeInterpreter-CL-34B online free url in huggingface.co:

https://huggingface.co/m-a-p/OpenCodeInterpreter-CL-34B

OpenCodeInterpreter-CL-34B install

OpenCodeInterpreter-CL-34B is an open source model from GitHub that offers a free installation service, and any user can find OpenCodeInterpreter-CL-34B on GitHub to install. At the same time, huggingface.co provides the effect of OpenCodeInterpreter-CL-34B install, users can directly use OpenCodeInterpreter-CL-34B installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

OpenCodeInterpreter-CL-34B install url in huggingface.co:

https://huggingface.co/m-a-p/OpenCodeInterpreter-CL-34B

Url of OpenCodeInterpreter-CL-34B

OpenCodeInterpreter-CL-34B huggingface.co Url

Provider of OpenCodeInterpreter-CL-34B huggingface.co

m-a-p
ORGANIZATIONS

Other API from m-a-p

huggingface.co

Total runs: 121.8K
Run Growth: -26.5K
Growth Rate: -21.73%
Updated:May 25 2025
huggingface.co

Total runs: 45.3K
Run Growth: 16.6K
Growth Rate: 36.59%
Updated:May 25 2025
huggingface.co

Total runs: 1.3K
Run Growth: 755
Growth Rate: 56.09%
Updated:April 09 2024
huggingface.co

Total runs: 993
Run Growth: -54
Growth Rate: -5.44%
Updated:June 02 2023
huggingface.co

Total runs: 814
Run Growth: 787
Growth Rate: 96.68%
Updated:November 19 2024
huggingface.co

Total runs: 584
Run Growth: 256
Growth Rate: 43.84%
Updated:June 02 2023
huggingface.co

Total runs: 494
Run Growth: -163
Growth Rate: -33.00%
Updated:June 02 2023
huggingface.co

Total runs: 305
Run Growth: 163
Growth Rate: 53.44%
Updated:March 18 2025
huggingface.co

Total runs: 78
Run Growth: -64
Growth Rate: -82.05%
Updated:June 03 2024
huggingface.co

Total runs: 25
Run Growth: -20
Growth Rate: -80.00%
Updated:March 12 2025
huggingface.co

Total runs: 24
Run Growth: 10
Growth Rate: 41.67%
Updated:December 20 2023
huggingface.co

Total runs: 15
Run Growth: -3
Growth Rate: -20.00%
Updated:April 08 2024