m-a-p / OpenCodeInterpreter-SC2-3B

huggingface.co
Total runs: 12
24-hour runs: 0
7-day runs: -1
30-day runs: -6
Model's Last Updated: March 03 2024
text-generation

Introduction of OpenCodeInterpreter-SC2-3B

Model Details of OpenCodeInterpreter-SC2-3B

OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement

OpenCodeInterpreter

[🏠Homepage] | [🛠️Code]


Introduction

OpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.

For further information and related work, refer to our paper: "OpenCodeInterpreter: A System for Enhanced Code Generation and Execution" available on arXiv.

Model Information

This model is based on starcoder2-3b .

Benchmark Scores

The OpenCodeInterpreter Models series exemplifies the evolution of coding model performance, particularly highlighting the significant enhancements brought about by the integration of execution feedback. In an effort to quantify these improvements, we present a detailed comparison across two critical benchmarks: HumanEval and MBPP. This comparison not only showcases the individual performance metrics on each benchmark but also provides an aggregated view of the overall performance enhancement. The subsequent table succinctly encapsulates the performance data, offering a clear perspective on how execution feedback contributes to elevating the models' capabilities in code interpretation and execution tasks.

Benchmark HumanEval (+) MBPP (+) Average (+)
OpenCodeInterpreter-DS-1.3B 65.2 (61.0) 63.4 (52.4) 64.3 (56.7)
+ Execution Feedback 65.2 (62.2) 65.2 (55.6) 65.2 (58.9)
OpenCodeInterpreter-DS-6.7B 76.2 (72.0) 73.9 (63.7) 75.1 (67.9)
+ Execution Feedback 81.1 (78.7) 82.7 (72.4) 81.9 (75.6)
+ Synth. Human Feedback 87.2 (86.6) 86.2 (74.2) 86.7 (80.4)
+ Synth. Human Feedback (Oracle) 89.7 (86.6) 87.2 (75.2) 88.5 (80.9)
OpenCodeInterpreter-DS-33B 79.3 (74.3) 78.7 (66.4) 79.0 (70.4)
+ Execution Feedback 82.9 (80.5) 83.5 (72.2) 83.2 (76.4)
+ Synth. Human Feedback 88.4 (86.0) 87.5 (75.9) 88.0 (81.0)
+ Synth. Human Feedback (Oracle) 92.7 (89.7) 90.5 (79.5) 91.6 (84.6)
OpenCodeInterpreter-CL-7B 72.6 (67.7) 66.4 (55.4) 69.5 (61.6)
+ Execution Feedback 75.6 (70.1) 69.9 (60.7) 72.8 (65.4)
OpenCodeInterpreter-CL-13B 77.4 (73.8) 70.7 (59.2) 74.1 (66.5)
+ Execution Feedback 81.1 (76.8) 78.2 (67.2) 79.7 (72.0)
OpenCodeInterpreter-CL-34B 78.0 (72.6) 73.4 (61.4) 75.7 (67.0)
+ Execution Feedback 81.7 (78.7) 80.2 (67.9) 81.0 (73.3)
OpenCodeInterpreter-CL-70B 76.2 (70.7) 73.0 (61.9) 74.6 (66.3)
+ Execution Feedback 79.9 (77.4) 81.5 (69.9) 80.7 (73.7)
OpenCodeInterpreter-GM-7B 56.1 (50.0) 39.8 (34.6) 48.0 (42.3)
+ Execution Feedback 64.0 (54.3) 48.6 (40.9) 56.3 (47.6)
OpenCodeInterpreter-SC2-3B 65.2 (57.9) 62.7 (52.9) 64.0 (55.4)
+ Execution Feedback 67.1 (60.4) 63.4 (54.9) 65.3 (57.7)
OpenCodeInterpreter-SC2-7B 73.8 (68.9) 61.7 (51.1) 67.8 (60.0)
+ Execution Feedback 75.6 (69.5) 66.9 (55.4) 71.3 (62.5)
OpenCodeInterpreter-SC2-15B 75.6 (69.5) 71.2 (61.2) 73.4 (65.4)
+ Execution Feedback 77.4 (72.0) 74.2 (63.4) 75.8 (67.7)

Note: The "(+)" notation represents scores from extended versions of the HumanEval and MBPP benchmarks. To ensure a fair comparison, the results shown for adding execution feedback are based on outcomes after just one iteration of feedback, without unrestricted iterations. This approach highlights the immediate impact of execution feedback on performance improvements across benchmarks.

Model Usage
Inference
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_path="m-a-p/OpenCodeInterpreter-SC2-3B"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)
model.eval()

prompt = "Write a function to find the shared elements from the given two lists."
inputs = tokenizer.apply_chat_template(
        [{'role': 'user', 'content': prompt }],
        return_tensors="pt"
    ).to(model.device)
outputs = model.generate(
    inputs, 
    max_new_tokens=1024,
    do_sample=False,
    pad_token_id=tokenizer.eos_token_id,
    eos_token_id=tokenizer.eos_token_id,
)
print(tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True))
Contact

If you have any inquiries, please feel free to raise an issue or reach out to us via email at: [email protected] , [email protected] . We're here to assist you!"

Runs of m-a-p OpenCodeInterpreter-SC2-3B on huggingface.co

12
Total runs
0
24-hour runs
-1
3-day runs
-1
7-day runs
-6
30-day runs

More Information About OpenCodeInterpreter-SC2-3B huggingface.co Model

More OpenCodeInterpreter-SC2-3B license Visit here:

https://choosealicense.com/licenses/apache-2.0

OpenCodeInterpreter-SC2-3B huggingface.co

OpenCodeInterpreter-SC2-3B huggingface.co is an AI model on huggingface.co that provides OpenCodeInterpreter-SC2-3B's model effect (), which can be used instantly with this m-a-p OpenCodeInterpreter-SC2-3B model. huggingface.co supports a free trial of the OpenCodeInterpreter-SC2-3B model, and also provides paid use of the OpenCodeInterpreter-SC2-3B. Support call OpenCodeInterpreter-SC2-3B model through api, including Node.js, Python, http.

OpenCodeInterpreter-SC2-3B huggingface.co Url

https://huggingface.co/m-a-p/OpenCodeInterpreter-SC2-3B

m-a-p OpenCodeInterpreter-SC2-3B online free

OpenCodeInterpreter-SC2-3B huggingface.co is an online trial and call api platform, which integrates OpenCodeInterpreter-SC2-3B's modeling effects, including api services, and provides a free online trial of OpenCodeInterpreter-SC2-3B, you can try OpenCodeInterpreter-SC2-3B online for free by clicking the link below.

m-a-p OpenCodeInterpreter-SC2-3B online free url in huggingface.co:

https://huggingface.co/m-a-p/OpenCodeInterpreter-SC2-3B

OpenCodeInterpreter-SC2-3B install

OpenCodeInterpreter-SC2-3B is an open source model from GitHub that offers a free installation service, and any user can find OpenCodeInterpreter-SC2-3B on GitHub to install. At the same time, huggingface.co provides the effect of OpenCodeInterpreter-SC2-3B install, users can directly use OpenCodeInterpreter-SC2-3B installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

OpenCodeInterpreter-SC2-3B install url in huggingface.co:

https://huggingface.co/m-a-p/OpenCodeInterpreter-SC2-3B

Url of OpenCodeInterpreter-SC2-3B

OpenCodeInterpreter-SC2-3B huggingface.co Url

Provider of OpenCodeInterpreter-SC2-3B huggingface.co

m-a-p
ORGANIZATIONS

Other API from m-a-p

huggingface.co

Total runs: 117.7K
Run Growth: -26.5K
Growth Rate: -21.73%
Updated:May 25 2025
huggingface.co

Total runs: 45.1K
Run Growth: 16.6K
Growth Rate: 36.59%
Updated:May 25 2025
huggingface.co

Total runs: 1.4K
Run Growth: 755
Growth Rate: 56.09%
Updated:April 09 2024
huggingface.co

Total runs: 980
Run Growth: -54
Growth Rate: -5.44%
Updated:June 02 2023
huggingface.co

Total runs: 817
Run Growth: 787
Growth Rate: 96.68%
Updated:November 19 2024
huggingface.co

Total runs: 582
Run Growth: 256
Growth Rate: 43.84%
Updated:June 02 2023
huggingface.co

Total runs: 452
Run Growth: -163
Growth Rate: -33.00%
Updated:June 02 2023
huggingface.co

Total runs: 275
Run Growth: 163
Growth Rate: 53.44%
Updated:March 18 2025
huggingface.co

Total runs: 78
Run Growth: -64
Growth Rate: -82.05%
Updated:June 03 2024
huggingface.co

Total runs: 28
Run Growth: -20
Growth Rate: -80.00%
Updated:March 12 2025
huggingface.co

Total runs: 25
Run Growth: 10
Growth Rate: 41.67%
Updated:December 20 2023
huggingface.co

Total runs: 16
Run Growth: -3
Growth Rate: -20.00%
Updated:April 08 2024