m-a-p / OpenCodeInterpreter-SC2-15B

huggingface.co
Total runs: 10
24-hour runs: 0
7-day runs: -1
30-day runs: 3
Model's Last Updated: March 03 2024
text-generation

Introduction of OpenCodeInterpreter-SC2-15B

Model Details of OpenCodeInterpreter-SC2-15B

OpenCodeInterpreter: Integrating Code Generation with Execution and Refinement

OpenCodeInterpreter

[🏠Homepage] | [🛠️Code]


Introduction

OpenCodeInterpreter is a family of open-source code generation systems designed to bridge the gap between large language models and advanced proprietary systems like the GPT-4 Code Interpreter. It significantly advances code generation capabilities by integrating execution and iterative refinement functionalities.

For further information and related work, refer to our paper: "OpenCodeInterpreter: A System for Enhanced Code Generation and Execution" available on arXiv.

Model Information

This model is based on starcoder2-15b .

Benchmark Scores

The OpenCodeInterpreter Models series exemplifies the evolution of coding model performance, particularly highlighting the significant enhancements brought about by the integration of execution feedback. In an effort to quantify these improvements, we present a detailed comparison across two critical benchmarks: HumanEval and MBPP. This comparison not only showcases the individual performance metrics on each benchmark but also provides an aggregated view of the overall performance enhancement. The subsequent table succinctly encapsulates the performance data, offering a clear perspective on how execution feedback contributes to elevating the models' capabilities in code interpretation and execution tasks.

Benchmark HumanEval (+) MBPP (+) Average (+)
OpenCodeInterpreter-DS-1.3B 65.2 (61.0) 63.4 (52.4) 64.3 (56.7)
+ Execution Feedback 65.2 (62.2) 65.2 (55.6) 65.2 (58.9)
OpenCodeInterpreter-DS-6.7B 76.2 (72.0) 73.9 (63.7) 75.1 (67.9)
+ Execution Feedback 81.1 (78.7) 82.7 (72.4) 81.9 (75.6)
+ Synth. Human Feedback 87.2 (86.6) 86.2 (74.2) 86.7 (80.4)
+ Synth. Human Feedback (Oracle) 89.7 (86.6) 87.2 (75.2) 88.5 (80.9)
OpenCodeInterpreter-DS-33B 79.3 (74.3) 78.7 (66.4) 79.0 (70.4)
+ Execution Feedback 82.9 (80.5) 83.5 (72.2) 83.2 (76.4)
+ Synth. Human Feedback 88.4 (86.0) 87.5 (75.9) 88.0 (81.0)
+ Synth. Human Feedback (Oracle) 92.7 (89.7) 90.5 (79.5) 91.6 (84.6)
OpenCodeInterpreter-CL-7B 72.6 (67.7) 66.4 (55.4) 69.5 (61.6)
+ Execution Feedback 75.6 (70.1) 69.9 (60.7) 72.8 (65.4)
OpenCodeInterpreter-CL-13B 77.4 (73.8) 70.7 (59.2) 74.1 (66.5)
+ Execution Feedback 81.1 (76.8) 78.2 (67.2) 79.7 (72.0)
OpenCodeInterpreter-CL-34B 78.0 (72.6) 73.4 (61.4) 75.7 (67.0)
+ Execution Feedback 81.7 (78.7) 80.2 (67.9) 81.0 (73.3)
OpenCodeInterpreter-CL-70B 76.2 (70.7) 73.0 (61.9) 74.6 (66.3)
+ Execution Feedback 79.9 (77.4) 81.5 (69.9) 80.7 (73.7)
OpenCodeInterpreter-GM-7B 56.1 (50.0) 39.8 (34.6) 48.0 (42.3)
+ Execution Feedback 64.0 (54.3) 48.6 (40.9) 56.3 (47.6)
OpenCodeInterpreter-SC2-3B 65.2 (57.9) 62.7 (52.9) 64.0 (55.4)
+ Execution Feedback 67.1 (60.4) 63.4 (54.9) 65.3 (57.7)
OpenCodeInterpreter-SC2-7B 73.8 (68.9) 61.7 (51.1) 67.8 (60.0)
+ Execution Feedback 75.6 (69.5) 66.9 (55.4) 71.3 (62.5)
OpenCodeInterpreter-SC2-15B 75.6 (69.5) 71.2 (61.2) 73.4 (65.4)
+ Execution Feedback 77.4 (72.0) 74.2 (63.4) 75.8 (67.7)

Note: The "(+)" notation represents scores from extended versions of the HumanEval and MBPP benchmarks. To ensure a fair comparison, the results shown for adding execution feedback are based on outcomes after just one iteration of feedback, without unrestricted iterations. This approach highlights the immediate impact of execution feedback on performance improvements across benchmarks.

Model Usage
Inference
import torch
from transformers import AutoTokenizer, AutoModelForCausalLM
model_path="m-a-p/OpenCodeInterpreter-SC2-3B"

tokenizer = AutoTokenizer.from_pretrained(model_path)
model = AutoModelForCausalLM.from_pretrained(
    model_path,
    torch_dtype=torch.bfloat16,
    device_map="auto",
)
model.eval()

prompt = "Write a function to find the shared elements from the given two lists."
inputs = tokenizer.apply_chat_template(
        [{'role': 'user', 'content': prompt }],
        return_tensors="pt"
    ).to(model.device)
outputs = model.generate(
    inputs, 
    max_new_tokens=1024,
    do_sample=False,
    pad_token_id=tokenizer.eos_token_id,
    eos_token_id=tokenizer.eos_token_id,
)
print(tokenizer.decode(outputs[0][len(inputs[0]):], skip_special_tokens=True))
Contact

If you have any inquiries, please feel free to raise an issue or reach out to us via email at: [email protected] , [email protected] . We're here to assist you!"

Runs of m-a-p OpenCodeInterpreter-SC2-15B on huggingface.co

10
Total runs
0
24-hour runs
-1
3-day runs
-1
7-day runs
3
30-day runs

More Information About OpenCodeInterpreter-SC2-15B huggingface.co Model

More OpenCodeInterpreter-SC2-15B license Visit here:

https://choosealicense.com/licenses/apache-2.0

OpenCodeInterpreter-SC2-15B huggingface.co

OpenCodeInterpreter-SC2-15B huggingface.co is an AI model on huggingface.co that provides OpenCodeInterpreter-SC2-15B's model effect (), which can be used instantly with this m-a-p OpenCodeInterpreter-SC2-15B model. huggingface.co supports a free trial of the OpenCodeInterpreter-SC2-15B model, and also provides paid use of the OpenCodeInterpreter-SC2-15B. Support call OpenCodeInterpreter-SC2-15B model through api, including Node.js, Python, http.

OpenCodeInterpreter-SC2-15B huggingface.co Url

https://huggingface.co/m-a-p/OpenCodeInterpreter-SC2-15B

m-a-p OpenCodeInterpreter-SC2-15B online free

OpenCodeInterpreter-SC2-15B huggingface.co is an online trial and call api platform, which integrates OpenCodeInterpreter-SC2-15B's modeling effects, including api services, and provides a free online trial of OpenCodeInterpreter-SC2-15B, you can try OpenCodeInterpreter-SC2-15B online for free by clicking the link below.

m-a-p OpenCodeInterpreter-SC2-15B online free url in huggingface.co:

https://huggingface.co/m-a-p/OpenCodeInterpreter-SC2-15B

OpenCodeInterpreter-SC2-15B install

OpenCodeInterpreter-SC2-15B is an open source model from GitHub that offers a free installation service, and any user can find OpenCodeInterpreter-SC2-15B on GitHub to install. At the same time, huggingface.co provides the effect of OpenCodeInterpreter-SC2-15B install, users can directly use OpenCodeInterpreter-SC2-15B installed effect in huggingface.co for debugging and trial. It also supports api for free installation.

OpenCodeInterpreter-SC2-15B install url in huggingface.co:

https://huggingface.co/m-a-p/OpenCodeInterpreter-SC2-15B

Url of OpenCodeInterpreter-SC2-15B

OpenCodeInterpreter-SC2-15B huggingface.co Url

Provider of OpenCodeInterpreter-SC2-15B huggingface.co

m-a-p
ORGANIZATIONS

Other API from m-a-p

huggingface.co

Total runs: 117.7K
Run Growth: -30.6K
Growth Rate: -26.00%
Updated:May 25 2025
huggingface.co

Total runs: 45.1K
Run Growth: 16.4K
Growth Rate: 36.38%
Updated:May 25 2025
huggingface.co

Total runs: 1.4K
Run Growth: 782
Growth Rate: 56.96%
Updated:April 09 2024
huggingface.co

Total runs: 980
Run Growth: -67
Growth Rate: -6.84%
Updated:June 02 2023
huggingface.co

Total runs: 817
Run Growth: 787
Growth Rate: 96.68%
Updated:November 19 2024
huggingface.co

Total runs: 582
Run Growth: 254
Growth Rate: 43.64%
Updated:June 02 2023
huggingface.co

Total runs: 452
Run Growth: -205
Growth Rate: -45.35%
Updated:June 02 2023
huggingface.co

Total runs: 275
Run Growth: 163
Growth Rate: 53.44%
Updated:March 18 2025
huggingface.co

Total runs: 78
Run Growth: -64
Growth Rate: -82.05%
Updated:June 03 2024
huggingface.co

Total runs: 28
Run Growth: -20
Growth Rate: -80.00%
Updated:March 12 2025
huggingface.co

Total runs: 25
Run Growth: 11
Growth Rate: 44.00%
Updated:December 20 2023
huggingface.co

Total runs: 16
Run Growth: -2
Growth Rate: -12.50%
Updated:April 08 2024