Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ tags:
|
|
13 |
pipeline_tag: text-generation
|
14 |
---
|
15 |
|
16 |
-
# OpenCodeReasoning-Nemotron-7B Overview
|
17 |
|
18 |
## Description: <br>
|
19 |
OpenCodeReasoning-Nemotron-7B-v1.1 is a large language model (LLM) which is a derivative of Qwen2.5-7B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 64k tokens. <br>
|
@@ -27,37 +27,35 @@ This model is ready for commercial/non-commercial use. <br>
|
|
27 |
|
28 |
Below results are the average of **64 evaluations** on LiveCodeBench (v5) [2408-2501].
|
29 |
|
30 |
-
| Model
|
31 |
-
|
32 |
-
| DeepSeek-R1
|
33 |
-
|
|
34 |
-
|
|
35 |
-
|
|
36 |
-
|
|
37 |
-
|
|
38 |
-
|
|
39 |
-
|
|
40 |
-
|
|
41 |
-
|
|
42 |
-
| **
|
43 |
-
| **
|
44 |
-
|
|
45 |
-
| **Distilled 14B+ Models**
|
46 |
-
|
|
47 |
-
| R1-Distill-Qwen-14B
|
48 |
-
| **
|
49 |
-
| **
|
50 |
-
|
|
51 |
-
|
|
52 |
-
|
|
53 |
-
|
|
54 |
-
|
|
55 |
-
|
|
56 |
-
|
|
57 |
-
|
|
58 |
-
| **
|
59 |
-
| **OCR-Qwen-32B-Instruct** | **61.7** |
|
60 |
-
| **OCR-Qwen-32B-Instruct-v1.1**| **?** |
|
61 |
|
62 |
|
63 |
## Reproducing our results
|
|
|
13 |
pipeline_tag: text-generation
|
14 |
---
|
15 |
|
16 |
+
# OpenCodeReasoning-Nemotron-7B-v1.1 Overview
|
17 |
|
18 |
## Description: <br>
|
19 |
OpenCodeReasoning-Nemotron-7B-v1.1 is a large language model (LLM) which is a derivative of Qwen2.5-7B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 64k tokens. <br>
|
|
|
27 |
|
28 |
Below results are the average of **64 evaluations** on LiveCodeBench (v5) [2408-2501].
|
29 |
|
30 |
+
| Model | Pass@1 |
|
31 |
+
|----------------------------------------|--------------------|
|
32 |
+
| DeepSeek-R1-0528 | 65.6 |
|
33 |
+
| DeepSeek-R1 | 65.6 |
|
34 |
+
| QwQ-32B | 61.3 |
|
35 |
+
| | |
|
36 |
+
| **Distilled 7B+ Models** | |
|
37 |
+
| | |
|
38 |
+
| Bespoke-Stratos-7B | 14.7 |
|
39 |
+
| OpenThinker-7B | 25.5 |
|
40 |
+
| R1-Distill-Qwen-7B | 38.0 |
|
41 |
+
| OlympicCoder-7B | 40.9 |
|
42 |
+
| **OpenCodeReasoning-Nemotron-7B** | **51.3** |
|
43 |
+
| **OpenCodeReasoning-Nemotron-7B-v1.1** | **?** |
|
44 |
+
| | |
|
45 |
+
| **Distilled 14B+ Models** | |
|
46 |
+
| | |
|
47 |
+
| R1-Distill-Qwen-14B | 51.3 |
|
48 |
+
| **OpenCodeReasoning-Nemotron-14B** | **59.4** |
|
49 |
+
| **OpenCodeReasoning-Nemotron-14B-v1.1**| **?** |
|
50 |
+
| | |
|
51 |
+
| **Distilled 32B+ Models** | |
|
52 |
+
| | |
|
53 |
+
| Bespoke-Stratos-32B | 30.1 |
|
54 |
+
| OpenThinker-32B | 54.1 |
|
55 |
+
| R1-Distill-Qwen-32B | 58.1 |
|
56 |
+
| OlympicCoder-32B | 57.4 |
|
57 |
+
| **OpenCodeReasoning-Nemotron-32B** | **61.7** |
|
58 |
+
| **OpenCodeReasoning-Nemotron-32B-v1.1**| **?** |
|
|
|
|
|
59 |
|
60 |
|
61 |
## Reproducing our results
|