Update README.md
Browse files
README.md
CHANGED
@@ -13,10 +13,10 @@ tags:
|
|
13 |
pipeline_tag: text-generation
|
14 |
---
|
15 |
|
16 |
-
# OpenCodeReasoning-Nemotron-
|
17 |
|
18 |
## Description: <br>
|
19 |
-
OpenCodeReasoning-Nemotron-
|
20 |
|
21 |
This model is ready for commercial/non-commercial use. <br>
|
22 |
|
@@ -131,7 +131,7 @@ Architecture Type: Dense decoder-only Transformer model
|
|
131 |
Network Architecture: Qwen-7B-Instruct
|
132 |
<br>
|
133 |
**This model was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
|
134 |
-
**OpenCodeReasoning-Nemotron-
|
135 |
|
136 |
## Input: <br>
|
137 |
**Input Type(s):** Text <br>
|
@@ -156,30 +156,30 @@ NVIDIA Hopper <br>
|
|
156 |
|
157 |
## Model Version(s):
|
158 |
1.1 (6/20/2025) <br>
|
159 |
-
OpenCodeReasoning-Nemotron-
|
160 |
-
OpenCodeReasoning-Nemotron-
|
161 |
-
OpenCodeReasoning-Nemotron-
|
162 |
|
163 |
|
164 |
# Training and Evaluation Datasets: <br>
|
165 |
|
166 |
## Training Dataset:
|
167 |
|
168 |
-
The training corpus for OpenCodeReasoning-Nemotron-
|
169 |
|
170 |
Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
|
171 |
Labeling Method: Hybrid: Automated, Human, Synthetic <br>
|
172 |
Properties: 1.165M samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
|
173 |
|
174 |
## Evaluation Dataset:
|
175 |
-
We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-
|
176 |
**Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
|
177 |
**Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
|
178 |
|
179 |
|
180 |
|
181 |
### License/Terms of Use: <br>
|
182 |
-
GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-
|
183 |
|
184 |
### Deployment Geography:
|
185 |
Global<br>
|
@@ -188,7 +188,7 @@ Global<br>
|
|
188 |
This model is intended for developers and researchers building LLMs. <br>
|
189 |
|
190 |
### Release Date: <br>
|
191 |
-
Huggingface [06/20/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-
|
192 |
|
193 |
## Reference(s):
|
194 |
[2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
|
|
|
13 |
pipeline_tag: text-generation
|
14 |
---
|
15 |
|
16 |
+
# OpenCodeReasoning-Nemotron-1.1-7B Overview
|
17 |
|
18 |
## Description: <br>
|
19 |
+
OpenCodeReasoning-Nemotron-1.1-7B is a large language model (LLM) which is a derivative of Qwen2.5-7B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 64k tokens. <br>
|
20 |
|
21 |
This model is ready for commercial/non-commercial use. <br>
|
22 |
|
|
|
131 |
Network Architecture: Qwen-7B-Instruct
|
132 |
<br>
|
133 |
**This model was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
|
134 |
+
**OpenCodeReasoning-Nemotron-1.1-7B was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
|
135 |
|
136 |
## Input: <br>
|
137 |
**Input Type(s):** Text <br>
|
|
|
156 |
|
157 |
## Model Version(s):
|
158 |
1.1 (6/20/2025) <br>
|
159 |
+
OpenCodeReasoning-Nemotron-1.1-7B<br>
|
160 |
+
OpenCodeReasoning-Nemotron-1.1-14B<br>
|
161 |
+
OpenCodeReasoning-Nemotron-1.1-32B<br>
|
162 |
|
163 |
|
164 |
# Training and Evaluation Datasets: <br>
|
165 |
|
166 |
## Training Dataset:
|
167 |
|
168 |
+
The training corpus for OpenCodeReasoning-Nemotron-1.1-7B is [OpenCodeReasoning](https://huggingface.co/datasets/nvidia/OpenCodeReasoning) dataset, which is composed of competitive programming questions and DeepSeek-R1 generated responses.
|
169 |
|
170 |
Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
|
171 |
Labeling Method: Hybrid: Automated, Human, Synthetic <br>
|
172 |
Properties: 1.165M samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
|
173 |
|
174 |
## Evaluation Dataset:
|
175 |
+
We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-1.1-7B. <br>
|
176 |
**Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
|
177 |
**Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
|
178 |
|
179 |
|
180 |
|
181 |
### License/Terms of Use: <br>
|
182 |
+
GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-1.1-7B/blob/main/LICENSE).
|
183 |
|
184 |
### Deployment Geography:
|
185 |
Global<br>
|
|
|
188 |
This model is intended for developers and researchers building LLMs. <br>
|
189 |
|
190 |
### Release Date: <br>
|
191 |
+
Huggingface [06/20/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-1.1-7B/ <br>
|
192 |
|
193 |
## Reference(s):
|
194 |
[2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
|