wasiuddina commited on
Commit
8222feb
·
verified ·
1 Parent(s): 6166322

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +10 -10
README.md CHANGED
@@ -13,10 +13,10 @@ tags:
13
  pipeline_tag: text-generation
14
  ---
15
 
16
- # OpenCodeReasoning-Nemotron-7B-v1.1 Overview
17
 
18
  ## Description: <br>
19
- OpenCodeReasoning-Nemotron-7B-v1.1 is a large language model (LLM) which is a derivative of Qwen2.5-7B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 64k tokens. <br>
20
 
21
  This model is ready for commercial/non-commercial use. <br>
22
 
@@ -131,7 +131,7 @@ Architecture Type: Dense decoder-only Transformer model
131
  Network Architecture: Qwen-7B-Instruct
132
  <br>
133
  **This model was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
134
- **OpenCodeReasoning-Nemotron-7B-v1.1 was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
135
 
136
  ## Input: <br>
137
  **Input Type(s):** Text <br>
@@ -156,30 +156,30 @@ NVIDIA Hopper <br>
156
 
157
  ## Model Version(s):
158
  1.1 (6/20/2025) <br>
159
- OpenCodeReasoning-Nemotron-7B-v1.1<br>
160
- OpenCodeReasoning-Nemotron-14B-v1.1<br>
161
- OpenCodeReasoning-Nemotron-32B-v1.1<br>
162
 
163
 
164
  # Training and Evaluation Datasets: <br>
165
 
166
  ## Training Dataset:
167
 
168
- The training corpus for OpenCodeReasoning-Nemotron-7B-v1.1 is [OpenCodeReasoning](https://huggingface.co/datasets/nvidia/OpenCodeReasoning) dataset, which is composed of competitive programming questions and DeepSeek-R1 generated responses.
169
 
170
  Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
171
  Labeling Method: Hybrid: Automated, Human, Synthetic <br>
172
  Properties: 1.165M samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
173
 
174
  ## Evaluation Dataset:
175
- We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-7B-v1.1. <br>
176
  **Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
177
  **Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
178
 
179
 
180
 
181
  ### License/Terms of Use: <br>
182
- GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-7B-v1.1/blob/main/LICENSE).
183
 
184
  ### Deployment Geography:
185
  Global<br>
@@ -188,7 +188,7 @@ Global<br>
188
  This model is intended for developers and researchers building LLMs. <br>
189
 
190
  ### Release Date: <br>
191
- Huggingface [06/20/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-7B-v1.1/ <br>
192
 
193
  ## Reference(s):
194
  [2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding
 
13
  pipeline_tag: text-generation
14
  ---
15
 
16
+ # OpenCodeReasoning-Nemotron-1.1-7B Overview
17
 
18
  ## Description: <br>
19
+ OpenCodeReasoning-Nemotron-1.1-7B is a large language model (LLM) which is a derivative of Qwen2.5-7B-Instruct (AKA the reference model). It is a reasoning model that is post-trained for reasoning for code generation. The model supports a context length of 64k tokens. <br>
20
 
21
  This model is ready for commercial/non-commercial use. <br>
22
 
 
131
  Network Architecture: Qwen-7B-Instruct
132
  <br>
133
  **This model was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
134
+ **OpenCodeReasoning-Nemotron-1.1-7B was developed based on Qwen2.5-7B-Instruct and has 7B model parameters. <br>**
135
 
136
  ## Input: <br>
137
  **Input Type(s):** Text <br>
 
156
 
157
  ## Model Version(s):
158
  1.1 (6/20/2025) <br>
159
+ OpenCodeReasoning-Nemotron-1.1-7B<br>
160
+ OpenCodeReasoning-Nemotron-1.1-14B<br>
161
+ OpenCodeReasoning-Nemotron-1.1-32B<br>
162
 
163
 
164
  # Training and Evaluation Datasets: <br>
165
 
166
  ## Training Dataset:
167
 
168
+ The training corpus for OpenCodeReasoning-Nemotron-1.1-7B is [OpenCodeReasoning](https://huggingface.co/datasets/nvidia/OpenCodeReasoning) dataset, which is composed of competitive programming questions and DeepSeek-R1 generated responses.
169
 
170
  Data Collection Method: Hybrid: Automated, Human, Synthetic <br>
171
  Labeling Method: Hybrid: Automated, Human, Synthetic <br>
172
  Properties: 1.165M samples from OpenCodeReasoning (https://huggingface.co/datasets/nvidia/OpenCodeReasoning)
173
 
174
  ## Evaluation Dataset:
175
+ We used the datasets listed in the next section to evaluate OpenCodeReasoning-Nemotron-1.1-7B. <br>
176
  **Data Collection Method: Hybrid: Automated, Human, Synthetic <br>**
177
  **Labeling Method: Hybrid: Automated, Human, Synthetic <br>**
178
 
179
 
180
 
181
  ### License/Terms of Use: <br>
182
+ GOVERNING TERMS: Use of this model is governed by [Apache 2.0](https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-1.1-7B/blob/main/LICENSE).
183
 
184
  ### Deployment Geography:
185
  Global<br>
 
188
  This model is intended for developers and researchers building LLMs. <br>
189
 
190
  ### Release Date: <br>
191
+ Huggingface [06/20/2025] via https://huggingface.co/nvidia/OpenCodeReasoning-Nemotron-1.1-7B/ <br>
192
 
193
  ## Reference(s):
194
  [2504.01943] OpenCodeReasoning: Advancing Data Distillation for Competitive Coding