Improve model card: Add paper URL, pipeline tag and Github URL

#6
by nielsr HF staff - opened
Files changed (1) hide show
  1. README.md +9 -5
README.md CHANGED
@@ -1,14 +1,15 @@
1
  ---
2
- library_name: transformers
 
 
3
  datasets:
4
  - BAAI/TACO
5
  - tasksource/PRM800K
6
  language:
7
  - en
8
- base_model:
9
- - Qwen/Qwen2.5-32B-Instruct
10
- - NovaSky-AI/Sky-T1-32B-Preview
11
  license: apache-2.0
 
12
  ---
13
 
14
  ## Model Details
@@ -18,9 +19,11 @@ license: apache-2.0
18
  <!-- Provide a longer summary of what this model is. -->
19
 
20
  This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
21
- Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) for more details.
22
 
23
  - **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
 
 
24
 
25
  ## Training Details
26
 
@@ -71,3 +74,4 @@ Please considering citing our blog post if you found it useful for your research
71
  note = {Accessed: 2025-01-23},
72
  year = {2025}
73
  }
 
 
1
  ---
2
+ base_model:
3
+ - Qwen/Qwen2.5-32B-Instruct
4
+ - NovaSky-AI/Sky-T1-32B-Preview
5
  datasets:
6
  - BAAI/TACO
7
  - tasksource/PRM800K
8
  language:
9
  - en
10
+ library_name: transformers
 
 
11
  license: apache-2.0
12
+ pipeline_tag: text-generation
13
  ---
14
 
15
  ## Model Details
 
19
  <!-- Provide a longer summary of what this model is. -->
20
 
21
  This is a 32B reasoning model preference optimized on top of Sky-T1-32B-Preview to significantly reduce generation lengths while maintaining accuracy. The performance is on par with o1-preview model in both math and coding, while reducing generation lengths by up to 57% relative to Sky-T1-32B-Preview.
22
+ Please see our [blog post](https://novasky-ai.github.io/posts/reduce-overthinking/) and [Sky-T1 blog post](https://novasky-ai.github.io/posts/sky-t1/) for more details.
23
 
24
  - **Developed by:** NovaSky Team from Sky Computing Lab at UC Berkeley.
25
+ - **Paper:** [LLMs Can Easily Learn to Reason from Demonstrations Structure, not content, is what matters!](https://hf.co/papers/2502.07374)
26
+ - **Code:** [https://github.com/NovaSky-AI/SkyThought](https://github.com/NovaSky-AI/SkyThought)
27
 
28
  ## Training Details
29
 
 
74
  note = {Accessed: 2025-01-23},
75
  year = {2025}
76
  }
77
+ ```