Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,59 @@
|
|
1 |
-
---
|
2 |
-
license: apache-2.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
datasets:
|
4 |
+
- Minami-su/Amara-o1-dataset
|
5 |
+
---
|
6 |
+
|
7 |
+
<div align="left">
|
8 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/62d7f90b102d144db4b4245b/WIGeEeS5gBvATkSf2GSB-.png"
|
9 |
+
alt="Model Illustration"
|
10 |
+
style="width:50%; max-width:none;" />
|
11 |
+
</div>
|
12 |
+
<blockquote> “何が綴られていたのか、私たちの文明では到底理解できない” <br/> (所阐述的内容超出了我们文明的理解范围) <br/> — sasakure.UK </blockquote>
|
13 |
+
|
14 |
+
# How to use
|
15 |
+
|
16 |
+
|
17 |
+
微调基于Qwen2.5-7B-Instruct
|
18 |
+
|
19 |
+
|
20 |
+
```python
|
21 |
+
|
22 |
+
# Use a pipeline as a high-level helper
|
23 |
+
|
24 |
+
from transformers import pipeline
|
25 |
+
|
26 |
+
messages = [
|
27 |
+
{"role": "user", "content": "Who are you?"},
|
28 |
+
]
|
29 |
+
pipe = pipeline("text-generation", model="Minami-su/Amara-o1-7B-Qwen")
|
30 |
+
pipe(messages)
|
31 |
+
|
32 |
+
|
33 |
+
# Load model directly
|
34 |
+
|
35 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
36 |
+
|
37 |
+
tokenizer = AutoTokenizer.from_pretrained("Minami-su/Amara-o1-7B-Qwen")
|
38 |
+
model = AutoModelForCausalLM.from_pretrained("Minami-su/Amara-o1-7B-Qwen")
|
39 |
+
```
|
40 |
+
|
41 |
+
#### Open Ended Generation Evaluation
|
42 |
+
|
43 |
+
<div align="left">
|
44 |
+
|
45 |
+
|
46 |
+
|
47 |
+
| Model | Arena-Hard | AlpacaEval 2.0 |
|
48 |
+
|-------|------------|----------------|
|
49 |
+
| DeepSeek-V2.5-0905 | 76.2 | 50.5 |
|
50 |
+
| Qwen2.5-72B-Instruct | 81.2 | 49.1 |
|
51 |
+
| LLaMA-3.1 405B | 69.3 | 40.5 |
|
52 |
+
| **Amara-o1-7B-Qwen** | ? | 42.12 |
|
53 |
+
| **Amara-o2-7B-Qwen** | ? | **51.33** |
|
54 |
+
| GPT-4o-0513 | 80.4 | 51.1 |
|
55 |
+
| Claude-Sonnet-3.5-1022 | 85.2 | 52.0 |
|
56 |
+
| DeepSeek-V3 | **85.5** | **70.0** |
|
57 |
+
|
58 |
+
Note: English open-ended conversation evaluations. For AlpacaEval 2.0, we use the length-controlled win rate as the metric.
|
59 |
+
</div>
|