Update README.md
Browse files
README.md
CHANGED
@@ -20,12 +20,11 @@ Our ongoing projects include:
|
|
20 |
|
21 |
## Models and Performance
|
22 |
|
23 |
-
We have developed several models, including AstroSage-8B, AstroLLaMA-2-70B, and AstroLLaMA-3-8B. Our AstroSage-8B model has demonstrated strong performance in astronomy Q&A tasks ([Ting et al. 2024](https://arxiv.org/abs/2407.11194), Pan et al. 2024):
|
24 |
|
25 |
| Model | Score (%) |
|
26 |
|-------|-----------|
|
27 |
-
| **AstroSage-8B (AstroMLab)** | **80.9** |
|
28 |
-
| **AstroLLaMA-2-70B-Base (AstroMLab)** | **76.0** |
|
29 |
| LLaMA-3.1-8B | 73.7 |
|
30 |
| Phi-3.5-4B | 72.8 |
|
31 |
| Gemma-2-9B | 71.5 |
|
@@ -37,7 +36,7 @@ We have developed several models, including AstroSage-8B, AstroLLaMA-2-70B, and
|
|
37 |
| ChatGLM3-6B | 50.4 |
|
38 |
| AstroLLaMA-2-7B (UniverseTBD) | 44.3 |
|
39 |
|
40 |
-
AstroSage-8B, our lightweight model, currently achieves the highest score among the ~
|
41 |
|
42 |

|
43 |
|
|
|
20 |
|
21 |
## Models and Performance
|
22 |
|
23 |
+
We have developed several models, including AstroSage-LLaMA-3.1-8B, AstroLLaMA-2-70B, and AstroLLaMA-3-8B. Our AstroSage-8B model has demonstrated strong performance in astronomy Q&A tasks ([Ting et al. 2024](https://arxiv.org/abs/2407.11194), Pan et al. 2024):
|
24 |
|
25 |
| Model | Score (%) |
|
26 |
|-------|-----------|
|
27 |
+
| **AstroSage-LLaMA-3.1-8B (AstroMLab)** | **80.9** |
|
|
|
28 |
| LLaMA-3.1-8B | 73.7 |
|
29 |
| Phi-3.5-4B | 72.8 |
|
30 |
| Gemma-2-9B | 71.5 |
|
|
|
36 |
| ChatGLM3-6B | 50.4 |
|
37 |
| AstroLLaMA-2-7B (UniverseTBD) | 44.3 |
|
38 |
|
39 |
+
AstroSage-8B, our lightweight model, currently achieves the highest score among the ~8B parameter models in its astronomy knowledge recall ability.
|
40 |
|
41 |

|
42 |
|