create model card
Browse files
README.md
CHANGED
@@ -4,12 +4,27 @@ tags:
|
|
4 |
- axolotl
|
5 |
- generated_from_trainer
|
6 |
base_model: mistralai/Mistral-7B-v0.1
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
7 |
---
|
8 |
|
9 |
-
|
10 |
-
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
-
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
13 |
<details><summary>See axolotl config</summary>
|
14 |
|
15 |
axolotl version: `0.3.0`
|
@@ -137,59 +152,54 @@ special_tokens:
|
|
137 |
|
138 |
</details><br>
|
139 |
|
140 |
-
#
|
|
|
|
|
|
|
|
|
141 |
|
142 |
-
|
143 |
|
144 |
-
|
145 |
|
146 |
-
|
147 |
|
148 |
-
|
149 |
|
150 |
-
|
151 |
|
152 |
-
|
153 |
|
154 |
-
|
155 |
|
156 |
-
|
157 |
|
|
|
158 |
|
159 |
-
|
160 |
-
- quant_method: bitsandbytes
|
161 |
-
- load_in_8bit: False
|
162 |
-
- load_in_4bit: True
|
163 |
-
- llm_int8_threshold: 6.0
|
164 |
-
- llm_int8_skip_modules: None
|
165 |
-
- llm_int8_enable_fp32_cpu_offload: False
|
166 |
-
- llm_int8_has_fp16_weight: False
|
167 |
-
- bnb_4bit_quant_type: nf4
|
168 |
-
- bnb_4bit_use_double_quant: True
|
169 |
-
- bnb_4bit_compute_dtype: bfloat16
|
170 |
|
171 |
-
|
172 |
|
173 |
-
|
174 |
-
|
175 |
-
|
176 |
-
|
177 |
-
|
178 |
-
|
179 |
-
|
180 |
-
|
181 |
-
|
182 |
-
|
183 |
-
|
|
|
|
|
184 |
|
185 |
-
|
186 |
|
|
|
187 |
|
|
|
188 |
|
189 |
-
|
190 |
|
191 |
-
-
|
192 |
-
- Transformers 4.37.0
|
193 |
-
- Pytorch 2.0.1+cu118
|
194 |
-
- Datasets 2.16.1
|
195 |
-
- Tokenizers 0.15.0
|
|
|
4 |
- axolotl
|
5 |
- generated_from_trainer
|
6 |
base_model: mistralai/Mistral-7B-v0.1
|
7 |
+
datasets:
|
8 |
+
- allenai/ai2_arc
|
9 |
+
- camel-ai/physics
|
10 |
+
- camel-ai/chemistry
|
11 |
+
- camel-ai/biology
|
12 |
+
- metaeval/reclor
|
13 |
+
- openbookqa
|
14 |
+
- mandyyyyii/scibench
|
15 |
+
- derek-thomas/ScienceQA
|
16 |
+
- wenhu/TheoremQA
|
17 |
+
- TIGER-Lab/ScienceEval
|
18 |
---
|
19 |
|
20 |
+
# 🔬 Einstein-7B
|
21 |
+
|
22 |
+
This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on datasets related to science.
|
23 |
+
|
24 |
+
This model is fine-tuned using [QLoRa](https://arxiv.org/abs/2305.14314) and [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
|
25 |
+
|
26 |
+
This model's training was sponsored by [sablo.ai](https://sablo.ai).
|
27 |
|
|
|
28 |
<details><summary>See axolotl config</summary>
|
29 |
|
30 |
axolotl version: `0.3.0`
|
|
|
152 |
|
153 |
</details><br>
|
154 |
|
155 |
+
# 📊 Datasets
|
156 |
+
|
157 |
+
Following datasets were used in this model:
|
158 |
+
|
159 |
+
- [ARC](https://huggingface.co/datasets/allenai/ai2_arc) (Note: Only **train** part)
|
160 |
|
161 |
+
- [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
|
162 |
|
163 |
+
- [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
|
164 |
|
165 |
+
- [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
|
166 |
|
167 |
+
- [openbookqa](https://huggingface.co/datasets/openbookqa)
|
168 |
|
169 |
+
- [reclor](https://huggingface.co/datasets/metaeval/reclor)
|
170 |
|
171 |
+
- [scibench](https://github.com/mandyyyyii/scibench)
|
172 |
|
173 |
+
- [ScienceQA](https://huggingface.co/datasets/derek-thomas/ScienceQA)
|
174 |
|
175 |
+
- [TheoremQA](https://huggingface.co/datasets/wenhu/TheoremQA)
|
176 |
|
177 |
+
- [ScienceEval](https://huggingface.co/datasets/TIGER-Lab/ScienceEval)
|
178 |
|
179 |
+
# 💬 Prompt Template
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
180 |
|
181 |
+
You can use this prompt template while using the model:
|
182 |
|
183 |
+
### Alpaca
|
184 |
+
|
185 |
+
```
|
186 |
+
Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
|
187 |
+
|
188 |
+
### Instruction:
|
189 |
+
{instruction}
|
190 |
+
|
191 |
+
### Input:
|
192 |
+
{input}
|
193 |
+
|
194 |
+
### Response:
|
195 |
+
```
|
196 |
|
197 |
+
# 🤝 Acknowledgments
|
198 |
|
199 |
+
Thanks to Platypus for providing scripts to convert some of the datasets to Alpaca format: [Platypus/data_pipeline](https://github.com/arielnlee/Platypus/tree/main/data_pipeline)
|
200 |
|
201 |
+
Thanks to all the dataset authors mentioned in the datasets section.
|
202 |
|
203 |
+
Thanks to [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) for making the repository I used to make this model.
|
204 |
|
205 |
+
[<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
|
|
|
|
|
|
|
|