Weyaxi commited on
Commit
b882398
·
verified ·
1 Parent(s): 36f6450

create model card

Browse files
Files changed (1) hide show
  1. README.md +52 -42
README.md CHANGED
@@ -4,12 +4,27 @@ tags:
4
  - axolotl
5
  - generated_from_trainer
6
  base_model: mistralai/Mistral-7B-v0.1
 
 
 
 
 
 
 
 
 
 
 
7
  ---
8
 
9
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
- should probably proofread and complete it, then remove this comment. -->
 
 
 
 
 
11
 
12
- [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
13
  <details><summary>See axolotl config</summary>
14
 
15
  axolotl version: `0.3.0`
@@ -137,59 +152,54 @@ special_tokens:
137
 
138
  </details><br>
139
 
140
- # science-mistral
 
 
 
 
141
 
142
- This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on the None dataset.
143
 
144
- ## Model description
145
 
146
- More information needed
147
 
148
- ## Intended uses & limitations
149
 
150
- More information needed
151
 
152
- ## Training and evaluation data
153
 
154
- More information needed
155
 
156
- ## Training procedure
157
 
 
158
 
159
- The following `bitsandbytes` quantization config was used during training:
160
- - quant_method: bitsandbytes
161
- - load_in_8bit: False
162
- - load_in_4bit: True
163
- - llm_int8_threshold: 6.0
164
- - llm_int8_skip_modules: None
165
- - llm_int8_enable_fp32_cpu_offload: False
166
- - llm_int8_has_fp16_weight: False
167
- - bnb_4bit_quant_type: nf4
168
- - bnb_4bit_use_double_quant: True
169
- - bnb_4bit_compute_dtype: bfloat16
170
 
171
- ### Training hyperparameters
172
 
173
- The following hyperparameters were used during training:
174
- - learning_rate: 0.0002
175
- - train_batch_size: 6
176
- - eval_batch_size: 6
177
- - seed: 42
178
- - gradient_accumulation_steps: 12
179
- - total_train_batch_size: 72
180
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
181
- - lr_scheduler_type: cosine
182
- - lr_scheduler_warmup_steps: 10
183
- - num_epochs: 2
 
 
184
 
185
- ### Training results
186
 
 
187
 
 
188
 
189
- ### Framework versions
190
 
191
- - PEFT 0.7.0
192
- - Transformers 4.37.0
193
- - Pytorch 2.0.1+cu118
194
- - Datasets 2.16.1
195
- - Tokenizers 0.15.0
 
4
  - axolotl
5
  - generated_from_trainer
6
  base_model: mistralai/Mistral-7B-v0.1
7
+ datasets:
8
+ - allenai/ai2_arc
9
+ - camel-ai/physics
10
+ - camel-ai/chemistry
11
+ - camel-ai/biology
12
+ - metaeval/reclor
13
+ - openbookqa
14
+ - mandyyyyii/scibench
15
+ - derek-thomas/ScienceQA
16
+ - wenhu/TheoremQA
17
+ - TIGER-Lab/ScienceEval
18
  ---
19
 
20
+ # 🔬 Einstein-7B
21
+
22
+ This model is a fine-tuned version of [mistralai/Mistral-7B-v0.1](https://huggingface.co/mistralai/Mistral-7B-v0.1) on datasets related to science.
23
+
24
+ This model is fine-tuned using [QLoRa](https://arxiv.org/abs/2305.14314) and [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl).
25
+
26
+ This model's training was sponsored by [sablo.ai](https://sablo.ai).
27
 
 
28
  <details><summary>See axolotl config</summary>
29
 
30
  axolotl version: `0.3.0`
 
152
 
153
  </details><br>
154
 
155
+ # 📊 Datasets
156
+
157
+ Following datasets were used in this model:
158
+
159
+ - [ARC](https://huggingface.co/datasets/allenai/ai2_arc) (Note: Only **train** part)
160
 
161
+ - [camel-ai/physics](https://huggingface.co/datasets/camel-ai/physics)
162
 
163
+ - [camel-ai/chemistry](https://huggingface.co/datasets/camel-ai/chemistry)
164
 
165
+ - [camel-ai/biology](https://huggingface.co/datasets/camel-ai/biology)
166
 
167
+ - [openbookqa](https://huggingface.co/datasets/openbookqa)
168
 
169
+ - [reclor](https://huggingface.co/datasets/metaeval/reclor)
170
 
171
+ - [scibench](https://github.com/mandyyyyii/scibench)
172
 
173
+ - [ScienceQA](https://huggingface.co/datasets/derek-thomas/ScienceQA)
174
 
175
+ - [TheoremQA](https://huggingface.co/datasets/wenhu/TheoremQA)
176
 
177
+ - [ScienceEval](https://huggingface.co/datasets/TIGER-Lab/ScienceEval)
178
 
179
+ # 💬 Prompt Template
 
 
 
 
 
 
 
 
 
 
180
 
181
+ You can use this prompt template while using the model:
182
 
183
+ ### Alpaca
184
+
185
+ ```
186
+ Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request.
187
+
188
+ ### Instruction:
189
+ {instruction}
190
+
191
+ ### Input:
192
+ {input}
193
+
194
+ ### Response:
195
+ ```
196
 
197
+ # 🤝 Acknowledgments
198
 
199
+ Thanks to Platypus for providing scripts to convert some of the datasets to Alpaca format: [Platypus/data_pipeline](https://github.com/arielnlee/Platypus/tree/main/data_pipeline)
200
 
201
+ Thanks to all the dataset authors mentioned in the datasets section.
202
 
203
+ Thanks to [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) for making the repository I used to make this model.
204
 
205
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)