benhaotang
commited on
Update README.md
Browse files
README.md
CHANGED
@@ -8,9 +8,7 @@ datasets:
|
|
8 |
|
9 |
# Mistral Physics Fine-tuned Model
|
10 |
|
11 |
-
This model is a
|
12 |
-
|
13 |
-
Sorry for not having F16 version, there is no way to fit everything into VRAM or RAM at the same time in my current configuration.
|
14 |
|
15 |
## Model description
|
16 |
- Base model: [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409)
|
@@ -27,17 +25,7 @@ Sorry for not having F16 version, there is no way to fit everything into VRAM or
|
|
27 |
```python
|
28 |
from transformers import AutoModelForCausalLM, BitsAndBytesConfig
|
29 |
import torch
|
30 |
-
|
31 |
-
load_in_8bit=False,
|
32 |
-
llm_int8_enable_fp32_cpu_offload=True
|
33 |
-
)
|
34 |
-
model = AutoModelForCausalLM.from_pretrained(
|
35 |
-
"benhaotang/mistral-small-physics-finetuned-bnb-4bit",
|
36 |
-
device_map="auto",
|
37 |
-
torch_dtype=torch.float16,
|
38 |
-
offload_folder="offload_folder",
|
39 |
-
quantization_config=bnb_config
|
40 |
-
)
|
41 |
tokenizer = AutoTokenizer.from_pretrained("benhaotang/mistral-small-physics-finetuned-bnb-4bit")
|
42 |
|
43 |
# Example usage
|
|
|
8 |
|
9 |
# Mistral Physics Fine-tuned Model
|
10 |
|
11 |
+
This model is a Lora adapter to [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409) finetuned on [kejian/arxiv-physics-debug-v0](https://huggingface.co/datasets/kejian/arxiv-physics-debug-v0). Mostly for concept proofing, don't trust it for real physics (I mean, even Claude 3.5 can be wrong on graduate physics plenty of times, let alone a 22B model, but this hould perform a lot better than [benhaotang/llama3.2-1B-physics-finetuned](https://huggingface.co/benhaotang/llama3.2-1B-physics-finetuned))!
|
|
|
|
|
12 |
|
13 |
## Model description
|
14 |
- Base model: [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409)
|
|
|
25 |
```python
|
26 |
from transformers import AutoModelForCausalLM, BitsAndBytesConfig
|
27 |
import torch
|
28 |
+
model = AutoPeftModelForCausalLM.from_pretrained("benhaotang/mistral-small-physics-finetuned-adapter",device_map="auto",torch_dtype=torch.float16)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
tokenizer = AutoTokenizer.from_pretrained("benhaotang/mistral-small-physics-finetuned-bnb-4bit")
|
30 |
|
31 |
# Example usage
|