benhaotang commited on
Commit
27c8f63
·
verified ·
1 Parent(s): d5969b3

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -14
README.md CHANGED
@@ -8,9 +8,7 @@ datasets:
8
 
9
  # Mistral Physics Fine-tuned Model
10
 
11
- This model is a fine-tuned version of [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409) on [kejian/arxiv-physics-debug-v0](https://huggingface.co/datasets/kejian/arxiv-physics-debug-v0). Mostly for concept proofing, don't trust it for real physics (I mean, even Claude 3.5 can be wrong on graduate physics plenty of times, let alone a 22B model, but this hould perform a lot better than [benhaotang/llama3.2-1B-physics-finetuned](https://huggingface.co/benhaotang/llama3.2-1B-physics-finetuned))!
12
-
13
- Sorry for not having F16 version, there is no way to fit everything into VRAM or RAM at the same time in my current configuration.
14
 
15
  ## Model description
16
  - Base model: [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409)
@@ -27,17 +25,7 @@ Sorry for not having F16 version, there is no way to fit everything into VRAM or
27
  ```python
28
  from transformers import AutoModelForCausalLM, BitsAndBytesConfig
29
  import torch
30
- bnb_config = BitsAndBytesConfig(
31
- load_in_8bit=False,
32
- llm_int8_enable_fp32_cpu_offload=True
33
- )
34
- model = AutoModelForCausalLM.from_pretrained(
35
- "benhaotang/mistral-small-physics-finetuned-bnb-4bit",
36
- device_map="auto",
37
- torch_dtype=torch.float16,
38
- offload_folder="offload_folder",
39
- quantization_config=bnb_config
40
- )
41
  tokenizer = AutoTokenizer.from_pretrained("benhaotang/mistral-small-physics-finetuned-bnb-4bit")
42
 
43
  # Example usage
 
8
 
9
  # Mistral Physics Fine-tuned Model
10
 
11
+ This model is a Lora adapter to [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409) finetuned on [kejian/arxiv-physics-debug-v0](https://huggingface.co/datasets/kejian/arxiv-physics-debug-v0). Mostly for concept proofing, don't trust it for real physics (I mean, even Claude 3.5 can be wrong on graduate physics plenty of times, let alone a 22B model, but this hould perform a lot better than [benhaotang/llama3.2-1B-physics-finetuned](https://huggingface.co/benhaotang/llama3.2-1B-physics-finetuned))!
 
 
12
 
13
  ## Model description
14
  - Base model: [mistralai/Mistral-Small-Instruct-2409](https://huggingface.co/mistralai/Mistral-Small-Instruct-2409)
 
25
  ```python
26
  from transformers import AutoModelForCausalLM, BitsAndBytesConfig
27
  import torch
28
+ model = AutoPeftModelForCausalLM.from_pretrained("benhaotang/mistral-small-physics-finetuned-adapter",device_map="auto",torch_dtype=torch.float16)
 
 
 
 
 
 
 
 
 
 
29
  tokenizer = AutoTokenizer.from_pretrained("benhaotang/mistral-small-physics-finetuned-bnb-4bit")
30
 
31
  # Example usage