Commit
·
05c88c1
1
Parent(s):
45da339
disk
Browse files
app.py
CHANGED
@@ -114,14 +114,14 @@ def download_and_merge_model(base_model_name, lora_model_name, output_dir, devic
|
|
114 |
"""
|
115 |
os.makedirs("temp", exist_ok=True)
|
116 |
log.info("Loading base model...")
|
117 |
-
model = AutoModelForCausalLM.from_pretrained(base_model_name, low_cpu_mem_usage=True, trust_remote_code=True, offload_folder="temp")
|
118 |
log.info("Loading adapter tokenizer...")
|
119 |
-
adapter_tokenizer = AutoTokenizer.from_pretrained(lora_model_name, trust_remote_code=True, offload_folder="temp")
|
120 |
log.info("Resizing token embeddings...")
|
121 |
added_tokens_decoder = adapter_tokenizer.added_tokens_decoder
|
122 |
model.resize_token_embeddings(adapter_tokenizer.vocab_size + len(added_tokens_decoder))
|
123 |
log.info("Loading LoRA adapter...")
|
124 |
-
peft_model = PeftModel.from_pretrained(model, lora_model_name, low_cpu_mem_usage=True, trust_remote_code=True, offload_folder="temp")
|
125 |
log.info("Merging and unloading model...")
|
126 |
model = peft_model.merge_and_unload()
|
127 |
log.info("Saving model...")
|
@@ -337,7 +337,7 @@ def create_ui():
|
|
337 |
)
|
338 |
quant_method = gr.CheckboxGroup(
|
339 |
choices=["Q2_K", "Q4_K", "IQ4_NL", "Q5_K_M", "Q6_K", "Q8_0"],
|
340 |
-
value=["Q4_K", "Q8_0"],
|
341 |
label="Quantization Method"
|
342 |
)
|
343 |
hf_token = gr.Textbox(
|
|
|
114 |
"""
|
115 |
os.makedirs("temp", exist_ok=True)
|
116 |
log.info("Loading base model...")
|
117 |
+
model = AutoModelForCausalLM.from_pretrained(base_model_name, low_cpu_mem_usage=True, device_map="disk", trust_remote_code=True, offload_folder="temp")
|
118 |
log.info("Loading adapter tokenizer...")
|
119 |
+
adapter_tokenizer = AutoTokenizer.from_pretrained(lora_model_name, trust_remote_code=True, device_map="disk", offload_folder="temp")
|
120 |
log.info("Resizing token embeddings...")
|
121 |
added_tokens_decoder = adapter_tokenizer.added_tokens_decoder
|
122 |
model.resize_token_embeddings(adapter_tokenizer.vocab_size + len(added_tokens_decoder))
|
123 |
log.info("Loading LoRA adapter...")
|
124 |
+
peft_model = PeftModel.from_pretrained(model, lora_model_name, low_cpu_mem_usage=True, device_map="disk", trust_remote_code=True, offload_folder="temp")
|
125 |
log.info("Merging and unloading model...")
|
126 |
model = peft_model.merge_and_unload()
|
127 |
log.info("Saving model...")
|
|
|
337 |
)
|
338 |
quant_method = gr.CheckboxGroup(
|
339 |
choices=["Q2_K", "Q4_K", "IQ4_NL", "Q5_K_M", "Q6_K", "Q8_0"],
|
340 |
+
value=["Q2_K", "Q4_K", "IQ4_NL", "Q5_K_M", "Q6_K", "Q8_0"],
|
341 |
label="Quantization Method"
|
342 |
)
|
343 |
hf_token = gr.Textbox(
|