Spaces:
Paused
Paused
Update app.py
Browse files
app.py
CHANGED
@@ -1,15 +1,17 @@
|
|
|
|
1 |
from transformers import AutoModelForCausalLM, AutoTokenizer, Trainer, TrainingArguments
|
2 |
from peft import LoraConfig, get_peft_model, TaskType
|
3 |
from datasets import load_dataset
|
4 |
import torch
|
5 |
-
from accelerate import Accelerator
|
6 |
|
7 |
def main():
|
8 |
-
|
|
|
|
|
9 |
# 基础模型位置
|
10 |
model_name = "dushuai112233/Qwen2-1.5B-Instruct"
|
11 |
# 设备
|
12 |
-
device =
|
13 |
|
14 |
# 加载分词器和模型
|
15 |
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
@@ -54,6 +56,9 @@ def main():
|
|
54 |
save_total_limit=2, # 最大保存模型数
|
55 |
)
|
56 |
|
|
|
|
|
|
|
57 |
# Define the Trainer
|
58 |
trainer = Trainer(
|
59 |
model=model, # 训练的模型
|
@@ -61,7 +66,6 @@ def main():
|
|
61 |
train_dataset=train_dataset, # 训练数据集
|
62 |
eval_dataset=val_dataset, # 验证数据集
|
63 |
tokenizer=tokenizer, # 用于预处理的分词器
|
64 |
-
accelerator=accelerator,
|
65 |
)
|
66 |
|
67 |
# Start Training
|
|
|
1 |
+
from accelerate import Accelerator
|
2 |
from transformers import AutoModelForCausalLM, AutoTokenizer, Trainer, TrainingArguments
|
3 |
from peft import LoraConfig, get_peft_model, TaskType
|
4 |
from datasets import load_dataset
|
5 |
import torch
|
|
|
6 |
|
7 |
def main():
|
8 |
+
# 初始化 Accelerator
|
9 |
+
accelerator = Accelerator()
|
10 |
+
|
11 |
# 基础模型位置
|
12 |
model_name = "dushuai112233/Qwen2-1.5B-Instruct"
|
13 |
# 设备
|
14 |
+
device = accelerator.device
|
15 |
|
16 |
# 加载分词器和模型
|
17 |
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
|
|
56 |
save_total_limit=2, # 最大保存模型数
|
57 |
)
|
58 |
|
59 |
+
# 将模型移到设备
|
60 |
+
model.to(device)
|
61 |
+
|
62 |
# Define the Trainer
|
63 |
trainer = Trainer(
|
64 |
model=model, # 训练的模型
|
|
|
66 |
train_dataset=train_dataset, # 训练数据集
|
67 |
eval_dataset=val_dataset, # 验证数据集
|
68 |
tokenizer=tokenizer, # 用于预处理的分词器
|
|
|
69 |
)
|
70 |
|
71 |
# Start Training
|