|
import os
|
|
|
|
os.environ["CUDA_VISIBLE_DEVICES"] = "0"
|
|
|
|
|
|
import torch
|
|
import numpy as np
|
|
|
|
from transformers import AutoModelForSequenceClassification, AutoModelForMultipleChoice, AutoModel
|
|
from transformers import TrainingArguments, Trainer
|
|
from transformers import AutoTokenizer
|
|
from transformers import DataCollatorWithPadding
|
|
from datasets import load_dataset
|
|
from datasets import load_metric
|
|
import torchsnooper
|
|
|
|
model_name = "THUDM/chatglm-6b"
|
|
|
|
if __name__ == "__main__":
|
|
device = "cuda" if torch.cuda.is_available() else "cpu"
|
|
tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
|
model = AutoModel.from_pretrained(model_name, trust_remote_code=True).half().cuda()
|
|
response, history = model.chat(tokenizer, "你好", history=[])
|
|
print(response)
|
|
response, history = model.chat(tokenizer, "晚上睡不着应该怎么办", history=history)
|
|
print(response)
|
|
|
|
|
|
|