from unsloth import FastLanguageModel | |
model, tokenizer = FastLanguageModel.from_pretrained( | |
model_name = "aidando73/llama-3.3-70b-instruct-code-agent-fine-tune-v1", | |
max_seq_length = 2048, | |
dtype = "float16", | |
load_in_4bit = True, | |
) | |
FastLanguageModel.for_inference(model) | |