abdfajar707 commited on
Commit
b514c5e
·
verified ·
1 Parent(s): 6a5528b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -7
app.py CHANGED
@@ -17,7 +17,7 @@ alpaca_prompt = """Berikut adalah instruksi yang deskripsikan tugas dan sepasang
17
  if True:
18
  from app import FastLanguageModel
19
  model, tokenizer = FastLanguageModel.from_pretrained(
20
- model_name = "abdfajar707/llama3_8B_lora_model_rkp_v5", # YOUR MODEL YOU USED FOR TRAINING
21
  max_seq_length = max_seq_length,
22
  dtype = dtype,
23
  load_in_4bit = load_in_4bit,
@@ -27,14 +27,10 @@ if True:
27
 
28
 
29
  # Fungsi untuk menghasilkan respons
30
- def generate_response(prompt, max_length=1000):
31
  inputs = tokenizer(
32
  [
33
- alpaca_prompt.format(
34
- prompt, # instruction
35
- "", # input
36
- "", # output - leave this blank for generation!
37
- )
38
  ], return_tensors = "pt").to("cuda")
39
  outputs = model.generate(**inputs, max_length=max_length, pad_token_id=tokenizer.eos_token_id)
40
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)
 
17
  if True:
18
  from app import FastLanguageModel
19
  model, tokenizer = FastLanguageModel.from_pretrained(
20
+ model_name = "abdfajar707/llama3_8B_lora_model_rkp_v1.1", # YOUR MODEL YOU USED FOR TRAINING
21
  max_seq_length = max_seq_length,
22
  dtype = dtype,
23
  load_in_4bit = load_in_4bit,
 
27
 
28
 
29
  # Fungsi untuk menghasilkan respons
30
+ def generate_response(prompt, max_length=4096):
31
  inputs = tokenizer(
32
  [
33
+ prompt, # instruction
 
 
 
 
34
  ], return_tensors = "pt").to("cuda")
35
  outputs = model.generate(**inputs, max_length=max_length, pad_token_id=tokenizer.eos_token_id)
36
  response = tokenizer.decode(outputs[0], skip_special_tokens=True)