jatingocodeo commited on
Commit
0bdc84a
·
verified ·
1 Parent(s): 6ae1346

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +11 -6
app.py CHANGED
@@ -43,18 +43,23 @@ Description: """
43
  # Generate response
44
  with torch.no_grad():
45
  outputs = model.generate(
46
- **inputs,
 
47
  max_new_tokens=max_length,
48
  temperature=temperature,
49
  top_p=top_p,
50
- num_return_sequences=1,
51
- pad_token_id=tokenizer.eos_token_id,
52
- do_sample=True
 
 
 
 
53
  )
54
 
55
  # Decode and return the response
56
- full_text = tokenizer.decode(outputs[0], skip_special_tokens=True)
57
- return full_text.split("Description: ")[-1].strip()
58
 
59
  def create_demo(model_id):
60
  # Load model and tokenizer
 
43
  # Generate response
44
  with torch.no_grad():
45
  outputs = model.generate(
46
+ input_ids=inputs['input_ids'], # Explicitly use input_ids
47
+ attention_mask=inputs['attention_mask'], # Add attention mask
48
  max_new_tokens=max_length,
49
  temperature=temperature,
50
  top_p=top_p,
51
+ do_sample=True,
52
+ pad_token_id=tokenizer.pad_token_id,
53
+ bos_token_id=tokenizer.bos_token_id,
54
+ eos_token_id=tokenizer.eos_token_id,
55
+ use_cache=True, # Enable caching
56
+ return_dict_in_generate=True, # Return as dict
57
+ output_scores=True # Get scores
58
  )
59
 
60
  # Decode and return the response
61
+ generated_text = tokenizer.decode(outputs.sequences[0], skip_special_tokens=True)
62
+ return generated_text.split("Description: ")[-1].strip()
63
 
64
  def create_demo(model_id):
65
  # Load model and tokenizer