cedpsam commited on
Commit
27493f6
·
1 Parent(s): 093b59a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -4
app.py CHANGED
@@ -17,9 +17,7 @@ hf_hub_download(repo_id=repo_id,
17
 
18
  llm = LlamaCpp(
19
  model_path=model_name,
20
- temperature=temperature,
21
- max_tokens=2000,
22
- top_p=top_p,
23
  callback_manager=callback_manager,
24
  verbose=True, # Verbose is required to pass to the callback manager
25
  )
@@ -46,7 +44,10 @@ def generate(
46
 
47
  # stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
48
  output = ""
49
- output=llm(formatted_prompt)
 
 
 
50
  output="ans:"+output
51
  # for response in stream:
52
  # output += response.token.text
 
17
 
18
  llm = LlamaCpp(
19
  model_path=model_name,
20
+
 
 
21
  callback_manager=callback_manager,
22
  verbose=True, # Verbose is required to pass to the callback manager
23
  )
 
44
 
45
  # stream = client.text_generation(formatted_prompt, **generate_kwargs, stream=True, details=True, return_full_text=False)
46
  output = ""
47
+ output=llm(formatted_prompt,
48
+ temperature=temperature,
49
+ max_tokens=30,
50
+ top_p=top_p,)
51
  output="ans:"+output
52
  # for response in stream:
53
  # output += response.token.text