KushwanthK commited on
Commit
4ac09cb
·
verified ·
1 Parent(s): a8ddbea

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -109,7 +109,7 @@ def prompt_engineer(text, longtext, query):
109
  sllm = HuggingFaceHub(
110
  repo_id="facebook/bart-large-cnn", model_kwargs={"temperature": 0, "max_new_tokens": 256, "task":"text-generation"}
111
  )
112
- st.write("Chat llm connection started..")
113
  except Exception as e:
114
  st.error(f"Error invoke: {e}")
115
 
@@ -159,9 +159,9 @@ def prompt_engineer(text, longtext, query):
159
 
160
  try:
161
  llm = HuggingFaceHub(
162
- repo_id="meta-llama/Meta-Llama-3-8B-Instruct", model_kwargs={"temperature": 0, "max_new_tokens": 256, "task":"text-generation"}
163
  )
164
- st.write("summary llm connection started..")
165
  response_text = llm.invoke(prompt)
166
  escaped_query = re.escape(query)
167
  result = re.split(f'Answer the question based on the above context: {escaped_query}\n',response_text)[-1]
 
109
  sllm = HuggingFaceHub(
110
  repo_id="facebook/bart-large-cnn", model_kwargs={"temperature": 0, "max_new_tokens": 256, "task":"text-generation"}
111
  )
112
+ st.write("Summary Chat llm connection started..")
113
  except Exception as e:
114
  st.error(f"Error invoke: {e}")
115
 
 
159
 
160
  try:
161
  llm = HuggingFaceHub(
162
+ repo_id="meta-llama/Meta-Llama-3-8B-Instruct", model_kwargs={"temperature": 0, "task":"text-generation"}
163
  )
164
+ st.write("GEN llm connection started..")
165
  response_text = llm.invoke(prompt)
166
  escaped_query = re.escape(query)
167
  result = re.split(f'Answer the question based on the above context: {escaped_query}\n',response_text)[-1]