Sephfox commited on
Commit
dd1aea5
·
verified ·
1 Parent(s): e67a430

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -3
app.py CHANGED
@@ -111,8 +111,9 @@ class LLMResponder:
111
  self.llm = Llama(model_path="path/to/your/gguf/file.gguf", n_ctx=1024)
112
  self.backend = "llama_cpp"
113
  else:
114
- self.llm_tokenizer = AutoTokenizer.from_pretrained(model_name)
115
- self.llm_model = AutoModelForCausalLM.from_pretrained(model_name)
 
116
  self.backend = "transformers"
117
 
118
  def generate_response(self, prompt):
@@ -189,4 +190,5 @@ iface = gr.Interface(
189
  )
190
 
191
  if __name__ == "__main__":
192
- iface.launch(share=True)
 
 
111
  self.llm = Llama(model_path="path/to/your/gguf/file.gguf", n_ctx=1024)
112
  self.backend = "llama_cpp"
113
  else:
114
+ # Use trust_remote_code=True to load the model despite missing config keys.
115
+ self.llm_tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
116
+ self.llm_model = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
117
  self.backend = "transformers"
118
 
119
  def generate_response(self, prompt):
 
190
  )
191
 
192
  if __name__ == "__main__":
193
+ # In Hugging Face Spaces, remove share=True.
194
+ iface.launch()