Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -111,8 +111,9 @@ class LLMResponder:
|
|
111 |
self.llm = Llama(model_path="path/to/your/gguf/file.gguf", n_ctx=1024)
|
112 |
self.backend = "llama_cpp"
|
113 |
else:
|
114 |
-
|
115 |
-
self.
|
|
|
116 |
self.backend = "transformers"
|
117 |
|
118 |
def generate_response(self, prompt):
|
@@ -189,4 +190,5 @@ iface = gr.Interface(
|
|
189 |
)
|
190 |
|
191 |
if __name__ == "__main__":
|
192 |
-
|
|
|
|
111 |
self.llm = Llama(model_path="path/to/your/gguf/file.gguf", n_ctx=1024)
|
112 |
self.backend = "llama_cpp"
|
113 |
else:
|
114 |
+
# Use trust_remote_code=True to load the model despite missing config keys.
|
115 |
+
self.llm_tokenizer = AutoTokenizer.from_pretrained(model_name, trust_remote_code=True)
|
116 |
+
self.llm_model = AutoModelForCausalLM.from_pretrained(model_name, trust_remote_code=True)
|
117 |
self.backend = "transformers"
|
118 |
|
119 |
def generate_response(self, prompt):
|
|
|
190 |
)
|
191 |
|
192 |
if __name__ == "__main__":
|
193 |
+
# In Hugging Face Spaces, remove share=True.
|
194 |
+
iface.launch()
|