Spaces:
Sleeping
Sleeping
model bug fix
Browse files
app.py
CHANGED
@@ -28,10 +28,11 @@ def get_models(llama=False):
|
|
28 |
access_token = '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'
|
29 |
llama_tokenizer = AutoTokenizer.from_pretrained(model_name, use_auth_token=access_token, use_fast=True)#, use_fast=True)
|
30 |
llama_model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=access_token, device_map={'':0})#, load_in_4bit=True)
|
31 |
-
|
32 |
-
|
33 |
-
|
34 |
-
|
|
|
35 |
|
36 |
model, tokenizer = get_models()
|
37 |
|
|
|
28 |
access_token = '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'
|
29 |
llama_tokenizer = AutoTokenizer.from_pretrained(model_name, use_auth_token=access_token, use_fast=True)#, use_fast=True)
|
30 |
llama_model = AutoModelForCausalLM.from_pretrained(model_name, use_auth_token=access_token, device_map={'':0})#, load_in_4bit=True)
|
31 |
+
st.write("The assistant is loaded and ready to use!")
|
32 |
+
return model, tokenizer, llama_model, llama_tokenizer
|
33 |
+
|
34 |
+
else:
|
35 |
+
return model, tokenizer
|
36 |
|
37 |
model, tokenizer = get_models()
|
38 |
|