Spaces:
Runtime error
Runtime error
Commit
·
c13a07b
1
Parent(s):
b83c2a3
Update app.py
Browse files
app.py
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
# from transformers import AutoModelForCausalLM, AutoTokenizer
|
3 |
|
4 |
from gpt4all import GPT4All
|
5 |
-
model = GPT4All("
|
6 |
|
7 |
# #----------------------------------------------------------------------------------------------------------------------------
|
8 |
# # !pip install -q accelerate==0.21.0 peft==0.4.0 bitsandbytes==0.40.2 transformers==4.31.0 trl==0.4.7
|
@@ -119,13 +119,13 @@ model = GPT4All("./wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin")
|
|
119 |
|
120 |
|
121 |
def generate_text(prompt):
|
122 |
-
|
123 |
# pipe = pipeline(task="text-generation", model=model, tokenizer=tokenizer, max_length=200)
|
124 |
# result = pipe(f"<s>[INST] {prompt} [/INST]")
|
125 |
# # prompt = "What is a large language model?"
|
126 |
# # input_ids = tokenizer.encode(prompt, return_tensors="pt")
|
127 |
|
128 |
-
output = model.generate(input_ids, max_length=200, num_return_sequences=1)
|
129 |
# result = tokenizer.decode(output[0], skip_special_tokens=True)
|
130 |
return result
|
131 |
|
|
|
2 |
# from transformers import AutoModelForCausalLM, AutoTokenizer
|
3 |
|
4 |
from gpt4all import GPT4All
|
5 |
+
model = GPT4All("wizardlm-13b-v1.1-superhot-8k.ggmlv3.q4_0.bin")
|
6 |
|
7 |
# #----------------------------------------------------------------------------------------------------------------------------
|
8 |
# # !pip install -q accelerate==0.21.0 peft==0.4.0 bitsandbytes==0.40.2 transformers==4.31.0 trl==0.4.7
|
|
|
119 |
|
120 |
|
121 |
def generate_text(prompt):
|
122 |
+
result = model.generate(prompt)
|
123 |
# pipe = pipeline(task="text-generation", model=model, tokenizer=tokenizer, max_length=200)
|
124 |
# result = pipe(f"<s>[INST] {prompt} [/INST]")
|
125 |
# # prompt = "What is a large language model?"
|
126 |
# # input_ids = tokenizer.encode(prompt, return_tensors="pt")
|
127 |
|
128 |
+
# output = model.generate(input_ids, max_length=200, num_return_sequences=1)
|
129 |
# result = tokenizer.decode(output[0], skip_special_tokens=True)
|
130 |
return result
|
131 |
|