Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,6 +1,6 @@
|
|
1 |
from huggingface_hub import InferenceClient
|
2 |
import gradio as gr
|
3 |
-
|
4 |
client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
|
5 |
#client = InferenceClient("Trelis/Mistral-7B-Instruct-v0.1-Summarize-16k")
|
6 |
#client = InferenceClient("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
|
@@ -47,7 +47,8 @@ def generate(
|
|
47 |
top_p=top_p,
|
48 |
repetition_penalty=repetition_penalty,
|
49 |
do_sample=True,
|
50 |
-
seed=
|
|
|
51 |
)
|
52 |
|
53 |
formatted_prompt = format_prompt(f"{system}, {prompt}", history)
|
|
|
1 |
from huggingface_hub import InferenceClient
|
2 |
import gradio as gr
|
3 |
+
import random
|
4 |
client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
|
5 |
#client = InferenceClient("Trelis/Mistral-7B-Instruct-v0.1-Summarize-16k")
|
6 |
#client = InferenceClient("TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T")
|
|
|
47 |
top_p=top_p,
|
48 |
repetition_penalty=repetition_penalty,
|
49 |
do_sample=True,
|
50 |
+
seed=random.randint(1,99999999999)
|
51 |
+
#seed=42,
|
52 |
)
|
53 |
|
54 |
formatted_prompt = format_prompt(f"{system}, {prompt}", history)
|