Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -8,8 +8,9 @@ from transformers import pipeline
|
|
8 |
#pipe = pipeline("text-generation", model="microsoft/Phi-3-mini-128k-instruct", trust_remote_code=True)
|
9 |
|
10 |
#client = InferenceClient("microsoft/Phi-3-mini-128k-instruct")
|
11 |
-
client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
|
12 |
#client = InferenceClient("google/gemma-1.1-7b-it")
|
|
|
13 |
|
14 |
|
15 |
@spaces.GPU
|
@@ -33,7 +34,7 @@ def respond(
|
|
33 |
|
34 |
response = ""
|
35 |
|
36 |
-
for message in
|
37 |
messages,
|
38 |
max_tokens=max_tokens,
|
39 |
stream=True,
|
|
|
8 |
#pipe = pipeline("text-generation", model="microsoft/Phi-3-mini-128k-instruct", trust_remote_code=True)
|
9 |
|
10 |
#client = InferenceClient("microsoft/Phi-3-mini-128k-instruct")
|
11 |
+
#client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
|
12 |
#client = InferenceClient("google/gemma-1.1-7b-it")
|
13 |
+
pipe = pipeline("text-generation", model="segolilylabs/Lily-Cybersecurity-7B-v0.2")
|
14 |
|
15 |
|
16 |
@spaces.GPU
|
|
|
34 |
|
35 |
response = ""
|
36 |
|
37 |
+
for message in pipe.chat_completion(
|
38 |
messages,
|
39 |
max_tokens=max_tokens,
|
40 |
stream=True,
|