Update app.py
Browse files
app.py
CHANGED
@@ -14,8 +14,6 @@ def image_to_base64(image):
|
|
14 |
img_str = base64.b64encode(buffered.getvalue()).decode('utf-8')
|
15 |
return img_str
|
16 |
|
17 |
-
# Function to interact with LLAVA model
|
18 |
-
# Function to interact with LLAVA model
|
19 |
# Function to interact with LLAVA model
|
20 |
def respond(
|
21 |
message,
|
@@ -48,7 +46,6 @@ def respond(
|
|
48 |
messages.append({"role": "user", "content": "Image uploaded", "image": image_b64})
|
49 |
|
50 |
# Call Hugging Face model for response
|
51 |
-
response = ""
|
52 |
for message in client.chat_completion(
|
53 |
messages,
|
54 |
max_tokens=max_tokens,
|
@@ -58,9 +55,7 @@ def respond(
|
|
58 |
):
|
59 |
token = message.choices[0].delta.content
|
60 |
|
61 |
-
|
62 |
-
yield response
|
63 |
-
|
64 |
|
65 |
# Debugging print statements
|
66 |
print("Starting Gradio interface setup...")
|
|
|
14 |
img_str = base64.b64encode(buffered.getvalue()).decode('utf-8')
|
15 |
return img_str
|
16 |
|
|
|
|
|
17 |
# Function to interact with LLAVA model
|
18 |
def respond(
|
19 |
message,
|
|
|
46 |
messages.append({"role": "user", "content": "Image uploaded", "image": image_b64})
|
47 |
|
48 |
# Call Hugging Face model for response
|
|
|
49 |
for message in client.chat_completion(
|
50 |
messages,
|
51 |
max_tokens=max_tokens,
|
|
|
55 |
):
|
56 |
token = message.choices[0].delta.content
|
57 |
|
58 |
+
yield token
|
|
|
|
|
59 |
|
60 |
# Debugging print statements
|
61 |
print("Starting Gradio interface setup...")
|