Spaces:
Sleeping
Sleeping
Commit
·
d469f0d
1
Parent(s):
56599c7
error handling
Browse files
app.py
CHANGED
@@ -56,11 +56,14 @@ def generate_response(user_input, model_id):
|
|
56 |
truncation=True
|
57 |
)
|
58 |
|
59 |
-
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
|
|
|
|
|
|
|
64 |
|
65 |
def formatted_prompt(question) -> str:
|
66 |
return f"<|im_start|>user\n{question}<|im_end|>\n<|im_start|>assistant:"
|
@@ -79,13 +82,16 @@ def handle_post_request():
|
|
79 |
message = data.get("inputs", "No message provided.")
|
80 |
model_id = data.get("model_id", "YALCINKAYA/opsgenius-large") # Default model if not provided
|
81 |
|
82 |
-
|
83 |
-
|
84 |
-
|
85 |
-
|
86 |
-
|
87 |
-
|
88 |
-
|
|
|
|
|
|
|
89 |
|
90 |
if __name__ == '__main__':
|
91 |
app.run(host='0.0.0.0', port=7860)
|
|
|
56 |
truncation=True
|
57 |
)
|
58 |
|
59 |
+
try:
|
60 |
+
# Generate response
|
61 |
+
outputs = model.generate(**inputs, generation_config=generation_config)
|
62 |
+
response = tokenizer.decode(outputs[0], skip_special_tokens=True)
|
63 |
+
return response
|
64 |
+
except Exception as e:
|
65 |
+
print(f"Error generating response: {e}")
|
66 |
+
return "Error generating response."
|
67 |
|
68 |
def formatted_prompt(question) -> str:
|
69 |
return f"<|im_start|>user\n{question}<|im_end|>\n<|im_start|>assistant:"
|
|
|
82 |
message = data.get("inputs", "No message provided.")
|
83 |
model_id = data.get("model_id", "YALCINKAYA/opsgenius-large") # Default model if not provided
|
84 |
|
85 |
+
try:
|
86 |
+
# Generate a response from the model
|
87 |
+
model_response = generate_response(message, model_id)
|
88 |
+
return jsonify({
|
89 |
+
"received_message": model_response,
|
90 |
+
"status": "POST request successful!"
|
91 |
+
})
|
92 |
+
except Exception as e:
|
93 |
+
print(f"Error handling POST request: {e}")
|
94 |
+
return jsonify({"error": "An error occurred while processing your request."}), 500
|
95 |
|
96 |
if __name__ == '__main__':
|
97 |
app.run(host='0.0.0.0', port=7860)
|