Commit
·
a1800a5
1
Parent(s):
1e856c7
added custom handler v5
Browse files- handler.py +5 -4
handler.py
CHANGED
@@ -28,7 +28,7 @@ class EndpointHandler:
|
|
28 |
|
29 |
# Default generation parameters
|
30 |
self.default_params = {
|
31 |
-
"max_new_tokens":
|
32 |
"temperature": 0.0,
|
33 |
"top_p": 0.9,
|
34 |
"top_k": 50,
|
@@ -55,8 +55,8 @@ class EndpointHandler:
|
|
55 |
logger.info("Processing new request")
|
56 |
logger.info(f"Input data: {data}")
|
57 |
|
58 |
-
|
59 |
-
if not
|
60 |
logger.warning("No input messages provided")
|
61 |
return [{"role": "assistant", "content": "No input messages provided"}]
|
62 |
|
@@ -65,13 +65,14 @@ class EndpointHandler:
|
|
65 |
logger.info(f"Generation parameters: {gen_params}")
|
66 |
|
67 |
# Apply the chat template
|
|
|
68 |
logger.info("Applying chat template")
|
69 |
prompt = self.tokenizer.apply_chat_template(
|
70 |
messages,
|
71 |
tokenize=False,
|
72 |
add_generation_prompt=True
|
73 |
)
|
74 |
-
logger.info(f"Generated prompt: {prompt}")
|
75 |
|
76 |
# Tokenize the prompt
|
77 |
logger.info("Tokenizing input")
|
|
|
28 |
|
29 |
# Default generation parameters
|
30 |
self.default_params = {
|
31 |
+
"max_new_tokens": 1000,
|
32 |
"temperature": 0.0,
|
33 |
"top_p": 0.9,
|
34 |
"top_k": 50,
|
|
|
55 |
logger.info("Processing new request")
|
56 |
logger.info(f"Input data: {data}")
|
57 |
|
58 |
+
input_messages = data.get("inputs", [])
|
59 |
+
if not input_messages:
|
60 |
logger.warning("No input messages provided")
|
61 |
return [{"role": "assistant", "content": "No input messages provided"}]
|
62 |
|
|
|
65 |
logger.info(f"Generation parameters: {gen_params}")
|
66 |
|
67 |
# Apply the chat template
|
68 |
+
messages = [{"role": "user", "content": input_messages}]
|
69 |
logger.info("Applying chat template")
|
70 |
prompt = self.tokenizer.apply_chat_template(
|
71 |
messages,
|
72 |
tokenize=False,
|
73 |
add_generation_prompt=True
|
74 |
)
|
75 |
+
logger.info(f"Generated chat prompt: {prompt}")
|
76 |
|
77 |
# Tokenize the prompt
|
78 |
logger.info("Tokenizing input")
|