Spaces:
Running
on
T4
Running
on
T4
model changed to Haiku 3.5 for query re=write
Browse files- utilities/invoke_models.py +10 -3
utilities/invoke_models.py
CHANGED
@@ -59,13 +59,16 @@ def invoke_model_mm(text,img):
|
|
59 |
return response_body.get("embedding")
|
60 |
|
61 |
def invoke_llm_model(input,is_stream,model_id="us.anthropic.claude-3-5-haiku-20241022-v1:0"):
|
62 |
-
|
|
|
|
|
|
|
63 |
if(is_stream == False):
|
64 |
response = bedrock_runtime_client.invoke_model(
|
65 |
modelId= model_id,#"us.anthropic.claude-3-5-haiku-20241022-v1:0",#"anthropic.claude-3-5-haiku-20241022-v1:0",#"anthropic.claude-3-5-sonnet-20240620-v1:0",,
|
66 |
contentType = "application/json",
|
67 |
accept = "application/json",
|
68 |
-
performanceConfigLatency=
|
69 |
|
70 |
body = json.dumps({
|
71 |
"anthropic_version": "bedrock-2023-05-31",
|
@@ -128,12 +131,16 @@ def read_from_table(file,question):
|
|
128 |
return agent_res
|
129 |
|
130 |
def generate_image_captions_llm(base64_string,question,model_id="us.anthropic.claude-3-5-haiku-20241022-v1:0"):
|
131 |
-
|
|
|
|
|
|
|
132 |
|
133 |
response = bedrock_runtime_client.invoke_model(
|
134 |
modelId= model_id,
|
135 |
contentType = "application/json",
|
136 |
accept = "application/json",
|
|
|
137 |
|
138 |
body = json.dumps({
|
139 |
"anthropic_version": "bedrock-2023-05-31",
|
|
|
59 |
return response_body.get("embedding")
|
60 |
|
61 |
def invoke_llm_model(input,is_stream,model_id="us.anthropic.claude-3-5-haiku-20241022-v1:0"):
|
62 |
+
if(model_id == "us.anthropic.claude-3-5-haiku-20241022-v1:0"):
|
63 |
+
performanceConfigLatency = 'optimized'
|
64 |
+
else:
|
65 |
+
performanceConfigLatency = 'standard'
|
66 |
if(is_stream == False):
|
67 |
response = bedrock_runtime_client.invoke_model(
|
68 |
modelId= model_id,#"us.anthropic.claude-3-5-haiku-20241022-v1:0",#"anthropic.claude-3-5-haiku-20241022-v1:0",#"anthropic.claude-3-5-sonnet-20240620-v1:0",,
|
69 |
contentType = "application/json",
|
70 |
accept = "application/json",
|
71 |
+
performanceConfigLatency=performanceConfigLatency,
|
72 |
|
73 |
body = json.dumps({
|
74 |
"anthropic_version": "bedrock-2023-05-31",
|
|
|
131 |
return agent_res
|
132 |
|
133 |
def generate_image_captions_llm(base64_string,question,model_id="us.anthropic.claude-3-5-haiku-20241022-v1:0"):
|
134 |
+
if(model_id == "us.anthropic.claude-3-5-haiku-20241022-v1:0"):
|
135 |
+
performanceConfigLatency = 'optimized'
|
136 |
+
else:
|
137 |
+
performanceConfigLatency = 'standard'
|
138 |
|
139 |
response = bedrock_runtime_client.invoke_model(
|
140 |
modelId= model_id,
|
141 |
contentType = "application/json",
|
142 |
accept = "application/json",
|
143 |
+
performanceConfigLatency=performanceConfigLatency,
|
144 |
|
145 |
body = json.dumps({
|
146 |
"anthropic_version": "bedrock-2023-05-31",
|