Commit
·
7f4063e
1
Parent(s):
665dc6a
debug pixtral
Browse files- app.py +2 -1
- src/pipelines.py +3 -1
app.py
CHANGED
@@ -55,7 +55,8 @@ if "pipeline" not in st.session_state:
|
|
55 |
reranker_model="monovlm",
|
56 |
device=device,
|
57 |
invoice_json_path=invoice_json_path,
|
58 |
-
gpu_memory_utilization=0.
|
|
|
59 |
)
|
60 |
pipeline = st.session_state.pipeline
|
61 |
|
|
|
55 |
reranker_model="monovlm",
|
56 |
device=device,
|
57 |
invoice_json_path=invoice_json_path,
|
58 |
+
gpu_memory_utilization=0.8,
|
59 |
+
max_tokens=4096,
|
60 |
)
|
61 |
pipeline = st.session_state.pipeline
|
62 |
|
src/pipelines.py
CHANGED
@@ -19,7 +19,9 @@ class InvoiceGenerator:
|
|
19 |
reranker_model=None,
|
20 |
device="cuda",
|
21 |
invoice_json_path=None,
|
22 |
-
max_model_len=4096,
|
|
|
|
|
23 |
):
|
24 |
self.model = Pixtral(max_model_len=max_model_len, max_tokens=max_tokens,
|
25 |
gpu_memory_utilization=gpu_memory_utilization)
|
|
|
19 |
reranker_model=None,
|
20 |
device="cuda",
|
21 |
invoice_json_path=None,
|
22 |
+
max_model_len=4096,
|
23 |
+
max_tokens=2048,
|
24 |
+
gpu_memory_utilization=0.95
|
25 |
):
|
26 |
self.model = Pixtral(max_model_len=max_model_len, max_tokens=max_tokens,
|
27 |
gpu_memory_utilization=gpu_memory_utilization)
|