connect ai
Browse files- ocr/api/message/db_requests.py +1 -1
- ocr/api/message/views.py +10 -4
- ocr/api/openai_requests.py +48 -0
- ocr/api/prompts.py +86 -0
- ocr/api/report/db_requests.py +7 -2
- ocr/api/report/model.py +1 -1
- ocr/api/report/views.py +21 -5
- ocr/api/utils.py +2 -4
ocr/api/message/db_requests.py
CHANGED
@@ -38,4 +38,4 @@ async def save_assistant_user_message(user_message: str, assistant_message: str,
|
|
38 |
await settings.DB_CLIENT.messages.insert_one(user_message.to_mongo())
|
39 |
await settings.DB_CLIENT.messages.insert_one(assistant_message.to_mongo())
|
40 |
|
41 |
-
return
|
|
|
38 |
await settings.DB_CLIENT.messages.insert_one(user_message.to_mongo())
|
39 |
await settings.DB_CLIENT.messages.insert_one(assistant_message.to_mongo())
|
40 |
|
41 |
+
return assistant_message
|
ocr/api/message/views.py
CHANGED
@@ -1,7 +1,11 @@
|
|
|
|
|
|
1 |
from ocr.api.message import message_router
|
2 |
from ocr.api.message.db_requests import get_all_chat_messages_obj, save_assistant_user_message
|
3 |
from ocr.api.message.models import MessageModel
|
4 |
from ocr.api.message.schemas import AllMessageWrapper, AllMessageResponse, CreateMessageRequest
|
|
|
|
|
5 |
from ocr.api.report.dto import Paging
|
6 |
from ocr.api.utils import transform_messages_to_openai
|
7 |
from ocr.core.wrappers import OcrResponseWrapper
|
@@ -24,9 +28,11 @@ async def create_message(
|
|
24 |
reportId: str,
|
25 |
message_data: CreateMessageRequest,
|
26 |
) -> OcrResponseWrapper[MessageModel]:
|
27 |
-
messages = await
|
28 |
-
|
29 |
-
|
30 |
-
|
|
|
|
|
31 |
response = await save_assistant_user_message(message_data.text, response, reportId)
|
32 |
return OcrResponseWrapper(data=response)
|
|
|
1 |
+
import asyncio
|
2 |
+
|
3 |
from ocr.api.message import message_router
|
4 |
from ocr.api.message.db_requests import get_all_chat_messages_obj, save_assistant_user_message
|
5 |
from ocr.api.message.models import MessageModel
|
6 |
from ocr.api.message.schemas import AllMessageWrapper, AllMessageResponse, CreateMessageRequest
|
7 |
+
from ocr.api.openai_requests import generate_agent_response
|
8 |
+
from ocr.api.report.db_requests import get_report_obj_by_id
|
9 |
from ocr.api.report.dto import Paging
|
10 |
from ocr.api.utils import transform_messages_to_openai
|
11 |
from ocr.core.wrappers import OcrResponseWrapper
|
|
|
28 |
reportId: str,
|
29 |
message_data: CreateMessageRequest,
|
30 |
) -> OcrResponseWrapper[MessageModel]:
|
31 |
+
messages, report = await asyncio.gather(
|
32 |
+
get_all_chat_messages_obj(reportId),
|
33 |
+
get_report_obj_by_id(reportId)
|
34 |
+
)
|
35 |
+
message_history = transform_messages_to_openai(messages, message_data.text)
|
36 |
+
response = await generate_agent_response(message_history, report)
|
37 |
response = await save_assistant_user_message(message_data.text, response, reportId)
|
38 |
return OcrResponseWrapper(data=response)
|
ocr/api/openai_requests.py
CHANGED
@@ -0,0 +1,48 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
from ocr.api.prompts import OCRPrompts
|
2 |
+
from ocr.api.report.model import ReportModel
|
3 |
+
from ocr.core.wrappers import openai_wrapper
|
4 |
+
|
5 |
+
|
6 |
+
@openai_wrapper()
|
7 |
+
async def generate_report(request_content: list[dict]):
|
8 |
+
messages = [
|
9 |
+
{
|
10 |
+
"role": "system",
|
11 |
+
"content": OCRPrompts.generate_report
|
12 |
+
},
|
13 |
+
{
|
14 |
+
"role": "user",
|
15 |
+
"content": request_content
|
16 |
+
}
|
17 |
+
]
|
18 |
+
return messages
|
19 |
+
|
20 |
+
|
21 |
+
@openai_wrapper()
|
22 |
+
async def generate_changes(content: list[dict], previous_report: str):
|
23 |
+
messages = [
|
24 |
+
{
|
25 |
+
"role": "system",
|
26 |
+
"content": OCRPrompts.generate_changes
|
27 |
+
.replace("{previous_report}", previous_report)
|
28 |
+
},
|
29 |
+
{
|
30 |
+
"role": "user",
|
31 |
+
"content": content
|
32 |
+
}
|
33 |
+
]
|
34 |
+
return messages
|
35 |
+
|
36 |
+
|
37 |
+
@openai_wrapper()
|
38 |
+
async def generate_agent_response(messages: list[dict], report: ReportModel):
|
39 |
+
messages = [
|
40 |
+
{
|
41 |
+
"role": "system",
|
42 |
+
"content": OCRPrompts.generate_agent_response
|
43 |
+
.replace("{reports}", report.report)
|
44 |
+
.replace("{changes}", report.changes or 'There is no changes.')
|
45 |
+
},
|
46 |
+
*messages
|
47 |
+
]
|
48 |
+
return messages
|
ocr/api/prompts.py
CHANGED
@@ -0,0 +1,86 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
class OCRPrompts:
|
2 |
+
generate_report = """## Task
|
3 |
+
|
4 |
+
You must analyze the text extracted from medical document and generate a comprehensive report in **Markdown2** format. Ensure that every detail provided in the document is included, and do not omit or modify any information. Your output must strictly follow the required format.
|
5 |
+
|
6 |
+
## Report Structure
|
7 |
+
|
8 |
+
The report should be structured as follows, with each section containing only relevant information from the document:
|
9 |
+
|
10 |
+
```markdown
|
11 |
+
## Patient Information
|
12 |
+
|
13 |
+
- Name: [Patient Name]
|
14 |
+
- Age: [Patient Age]
|
15 |
+
- Date of Scan: [Date]
|
16 |
+
- Indication: [Reason for the CT scan]
|
17 |
+
|
18 |
+
## Findings
|
19 |
+
|
20 |
+
**Primary findings**:
|
21 |
+
[Describe significant abnormalities or findings relevant to the indication]
|
22 |
+
|
23 |
+
** Secondary findings**:
|
24 |
+
[List incidental findings, e.g., "Mild hepatic steatosis noted."]
|
25 |
+
**No abnormalities**:
|
26 |
+
[Mention organs or systems without abnormalities, e.g., "No evidence of lymphadenopathy or pleural effusion."]
|
27 |
+
|
28 |
+
## Impression
|
29 |
+
|
30 |
+
[Summarize the findings concisely, e.g., "Findings suggest a primary lung tumor. Biopsy recommended for further evaluation."]
|
31 |
+
|
32 |
+
## Recommendations
|
33 |
+
|
34 |
+
[Include next steps or further tests, e.g., "PET scan and consultation with oncology recommended."]
|
35 |
+
```
|
36 |
+
|
37 |
+
[INST]
|
38 |
+
|
39 |
+
## Instructions
|
40 |
+
|
41 |
+
- **Do not invent or infer any information.** Only use data provided in the user request.
|
42 |
+
- Ensure that the format is followed strictly, and the output is complete without any deviations.
|
43 |
+
|
44 |
+
[/INST]"""
|
45 |
+
generate_changes = """## Task
|
46 |
+
|
47 |
+
You must perform a detailed comparative analysis of the patient's new data from the attached user images against their previous data (`Previous Patient data`). Identify and explicitly highlight all differences, including but not limited to disease progression, remission, newly emerging conditions, and significant clinical changes. Your response must be formatted in **Markdown**.
|
48 |
+
|
49 |
+
## Data
|
50 |
+
|
51 |
+
**Previous Patient Data**:
|
52 |
+
```
|
53 |
+
{previous_report}
|
54 |
+
```
|
55 |
+
|
56 |
+
[INST]
|
57 |
+
|
58 |
+
## Mandatory Instructions
|
59 |
+
|
60 |
+
- Conduct a **meticulous** comparison of the new and old data, ensuring all discrepancies, updates, and changes in the patient's health status are clearly documented.
|
61 |
+
- Provide a structured, concise, and accurate Markdown report.
|
62 |
+
- Do **not** include any speculative analysis—only factual differences explicitly observed in the data.
|
63 |
+
|
64 |
+
[/INST]"""
|
65 |
+
generate_agent_response = """## Objective
|
66 |
+
|
67 |
+
You are an AI medical assistant. Your task is to provide **precise and direct** answers to the doctor's questions based **only** on the provided `Report`, `Patient changes`, and your **verified medical knowledge**. Your responses must be **brief, factual, and strictly to the point**.
|
68 |
+
|
69 |
+
## Data
|
70 |
+
|
71 |
+
**Report**:
|
72 |
+
```
|
73 |
+
{reports}
|
74 |
+
```
|
75 |
+
|
76 |
+
**Patient changes**:
|
77 |
+
```
|
78 |
+
{changes}
|
79 |
+
```
|
80 |
+
|
81 |
+
## Mandatory Instructions
|
82 |
+
|
83 |
+
- Do not elaborate or provide explanations unless explicitly requested.
|
84 |
+
- **Do not include unnecessary details.** Only provide **essential** information relevant to the doctor's question.
|
85 |
+
- **Format your response as plain text** without paragraphs, line breaks, or any additional formatting.
|
86 |
+
- **Do not speculate.** If the requested information is unavailable in the provided data, respond with: `"Insufficient data to answer."`"""
|
ocr/api/report/db_requests.py
CHANGED
@@ -21,7 +21,12 @@ async def get_report_obj_by_id(report_id: str) -> ReportModel:
|
|
21 |
return ReportModel.from_mongo(report)
|
22 |
|
23 |
|
24 |
-
async def save_report_obj(report: str, changes: str) -> ReportModel:
|
25 |
-
report = ReportModel(report=report, changes=changes, filename=
|
26 |
await settings.DB_CLIENT.reports.insert_one(report.to_mongo())
|
27 |
return report
|
|
|
|
|
|
|
|
|
|
|
|
21 |
return ReportModel.from_mongo(report)
|
22 |
|
23 |
|
24 |
+
async def save_report_obj(report: str, changes: str | None, filename: str) -> ReportModel:
|
25 |
+
report = ReportModel(report=report, changes=changes, filename=filename)
|
26 |
await settings.DB_CLIENT.reports.insert_one(report.to_mongo())
|
27 |
return report
|
28 |
+
|
29 |
+
|
30 |
+
async def get_last_report_obj() -> ReportModel | None:
|
31 |
+
report = await settings.DB_CLIENT.reports.find().sort("_id", -1).to_list(length=1)
|
32 |
+
return ReportModel.from_mongo(report[0]) if report else None
|
ocr/api/report/model.py
CHANGED
@@ -7,7 +7,7 @@ from ocr.core.database import MongoBaseModel
|
|
7 |
|
8 |
class ReportModel(MongoBaseModel):
|
9 |
report: str
|
10 |
-
changes: str
|
11 |
filename: str
|
12 |
datetimeInserted: datetime = Field(default_factory=datetime.now)
|
13 |
datetimeUpdated: datetime = Field(default_factory=datetime.now)
|
|
|
7 |
|
8 |
class ReportModel(MongoBaseModel):
|
9 |
report: str
|
10 |
+
changes: str | None = None
|
11 |
filename: str
|
12 |
datetimeInserted: datetime = Field(default_factory=datetime.now)
|
13 |
datetimeUpdated: datetime = Field(default_factory=datetime.now)
|
ocr/api/report/views.py
CHANGED
@@ -1,10 +1,15 @@
|
|
|
|
|
|
1 |
from fastapi import UploadFile, File
|
2 |
|
|
|
3 |
from ocr.api.report import report_router
|
4 |
-
from ocr.api.report.db_requests import get_all_reports_obj, delete_all_reports, get_report_obj_by_id, save_report_obj
|
|
|
5 |
from ocr.api.report.dto import Paging
|
6 |
from ocr.api.report.model import ReportModel
|
7 |
from ocr.api.report.schemas import AllReportResponse
|
|
|
8 |
from ocr.core.wrappers import OcrResponseWrapper
|
9 |
|
10 |
|
@@ -34,8 +39,19 @@ async def get_report(reportId: str) -> OcrResponseWrapper[ReportModel]:
|
|
34 |
async def create_report(
|
35 |
file: UploadFile = File(...),
|
36 |
) -> OcrResponseWrapper[ReportModel]:
|
37 |
-
|
38 |
-
|
39 |
-
|
40 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
41 |
return OcrResponseWrapper(data=report)
|
|
|
1 |
+
import asyncio
|
2 |
+
|
3 |
from fastapi import UploadFile, File
|
4 |
|
5 |
+
from ocr.api.openai_requests import generate_report, generate_changes
|
6 |
from ocr.api.report import report_router
|
7 |
+
from ocr.api.report.db_requests import get_all_reports_obj, delete_all_reports, get_report_obj_by_id, save_report_obj, \
|
8 |
+
get_last_report_obj
|
9 |
from ocr.api.report.dto import Paging
|
10 |
from ocr.api.report.model import ReportModel
|
11 |
from ocr.api.report.schemas import AllReportResponse
|
12 |
+
from ocr.api.utils import divide_images, prepare_request_content, clean_response
|
13 |
from ocr.core.wrappers import OcrResponseWrapper
|
14 |
|
15 |
|
|
|
39 |
async def create_report(
|
40 |
file: UploadFile = File(...),
|
41 |
) -> OcrResponseWrapper[ReportModel]:
|
42 |
+
try:
|
43 |
+
last_report, contents = await asyncio.gather(get_last_report_obj(), file.read())
|
44 |
+
report, changes = None, None
|
45 |
+
images = divide_images(contents)
|
46 |
+
content = prepare_request_content(images)
|
47 |
+
if last_report:
|
48 |
+
report, changes = await asyncio.gather(
|
49 |
+
generate_report(content),
|
50 |
+
generate_changes(content, last_report.report)
|
51 |
+
)
|
52 |
+
else:
|
53 |
+
report = await generate_report(content)
|
54 |
+
report = await save_report_obj(clean_response(report), clean_response(changes), file.filename)
|
55 |
+
finally:
|
56 |
+
await file.close()
|
57 |
return OcrResponseWrapper(data=report)
|
ocr/api/utils.py
CHANGED
@@ -6,10 +6,8 @@ import pytesseract
|
|
6 |
from PIL import Image
|
7 |
from pdf2image import convert_from_bytes
|
8 |
|
9 |
-
from ocr.api.message.models import MessageModel
|
10 |
|
11 |
-
|
12 |
-
def transform_messages_to_openai(messages: list[MessageModel]) -> list[dict]:
|
13 |
openai_messages = []
|
14 |
for message in messages:
|
15 |
content = message.text
|
@@ -17,7 +15,7 @@ def transform_messages_to_openai(messages: list[MessageModel]) -> list[dict]:
|
|
17 |
"role": message.author.value,
|
18 |
"content": content
|
19 |
})
|
20 |
-
|
21 |
return openai_messages
|
22 |
|
23 |
|
|
|
6 |
from PIL import Image
|
7 |
from pdf2image import convert_from_bytes
|
8 |
|
|
|
9 |
|
10 |
+
def transform_messages_to_openai(messages: list, user_query: str) -> list[dict]:
|
|
|
11 |
openai_messages = []
|
12 |
for message in messages:
|
13 |
content = message.text
|
|
|
15 |
"role": message.author.value,
|
16 |
"content": content
|
17 |
})
|
18 |
+
openai_messages.append({"role": "user", "content": user_query})
|
19 |
return openai_messages
|
20 |
|
21 |
|