switching to gpt-3.5-turbo
Browse files- app.py +11 -11
- requirements.txt +1 -0
app.py
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
import os
|
2 |
-
from langchain.llms import OpenAI
|
3 |
|
4 |
os.system("pip install -U gradio")
|
5 |
|
@@ -80,8 +80,8 @@ session_token = os.environ.get("SessionToken")
|
|
80 |
|
81 |
def generate_caption(object_list_str, api_key, temperature):
|
82 |
query = f"You are an intelligent image captioner. I will hand you the objects and their position, and you should give me a detailed description for the photo. In this photo we have the following objects\n{object_list_str}"
|
83 |
-
llm =
|
84 |
-
model_name="
|
85 |
)
|
86 |
|
87 |
try:
|
@@ -142,13 +142,13 @@ def inference(img, vocabulary, api_key, temperature):
|
|
142 |
|
143 |
with gr.Blocks() as demo:
|
144 |
with gr.Column():
|
145 |
-
gr.Markdown("# Image Captioning using LangChain
|
146 |
gr.Markdown(
|
147 |
-
"Use Detic to detect objects in an image and then use
|
148 |
)
|
149 |
|
150 |
-
with gr.
|
151 |
-
with gr.
|
152 |
inp = gr.Image(label="Input Image", type="filepath")
|
153 |
with gr.Column():
|
154 |
openai_api_key_textbox = gr.Textbox(
|
@@ -164,10 +164,10 @@ with gr.Blocks() as demo:
|
|
164 |
value="lvis",
|
165 |
)
|
166 |
|
167 |
-
|
168 |
-
|
169 |
-
|
170 |
-
|
171 |
|
172 |
btn_detic.click(
|
173 |
fn=inference,
|
|
|
1 |
import os
|
2 |
+
from langchain.llms import OpenAI, OpenAIChat
|
3 |
|
4 |
os.system("pip install -U gradio")
|
5 |
|
|
|
80 |
|
81 |
def generate_caption(object_list_str, api_key, temperature):
|
82 |
query = f"You are an intelligent image captioner. I will hand you the objects and their position, and you should give me a detailed description for the photo. In this photo we have the following objects\n{object_list_str}"
|
83 |
+
llm = OpenAIChat(
|
84 |
+
model_name="gpt-3.5-turbo", openai_api_key=api_key, temperature=temperature
|
85 |
)
|
86 |
|
87 |
try:
|
|
|
142 |
|
143 |
with gr.Blocks() as demo:
|
144 |
with gr.Column():
|
145 |
+
gr.Markdown("# Image Captioning using Detic and ChatGPT with LangChain 🦜️🔗")
|
146 |
gr.Markdown(
|
147 |
+
"Use Detic to detect objects in an image and then use `gpt-3.5-turbo` to describe the image."
|
148 |
)
|
149 |
|
150 |
+
with gr.Row():
|
151 |
+
with gr.Column():
|
152 |
inp = gr.Image(label="Input Image", type="filepath")
|
153 |
with gr.Column():
|
154 |
openai_api_key_textbox = gr.Textbox(
|
|
|
164 |
value="lvis",
|
165 |
)
|
166 |
|
167 |
+
btn_detic = gr.Button("Run Detic and ChatGPT")
|
168 |
+
with gr.Column():
|
169 |
+
output_desc = gr.Textbox(label="Description Description", lines=5)
|
170 |
+
outviz = gr.Image(label="Visualization", type="pil")
|
171 |
|
172 |
btn_detic.click(
|
173 |
fn=inference,
|
requirements.txt
CHANGED
@@ -35,5 +35,6 @@ nltk
|
|
35 |
|
36 |
git+https://github.com/openai/CLIP.git
|
37 |
|
|
|
38 |
openai
|
39 |
langchain
|
|
|
35 |
|
36 |
git+https://github.com/openai/CLIP.git
|
37 |
|
38 |
+
|
39 |
openai
|
40 |
langchain
|