amirkhanbloch commited on
Commit
bba05e0
·
verified ·
1 Parent(s): d4004ce

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +167 -48
app.py CHANGED
@@ -1,50 +1,169 @@
1
- import os
2
- from huggingface_hub import login
3
- import torch
4
  import gradio as gr
5
- from diffusers import StableDiffusion3Pipeline
6
-
7
- # Get Hugging Face token from environment variables
8
- hf_token = os.getenv("HF_API_TOKEN")
9
-
10
- if hf_token:
11
- login(token=hf_token)
12
- print("Login successful")
13
- else:
14
- raise ValueError("Hugging Face token is missing. Please set it in the environment variables.")
15
-
16
- def image_generator(prompt):
17
- device = "cuda" if torch.cuda.is_available() else "cpu"
18
- pipeline = StableDiffusion3Pipeline.from_pretrained(
19
- "stabilityai/stable-diffusion-3-medium-diffusers",
20
- torch_dtype=torch.float16 if device == "cuda" else torch.float32,
21
- text_encoder_3=None,
22
- tokenizer_3=None
 
 
 
23
  )
24
- # Move the pipeline to the appropriate device
25
- pipeline.to(device)
26
-
27
- # Generate the image
28
- image = pipeline(
29
- prompt=prompt,
30
- negative_prompt="blurred, ugly, watermark, low, resolution, blurry",
31
- num_inference_steps=40,
32
- height=1024,
33
- width=1024,
34
- guidance_scale=9.0
35
- ).images[0]
36
-
37
- return image
38
-
39
- # Create a Gradio interface
40
- interface = gr.Interface(
41
- fn=image_generator,
42
- inputs=gr.Textbox(lines=2, placeholder="Enter your prompt..."),
43
- outputs=gr.Image(type="pil"),
44
- title="Image Generator App",
45
- description="This is a simple image generator app using HuggingFace's Stable Diffusion 3 model."
46
- )
47
-
48
- # Launch the interface
49
- interface.launch()
50
- print(interface)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  import gradio as gr
2
+ import os
3
+
4
+ os.system('pip install dashscope -U')
5
+ import tempfile
6
+ from pathlib import Path
7
+ import secrets
8
+ import dashscope
9
+ from dashscope import MultiModalConversation, Generation
10
+ from PIL import Image
11
+
12
+
13
+ # 设置API密钥
14
+ YOUR_API_TOKEN = os.getenv('YOUR_API_TOKEN')
15
+ dashscope.api_key = YOUR_API_TOKEN
16
+ math_messages = []
17
+ def process_image(image, shouldConvert=False):
18
+ # 获取上传文件的目录
19
+ global math_messages
20
+ math_messages = [] # reset when upload image
21
+ uploaded_file_dir = os.environ.get("GRADIO_TEMP_DIR") or str(
22
+ Path(tempfile.gettempdir()) / "gradio"
23
  )
24
+ os.makedirs(uploaded_file_dir, exist_ok=True)
25
+
26
+ # 创建临时文件路径
27
+ name = f"tmp{secrets.token_hex(20)}.jpg"
28
+ filename = os.path.join(uploaded_file_dir, name)
29
+ # 保存上传的图片
30
+ if shouldConvert:
31
+ new_img = Image.new('RGB', size=(image.width, image.height), color=(255, 255, 255))
32
+ new_img.paste(image, (0, 0), mask=image)
33
+ image = new_img
34
+ image.save(filename)
35
+
36
+ # 调用qwen-vl-max-0809模型处理图片
37
+ messages = [{
38
+ 'role': 'system',
39
+ 'content': [{'text': 'You are a helpful assistant.'}]
40
+ }, {
41
+ 'role': 'user',
42
+ 'content': [
43
+ {'image': f'file://{filename}'},
44
+ {'text': 'Please describe the math-related content in this image, ensuring that any LaTeX formulas are correctly transcribed. Non-mathematical details do not need to be described.'}
45
+ ]
46
+ }]
47
+
48
+ response = MultiModalConversation.call(model='qwen-vl-max-0809', messages=messages)
49
+
50
+ # 清理临时文件
51
+ os.remove(filename)
52
+
53
+ return response.output.choices[0]["message"]["content"]
54
+
55
+ def get_math_response(image_description, user_question):
56
+ global math_messages
57
+ if not math_messages:
58
+ math_messages.append({'role': 'system', 'content': 'You are a helpful math assistant.'})
59
+ math_messages = math_messages[:1]
60
+ if image_description is not None:
61
+ content = f'Image description: {image_description}\n\n'
62
+ else:
63
+ content = ''
64
+ query = f"{content}User question: {user_question}"
65
+ math_messages.append({'role': 'user', 'content': query})
66
+ response = Generation.call(
67
+ model="qwen2-math-72b-instruct",
68
+ messages=math_messages,
69
+ result_format='message',
70
+ stream=True
71
+ )
72
+ answer = None
73
+ for resp in response:
74
+ if resp.output is None:
75
+ continue
76
+ answer = resp.output.choices[0].message.content
77
+ yield answer.replace("\\", "\\\\")
78
+ print(f'query: {query}\nanswer: {answer}')
79
+ if answer is None:
80
+ math_messages.pop()
81
+ else:
82
+ math_messages.append({'role': 'assistant', 'content': answer})
83
+
84
+
85
+ def math_chat_bot(image, sketchpad, question, state):
86
+ current_tab_index = state["tab_index"]
87
+ image_description = None
88
+ # Upload
89
+ if current_tab_index == 0:
90
+ if image is not None:
91
+ image_description = process_image(image)
92
+ # Sketch
93
+ elif current_tab_index == 1:
94
+ print(sketchpad)
95
+ if sketchpad and sketchpad["composite"]:
96
+ image_description = process_image(sketchpad["composite"], True)
97
+ yield from get_math_response(image_description, question)
98
+
99
+ css = """
100
+ #qwen-md .katex-display { display: inline; }
101
+ #qwen-md .katex-display>.katex { display: inline; }
102
+ #qwen-md .katex-display>.katex>.katex-html { display: inline; }
103
+ """
104
+
105
+ def tabs_select(e: gr.SelectData, _state):
106
+ _state["tab_index"] = e.index
107
+
108
+
109
+ # 创建Gradio接口
110
+ with gr.Blocks(css=css) as demo:
111
+ gr.HTML("""\
112
+ <p align="center"><img src="https://modelscope.oss-cn-beijing.aliyuncs.com/resource/qwen.png" style="height: 60px"/><p>"""
113
+ """<center><font size=8>📖 Qwen2-Math Demo</center>"""
114
+ """\
115
+ <center><font size=3>This WebUI is based on Qwen2-VL for OCR and Qwen2-Math for mathematical reasoning. You can input either images or texts of mathematical or arithmetic problems.</center>"""
116
+ )
117
+ state = gr.State({"tab_index": 0})
118
+ with gr.Row():
119
+ with gr.Column():
120
+ with gr.Tabs() as input_tabs:
121
+ with gr.Tab("Upload"):
122
+ input_image = gr.Image(type="pil", label="Upload"),
123
+ with gr.Tab("Sketch"):
124
+ input_sketchpad = gr.Sketchpad(type="pil", label="Sketch", layers=False)
125
+ input_tabs.select(fn=tabs_select, inputs=[state])
126
+ input_text = gr.Textbox(label="input your question")
127
+ with gr.Row():
128
+ with gr.Column():
129
+ clear_btn = gr.ClearButton(
130
+ [*input_image, input_sketchpad, input_text])
131
+ with gr.Column():
132
+ submit_btn = gr.Button("Submit", variant="primary")
133
+ with gr.Column():
134
+ output_md = gr.Markdown(label="answer",
135
+ latex_delimiters=[{
136
+ "left": "\\(",
137
+ "right": "\\)",
138
+ "display": True
139
+ }, {
140
+ "left": "\\begin\{equation\}",
141
+ "right": "\\end\{equation\}",
142
+ "display": True
143
+ }, {
144
+ "left": "\\begin\{align\}",
145
+ "right": "\\end\{align\}",
146
+ "display": True
147
+ }, {
148
+ "left": "\\begin\{alignat\}",
149
+ "right": "\\end\{alignat\}",
150
+ "display": True
151
+ }, {
152
+ "left": "\\begin\{gather\}",
153
+ "right": "\\end\{gather\}",
154
+ "display": True
155
+ }, {
156
+ "left": "\\begin\{CD\}",
157
+ "right": "\\end\{CD\}",
158
+ "display": True
159
+ }, {
160
+ "left": "\\[",
161
+ "right": "\\]",
162
+ "display": True
163
+ }],
164
+ elem_id="qwen-md")
165
+ submit_btn.click(
166
+ fn=math_chat_bot,
167
+ inputs=[*input_image, input_sketchpad, input_text, state],
168
+ outputs=output_md)
169
+ demo.launch()