michaelmc1618 commited on
Commit
d93fd93
·
verified ·
1 Parent(s): 0784dbb

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +89 -101
app.py CHANGED
@@ -22,14 +22,7 @@ mask_filling_pipeline = pipeline("fill-mask", model="nlpaueb/legal-bert-base-unc
22
  # Inference client for chat completion
23
  client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
24
 
25
- def respond(
26
- message,
27
- history: list[tuple[str, str]],
28
- system_message,
29
- max_tokens,
30
- temperature,
31
- top_p,
32
- ):
33
  messages = [{"role": "system", "content": system_message}]
34
 
35
  for val in history:
@@ -51,13 +44,13 @@ def respond(
51
  token = message.choices[0].delta.content
52
  if token is not None:
53
  response += token
54
- yield response, history + [(message, response)]
55
 
56
  def generate_case_outcome(prosecutor_response, defense_response):
57
- prompt = f"Give the real case details from a verified source. Detail the outcome and why."
58
  evaluation = ""
59
  for message in client.chat_completion(
60
- [{"role": "system", "content": "Give the real case details from a verified source. Detail the outcome and why."},
61
  {"role": "user", "content": prompt}],
62
  max_tokens=512,
63
  stream=True,
@@ -82,98 +75,16 @@ def determine_winner(outcome):
82
  else:
83
  return "No clear winner"
84
 
85
- # Custom CSS for a clean layout
86
- custom_css = """
87
- body {
88
- background-color: #ffffff;
89
- color: #000000;
90
- font-family: Arial, sans-serif;
91
- }
92
- .gradio-container {
93
- max-width: 1000px;
94
- margin: 0 auto;
95
- padding: 20px;
96
- background-color: #ffffff;
97
- border: 1px solid #e0e0e0;
98
- border-radius: 8px;
99
- box-shadow: 0 2px 5px rgba(0, 0, 0, 0.1);
100
- }
101
- .gr-button {
102
- background-color: #ffffff !important;
103
- border-color: #ffffff !important;
104
- color: #000000 !important;
105
- margin: 5px;
106
- }
107
- .gr-button:hover {
108
- background-color: #ffffff !important;
109
- border-color: #004085 !important;
110
- }
111
- .gr-input, .gr-textbox, .gr-slider, .gr-markdown, .gr-chatbox {
112
- border-radius: 4px;
113
- border: 1px solid #ced4da;
114
- background-color: #ffffff !important;
115
- color: #000000 !important;
116
- }
117
- .gr-input:focus, .gr-textbox:focus, .gr-slider:focus {
118
- border-color: #ffffff;
119
- outline: 0;
120
- box-shadow: 0 0 0 0.2rem rgba(255, 255, 255, 1.0);
121
- }
122
- #flagging-button {
123
- display: none;
124
- }
125
- footer {
126
- display: none;
127
- }
128
- .chatbox .chat-container .chat-message {
129
- background-color: #ffffff !important;
130
- color: #000000 !important;
131
- }
132
- .chatbox .chat-container .chat-message-input {
133
- background-color: #ffffff !important;
134
- color: #000000 !important;
135
- }
136
- .gr-markdown {
137
- background-color: #ffffff !important;
138
- color: #000000 !important;
139
- }
140
- .gr-markdown h1, .gr-markdown h2, .gr-markdown h3, .gr-markdown h4, .gr-markdown h5, .gr-markdown h6, .gr-markdown p, .gr-markdown ul, .gr-markdown ol, .gr-markdown li {
141
- color: #000000 !important;
142
- }
143
- .score-box {
144
- width: 60px;
145
- height: 60px;
146
- display: flex;
147
- align-items: center;
148
- justify-content: center;
149
- font-size: 12px;
150
- font-weight: bold;
151
- color: black;
152
- margin: 5px;
153
- }
154
- .scroll-box {
155
- max-height: 200px;
156
- overflow-y: scroll;
157
- border: 1px solid #ced4da;
158
- padding: 10px;
159
- border-radius: 4px;
160
- }
161
- """
162
-
163
  def chat_between_bots(system_message1, system_message2, max_tokens, temperature, top_p, history1, history2, shared_history, message):
164
- response1, history1 = list(respond(message, history1, system_message1, max_tokens, temperature, top_p))[-1]
165
- response2, history2 = list(respond(message, history2, system_message2, max_tokens, temperature, top_p))[-1]
166
- shared_history.append(f"Prosecutor: {response1}")
167
- shared_history.append(f"Defense Attorney: {response2}")
168
-
169
- max_length = max(len(response1), len(response2))
170
- response1 = response1[:max_length]
171
- response2 = response2[:max_length]
172
 
173
- outcome = generate_case_outcome(response1, response2)
174
  winner = determine_winner(outcome)
175
 
176
- return response1, response2, history1, history2, shared_history, outcome, winner
177
 
178
  def extract_text_from_pdf(pdf_file):
179
  text = ""
@@ -194,7 +105,6 @@ def update_pdf_gallery_and_extract_text(pdf_files):
194
  return pdf_files, pdf_text
195
 
196
  def get_top_10_cases():
197
- # Here, I'm generating a list of 10 example cases. In a real-world scenario, you'd fetch this data from a database or another source.
198
  cases = [
199
  {"name": "Smith v. Jones", "number": "CA12345"},
200
  {"name": "Johnson v. State", "number": "CA67890"},
@@ -236,7 +146,7 @@ def bot(history):
236
  if token is not None:
237
  response += token
238
  history[-1][1] = response
239
- yield history
240
 
241
  def print_like_dislike(x: gr.LikeData):
242
  print(x.index, x.value, x.liked)
@@ -251,6 +161,84 @@ def ask_about_case_outcome(shared_history, question):
251
  result = qa_pipeline(question=question, context=shared_history)
252
  return result['answer']
253
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
254
  with gr.Blocks(css=custom_css) as demo:
255
  history1 = gr.State([])
256
  history2 = gr.State([])
 
22
  # Inference client for chat completion
23
  client = InferenceClient("HuggingFaceH4/zephyr-7b-beta")
24
 
25
+ def respond(message, history, system_message, max_tokens, temperature, top_p):
 
 
 
 
 
 
 
26
  messages = [{"role": "system", "content": system_message}]
27
 
28
  for val in history:
 
44
  token = message.choices[0].delta.content
45
  if token is not None:
46
  response += token
47
+ return response, history + [(message, response)]
48
 
49
  def generate_case_outcome(prosecutor_response, defense_response):
50
+ prompt = f"Prosecutor's argument: {prosecutor_response}\nDefense Attorney's argument: {defense_response}\nProvide the case details from verified sources and give the outcome along with reasons."
51
  evaluation = ""
52
  for message in client.chat_completion(
53
+ [{"role": "system", "content": "Analyze the case and provide the outcome based on verified sources."},
54
  {"role": "user", "content": prompt}],
55
  max_tokens=512,
56
  stream=True,
 
75
  else:
76
  return "No clear winner"
77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
  def chat_between_bots(system_message1, system_message2, max_tokens, temperature, top_p, history1, history2, shared_history, message):
79
+ prosecutor_response, history1 = respond(message, history1, system_message1, max_tokens, temperature, top_p)
80
+ defense_response, history2 = respond(message, history2, system_message2, max_tokens, temperature, top_p)
81
+ shared_history.append(f"Prosecutor: {prosecutor_response}")
82
+ shared_history.append(f"Defense Attorney: {defense_response}")
 
 
 
 
83
 
84
+ outcome = generate_case_outcome(prosecutor_response, defense_response)
85
  winner = determine_winner(outcome)
86
 
87
+ return prosecutor_response, defense_response, history1, history2, shared_history, outcome, winner
88
 
89
  def extract_text_from_pdf(pdf_file):
90
  text = ""
 
105
  return pdf_files, pdf_text
106
 
107
  def get_top_10_cases():
 
108
  cases = [
109
  {"name": "Smith v. Jones", "number": "CA12345"},
110
  {"name": "Johnson v. State", "number": "CA67890"},
 
146
  if token is not None:
147
  response += token
148
  history[-1][1] = response
149
+ return history
150
 
151
  def print_like_dislike(x: gr.LikeData):
152
  print(x.index, x.value, x.liked)
 
161
  result = qa_pipeline(question=question, context=shared_history)
162
  return result['answer']
163
 
164
+ # Custom CSS for a clean layout
165
+ custom_css = """
166
+ body {
167
+ background-color: #ffffff;
168
+ color: #000000;
169
+ font-family: Arial, sans-serif;
170
+ }
171
+ .gradio-container {
172
+ max-width: 1000px;
173
+ margin: 0 auto;
174
+ padding: 20px;
175
+ background-color: #ffffff;
176
+ border: 1px solid #e0e0e0;
177
+ border-radius: 8px;
178
+ box-shadow: 0 2px 5px rgba(0, 0, 0, 0.1);
179
+ }
180
+ .gr-button {
181
+ background-color: #ffffff !important;
182
+ border-color: #ffffff !important;
183
+ color: #000000 !important;
184
+ margin: 5px;
185
+ }
186
+ .gr-button:hover {
187
+ background-color: #ffffff !important;
188
+ border-color: #004085 !important;
189
+ }
190
+ .gr-input, .gr-textbox, .gr-slider, .gr-markdown, .gr-chatbox {
191
+ border-radius: 4px;
192
+ border: 1px solid #ced4da;
193
+ background-color: #ffffff !important;
194
+ color: #000000 !important;
195
+ }
196
+ .gr-input:focus, .gr-textbox:focus, .gr-slider:focus {
197
+ border-color: #ffffff;
198
+ outline: 0;
199
+ box-shadow: 0 0 0 0.2rem rgba(255, 255, 255, 1.0);
200
+ }
201
+ #flagging-button {
202
+ display: none;
203
+ }
204
+ footer {
205
+ display: none;
206
+ }
207
+ .chatbox .chat-container .chat-message {
208
+ background-color: #ffffff !important;
209
+ color: #000000 !important;
210
+ }
211
+ .chatbox .chat-container .chat-message-input {
212
+ background-color: #ffffff !important;
213
+ color: #000000 !important;
214
+ }
215
+ .gr-markdown {
216
+ background-color: #ffffff !important;
217
+ color: #000000 !important;
218
+ }
219
+ .gr-markdown h1, .gr-markdown h2, .gr-markdown h3, .gr-markdown h4, .gr-markdown h5, .gr-markdown h6, .gr-markdown p, .gr-markdown ul, .gr-markdown ol, .gr-markdown li {
220
+ color: #000000 !important;
221
+ }
222
+ .score-box {
223
+ width: 60px;
224
+ height: 60px;
225
+ display: flex;
226
+ align-items: center;
227
+ justify-content: center;
228
+ font-size: 12px;
229
+ font-weight: bold;
230
+ color: black;
231
+ margin: 5px;
232
+ }
233
+ .scroll-box {
234
+ max-height: 200px;
235
+ overflow-y: scroll;
236
+ border: 1px solid #ced4da;
237
+ padding: 10px;
238
+ border-radius: 4px;
239
+ }
240
+ """
241
+
242
  with gr.Blocks(css=custom_css) as demo:
243
  history1 = gr.State([])
244
  history2 = gr.State([])