Omnibus commited on
Commit
da3f5e6
·
verified ·
1 Parent(s): e51947e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -117
app.py CHANGED
@@ -1,22 +1,16 @@
1
  import gradio as gr
2
  import base64
3
  import random
4
- import gradio as gr
5
- #import urllib.request
6
  import requests
7
  import bs4
8
  import lxml
9
  import os
10
- #import subprocess
11
  from huggingface_hub import InferenceClient,HfApi
12
  import random
13
  import json
14
  import datetime
15
  from pypdf import PdfReader
16
  import uuid
17
- #from query import tasks
18
- from gradio_client import Client
19
- import cairosvg
20
 
21
  from agent import (
22
  PREFIX,
@@ -27,19 +21,9 @@ from agent import (
27
  LOG_RESPONSE,
28
  )
29
  api=HfApi()
30
-
31
  client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
32
 
33
-
34
- def sort_fn(inp):
35
-
36
- client_sort = Client("Omnibus/sort_document")
37
- sen,nouns = client_sort.predict(
38
- f"{inp}", # str in 'Paste Text' Textbox component
39
- api_name="/sort_doc"
40
- )
41
- return nouns
42
-
43
  def find_all(url):
44
  return_list=[]
45
  print (url)
@@ -68,51 +52,8 @@ def find_all(url):
68
  except Exception as e:
69
  print (e)
70
  return False, f'Error: {e}'
71
-
72
- #else:
73
- # history = "observation: The search query I used did not return a valid response"
74
-
75
  return "MAIN", None, history, task
76
 
77
- FIND_KEYWORDS="""Find keywords from the dictionary of provided keywords that are relevant to the users query.
78
- Return the keyword:value pairs from the list in the form of a JSON file output.
79
- dictionary:
80
- {keywords}
81
- user query:
82
- """
83
-
84
- def find_keyword_fn(c,inp,data):
85
-
86
- data=str(data)
87
- seed=random.randint(1,1000000000)
88
- divr=int(c)/20000
89
- divi=int(divr)+1 if divr != int(divr) else int(divr)
90
- chunk = int(int(c)/divr)
91
- out = []
92
- s=0
93
- e=chunk
94
- print(f'e:: {e}')
95
- #task = f'Compile this data to fulfill the task: {task}, and complete the purpose: {purpose}\n'
96
- for z in range(divi):
97
- print(f's:e :: {s}:{e}')
98
-
99
- hist = data[s:e]
100
- resp = run_gpt(
101
- FIND_KEYWORDS,
102
- stop_tokens=[],
103
- max_tokens=8192,
104
- seed=seed,
105
- keywords=data,
106
- ).strip("\n")
107
- out.append(resp)
108
- #new_history = resp
109
- print (resp)
110
- #out+=resp
111
- e=e+chunk
112
- s=s+chunk
113
- return out
114
-
115
-
116
  def read_txt(txt_path):
117
  text=""
118
  with open(txt_path,"r") as f:
@@ -138,9 +79,6 @@ def read_pdf_online(url):
138
  response = requests.get(url, stream=True)
139
  print(response.status_code)
140
  text=""
141
- #################
142
-
143
- #####################
144
  try:
145
  if response.status_code == 200:
146
  with open("test.pdf", "wb") as f:
@@ -300,60 +238,7 @@ def compress_data(c, instruct, history):
300
  s=s+chunk
301
  return out
302
 
303
-
304
- def compress_data_og(c, instruct, history):
305
- seed=random.randint(1,1000000000)
306
-
307
- print (c)
308
- #tot=len(purpose)
309
- #print(tot)
310
- divr=int(c)/MAX_DATA
311
- divi=int(divr)+1 if divr != int(divr) else int(divr)
312
- chunk = int(int(c)/divr)
313
- print(f'chunk:: {chunk}')
314
- print(f'divr:: {divr}')
315
- print (f'divi:: {divi}')
316
- out = []
317
- #out=""
318
- s=0
319
- e=chunk
320
- print(f'e:: {e}')
321
- new_history=""
322
- #task = f'Compile this data to fulfill the task: {task}, and complete the purpose: {purpose}\n'
323
- for z in range(divi):
324
- print(f's:e :: {s}:{e}')
325
-
326
- hist = history[s:e]
327
-
328
- resp = run_gpt(
329
- COMPRESS_DATA_PROMPT,
330
- stop_tokens=["observation:", "task:", "action:", "thought:"],
331
- max_tokens=8192,
332
- seed=seed,
333
- direction=instruct,
334
- knowledge=new_history,
335
- history=hist,
336
- ).strip("\n")
337
-
338
- new_history = resp
339
- print (resp)
340
- out+=resp
341
- e=e+chunk
342
- s=s+chunk
343
- '''
344
- resp = run_gpt(
345
- COMPRESS_DATA_PROMPT,
346
- stop_tokens=["observation:", "task:", "action:", "thought:"],
347
- max_tokens=8192,
348
- seed=seed,
349
- direction=instruct,
350
- knowledge=new_history,
351
- history="All data has been recieved.",
352
- )'''
353
- print ("final" + resp)
354
- #history = "observation: {}\n".format(resp)
355
- return resp
356
-
357
  def get_chart(inp):
358
  seed=random.randint(1,1000000000)
359
  try:
 
1
  import gradio as gr
2
  import base64
3
  import random
 
 
4
  import requests
5
  import bs4
6
  import lxml
7
  import os
 
8
  from huggingface_hub import InferenceClient,HfApi
9
  import random
10
  import json
11
  import datetime
12
  from pypdf import PdfReader
13
  import uuid
 
 
 
14
 
15
  from agent import (
16
  PREFIX,
 
21
  LOG_RESPONSE,
22
  )
23
  api=HfApi()
 
24
  client = InferenceClient("mistralai/Mixtral-8x7B-Instruct-v0.1")
25
 
26
+ ############ Document Functions #################
 
 
 
 
 
 
 
 
 
27
  def find_all(url):
28
  return_list=[]
29
  print (url)
 
52
  except Exception as e:
53
  print (e)
54
  return False, f'Error: {e}'
 
 
 
 
55
  return "MAIN", None, history, task
56
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
  def read_txt(txt_path):
58
  text=""
59
  with open(txt_path,"r") as f:
 
79
  response = requests.get(url, stream=True)
80
  print(response.status_code)
81
  text=""
 
 
 
82
  try:
83
  if response.status_code == 200:
84
  with open("test.pdf", "wb") as f:
 
238
  s=s+chunk
239
  return out
240
 
241
+
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
242
  def get_chart(inp):
243
  seed=random.randint(1,1000000000)
244
  try: