File size: 3,635 Bytes
4e0c974 cdef94b 4e0c974 cdef94b 8ba499f cdef94b 8ba499f 4e0c974 cdef94b 4e0c974 cdef94b 8ba499f cdef94b 4e0c974 9b5be52 4e0c974 9b5be52 4e0c974 9b5be52 4e0c974 cdef94b 88dc71c 8ba499f 88dc71c 8ba499f 88dc71c 8ba499f 88dc71c cdef94b 4e0c974 8ba499f cdef94b 4e0c974 cdef94b 4e0c974 cdef94b 4e0c974 cdef94b 4e0c974 cdef94b d662165 4e0c974 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 |
import aiohttp
import asyncio, pprint
import google.generativeai as palm
from langchain.text_splitter import RecursiveCharacterTextSplitter
from langchain import PromptTemplate
import os
from poe_api_wrapper import PoeApi
import pprint
client = PoeApi("sXvCnfYy8CHnXNTRlxhmVg==")
bot = "Assistant"
CHAT_CODE = ""
PALM_API = ""
API_KEY = os.environ.get("PALM_API", PALM_API)
palm.configure(api_key=API_KEY)
text_splitter = RecursiveCharacterTextSplitter(
separators=["\n\n", "\n", "."],
chunk_size=1500,
length_function=len,
chunk_overlap=100,
)
map_prompt = """
Write a verbose summary like a masters student of the following:
"{text}"
CONCISE SUMMARY:
"""
combine_prompt = """
Write a concise summary of the following text delimited by triple backquotes.
Return your response in a detailed verbose paragraph which covers the text. Make it as insightful to the reader as possible, write like a masters student.
```{text}```
SUMMARY:
"""
def count_tokens(text):
return palm.count_message_tokens(prompt=text)["token_count"]
# async def PalmTextModel(text, candidates=1):
# url = f"https://generativelanguage.googleapis.com/v1beta2/models/text-bison-001:generateText?key={API_KEY}"
# headers = {
# "Content-Type": "application/json",
# }
# data = {
# "prompt": {"text": text},
# "temperature": 0.95,
# "top_k": 100,
# "top_p": 0.95,
# "candidate_count": candidates,
# "max_output_tokens": 1024,
# "stop_sequences": ["</output>"],
# "safety_settings": [
# {"category": "HARM_CATEGORY_DEROGATORY", "threshold": 4},
# {"category": "HARM_CATEGORY_TOXICITY", "threshold": 4},
# {"category": "HARM_CATEGORY_VIOLENCE", "threshold": 4},
# {"category": "HARM_CATEGORY_SEXUAL", "threshold": 4},
# {"category": "HARM_CATEGORY_MEDICAL", "threshold": 4},
# {"category": "HARM_CATEGORY_DANGEROUS", "threshold": 4},
# ],
# }
# async with aiohttp.ClientSession() as session:
# async with session.post(url, json=data, headers=headers) as response:
# if response.status == 200:
# result = await response.json()
# # print(result)
# if candidates > 1:
# temp = [candidate["output"] for candidate in result["candidates"]]
# return temp
# temp = result["candidates"][0]["output"]
# return temp
# else:
# print(f"Error: {response.status}\n{await response.text()}")
async def PalmTextModel(message):
global CHAT_CODE
if CHAT_CODE == "":
for chunk in client.send_message(bot, message):
pass
CHAT_CODE = chunk["chatCode"]
else:
for chunk in client.send_message(bot, message, chatCode=CHAT_CODE):
pass
return chunk["text"]
async def Summarizer(essay):
docs = text_splitter.create_documents([essay])
# for 1 large document
if len(docs) == 1:
tasks = [
PalmTextModel(combine_prompt.format(text=doc.page_content)) for doc in docs
]
# Gather and execute the tasks concurrently
responses = await asyncio.gather(*tasks)
ans = " ".join(responses)
return ans
tasks = [PalmTextModel(map_prompt.format(text=doc.page_content)) for doc in docs]
# Gather and execute the tasks concurrently
responses = await asyncio.gather(*tasks)
main = " ".join(responses)
ans = await PalmTextModel(combine_prompt.format(text=main))
return ans
|