date_collected
stringclasses
1 value
repo_name
stringlengths
6
116
file_name
stringlengths
2
220
file_contents
stringlengths
13
357k
prompts
sequence
2024-01-10
qwopp/ai-wardrobe
wardrobe~api~posts.py
"""REST API for wardrobe.""" import hashlib import uuid import flask import wardrobe import openai openai.api_type = "azure" openai.api_key = '07c5f40ba37b4b40863a80101eaf2105' openai.api_base = 'https://api.umgpt.umich.edu/azure-openai-api/ptu' openai.api_version = '2023-03-15-preview' import matplotlib.pyplot as plt import numpy as np import PIL import tensorflow as tf from tensorflow import keras from tensorflow.keras import layers from tensorflow.keras.models import Sequential import pathlib import os def get_most_recent_clothesid(logname): """Return most recent clothesid by user""" connection = wardrobe.model.get_db() cur = connection.execute( """ SELECT clothing.clothesid FROM clothing WHERE clothing.owner = ? ORDER BY clothing.clothesid DESC LIMIT 1 """, (logname,), ) m_r = cur.fetchone() if m_r: return m_r['clothesid'] return 0 def not_logged(): """Return 403 if user is not logged in.""" print("WE AINT LOGGED IN!") context1 = { "message": "Forbidden", "status_code": 403 } return flask.jsonify(**context1), 403 def verify_pass(password, storedhash): """Wardrobe check if password matches.""" algo, salt2, passhash = storedhash.split('$') hash_obj58 = hashlib.new(algo) saltedpass = salt2 + password hash_obj58.update(saltedpass.encode('utf-8')) print(passhash) print(hash_obj58.hexdigest()) return passhash == hash_obj58.hexdigest() def check_logged(): """Check if user is logged in.""" logname = flask.session.get('username') if 'username' not in flask.session: auth = flask.request.authorization if auth is not None and 'username' in auth and 'password' in auth: username = flask.request.authorization['username'] password = flask.request.authorization['password'] else: return "notloggedin" connection = wardrobe.model.get_db() if not username or not password: return "notloggedin" cur = connection.execute( """ SELECT username, password FROM users WHERE username = ? """, (username,), ) user = cur.fetchone() if user and verify_pass(password, user['password']): flask.session['username'] = user['username'] logname = user['username'] else: return "notloggedin" return logname @wardrobe.app.route('/api/v1/', methods=["GET"]) def get_api(): """Return API services.""" context = { "clothes": "/api/v1/clothes/", "url": flask.request.path, } return flask.jsonify(**context) @wardrobe.app.route('/api/v1/clothing/', methods=["GET"]) def get_clothing(): """Return the clothes stored on the logged-in user's account.""" logname = check_logged() if logname == "notloggedin": print("NOT LOGGED IN") return not_logged() size = flask.request.args.get('size', default=12, type=int) page = flask.request.args.get('page', default=0, type=int) m_r = get_most_recent_clothesid(logname) clothesid_lte = flask.request.args.get('clothesid_lte', default=m_r, type=int) newp2 = flask.request.full_path if newp2.endswith('?'): newp2 = newp2[:-1] if size <= 0 or page < 0 or (clothesid_lte is not None and clothesid_lte < 0): context = { "message": "Bad Request", "status_code": 400 } return flask.jsonify(**context), 400 connection = wardrobe.model.get_db() cur = connection.execute( """ SELECT clothing.clothesid, clothing.filename, clothing.owner, clothing.article, clothing.confidence FROM clothing WHERE clothing.owner = ? AND (clothing.clothesid <= ? OR ? IS NULL) ORDER BY clothing.clothesid DESC LIMIT ? OFFSET ? """, (logname, clothesid_lte, clothesid_lte, size, page * size), ) clothes_data = cur.fetchall() next_page_url = "" if len(clothes_data) >= size: npu = "/api/v1/clothing/?size={}&page={}&clothesid_lte={}" next_page_url = npu.format(size, page + 1, clothesid_lte) response = { "next": next_page_url, "results": [ { "clothesid": clothing['clothesid'], "filename": "/uploads/" + clothing['filename'], "owner": clothing['owner'], "article": clothing['article'], "confidence": clothing['confidence'], "url": f"/api/v1/clothing/{clothing['clothesid']}/" } for clothing in clothes_data ], "url": newp2, } return flask.jsonify(**response) @wardrobe.app.route('/api/v1/prompt/', methods=["POST"]) def prompt_to_output(): logname = check_logged() data = flask.request.json prompt = data.get('inputData') # Get prompt input # Get all articles from DB connection = wardrobe.model.get_db() cur = connection.execute( """ SELECT clothing.article FROM clothing WHERE clothing.owner = ? ORDER BY clothing.clothesid DESC; """, (logname,), ) dats2 = cur.fetchall() dats_prompt = "List of clothing in closet: " gpt_prompt = "Using the list given, give me an outfit based off of the following prompt: (" + prompt + "). You should only return the required pieces of the outfit. You should only include 1 shirt maximum, and 1 shorts maximum. The output should be sorted from head to toe, separated by commas with spaces after the commas. You should under no circumstance include text before or after the outfit." gbt_output = "" #print(dats2) for x in dats2: dats_prompt += x['article'] + "," #print(dats_prompt) #print(gpt_prompt) #print(dats_prompt) #print(gpt_prompt) try: response = openai.ChatCompletion.create( engine='gpt-4', messages=[ {"role": "system", "content": dats_prompt}, {"role": "user", "content": gpt_prompt} ] ) gbt_output = response['choices'][0]['message']['content'] print(gbt_output) # Parse GPT output gbt_output = gbt_output.split(",") gbt_output = [item.strip() for item in gbt_output] except openai.error.APIError as e: # Handle API error here, e.g. retry or log print(f"OpenAI API returned an API Error: {e}") except openai.error.AuthenticationError as e: # Handle Authentication error here, e.g. invalid API key print(f"OpenAI API returned an Authentication Error: {e}") except openai.error.APIConnectionError as e: # Handle connection error here print(f"Failed to connect to OpenAI API: {e}") except openai.error.InvalidRequestError as e: # Handle connection error here print(f"Invalid Request Error: {e}") except openai.error.RateLimitError as e: # Handle rate limit error print(f"OpenAI API request exceeded rate limit: {e}") except openai.error.ServiceUnavailableError as e: # Handle Service Unavailable error print(f"Service Unavailable: {e}") except openai.error.Timeout as e: # Handle request timeout print(f"Request timed out: {e}") except: print("An exception has occured.") print(gbt_output) connection = wardrobe.model.get_db() file_image_data = {} for article in gbt_output: cur = connection.execute( """ SELECT filename FROM clothing WHERE article = ? LIMIT 1 """, (article,) ) files = cur.fetchall() file_image_data[article] = [file_data['filename'] for file_data in files] ordered_file_image_data = [file_image_data[article] for article in gbt_output] ordered_file_image_data_flat = [f"/uploads/{filename}" for sublist in ordered_file_image_data for filename in sublist] response_data = { "imageFiles": ordered_file_image_data_flat } return flask.jsonify(response_data), 200 @wardrobe.app.route('/api/v1/upload/', methods=['POST']) def upload_file(): logname = check_logged() if 'username' not in flask.session: flask.abort(403) file = flask.request.files['file'] if file.filename == '': return flask.jsonify({"message": "No file selected."}), 400 img_height = 180 img_width = 180 # verify logged in user filename = file.filename stem = uuid.uuid4().hex suffix = pathlib.Path(filename).suffix.lower() uuid_basename = f"{stem}{suffix}" if not file: flask.abort(400) path = wardrobe.app.config["UPLOAD_FOLDER"] / uuid_basename file.save(path) model = tf.keras.models.load_model('my_model_RMSprop.keras') new_image_path = path img = tf.keras.utils.load_img( new_image_path, target_size=(img_height, img_width) ) img_array = tf.keras.utils.img_to_array(img) img_array = tf.expand_dims(img_array, 0) # Create a batch predictions = model.predict(img_array) score = tf.nn.softmax(predictions[0]) class_names = ['black_dress', 'black_pants', 'black_shirt', 'black_shoes', 'black_shorts', 'black_suit', 'blue_dress', 'blue_pants', 'blue_shirt', 'blue_shoes', 'blue_shorts', 'brown_hoodie', 'brown_pants', 'brown_shoes', 'green_pants', 'green_shirt', 'green_shoes', 'green_shorts', 'green_suit', 'pink_hoodie', 'pink_pants', 'pink_skirt', 'red_dress', 'red_hoodie', 'red_pants', 'red_shirt', 'red_shoes', 'silver_shoes', 'silver_skirt', 'white_dress', 'white_pants', 'white_shoes', 'white_shorts', 'white_suit', 'yellow_dress', 'yellow_shorts', 'yellow_skirt'] article = class_names[np.argmax(score)] confidence = 100 * np.max(score) print(article) print(confidence) connection = wardrobe.model.get_db() connection.execute( """ INSERT INTO clothing (filename, owner, article, confidence) VALUES (?, ?, ?, ?) """, (uuid_basename, logname, article, confidence), ) connection.commit() return flask.jsonify({"message": "File uploaded successfully."}), 200
[ "Using the list given, give me an outfit based off of the following prompt: (PLACEHOLDER). You should only return the required pieces of the outfit. You should only include 1 shirt maximum, and 1 shorts maximum. The output should be sorted from head to toe, separated by commas with spaces after the commas. You should under no circumstance include text before or after the outfit.", "PLACEHOLDER,", "inputData", "List of clothing in closet: " ]
2024-01-10
Watchfulio/guidance
guidance~llms~caches~_diskcache.py
import os import diskcache import platformdirs from guidance.llms.caches import Cache class DiskCache(Cache): """DiskCache is a cache that uses diskcache lib.""" def __init__(self, llm_name: str): self._diskcache = diskcache.Cache( os.path.join( platformdirs.user_cache_dir("guidance"), f"_{llm_name}.diskcache" ) ) def __getitem__(self, key: str) -> str: return self._diskcache[key] def __setitem__(self, key: str, value: str) -> None: self._diskcache[key] = value def __contains__(self, key: str) -> bool: return key in self._diskcache def clear(self): self._diskcache.clear()
[]
2024-01-10
swathipil/azure-sdk-for-python
sdk~ai~azure-ai-generative~azure~ai~generative~index~_mlindex.py
# --------------------------------------------------------- # Copyright (c) Microsoft Corporation. All rights reserved. # --------------------------------------------------------- """MLIndex class for interacting with MLIndex assets.""" import os import tempfile import uuid from datetime import datetime from pathlib import Path from typing import Any, Dict, Iterator, Optional, Union import yaml from azure.core.credentials import TokenCredential from azure.ai.generative.index._documents import Document, DocumentChunksIterator from azure.ai.generative.index._embeddings import EmbeddingsContainer from azure.ai.generative.index._utils.connections import ( WorkspaceConnection, get_connection_credential, get_connection_by_id_v2, get_id_from_connection, get_target_from_connection, ) from azure.ai.generative.index._utils.logging import ( get_logger, langchain_version, packages_versions_for_compatibility, track_activity, version, ) try: from langchain.document_loaders.base import BaseLoader except ImportError: BaseLoader = Iterator[Document] logger = get_logger("mlindex") class MLIndex: """MLIndex class for interacting with MLIndex assets.""" INDEX_FIELD_MAPPING_KEY = "field_mapping" INDEX_FIELD_MAPPING_TYPES = { "content": "Raw data content of indexed document", "embedding": "Embedding of indexed document content", "metadata": "Metadata of indexed document, must be a JSON string", "filename": "Filename of indexed document, relative to data source root", "title": "Title of indexed document", "url": "User facing citation URL for indexed document", } base_uri: str index_config: dict embeddings_config: dict def __init__( self, uri: Optional[Union[str, Path, object]] = None, mlindex_config: Optional[dict] = None ): """ Initialize MLIndex from a URI or AzureML Data Asset. Args: ---- uri: URI to MLIndex asset folder (remote or local) mlindex_config: MLIndex config dictionary credential: Credential to use for talking to Azure resources """ with track_activity(logger, "MLIndex.__init__") as activity_logger: if uri is not None: if isinstance(uri, str): uri = str(uri) elif isinstance(uri, Path): uri = str(uri) else: # Assume given AzureML Data Asset uri = uri.path try: import fsspec except ImportError: raise ValueError( "Could not import fsspec python package. " "Please install it with `pip install fsspec`." ) try: import azureml.fsspec # TODO: Patch azureml.dataprep auth logic to use given credential for loading MLIndex except ImportError: logger.warning( "Could not import azureml-fsspec python package. " "Loading from remote filesystems supported by AzureML will not work. " "Please install it with `pip install azureml-fsspec`." ) self.base_uri = uri mlindex_config = None uri = uri.rstrip("/") mlindex_uri = f"{uri}/MLIndex" if not uri.endswith("MLIndex") else uri try: mlindex_file = fsspec.open(mlindex_uri, "r") if hasattr(mlindex_file.fs, "_path"): # File on azureml filesystem has path relative to container root so need to get underlying fs path self.base_uri = mlindex_file.fs._path.split("/MLIndex")[0] else: self.base_uri = mlindex_file.path.split("/MLIndex")[0] with mlindex_file as f: mlindex_config = yaml.safe_load(f) except Exception as e: raise ValueError(f"Could not find MLIndex: {e}") from e elif mlindex_config is None: raise ValueError("Must provide either uri or mlindex_config") self.index_config = mlindex_config.get("index", {}) if self.index_config is None: raise ValueError("Could not find index config in MLIndex yaml") activity_logger.activity_info["index_kind"] = self.index_config.get("kind", "none") self.embeddings_config = mlindex_config.get("embeddings", {}) if self.embeddings_config is None: raise ValueError("Could not find embeddings config in MLIndex yaml") activity_logger.activity_info["embeddings_kind"] = self.embeddings_config.get("kind", "none") activity_logger.activity_info["embeddings_api_type"] = self.embeddings_config.get("api_type", "none") @property def name(self) -> str: """Returns the name of the MLIndex.""" return self.index_config.get("name", "") @name.setter def name(self, value: str): """Sets the name of the MLIndex.""" self.index_config["name"] = value @property def description(self) -> str: """Returns the description of the MLIndex.""" return self.index_config.get("description", "") @description.setter def description(self, value: str): """Sets the description of the MLIndex.""" self.index_config["description"] = value def get_langchain_embeddings(self, credential: Optional[TokenCredential] = None): """Get the LangChainEmbeddings from the MLIndex.""" embeddings = EmbeddingsContainer.from_metadata(self.embeddings_config.copy()) return embeddings.as_langchain_embeddings(credential=credential) def as_langchain_vectorstore(self, credential: Optional[TokenCredential] = None): """Converts MLIndex to a retriever object that can be used with langchain, may download files.""" with track_activity(logger, "MLIndex.as_langchain_vectorstore") as activity_logger: index_kind = self.index_config.get("kind", "none") activity_logger.activity_info["index_kind"] = index_kind activity_logger.activity_info["embeddings_kind"] = self.embeddings_config.get("kind", "none") activity_logger.activity_info["embeddings_api_type"] = self.embeddings_config.get("api_type", "none") if index_kind == "acs": from azure.ai.generative.index._indexes.azure_search import import_azure_search_or_so_help_me import_azure_search_or_so_help_me() if self.index_config.get("field_mapping", {}).get("embedding", None) is None: raise ValueError("field_mapping.embedding must be set in MLIndex config for acs index, try `.as_langchain_retriever()` instead.") try: connection_credential = get_connection_credential(self.index_config, credential=credential) except Exception as e: # azure.ai.generative has workflow where env vars are set before doing stuff. if "AZURE_AI_SEARCH_KEY" in os.environ or "AZURE_COGNITIVE_SEARCH_KEY" in os.environ: from azure.core.credentials import AzureKeyCredential logger.warning(f"Failed to get credential for ACS with {e}, falling back to env vars.") connection_credential = AzureKeyCredential(os.environ["AZURE_AI_SEARCH_KEY"] if "AZURE_AI_SEARCH_KEY" in os.environ else os.environ["AZURE_COGNITIVE_SEARCH_KEY"]) else: raise e from packaging import version as pkg_version azure_search_documents_version = packages_versions_for_compatibility["azure-search-documents"] search_client_version = pkg_version.parse(azure_search_documents_version) langchain_pkg_version = pkg_version.parse(langchain_version) if (search_client_version > pkg_version.parse("11.4.0b6") and langchain_pkg_version > pkg_version.parse("0.0.273")) \ or (search_client_version == pkg_version.parse("11.4.0b6") and langchain_pkg_version < pkg_version.parse("0.0.273") and pkg_version.parse(langchain_pkg_version >= "0.0.198")): from langchain.vectorstores import azuresearch # TODO: These fields effect all ACS retrievers in the same process, should change class so it can # use these as defaults but uses names passed in as args preferentially azuresearch.FIELDS_ID = self.index_config.get("field_mapping", {}).get("id", "id") azuresearch.FIELDS_CONTENT = self.index_config.get("field_mapping", {}).get("content", "content") azuresearch.FIELDS_CONTENT_VECTOR = self.index_config.get("field_mapping", {}).get("embedding", "content_vector_open_ai") azuresearch.FIELDS_METADATA = self.index_config.get("field_mapping", {}).get("metadata", "meta_json_string") from azure.core.credentials import AzureKeyCredential from langchain.vectorstores.azuresearch import AzureSearch return AzureSearch( azure_search_endpoint=self.index_config.get( "endpoint", get_target_from_connection( get_connection_by_id_v2( self.index_config["connection"]["id"], credential=credential ) ) ), azure_search_key=connection_credential.key if isinstance(connection_credential, AzureKeyCredential) else None, index_name=self.index_config.get("index"), embedding_function=self.get_langchain_embeddings(credential=credential).embed_query, search_type="hybrid", semantic_configuration_name=self.index_config.get("semantic_configuration_name", "azureml-default"), user_agent=f"azureml-rag=={version}/mlindex,langchain=={langchain_version}", ) else: from azure.ai.generative.index._langchain.acs import AzureCognitiveSearchVectorStore logger.warning(f"azure-search-documents=={azure_search_documents_version} not compatible langchain.vectorstores.azuresearch yet, using REST client based VectorStore.") return AzureCognitiveSearchVectorStore( index_name=self.index_config.get("index"), endpoint=self.index_config.get( "endpoint", get_target_from_connection( get_connection_by_id_v2( self.index_config["connection"]["id"], credential=credential ) ) ), embeddings=self.get_langchain_embeddings(credential=credential), field_mapping=self.index_config.get("field_mapping", {}), credential=connection_credential, ) elif index_kind == "faiss": from fsspec.core import url_to_fs store = None engine = self.index_config.get("engine") if engine == "langchain.vectorstores.FAISS": from azure.ai.generative.index._langchain.vendor.vectorstores.faiss import FAISS embeddings = EmbeddingsContainer.from_metadata(self.embeddings_config.copy()).as_langchain_embeddings(credential=credential) fs, uri = url_to_fs(self.base_uri) with tempfile.TemporaryDirectory() as tmpdir: fs.download(f"{uri.rstrip('/')}/index.pkl", f"{str(tmpdir)}") fs.download(f"{uri.rstrip('/')}/index.faiss", f"{str(tmpdir)}") store = FAISS.load_local(str(tmpdir), embeddings) elif engine.endswith("indexes.faiss.FaissAndDocStore"): from azure.ai.generative.index._indexes.faiss import FaissAndDocStore error_fmt_str = """Failed to import langchain faiss bridge module with: {e}\n" This could be due to an incompatible change in langchain since this bridge was implemented. If you understand what has changed you could implement your own wrapper of azure.ai.tools.mlindex._indexes.faiss.FaissAndDocStore. """ try: from azure.ai.generative.index._langchain.faiss import azureml_faiss_as_langchain_faiss except Exception as e: logger.warning(error_fmt_str.format(e=e)) azureml_faiss_as_langchain_faiss = None embeddings = EmbeddingsContainer.from_metadata(self.embeddings_config.copy()).as_langchain_embeddings(credential=credential) store = FaissAndDocStore.load(self.base_uri, embeddings.embed_query) if azureml_faiss_as_langchain_faiss is not None: try: store = azureml_faiss_as_langchain_faiss(FaissAndDocStore.load(self.base_uri, embeddings.embed_query)) except Exception as e: logger.error(error_fmt_str.format(e=e)) raise else: raise ValueError(f"Unknown engine: {engine}") return store else: raise ValueError(f"Unknown index kind: {index_kind}") def as_langchain_retriever(self, credential: Optional[TokenCredential] = None, **kwargs): """Converts MLIndex to a retriever object that can be used with langchain, may download files.""" index_kind = self.index_config.get("kind", None) if index_kind == "acs": if self.index_config.get("field_mapping", {}).get("embedding", None) is None: from azure.ai.generative.index._langchain.acs import AzureCognitiveSearchVectorStore connection_credential = get_connection_credential(self.index_config, credential=credential) endpoint = self.index_config.get("endpoint", None) if not endpoint: endpoint = get_target_from_connection( get_connection_by_id_v2( self.index_config["connection"]["id"], credential=credential ) ) return AzureCognitiveSearchVectorStore( index_name=self.index_config.get("index"), endpoint=endpoint, embeddings=self.get_langchain_embeddings(credential=credential), field_mapping=self.index_config.get("field_mapping", {}), credential=connection_credential, ).as_retriever(**kwargs) return self.as_langchain_vectorstore(credential=credential).as_retriever(**kwargs) elif index_kind == "faiss": return self.as_langchain_vectorstore(credential=credential).as_retriever(*kwargs) else: raise ValueError(f"Unknown index kind: {index_kind}") def as_native_index_client(self, credential: Optional[TokenCredential] = None): """ Converts MLIndex config into a client for the underlying Index, may download files. An azure.search.documents.SearchClient for acs indexes or an azure.ai.generative.index._indexes.indexFaissAndDocStore for faiss indexes. """ index_kind = self.index_config.get("kind", None) if index_kind == "acs": connection_credential = get_connection_credential(self.index_config, credential=credential) from azure.search.documents import SearchClient return SearchClient( endpoint=self.index_config.get("endpoint"), index_name=self.index_config.get("index"), credential=connection_credential, user_agent=f"azureml-rag=={version}/mlindex" ) elif index_kind == "faiss": from azure.ai.generative.index._indexes.faiss import FaissAndDocStore embeddings = self.get_langchain_embeddings(credential=credential) return FaissAndDocStore.load(self.base_uri, embeddings.embed_query) else: raise ValueError(f"Unknown index kind: {index_kind}") def __repr__(self): """Returns a string representation of the MLIndex object.""" return yaml.dump({ "index": self.index_config, "embeddings": self.embeddings_config, }) def override_connections( self, embedding_connection: Optional[Union[str, WorkspaceConnection]] = None, index_connection: Optional[Union[str, WorkspaceConnection]] = None, credential: Optional[TokenCredential] = None ) -> "MLIndex": """ Override the connections used by the MLIndex. Args: ---- embedding_connection: Optional connection to use for embeddings model index_connection: Optional connection to use for index credential: Optional credential to use when resolving connection information """ if embedding_connection: if self.embeddings_config.get("key") is not None: self.embeddings_config.pop("key") if embedding_connection.__class__.__name__ == "AzureOpenAIConnection": # PromptFlow Connection self.embeddings_config["connection_type"] = "inline" self.embeddings_config["connection"] = { "key": embedding_connection.secrets.get("api_key"), "api_base": embedding_connection.api_base, "api_type": embedding_connection.api_type, "api_version": embedding_connection.api_version } else: self.embeddings_config["connection_type"] = "workspace_connection" if isinstance(embedding_connection, str): from azure.ai.generative.index._utils.connections import get_connection_by_id_v2 embedding_connection = get_connection_by_id_v2(embedding_connection, credential=credential) self.embeddings_config["connection"] = {"id": get_id_from_connection(embedding_connection)} if index_connection: if self.index_config["kind"] != "acs": print("Index kind is not acs, ignoring override for connection") else: self.index_config["connection_type"] = "workspace_connection" if isinstance(index_connection, str): from azure.ai.generative.index._utils.connections import get_connection_by_id_v2 index_connection = get_connection_by_id_v2(index_connection, credential=credential) self.index_config["connection"] = {"id": get_id_from_connection(index_connection)} self.save(just_config=True) return self def set_embeddings_connection(self, connection: Optional[Union[str, WorkspaceConnection]], credential: Optional[TokenCredential] = None) -> "MLIndex": """Set the embeddings connection used by the MLIndex.""" return self.override_connections(embedding_connection=connection) @staticmethod def from_files( source_uri: str, source_glob: str = "**/*", chunk_size: int = 1000, chunk_overlap: int = 0, citation_url: Optional[str] = None, citation_replacement_regex: Optional[Dict[str, str]] = None, embeddings_model: str = "hugging_face://model/sentence-transformers/all-mpnet-base-v2", embeddings_connection: Optional[str] = None, embeddings_container: Optional[Union[str, Path]] = None, index_type: str = "faiss", index_connection: Optional[str] = None, index_config: Dict[str, Any] = {}, output_path: Optional[Union[str, Path]] = None, credential: Optional[TokenCredential] = None ) -> "MLIndex": r""" Create a new MLIndex from a repo. Args: ---- source_uri: Iterator of documents to index source_glob: Glob pattern to match files to index chunk_size: Size of chunks to split documents into chunk_overlap: Size of overlap between chunks citation_url: Optional url to replace citation urls with citation_replacement_regex: Optional regex to use to replace citation urls, e.g. `{"match_pattern": "(.*)/articles/(.*)(\.[^.]+)$", "replacement_pattern": "\1/\2"}` embeddings_model: Name of embeddings model to use, expected format `azure_open_ai://deployment/.../model/text-embedding-ada-002` or `hugging_face://model/all-mpnet-base-v2` embeddings_connection: Optional connection to use for embeddings model embeddings_container: Optional path to location where un-indexed embeddings can be saved/loaded. index_type: Type of index to use, e.g. faiss index_connection: Optional connection to use for index index_config: Config for index, e.g. index_name or field_mapping for acs Returns: ------- MLIndex """ from azure.ai.generative.index._documents import DocumentChunksIterator, split_documents with track_activity(logger, "MLIndex.from_files"): chunked_documents = DocumentChunksIterator( files_source=source_uri, glob=source_glob, base_url=citation_url, document_path_replacement_regex=citation_replacement_regex, chunked_document_processors=[lambda docs: split_documents(docs, splitter_args={"chunk_size": chunk_size, "chunk_overlap": chunk_overlap, "use_rcts": False})], ) mlindex = MLIndex.from_documents( chunked_documents, embeddings_model=embeddings_model, embeddings_connection=embeddings_connection, embeddings_container=embeddings_container, index_type=index_type, index_connection=index_connection, index_config=index_config, output_path=output_path, credential=credential ) return mlindex @staticmethod def from_documents( documents: Union[Iterator[Document], BaseLoader, DocumentChunksIterator], embeddings_model: str = "hugging_face://model/sentence-transformers/all-mpnet-base-v2", embeddings_connection: Optional[str] = None, embeddings_container: Optional[Union[str, Path]] = None, index_type: str = "faiss", index_connection: Optional[str] = None, index_config: Dict[str, Any] = {}, output_path: Optional[Union[str, Path]] = None, credential: Optional[TokenCredential] = None ) -> "MLIndex": """ Create a new MLIndex from documents. Args: ---- documents: Iterator of documents to index index_kind: Kind of index to use embeddings_model: Name of embeddings model to use, expected format `azure_open_ai://deployment/.../model/text-embedding-ada-002` or `hugging_face://model/all-mpnet-base-v2` embeddings_container: Optional path to location where un-indexed embeddings can be saved/loaded. index_type: Type of index to use, e.g. faiss index_connection: Optional connection to use for index index_config: Config for index, e.g. index_name or field_mapping for acs output_path: Optional path to save index to Returns: ------- MLIndex """ import time embeddings = None # TODO: Move this logic to load from embeddings_container into EmbeddingsContainer try: if embeddings_container is not None: if isinstance(embeddings_container, str) and "://" in embeddings_container: from fsspec.core import url_to_fs fs, uri = url_to_fs(embeddings_container) else: embeddings_container = Path(embeddings_container) previous_embeddings_dir_name = None try: previous_embeddings_dir_name = str(max([dir for dir in embeddings_container.glob("*") if dir.is_dir()], key=os.path.getmtime).name) except Exception as e: logger.warning( f"failed to get latest folder from {embeddings_container} with {e}.", extra={"print": True}) pass if previous_embeddings_dir_name is not None: try: embeddings = EmbeddingsContainer.load(previous_embeddings_dir_name, embeddings_container) except Exception as e: logger.warning( f"failed to load embeddings from {embeddings_container} with {e}.", extra={"print": True}) pass finally: if embeddings is None: logger.info("Creating new EmbeddingsContainer") if isinstance(embeddings_model, str): connection_args = {} if "open_ai" in embeddings_model: from azure.ai.generative.index._utils.connections import get_connection_by_id_v2 if embeddings_connection: if isinstance(embeddings_connection, str): embeddings_connection = get_connection_by_id_v2(embeddings_connection, credential=credential) connection_args = { "connection_type": "workspace_connection", "connection": {"id": get_id_from_connection(embeddings_connection)}, "endpoint": embeddings_connection.target if hasattr(embeddings_connection, "target") else embeddings_connection["properties"]["target"], } else: connection_args = { "connection_type": "environment", "connection": {"key": "OPENAI_API_KEY"}, "endpoint": os.getenv("OPENAI_API_BASE"), } if os.getenv("OPENAI_API_TYPE"): connection_args["api_type"] = os.getenv("OPENAI_API_TYPE") if os.getenv("OPENAI_API_VERSION"): connection_args["api_version"] = os.getenv("OPENAI_API_VERSION") embeddings = EmbeddingsContainer.from_uri(embeddings_model, credential=credential, **connection_args) else: raise ValueError(f"Unknown embeddings model: {embeddings_model}") # try: # import sentence_transformers # if isinstance(embeddings_model, sentence_transformers.SentenceTransformer): # embeddings = EmbeddingsContainer.from_sentence_transformer(embeddings_model) # except Exception as e: # logger.warning(f"Failed to load sentence_transformers with {e}.") pre_embed = time.time() embeddings = embeddings.embed(documents) post_embed = time.time() logger.info(f"Embedding took {post_embed - pre_embed} seconds") if embeddings_container is not None: now = datetime.datetime.now() # TODO: This means new snapshots will be created for every run, ideally there'd be a use container as readonly vs persist snapshot option embeddings.save(str(embeddings_container / f"{now.strftime('%Y%m%d')}_{now.strftime('%H%M%S')}_{str(uuid.uuid4()).split('-')[0]}")) mlindex = MLIndex.from_embeddings_container( embeddings, index_type=index_type, index_connection=index_connection, index_config=index_config, output_path=output_path, credential=credential ) return mlindex @staticmethod def from_embeddings_container( embeddings: EmbeddingsContainer, index_type: str, index_connection: Optional[str] = None, index_config: Dict[str, Any] = {}, output_path: Optional[Union[str, Path]] = None, credential: Optional[TokenCredential] = None ) -> "MLIndex": """ Create a new MLIndex from embeddings. Args ---- embeddings: EmbeddingsContainer to index index_type: Type of index to use, e.g. faiss index_connection: Optional connection to use for index index_config: Config for index, e.g. index_name or field_mapping for acs output_path: Optional path to save index to credential: Optional credential to use when resolving connection information Returns ------- MLIndex """ if output_path is None: output_path = Path.cwd() / f"{index_type}_{embeddings.kind}_index" if index_type == "faiss": embeddings.write_as_faiss_mlindex(output_path=output_path, engine="indexes.faiss.FaissAndDocStore") mlindex = MLIndex( uri=Path(output_path), ) elif index_type == "acs": from azure.ai.generative.index._tasks.update_acs import create_index_from_raw_embeddings from azure.ai.generative.index._utils.connections import get_connection_by_id_v2 if not index_connection: index_config = { **index_config, **{ "endpoint": os.getenv("AZURE_AI_SEARCH_ENDPOINT"), "api_version": "2023-07-01-preview", } } connection_args = { "connection_type": "environment", "connection": {"key": "AZURE_AI_SEARCH_KEY"} } else: if isinstance(index_connection, str): index_connection = get_connection_by_id_v2(index_connection, credential=credential) index_config = { **index_config, **{ "endpoint": index_connection.target if hasattr(index_connection, "target") else index_connection["properties"]["target"], "api_version": index_connection.metadata.get("apiVersion", "2023-07-01-preview") if hasattr(index_connection, "metadata") else index_connection["properties"]["metadata"].get("apiVersion", "2023-07-01-preview"), } } connection_args = { "connection_type": "workspace_connection", "connection": {"id": get_id_from_connection(index_connection)} } mlindex = create_index_from_raw_embeddings( embeddings, index_config, connection=connection_args, output_path=str(output_path), credential=credential ) else: raise ValueError(f"Unknown index type: {index_type}") return mlindex def save(self, output_uri: Optional[str], just_config: bool = False): """ Save the MLIndex to a uri. Will use uri MLIndex was loaded from if `output_uri` not set. """ # Use fsspec to create MLIndex yaml file at output_uri and call save on _underlying_index if present try: import fsspec mlindex_file = fsspec.open(f"{output_uri.rstrip('/')}/MLIndex", "w") # parse yaml to dict with mlindex_file as f: yaml.safe_dump({ "embeddings": self.embeddings_config, "index": self.index_config }, f) if not just_config: files = fsspec.open_files(f"{self.base_uri}/*") files += fsspec.open_files(f"{self.base_uri}/**/*") for file in files: if file.path.endswith("MLIndex"): continue with file.open() as src: with fsspec.open(f"{output_uri.rstrip('/')}/{file.path.replace(self.base_uri, '').lstrip('/')}", "wb") as dest: dest.write(src.read()) except Exception as e: raise ValueError(f"Could not save MLIndex: {e}") from e
[ "Raw data content of indexed document" ]
2024-01-10
kerberosmansour/TLA_AutoRepair
autorepair.py
import difflib import json import os import platform import shutil import subprocess import sys import urllib.request import argparse import ssl import urllib.error import requests import os import json import os import json import openai from dotenv import load_dotenv from termcolor import cprint from typing import List, Dict # Load environment variables load_dotenv() openai.api_key = os.getenv("OPENAI_API_KEY") # Constants DEFAULT_MODEL = os.getenv("DEFAULT_MODEL", "gpt-4") VALIDATE_JSON_RETRY = int(os.getenv("VALIDATE_JSON_RETRY", -1)) PROMPT_FILE_PATH = os.path.join(os.path.dirname(__file__), "prompt.txt") # Read the system prompt with open(PROMPT_FILE_PATH, "r") as prompt_file: SYSTEM_PROMPT = prompt_file.read() # Define standard directories for TLA+ tools based on the platform STANDARD_DIRS = { "Windows": "C:/Program Files/TLA+", "Darwin": "/Users/Shared/TLA+", "Linux": "/usr/local/share/TLA+", } def get_standard_dir(): """ Get the standard directory for TLA+ tools based on the platform """ return STANDARD_DIRS.get(platform.system(), "/usr/local/share/TLA+") def run_tla_spec(spec_name: str, tla_tools_path: str) -> str: """ Run TLC Model Checker on the given TLA+ specification. Parameters: spec_name: The name of the TLA+ specification to run. tla_tools_path: The file path to the TLA+ tools. Returns: A string containing the output of the TLC Model Checker. """ subprocess_args = ["java", "-cp", tla_tools_path, "tlc2.TLC", spec_name] try: result = subprocess.check_output(subprocess_args, stderr=subprocess.STDOUT) except subprocess.CalledProcessError as error: return error.output.decode("utf-8"), error.returncode return result.decode("utf-8"), 0 def json_validated_response( model: str, messages: List[Dict], nb_retry: int = VALIDATE_JSON_RETRY ) -> Dict: """ This function is needed because the API can return a non-json response. This will run recursively VALIDATE_JSON_RETRY times. If VALIDATE_JSON_RETRY is -1, it will run recursively until a valid json response is returned. """ json_response = {} if nb_retry != 0: response = openai.ChatCompletion.create( model=model, messages=messages, temperature=0.5, ) messages.append(response.choices[0].message) content = response.choices[0].message.content # see if json can be parsed try: json_start_index = content.index( "[" ) # find the starting position of the JSON data json_data = content[ json_start_index: ] # extract the JSON data from the response string json_response = json.loads(json_data) return json_response except (json.decoder.JSONDecodeError, ValueError) as e: cprint(f"{e}. Re-running the query.", "red") # debug cprint(f"\nGPT RESPONSE:\n\n{content}\n\n", "yellow") # append a user message that says the json is invalid messages.append( { "role": "user", "content": ( "Your response could not be parsed by json.loads. " "Please restate your last message as pure JSON." ), } ) # dec nb_retry nb_retry -= 1 # rerun the api call return json_validated_response(model, messages, nb_retry) except Exception as e: cprint(f"Unknown error: {e}", "red") cprint(f"\nGPT RESPONSE:\n\n{content}\n\n", "yellow") raise e raise Exception( f"No valid json response found after {VALIDATE_JSON_RETRY} tries. Exiting." ) def send_error_to_gpt( spec_path: str, error_message: str, model: str = DEFAULT_MODEL ) -> Dict: # Read the TLA+ specification with open(spec_path, "r") as f: spec_lines = f.readlines() # Assume the .CFG file has the same name as the TLA+ specification but with a .cfg extension cfg_path = spec_path.rsplit(".", 1)[0] + ".cfg" # Read the .CFG file with open(cfg_path, "r") as f: model_cfg = f.read() # Full spec for context full_spec = "".join(spec_lines) # Spec lines with line numbers for reference spec_with_lines = [] for i, line in enumerate(spec_lines): spec_with_lines.append(str(i + 1) + ": " + line) spec_with_lines = "".join(spec_with_lines) # Create the prompt for the AI model prompt = ( "Here is the TLA+ spec that has errors that need fixing:\n\n" f"{full_spec}\n\n" "Here is the TLA+ model requirements that need to be met for more context:\n\n" f"{model_cfg}\n\n" "Here is the TLA+ spec lines that needs fixing:\n\n" f"{spec_with_lines}\n\n" "Here is the error message:\n\n" f"{error_message}\n" "Please provide your suggested changes, and remember to stick to the " "exact format as described above." ) # Send the prompt to the AI model messages = [ { "role": "system", "content": SYSTEM_PROMPT, }, { "role": "user", "content": prompt, }, ] return json_validated_response(model, messages) def apply_changes(file_path: str, changes: List, confirm: bool = False): with open(file_path) as f: original_file_lines = f.readlines() # Filter out explanation elements operation_changes = [change for change in changes if "operation" in change] explanations = [ change["explanation"] for change in changes if "explanation" in change ] # Sort the changes in reverse line order operation_changes.sort(key=lambda x: x["line"], reverse=True) file_lines = original_file_lines.copy() for change in operation_changes: operation = change["operation"] line = change["line"] content = change["content"] if operation == "Replace": file_lines[line - 1] = content + "\n" elif operation == "Delete": del file_lines[line - 1] elif operation == "InsertAfter": file_lines.insert(line, content + "\n") # Print explanations cprint("Explanations:", "blue") for explanation in explanations: cprint(f"- {explanation}", "blue") # Display changes diff print("\nChanges to be made:") diff = difflib.unified_diff(original_file_lines, file_lines, lineterm="") for line in diff: if line.startswith("+"): cprint(line, "green", end="") elif line.startswith("-"): cprint(line, "red", end="") else: print(line, end="") if confirm: # check if user wants to apply changes or exit confirmation = input("Do you want to apply these changes? (y/n): ") if confirmation.lower() != "y": print("Changes not applied") sys.exit(0) with open(file_path, "w") as f: f.writelines(file_lines) print("Changes applied.") def check_model_availability(model): available_models = [x["id"] for x in openai.Model.list()["data"]] if model not in available_models: print( f"Model {model} is not available. Perhaps try running with " "`--model=gpt-3.5-turbo` instead? You can also configure a " "default model in the .env" ) exit() def find_tla_tools_path(): # Run the locate command to find the TLA+ tools jar file try: locate_output = ( subprocess.check_output(["locate", "tla2tools.jar"]) .decode("utf-8") .split("\n") ) except subprocess.CalledProcessError: print( "The locate command failed. Please make sure that the locate database is up to date." ) return None # Filter out any empty lines or lines that don't end with "tla2tools.jar" jar_files = [path for path in locate_output if path.endswith("tla2tools.jar")] if not jar_files: print("Could not find the TLA+ tools jar file.") return None # Find the jar file with the latest modification time latest_jar_file = max(jar_files, key=os.path.getmtime) return latest_jar_file def install_tla_plus(disable_ssl_verification=False): """Downloads and sets up TLA+.""" print("TLA+ not found. Attempting to install...") # Download the tla2tools.jar file url = "https://github.com/tlaplus/tlaplus/releases/latest/download/tla2tools.jar" tla2tools_path = os.path.join(get_standard_dir(), "tla2tools.jar") try: # Attempt to download with SSL verification response = requests.get(url, verify=not disable_ssl_verification) response.raise_for_status() with open(tla2tools_path, "wb") as f: f.write(response.content) except requests.exceptions.RequestException as e: print(f"Error occurred during download: {e}") return None print("TLA+ has been downloaded.") return tla2tools_path def check_tla_tools_availability(): """ Checks if the TLC model checker is available in the system path and functioning correctly. """ # The name of the TLA+ tools jar file tla_tools_jar = "tla2tools.jar" # Check if TLA+ tools are available in system PATH tla_tools_path = shutil.which(tla_tools_jar) # If not found in PATH, check in our standard directory if not tla_tools_path: tla_tools_path = os.path.join(get_standard_dir(), tla_tools_jar) # If not found in standard directory, attempt to install if not os.path.isfile(tla_tools_path): tla_tools_path = install_tla_plus() # If still not found, use the locate command to find the most recent jar file if not tla_tools_path: tla_tools_path = find_tla_tools_path() # Define a simple TLA+ spec with a bound on x trivial_spec = """ --------------------------- MODULE TempTestSpec --------------------------- EXTENDS Naturals VARIABLE x Init == x = 0 Next == IF x < 5 THEN x' = x + 1 ELSE x' = x ============================================================================== """ # Write the spec to a temporary file with open("TempTestSpec.tla", "w") as f: f.write(trivial_spec) # Define a simple TLA+ config trivial_config = """ INIT Init NEXT Next """ # Write the config to a temporary file with open("TempTestSpec.cfg", "w") as f: f.write(trivial_config) # Run TLC on the temporary spec try: result = subprocess.run( ["java", "-cp", tla_tools_path, "tlc2.TLC", "TempTestSpec"], stderr=subprocess.PIPE, ) except FileNotFoundError: print( "TLA+ Tools are not available. Please follow these steps to install:\n" "1. Download the TLA+ tools JAR file from https://github.com/tlaplus/tlaplus/releases\n" "2. Place the JAR file in a directory, such as /Users/yourusername/Dev/TLA+\n" "3. Make sure Java is installed on your system. You can download Java from https://www.java.com/en/download/\n" "4. Ensure that the TLA+ tools are accessible in your Java classpath.\n" ) sys.exit(1) # If there was an error running TLC, print the error and exit if result.returncode != 0: print( "There was an error running the TLC model checker. Error details:\n" f"{result.stderr.decode('utf-8')}" ) sys.exit(1) # Clean up the temporary spec and config files os.remove("TempTestSpec.tla") os.remove("TempTestSpec.cfg") return tla_tools_path def provide_detailed_comments(spec_path: str, model: str = DEFAULT_MODEL): cfg_path = os.path.splitext(spec_path)[0] + ".cfg" # Step 1: Load both the .cfg and .tla files and send them in a prompt cfg_file_content = "" if os.path.exists(cfg_path): with open(cfg_path, "r") as f: cfg_file_content = f.read() with open(spec_path, "r") as f: spec_lines = f.readlines() prompt = ( "Here is a TLA+ specification:\n\n" "tla\n" "".join(spec_lines) + "\n" "\n\n" "And here is its .cfg file (the model to check):\n\n" "Here is the .cfg file \n" + cfg_file_content + "\n" "\n\n" "Please rewrite the TLA+ file to include detailed comments " "that are readable and useful for developers and people new to TLA+. " "Ensure the code is enclosed within three backticks (```)." ) response = openai.ChatCompletion.create( model=model, messages=[{"role": "user", "content": prompt}], temperature=0.5, ) # Step 3: Take the response and pull the TLA+ file and rewrite the file # We expect the TLA+ code to be in a code block in the response response_lines = response.choices[0].message["content"].split("\n") response_code_start = next( i for i, line in enumerate(response_lines) if line.startswith("```") ) response_code_end = next( i for i, line in enumerate( response_lines[response_code_start + 1 :], response_code_start + 1 ) if line.startswith("```") ) response_code = response_lines[response_code_start + 1 : response_code_end] response_code = [line.replace("\\\\", "\\") for line in response_code] # Step 4: Run a sanitation check on the TLA+ file that there are not comments on the header line if response_code[0].startswith("----------------------------- MODULE"): response_code[0] += "\n\n" # Step 5: Make sure there are no generic or empty comments response_code = [ line for line in response_code if "\\*" not in line or line.split("\\*")[1].strip() not in {"", "this is a specification"} ] # Step 6: Check there are no comments in the footer footer_index = next( (i for i, line in enumerate(response_code) if line.startswith("====")), None ) if footer_index is not None: response_code = response_code[:footer_index] + [ line for line in response_code[footer_index:] if "\\*" not in line ] # Step 7: Write the sanitized lines back to the file with open(spec_path, "w") as f: f.write("\n".join(response_code)) print( "\nDetailed comments for the specification have been added to the TLA+ file.\n" ) def main(spec_name, revert=False, model=DEFAULT_MODEL, confirm=False): if revert: backup_file = spec_name + ".bak" if os.path.exists(backup_file): shutil.copy(backup_file, spec_name) print(f"Reverted changes to {spec_name}") sys.exit(0) else: print(f"No backup file found for {spec_name}") sys.exit(1) tla_tools_path = check_tla_tools_availability() while True: output, returncode = run_tla_spec(spec_name, tla_tools_path) if returncode != 0: print( f"An error occurred when checking {spec_name} with TLC. Error message:\n{output}" ) # Make a backup of the spec file shutil.copy(spec_name, spec_name + ".bak") # Send error to GPT and get suggestions changes_suggestion = send_error_to_gpt(spec_name, output, model) # Apply the changes apply_changes(spec_name, changes_suggestion, confirm) print( f"Changes applied to {spec_name}. Please check the spec and rerun the tool if necessary." ) else: print(f"No errors detected in {spec_name}.") print("Getting detailed comments for the specification...") provide_detailed_comments(spec_name, model) # Run TLC again to check if the comments introduced any errors print("Running TLC again to check if the comments introduced any errors...") output, returncode = run_tla_spec(spec_name, tla_tools_path) if returncode != 0: print( f"An error occurred after adding comments to {spec_name}. Error message:\n{output}" ) print("Removing comments...") shutil.copy(spec_name + ".bak", spec_name) else: print("No errors detected after adding comments. Exiting...") break if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument( "spec_name", type=str, help="The name of the TLA+ spec to check" ) parser.add_argument( "--revert", action="store_true", help="Revert the spec to its previous state" ) parser.add_argument( "--model", type=str, default=DEFAULT_MODEL, help="The name of the GPT model to use", ) parser.add_argument( "--confirm", action="store_true", help="Ask for confirmation before each change" ) args = parser.parse_args() main(args.spec_name, args.revert, args.model, args.confirm)
[ "\n\n", "tla\n", "\n", "Please rewrite the TLA+ file to include detailed comments ", "that are readable and useful for developers and people new to TLA+. ", "Ensure the code is enclosed within three backticks (```).", "And here is its .cfg file (the model to check):\n\n", "Here is a TLA+ specification:\n\n", "prompt.txt", "Here is the TLA+ spec that has errors that need fixing:\n\nPLACEHOLDER\n\nHere is the TLA+ model requirements that need to be met for more context:\n\nPLACEHOLDER\n\nHere is the TLA+ spec lines that needs fixing:\n\nPLACEHOLDER\n\nHere is the error message:\n\nPLACEHOLDER\nPlease provide your suggested changes, and remember to stick to the exact format as described above.", "Your response could not be parsed by json.loads. Please restate your last message as pure JSON.", "Here is the .cfg file \n" ]
2024-01-10
alliraine/Asterdroid
actions~aster.py
from signalbot import Command, Context import openai import os class AsterCommand(Command): def describe(self) -> str: return "OpenAI Trigger" async def handle(self, c: Context): command = c.message.text print(c.message.mentions) if len(c.message.mentions) > 0: if "2219a1a4-828b-4af1-9804-ca839236d40f" in c.message.mentions[0]["uuid"]: await c.start_typing() await c.react('🤖') openai.api_key = os.getenv("OPEN_AI") response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ {"role": "system", "content": "You are a chat bot called Asterbot. You are here to assisted The Constellation a queer lesbian polycule which consists of Alli, Jen, Ellie, and Sae."}, {"role": "user", "content": command} ] ) await c.send(response.choices[0].message.content) await c.stop_typing() return
[ "You are a chat bot called Asterbot. You are here to assisted The Constellation a queer lesbian polycule which consists of Alli, Jen, Ellie, and Sae." ]
2024-01-10
pli2014/ChatGLM3
tool_agent_demo~main_agent_test_2.py
import os import re import json from typing import List, Union from langchain.agents import Tool, AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.prompts import StringPromptTemplate from langchain import OpenAI, SerpAPIWrapper, LLMChain from langchain.schema import AgentAction, AgentFinish os.environ["OPENAI_API_KEY"] = "1234" def find_person(name: str): """ 模拟本地数据库查询。 Args: name (str): 人物名称,由LLM提取。 Returns: _type_: _description_ """ info = { '张三': '男', '小红': '女' } return info.get(name, f'未找到{name}的性别信息,我应该直接返回 Observation: 未知') def recommend_item(gender: str): """ 根据人物性别推荐不同的商品。 Args: gender (str): 人物的性别,由 LLM 提取。 Returns: _type_: _description_ """ recommend = { '男': ['Steam爆款', 'RTX-9090', 'iPhone 80'], '女': ['夏季碎花裙', '轻盈帆布包', '琉璃唇釉'], '未知': ['AJ新品', '手冲咖啡'] } return recommend.get(gender, f'未找到合适的推荐商品,我应该返回 Final Answer: 随便买些什么吧,只要消费就能快乐!') tools = [ Tool( name = "查询人物性别", func=find_person, description="通过人名查找该人物的性别时用的工具,输入应该是人物的名字" ), Tool( name = "根据性别推荐商品", func=recommend_item, description="当知道了一个人性别后想进一步获得他可能感兴趣的商品时用的工具,输入应该是人物的性别" ) ] template_zh = """按照给定的格式回答以下问题。你可以使用下面这些工具: {tools} 回答时需要遵循以下用---括起来的格式: --- Question: 我需要回答的问题 Thought: 回答这个上述我需要做些什么 Action: `{tool_names}` 中的其中一个工具名 Action Input: 选择工具所需要的输入 Observation: 选择工具返回的结果 ...(这个思考/行动/行动输入/观察可以重复N次) Thought: 我现在知道最终答案 Final Answer: 原始输入问题的最终答案 --- 现在开始回答,记得在给出最终答案前多按照指定格式进行一步一步的推理。 Question: {input} {agent_scratchpad} """ class CustomPromptTemplate(StringPromptTemplate): template: str # 标准模板 tools: List[Tool] # 可使用工具集合 def format( self, **kwargs ) -> str: """ 按照定义的 template,将需要的值都填写进去。 Returns: str: 填充好后的 template。 """ intermediate_steps = kwargs.pop("intermediate_steps") # 取出中间步骤并进行执行 thoughts = "" for action, observation in intermediate_steps: thoughts += action.log thoughts += f"\nObservation: {observation}\nThought: " kwargs["agent_scratchpad"] = thoughts # 记录下当前想法 kwargs["tools"] = "\n".join([f"{tool.name}: {tool.description}" for tool in self.tools]) # 枚举所有可使用的工具名+工具描述 kwargs["tool_names"] = ", ".join([tool.name for tool in self.tools]) # 枚举所有的工具名称 cur_prompt = self.template.format(**kwargs) print(cur_prompt) return cur_prompt prompt = CustomPromptTemplate( template=template_zh, tools=tools, input_variables=["input", "intermediate_steps"] ) class CustomOutputParser(AgentOutputParser): def parse( self, llm_output: str ) -> Union[AgentAction, AgentFinish]: """ 解析 llm 的输出,根据输出文本找到需要执行的决策。 Args: llm_output (str): _description_ Raises: ValueError: _description_ Returns: Union[AgentAction, AgentFinish]: _description_ """ if "Final Answer:" in llm_output: # 如果句子中包含 Final Answer 则代表已经完成 return AgentFinish( return_values={"output": llm_output.split("Final Answer:")[-1].strip()}, log=llm_output, ) # 'Thought: 首先我需要知道张三的性别,然后才能根据性别推荐合适的商品。\nAction: 查询人物性别\nInput: 张三\nAction Output: 男\n\n Thought: 既然张三是男性,我会推荐一些男性可能感兴趣的商品。\n # Action: 根据性别推荐商品\nInput: 男\nAction Output: 请稍等,我需要调用相关工具来获取推荐结果' print(f"llm_output:{llm_output}") regex = r"Thought.*Action:(.*?)\nAction Input:[\s]*(.*?)\n" # 解析 action_input 和 action match = re.search(regex, llm_output, re.DOTALL) if not match: raise ValueError(f"Could not parse LLM output: `{llm_output}`") action = match.group(1).strip() action_input = match.group(2) return AgentAction(tool=action, tool_input=action_input.strip(" ").strip('"'), log=llm_output) output_parser = CustomOutputParser() #llm = OpenAI(temperature=0) #指定ChatGLM2-6B的API endpoint url,用langchain的ChatOpanAI类初始化一个ChatGLM的chat模型 from langchain.chat_models import ChatOpenAI llm = ChatOpenAI( model_name="chatglm", openai_api_base="http://127.0.0.1:6006/v1", openai_api_key="1234", streaming=False, temperature=0 ) llm_chain = LLMChain( llm=llm, prompt=prompt ) tool_names = [tool.name for tool in tools] agent = LLMSingleActionAgent( llm_chain=llm_chain, output_parser=output_parser, stop=["\nObservation:"], allowed_tools=tool_names ) agent_executor = AgentExecutor.from_agent_and_tools( agent=agent, tools=tools, verbose=True ) res = agent_executor.run( "我想送点礼物给小红" ) print(res)
[ "input", "intermediate_steps", "按照给定的格式回答以下问题。你可以使用下面这些工具:\n\n{tools}\n\n回答时需要遵循以下用---括起来的格式:\n\n---\nQuestion: 我需要回答的问题\nThought: 回答这个上述我需要做些什么\nAction: `{tool_names}` 中的其中一个工具名\nAction Input: 选择工具所需要的输入\nObservation: 选择工具返回的结果\n...(这个思考/行动/行动输入/观察可以重复N次)\nThought: 我现在知道最终答案\nFinal Answer: 原始输入问题的最终答案\n---\n\n现在开始回答,记得在给出最终答案前多按照指定格式进行一步一步的推理。\n\nQuestion: {input}\n{agent_scratchpad}\n" ]
2024-01-10
pli2014/ChatGLM3
tool_agent_demo~main_agent_test.py
from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.agents import AgentType from langchain.agents import Tool from langchain.chains import LLMChain, LLMMathChain from langchain_experimental.utilities import PythonREPL from chatglm3_6b_llm import Chatglm3_6b_LLM from pydantic import BaseModel, Field from langchain.tools import tool from langchain.prompts.chat import ( HumanMessagePromptTemplate, ChatPromptTemplate ) #指定ChatGLM2-6B的API endpoint url,用langchain的ChatOpanAI类初始化一个ChatGLM的chat模型 from langchain.chat_models import ChatOpenAI llm = ChatOpenAI( model_name="chatglm", openai_api_base="http://127.0.0.1:6006/v1", openai_api_key="1234", streaming=False, ) class WeatherInput(BaseModel): city_name: str = Field(description="城市") @tool("get_weather_tool", return_direct=True, args_schema=WeatherInput) def get_weather_tool(city_name: str) -> str: """ get_weather_tool 根据城市获取当地的天气""" print(f"get_weather_tool the current weather for:{city_name}") return str({"city":city_name,"weather":"多云 23°C"}) tools = [get_weather_tool] agent = initialize_agent(tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True, handle_parsing_errors=True, max_iterations=3, early_stopping_method="generate") user_input = "杭州的天气如何?" answer = agent.run(user_input) print(answer)
[]
2024-01-10
pli2014/ChatGLM3
composite_demo~remote_call_llm_demo.py
#指定ChatGLM2-6B的API endpoint url,用langchain的ChatOpanAI类初始化一个ChatGLM的chat模型 from langchain.chat_models import ChatOpenAI llm = ChatOpenAI( model_name="chatglm", openai_api_base="http://127.0.0.1:6006/v1", openai_api_key="1234", streaming=False, ) #使用会话实体内存,利用ChatGLM在会话过程中分析提到的实体(Entity) from langchain.chains.conversation.memory import ConversationEntityMemory from langchain.chains.conversation.prompt import ENTITY_MEMORY_CONVERSATION_TEMPLATE entity_memory = ConversationEntityMemory(llm=llm, k=5 ) #生成会话链 from langchain.chains import ConversationChain Conversation = ConversationChain( llm=llm, prompt=ENTITY_MEMORY_CONVERSATION_TEMPLATE, memory=entity_memory, verbose=True, ) #开始测试 resp = Conversation.run("你好,我名字叫Loui,在清华工作") print(resp)
[]
2024-01-10
pli2014/ChatGLM3
tool_agent_demo~main_agent_test%20zh.py
import os import re import json from typing import List, Union from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.agents import AgentType from langchain.agents import Tool from langchain.chains import LLMChain, LLMMathChain from langchain_experimental.utilities import PythonREPL from chatglm3_6b_llm import Chatglm3_6b_LLM from pydantic import BaseModel, Field from langchain.tools import tool from langchain.prompts import StringPromptTemplate from langchain.prompts.chat import HumanMessagePromptTemplate,ChatPromptTemplate from langchain.agents import Tool, AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.schema import AgentAction, AgentFinish #指定ChatGLM2-6B的API endpoint url,用langchain的ChatOpanAI类初始化一个ChatGLM的chat模型 from langchain.chat_models import ChatOpenAI llm = ChatOpenAI( model_name="chatglm3-6b", openai_api_base="http://127.0.0.1:6006/v1", openai_api_key="1234", streaming=False, ) template_zh = """按照给定的格式回答以下问题。你可以使用下面这些工具: {tools} 回答时需要遵循以下用---括起来的格式: --- Question: 我需要回答的问题 Thought: 回答这个上述我需要做些什么 Action: ”{tool_names}“ 中的其中一个工具名 Action Input: 选择工具所需要的输入 Observation: 选择工具返回的结果 ...(这个Thought/Action/Action Input/Observation可以重复N次) Thought: 我现在知道最终答案 Final Answer: 原始输入问题的最终答案 --- 现在开始回答,记得在给出最终答案前多按照指定格式进行一步一步的推理。 Question: {input} {agent_scratchpad} """ class CustomPromptTemplate(StringPromptTemplate): template: str # 标准模板 tools: List[Tool] # 可使用工具集合 def format( self, **kwargs ) -> str: """ 按照定义的 template,将需要的值都填写进去。 Returns: str: 填充好后的 template。 """ intermediate_steps = kwargs.pop("intermediate_steps") # 取出中间步骤并进行执行 thoughts = "" for action, observation in intermediate_steps: thoughts += action.log thoughts += f"\nObservation: {observation}\nThought: " kwargs["agent_scratchpad"] = thoughts # 记录下当前想法 kwargs["tools"] = "\n".join([f"{tool.name}: {tool.description}" for tool in self.tools]) # 枚举所有可使用的工具名+工具描述 kwargs["tool_names"] = ", ".join([tool.name for tool in self.tools]) # 枚举所有的工具名称 cur_prompt = self.template.format(**kwargs) print(cur_prompt) return cur_prompt class WeatherInput(BaseModel): city_name: str = Field(description="城市") @tool("get_weather_tool", return_direct=True, args_schema=WeatherInput) def get_weather_tool(city_name: str) -> str: """ get_weather_tool 根据城市获取当地的天气""" print(f"get_weather_tool the current weather for:{city_name}") return str({"city":city_name,"weather":"多云 23°C"}) tools = [get_weather_tool] prompt = CustomPromptTemplate( template=template_zh, tools=tools, input_variables=["input", "intermediate_steps"] ) llm_chain = LLMChain( llm=llm, prompt=prompt ) class CustomOutputParser(AgentOutputParser): def parse( self, llm_output: str ) -> Union[AgentAction, AgentFinish]: """ 解析 llm 的输出,根据输出文本找到需要执行的决策。 Args: llm_output (str): _description_ Raises: ValueError: _description_ Returns: Union[AgentAction, AgentFinish]: _description_ """ if "Final Answer:" in llm_output: # 如果句子中包含 Final Answer 则代表已经完成 return AgentFinish( return_values={"output": llm_output.split("Final Answer:")[-1].strip()}, log=llm_output, ) # 'Thought: 首先我需要知道张三的性别,然后才能根据性别推荐合适的商品。\nAction: 查询人物性别\nInput: 张三\nAction Output: 男\n\n Thought: 既然张三是男性,我会推荐一些男性可能感兴趣的商品。\n # Action: 根据性别推荐商品\nInput: 男\nAction Output: 请稍等,我需要调用相关工具来获取推荐结果' print(f"llm_output:{llm_output}") regex = r"Thought.*Action:(.*?)\nAction Input:[\s]*(.*?)\n" # 解析 action_input 和 action match = re.search(regex, llm_output, re.DOTALL) if not match: raise ValueError(f"Could not parse LLM output: `{llm_output}`") action = match.group(1).strip() action_input = match.group(2) return AgentAction(tool=action, tool_input=action_input.strip(" ").strip('"'), log=llm_output) output_parser = CustomOutputParser() tool_names = [tool.name for tool in tools] agent = LLMSingleActionAgent( llm_chain=llm_chain, output_parser=output_parser, stop=["\nObservation:"], allowed_tools=tool_names, ) agent_executor = AgentExecutor.from_agent_and_tools( agent=agent, tools=tools, verbose=True ) res = agent_executor.run( "杭州的气温怎样?" ) print(res)
[ "按照给定的格式回答以下问题。你可以使用下面这些工具:\n\n{tools}\n\n回答时需要遵循以下用---括起来的格式:\n\n---\nQuestion: 我需要回答的问题\nThought: 回答这个上述我需要做些什么\nAction: ”{tool_names}“ 中的其中一个工具名\nAction Input: 选择工具所需要的输入\nObservation: 选择工具返回的结果\n...(这个Thought/Action/Action Input/Observation可以重复N次)\nThought: 我现在知道最终答案\nFinal Answer: 原始输入问题的最终答案\n---\n\n现在开始回答,记得在给出最终答案前多按照指定格式进行一步一步的推理。\n\nQuestion: {input}\n{agent_scratchpad}\n", "input", "intermediate_steps" ]
2024-01-10
pli2014/ChatGLM3
composite_demo~chatglm3_6b_llm.py
#从typing库中导入必要的函数和类型声明 from typing import Any, List, Mapping, Optional #导入所需的类和接口 from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM import requests import json # 语法解释: # 1. 使用typing库中的相关类型进行类型声明 # 2. 使用继承实现自定义LLM类的功能扩展 # 3. 通过重写父类的方法以实现特定的功能需求 # 4. 使用@property装饰器很好地实现了对私有变量和方法的封装和保护 # 5. _identifying_params属性和_llm_type属性分别用于标识和记录各自对象的属性和类型信息。 #定义一个名为CustomLLM的子类,继承自LLM类 base_url = "http://localhost:6006" # 把这里改成你的IP地址 def create_chat_completion(model, messages, use_stream=False): data = { "model": model, # 模型名称 "messages": messages, # 会话历史 "stream": use_stream, # 是否流式响应 "max_tokens": 100, # 最多生成字数 "temperature": 0.8, # 温度 "top_p": 0.8, # 采样概率 } response = requests.post(f"{base_url}/v1/chat/completions", json=data, stream=use_stream) if response.status_code == 200: if use_stream: # 处理流式响应 for line in response.iter_lines(): if line: decoded_line = line.decode('utf-8')[6:] try: response_json = json.loads(decoded_line) content = response_json.get("choices", [{}])[0].get("delta", {}).get("content", "") print(content) except: print("Special Token:", decoded_line) else: # 处理非流式响应 decoded_line = response.json() content = decoded_line.get("choices", [{}])[0].get("message", "").get("content", "") return content else: print("Error:", response.status_code) return None class Chatglm3_6b_LLM(LLM): # 类的成员变量,类型为整型 n: int # 用于指定该子类对象的类型 @property def _llm_type(self) -> str: return "Chatglm3_6b_LLM" # 重写基类方法,根据用户输入的prompt来响应用户,返回字符串 def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> str: chat_messages = [ { "role": "system", "content": "从现在开始扮演一个agent智能体" } ] chat_messages.append({"role": "user", "content": prompt}) response = create_chat_completion("chatglm3-6b", chat_messages, use_stream=False) return response # 返回一个字典类型,包含LLM的唯一标识 @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {"n": self.n}
[ "从现在开始扮演一个agent智能体" ]
2024-01-10
pli2014/ChatGLM3
tool_agent_demo~chatglm3_6b_llm.py
#从typing库中导入必要的函数和类型声明 from typing import Any, List, Mapping, Optional #导入所需的类和接口 from langchain.callbacks.manager import CallbackManagerForLLMRun from langchain.llms.base import LLM import requests import json # 语法解释: # 1. 使用typing库中的相关类型进行类型声明 # 2. 使用继承实现自定义LLM类的功能扩展 # 3. 通过重写父类的方法以实现特定的功能需求 # 4. 使用@property装饰器很好地实现了对私有变量和方法的封装和保护 # 5. _identifying_params属性和_llm_type属性分别用于标识和记录各自对象的属性和类型信息。 #定义一个名为CustomLLM的子类,继承自LLM类 base_url = "http://localhost:6006" # 把这里改成你的IP地址 def create_chat_completion(model, messages, use_stream=False): data = { "model": model, # 模型名称 "messages": messages, # 会话历史 "stream": use_stream, # 是否流式响应 "max_tokens": 100, # 最多生成字数 "temperature": 0.8, # 温度 "top_p": 0.8, # 采样概率 } response = requests.post(f"{base_url}/v1/chat/completions", json=data, stream=use_stream) if response.status_code == 200: if use_stream: # 处理流式响应 for line in response.iter_lines(): if line: decoded_line = line.decode('utf-8')[6:] try: response_json = json.loads(decoded_line) content = response_json.get("choices", [{}])[0].get("delta", {}).get("content", "") #print(content) except: print("Special Token:", decoded_line) else: # 处理非流式响应 decoded_line = response.json() content = decoded_line.get("choices", [{}])[0].get("message", "").get("content", "") return content else: print("Error:", response.status_code) return None class Chatglm3_6b_LLM(LLM): # 用于指定该子类对象的类型 @property def _llm_type(self) -> str: return "Chatglm3_6b_LLM" # 重写基类方法,根据用户输入的prompt来响应用户,返回字符串 def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> str: chat_messages = [ { "role": "system", "content": "从现在开始扮演一个agent智能体" } ] chat_messages.append({"role": "user", "content": prompt}) response = create_chat_completion("chatglm3-6b", chat_messages, use_stream=False) print(f"===llm answer with response: {response}") return response # 返回一个字典类型,包含LLM的唯一标识 @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {"model": "Chatglm3_6b_LLM"}
[ "从现在开始扮演一个agent智能体" ]
2024-01-10
pli2014/ChatGLM3
composite_demo~FakeListLLM_test.py
from langchain.llms.fake import FakeListLLM from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.agents import AgentType from langchain.agents import Tool from langchain.chains import LLMMathChain from langchain_experimental.utilities import PythonREPL from chatglm3_6b_llm import Chatglm3_6b_LLM llm = Chatglm3_6b_LLM(n=10) # 定义函数系列工具 python_repl = PythonREPL() # 函数定义-python def get_python_repl(script:str)->str: """ 1, 不做安全验证 """ print(f"call this script:{script}") return python_repl.run(script) # 生成工具类 python_repl_tool = Tool( name="python_repl_tool", description="这是一个调用执行Python脚本", func=get_python_repl ) llm_math = LLMMathChain(llm = llm) # initialize the math tool llm_math_tool = Tool( name ='llm_math_tool', func = llm_math.run, description ='Useful for when you need to answer questions about math.' ) tools = [python_repl_tool, llm_math_tool] agent = initialize_agent(tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True, handle_parsing_errors=True, max_iterations=5, early_stopping_method="generate") user_input = "执行下python脚本: print('hello world')" answer = agent.run(user_input) print(answer)
[]
2024-01-10
pli2014/ChatGLM3
tool_agent_demo~main_agent_test_1.py
from langchain.agents import load_tools from langchain.agents import initialize_agent from langchain.agents import AgentType from langchain.agents import Tool from langchain.chains import LLMChain, LLMMathChain from langchain_experimental.utilities import PythonREPL from chatglm3_6b_llm import Chatglm3_6b_LLM from pydantic import BaseModel, Field from langchain.tools import tool from langchain.prompts.chat import ( HumanMessagePromptTemplate, ChatPromptTemplate ) #指定ChatGLM2-6B的API endpoint url,用langchain的ChatOpanAI类初始化一个ChatGLM的chat模型 from langchain.chat_models import ChatOpenAI llm = ChatOpenAI( model_name="chatglm", openai_api_base="http://127.0.0.1:6006/v1", openai_api_key="1234", streaming=False, ) #使用会话实体内存,利用ChatGLM在会话过程中分析提到的实体(Entity) from langchain.chains.conversation.memory import ConversationEntityMemory from langchain.chains.conversation.prompt import ENTITY_MEMORY_CONVERSATION_TEMPLATE entity_memory = ConversationEntityMemory(llm=llm, k=5 ) #生成会话链 # from langchain.chains import ConversationChain # conversation = ConversationChain( # llm=llm, # prompt=ENTITY_MEMORY_CONVERSATION_TEMPLATE, # memory=entity_memory, # verbose=True, # ) class WeatherInput(BaseModel): location: str = Field(description="The name of the city to be queried") @tool("get_weather_tool", return_direct=True, args_schema=WeatherInput) def get_weather_tool(location: str) -> str: """get_weathe_tool the current weather for `city_name`""" print(f"get_weather_tool the current weather for:{location}") return f"城市:{location},天气:多云 23°C " tools = [get_weather_tool] # agent = initialize_agent(tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True, # handle_parsing_errors=True, max_iterations=3, early_stopping_method="generate") # user_input = "杭州的天气如何?" # answer = agent.run(user_input) # print(answer) agent_prompt = """Answer the following questions as best you can. You have access to the following tools: Search Order: 一个能够查询订单信息,获得最新的订单情况的工具,参数是输入订单id Recommend product: 一个能够基于商品及用户 信息为用户进行商品推荐导购的工具,参数是输入要推荐的商品类型 The way you use the tools is by specifying a json blob. Specifically, this json should have a `action` key (with the name of the tool to use) and a `action_input` key (with the input to the tool going here). The only values that should be in the "action" field are: Search Order, Recommend product The $JSON_BLOB should only contain a SINGLE action, do NOT return a list of multiple actions. Here is an example of a valid $JSON_BLOB: \`\`\` {{ "action": $TOOL_NAME, "action_input": $INPUT }} \`\`\` ALWAYS use the following format: Question: the input question you must answer Thought: you should always think about what to do Action: \`\`\` $JSON_BLOB \`\`\` Observation: the result of the action ... (this Thought/Action/Observation can repeat N times) Thought: I now know the final answer Final Answer: the final answer to the original input question Begin! Reminder to always use the exact characters `Final Answer` when responding.' {question} """ question="我想卖一件衣服,但不知道哪款适合我,有什么好推荐吗" messages = [ HumanMessagePromptTemplate.from_template( template=agent_prompt, ).format(question=question), ] # 模拟电商订单 def search_order(input: str)->str: print("调用search_order:一个能够查询订单信息,获得最新的订单情况的工具:") return "{order},订单状态:已发货".format(order=input) # 模拟商品推荐 def recommend_product(input: str)->str: print("调用recommend_product:一个能够基于商品及用户信息为用户进行商品推荐导购的工具:") return "黑色连衣裙" tools = [ Tool( name="Search Order", func=search_order, description="""一个能够查询订单信息,获得最新的订单情况的工具,参数是输入订单id""" ), Tool( name="Recommend product", func=recommend_product, description="一个能够基于商品及用户信息为用户进行商品推荐导购的工具,参数是输入要推荐的商品类型" ) ] agent_tools = initialize_agent(tools=tools, llm=llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, handle_parsing_errors=True, verbose=True) result = agent_tools.run("我想卖一件衣服,但不知道哪款适合我,有什么好推荐吗") print(result)
[ "Answer the following questions as best you can. You have access to the following tools:\n\nSearch Order:\n一个能够查询订单信息,获得最新的订单情况的工具,参数是输入订单id\nRecommend product: 一个能够基于商品及用户\n信息为用户进行商品推荐导购的工具,参数是输入要推荐的商品类型\n\nThe way you use the tools is by specifying a json blob.\nSpecifically, this json should have a `action` key (with the name of the tool to use) and a `action_input` key (with the input to the tool going here).\n\nThe only values that should be in the \"action\" field are: Search Order, Recommend product\n\nThe $JSON_BLOB should only contain a SINGLE action, do NOT return a list of multiple actions. Here is an example of a valid $JSON_BLOB:\n\n\\`\\`\\`\n{{\n \"action\": $TOOL_NAME,\n \"action_input\": $INPUT\n}}\n\\`\\`\\`\n\nALWAYS use the following format:\n\nQuestion: the input question you\nmust answer\nThought: you should always think about what to do\nAction:\n\\`\\`\\`\n$JSON_BLOB\n\\`\\`\\`\nObservation:\nthe result of the action\n... (this Thought/Action/Observation can repeat N times)\nThought: I now know the\nfinal answer\nFinal Answer: the final answer to the original input question\n\nBegin! Reminder to always use the exact characters `Final Answer` when responding.'\n{question}\n" ]
2024-01-10
PacktPublishing/LangChain-MasterClass---Build-15-OpenAI-and-LLAMA-2-LLM-Apps-using-Python
S19%20-%20Project%2010%20-%20Automatic%20Ticket%20Classification%20Tool~Automatic%20Ticket%20Classification%20Tool~user_utils.py
import pinecone from langchain.vectorstores import Pinecone from langchain.embeddings.sentence_transformer import SentenceTransformerEmbeddings from langchain.llms import OpenAI from langchain.chains.question_answering import load_qa_chain from langchain.callbacks import get_openai_callback import joblib #Function to pull index data from Pinecone def pull_from_pinecone(pinecone_apikey,pinecone_environment,pinecone_index_name,embeddings): pinecone.init( api_key=pinecone_apikey, environment=pinecone_environment ) index_name = pinecone_index_name index = Pinecone.from_existing_index(index_name, embeddings) return index def create_embeddings(): embeddings = SentenceTransformerEmbeddings(model_name="all-MiniLM-L6-v2") return embeddings #This function will help us in fetching the top relevent documents from our vector store - Pinecone Index def get_similar_docs(index,query,k=2): similar_docs = index.similarity_search(query, k=k) return similar_docs def get_answer(docs,user_input): chain = load_qa_chain(OpenAI(), chain_type="stuff") with get_openai_callback() as cb: response = chain.run(input_documents=docs, question=user_input) return response def predict(query_result): Fitmodel = joblib.load('modelsvm.pk1') result=Fitmodel.predict([query_result]) return result[0]
[]
2024-01-10
pdjewell/steph_scope_demo
scripts~extractdata.py
import os import json import argparse from langchain.document_loaders import PyPDFLoader, PyMuPDFLoader, DirectoryLoader from langchain.text_splitter import CharacterTextSplitter, NLTKTextSplitter, RecursiveCharacterTextSplitter, SpacyTextSplitter import gspread from google.auth import default import pandas as pd from dotenv import load_dotenv from langchain.vectorstores import Chroma from langchain.embeddings.openai import OpenAIEmbeddings from langchain.embeddings import HuggingFaceInstructEmbeddings from langchain.chat_models import ChatOpenAI import openai import gspread import nltk # nltk.download('punkt') from google.oauth2.service_account import Credentials class DataExtractor: def __init__(self, dir_path): self.dir_path = dir_path def extract_data(self): with open(self.filename, 'r') as f: data = f.read() return data def load_pdfs(self): dir_loader = DirectoryLoader(self.dir_path, glob="**/*.pdf", loader_cls=PyPDFLoader, show_progress=True, use_multithreading=True, silent_errors=True) docs = dir_loader.load() print(f"\nNumber of docs after initial loading: {len(docs)}, from: {self.dir_path}") return docs def chunk_docs(self, docs, chunk_size=2000, nltk=True, spacy=False, recursive=False): if nltk: text_splitter = NLTKTextSplitter(chunk_size=chunk_size, chunk_overlap=0) elif spacy: text_splitter = SpacyTextSplitter(chunk_size=chunk_size) elif recursive: text_splitter = RecursiveCharacterTextSplitter( chunk_size=chunk_size, chunk_overlap=0, length_function=len, separators=["\n\n\n","\n\n", "\n", ". ", " ", ""],) else: text_splitter = CharacterTextSplitter( separator='\n', chunk_size=chunk_size, chunk_overlap=0, length_function=len,) all_text = [doc.page_content for doc in docs] metadatas = [doc.metadata for doc in docs] chunks = text_splitter.create_documents(all_text, metadatas=metadatas) print(f"Number of chunks: {len(chunks)}") return chunks def update_metadata_from_file(self, docs, file_path=None): if file_path is None: file_path = os.path.join(os.getcwd(), "guideline_metadata.csv") else: file_path = file_path metadata_df = pd.read_csv(file_path) metadata_df = metadata_df.fillna(" ") metadata_df.set_index('file_name', inplace=True) metadata_dict = metadata_df.to_dict(orient='index') for doc in docs: file_name = os.path.basename(doc.metadata['source']) try: merged_dict = metadata_dict[file_name].copy() merged_dict['broad_category'] = merged_dict['broad_category'].lower().strip() merged_dict['file_name'] = file_name merged_dict.update(doc.metadata) doc.metadata = merged_dict except KeyError: print(f"File name {file_name} not found in metadata file. Skipping...") print(f"Updated {len(docs)} docs metadata from: {file_path}") return docs ### NOT WORKING YET ## Problem with authentication def update_metadata_from_gsheets(self, docs, key=None): # Get google sheets key from env load_dotenv() key = os.environ.get('GOOGLE_SHEETS_KEY') credentials = os.environ.get('GOOGLE_SERVICE_ACCOUNT') # Google Sheets authorization. creds, _ = default() creds = Credentials.from_service_account_file(credentials) gc = gspread.authorize(creds) # Open the Google Spreadsheet by its title. workbook = gc.open_by_key(key) worksheet = workbook.sheet1 # Get all values in the sheet and convert it into DataFrame. values = worksheet.get_all_values() metadata_df = pd.DataFrame(values[1:], columns=values[0]) metadata_df.set_index('file_name', inplace=True) # Convert DataFrame to dictionary. metadata_dict = metadata_df.to_dict(orient='index') for doc in docs: file_name = os.path.basename(doc.metadata['source']) merged_dict = metadata_dict[file_name].copy() merged_dict['file_name'] = file_name merged_dict.update(doc.metadata) doc.metadata = merged_dict print(f"Updated {len(docs)} docs metadata from: {workbook}") return docs def add_chunk_index(self, chunks): sources = [] for chunk in chunks: sources.append(chunk.metadata['source']) list(set(sources)) for source in sources: chunk_index = 0 for chunk in chunks: if source == chunk.metadata['source']: chunk.metadata['chunk_index'] = chunk_index chunk_index += 1 else: continue total_chunks = chunk_index for chunk in chunks: if source == chunk.metadata['source']: chunk.metadata['last_chunk_index'] = total_chunks - 1 print(f"Added chunk_index to metadata of {len(chunks)} chunks") return chunks # Create chroma vectorstore def get_chroma_vectorstore(self, chunks, use_openai=True, persist_directory="chroma_db"): # load api keys load_dotenv() openai.api_key = os.environ.get('OPENAI_API_KEY') if use_openai: model_name = 'text-embedding-ada-002' embeddings = OpenAIEmbeddings(model=model_name, openai_api_key=openai.api_key) else: model_name = "hkunlp/instructor-xl" embed_instruction = "Represent the text from the clinical guidelines" query_instruction = "Query the most relevant text from clinical guidelines" embeddings = HuggingFaceInstructEmbeddings(model_name=model_name, embed_instruction=embed_instruction, query_instruction=query_instruction) # Create vectorstore vectorstore = Chroma.from_documents(documents=chunks, embedding=embeddings, persist_directory=persist_directory) print(f"Created chroma vectorscore, called: {persist_directory}") return vectorstore def __call__(self, persist_directory="chroma_db"): docs = self.load_pdfs() chunks = self.chunk_docs(docs, chunk_size=3000, nltk=True) chunks = self.update_metadata_from_file(chunks, file_path=None) chunks = self.add_chunk_index(chunks) vectorstore = self.get_chroma_vectorstore(chunks, use_openai=True, persist_directory=persist_directory) return vectorstore def main(dir_path=None): if dir_path is None: # take shell arguments for directory path parser = argparse.ArgumentParser() parser.add_argument("--dir_path", type=str, default=None) args = parser.parse_args() dir_path = args.dir_path else: dir_path = dir_path extractor = DataExtractor(dir_path) persist_directory = os.path.join(os.getcwd(), "vector_db", "chroma_db_ada_embeddings") vectorstore = extractor(persist_directory) return None if __name__ == "__main__": main(dir_path="./guidelines")
[]
2024-01-10
pdjewell/steph_scope_demo
scripts~loadvectorstore.py
import os import streamlit as st import openai from langchain.vectorstores import Chroma from langchain.embeddings.openai import OpenAIEmbeddings from dotenv import load_dotenv class VectorStoreLoader: def __init__(self, vector_db_directory=os.path.join(os.getcwd(), 'vector_db'), ): self.vector_db_directory = vector_db_directory # Load chroma db from file def load_chroma_db(self, persist_directory): # get openai api key from env load_dotenv() openai.api_key = os.environ.get('OPENAI_API_KEY') # Select embedding model model_name = 'text-embedding-ada-002' embeddings = OpenAIEmbeddings(model=model_name, openai_api_key=openai.api_key) # Load vectorstore vectorstore = Chroma(persist_directory=persist_directory, embedding_function=embeddings) return vectorstore def load_all_chroma_db(self): vectorstores = {} for vector_db in os.listdir(self.vector_db_directory): print(vector_db) vector_db_path = os.path.join(self.vector_db_directory, vector_db) print(vector_db_path) vectorstore = self.load_chroma_db(vector_db_path) vectorstores[vector_db] = vectorstore print(f"Successfully loaded: {len(vectorstores)} vectorstores") return vectorstores
[]
2024-01-10
ivanbelenky/jsonllm
jsonllm~completions.py
from typing import Dict, Union from openai import ChatCompletion, Completion from vertexai.preview.language_models import ChatModel, TextGenerationModel # type: ignore from .utils import no_tokens, OpenAIErrors from .constants import MAX_TOKENS, DEFAULT_MODEL_KWARGS, OPENAI_MODELS, GOOGLE_MODELS class _Completion: class ClientError(Exception): pass class ServerError(Exception): pass @staticmethod def _openai(prompt: str, model: str, **model_kwargs: Dict[str, Union[str, float, int]]) -> str: try: model_kwargs = model_kwargs or DEFAULT_MODEL_KWARGS[model] if MAX_TOKENS[model] - no_tokens(prompt) < 0: raise Exception("Failed to complete prompt, not enough tokens left. Try reducing prompt length.") if 'gpt-3.5-turbo' in model: return ChatCompletion.create(model=model, messages=[{'role':'user','content':prompt}], **model_kwargs).choices[0].message.content # type: ignore elif any([m in model for m in ['ada', 'babbage', 'curie', 'davinci']]): return Completion.create(model=model ,prompt=prompt, **model_kwargs).choices[0].text # type: ignore except OpenAIErrors as e: raise _Completion.ServerError(f"Failed to complete prompt: {e}") except Exception as e: raise _Completion.ClientError(f"Failed to complete prompt: {e}") raise ValueError(f"Model {model} not implemented") @staticmethod def _google(prompt: str, model: str, **model_kwargs: Dict[str, Union[str, float, int]]) -> str: '''Usable models (tested) are: chat-bison@001 & †ext-bison@001''' try: if 'chat' in model: return ChatModel.from_pretrained(model).start_chat(examples=[]).send_message(prompt, **model_kwargs).text # type: ignore elif 'text' in model: return TextGenerationModel.from_pretrained(model).predict(prompt, **model_kwargs).text # type: ignore # except GoogleErrors as e: TODO: what error Google is responsible for? except Exception as e: raise _Completion.ClientError(f"Failed to complete prompt: {e}") raise ValueError(f"Model {model} not implemented") @staticmethod def complete_prompt(prompt: str, model: str, **model_kwargs: Dict[str, Union[str, float, int]]) -> str: if model in OPENAI_MODELS: return _Completion._openai(prompt, model=model, **model_kwargs) elif model in GOOGLE_MODELS: return _Completion._google(prompt, model=model, **model_kwargs) else: raise NotImplementedError(f"Completion model {model} not implemented")
[]
2024-01-10
yuejunzhang/AudioGPT
audio-chatgpt.py
import sys import os sys.path.append(os.path.dirname(os.path.realpath(__file__))) sys.path.append(os.path.dirname(os.path.dirname(os.path.realpath(__file__)))) sys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'NeuralSeq')) sys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'text_to_audio/Make_An_Audio')) sys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'audio_detection')) sys.path.append(os.path.join(os.path.dirname(os.path.realpath(__file__)), 'mono2binaural')) import gradio as gr import matplotlib import librosa import torch from langchain.agents.initialize import initialize_agent from langchain.agents.tools import Tool from langchain.chains.conversation.memory import ConversationBufferMemory from langchain.llms.openai import OpenAI import re import uuid import soundfile from PIL import Image import numpy as np from omegaconf import OmegaConf from einops import repeat from ldm.util import instantiate_from_config from ldm.data.extract_mel_spectrogram import TRANSFORMS_16000 from vocoder.bigvgan.models import VocoderBigVGAN from ldm.models.diffusion.ddim import DDIMSampler import whisper from utils.hparams import set_hparams from utils.hparams import hparams as hp import scipy.io.wavfile as wavfile import librosa from audio_infer.utils import config as detection_config from audio_infer.pytorch.models import PVT import clip import numpy as np AUDIO_CHATGPT_PREFIX = """AudioGPT AudioGPT can not directly read audios, but it has a list of tools to finish different speech, audio, and singing voice tasks. Each audio will have a file name formed as "audio/xxx.wav". When talking about audios, AudioGPT is very strict to the file name and will never fabricate nonexistent files. AudioGPT is able to use tools in a sequence, and is loyal to the tool observation outputs rather than faking the audio content and audio file name. It will remember to provide the file name from the last tool observation, if a new audio is generated. Human may provide new audios to AudioGPT with a description. The description helps AudioGPT to understand this audio, but AudioGPT should use tools to finish following tasks, rather than directly imagine from the description. Overall, AudioGPT is a powerful audio dialogue assistant tool that can help with a wide range of tasks and provide valuable insights and information on a wide range of topics. TOOLS: ------ AudioGPT has access to the following tools:""" AUDIO_CHATGPT_FORMAT_INSTRUCTIONS = """To use a tool, please use the following format: ``` Thought: Do I need to use a tool? Yes Action: the action to take, should be one of [{tool_names}] Action Input: the input to the action Observation: the result of the action ``` When you have a response to say to the Human, or if you do not need to use a tool, you MUST use the format: ``` Thought: Do I need to use a tool? No {ai_prefix}: [your response here] ``` """ AUDIO_CHATGPT_SUFFIX = """You are very strict to the filename correctness and will never fake a file name if not exists. You will remember to provide the audio file name loyally if it's provided in the last tool observation. Begin! Previous conversation history: {chat_history} New input: {input} Thought: Do I need to use a tool? {agent_scratchpad}""" def cut_dialogue_history(history_memory, keep_last_n_words = 500): tokens = history_memory.split() n_tokens = len(tokens) print(f"history_memory:{history_memory}, n_tokens: {n_tokens}") if n_tokens < keep_last_n_words: return history_memory else: paragraphs = history_memory.split('\n') last_n_tokens = n_tokens while last_n_tokens >= keep_last_n_words: last_n_tokens = last_n_tokens - len(paragraphs[0].split(' ')) paragraphs = paragraphs[1:] return '\n' + '\n'.join(paragraphs) def merge_audio(audio_path_1, audio_path_2): merged_signal = [] sr_1, signal_1 = wavfile.read(audio_path_1) sr_2, signal_2 = wavfile.read(audio_path_2) merged_signal.append(signal_1) merged_signal.append(signal_2) merged_signal = np.hstack(merged_signal) merged_signal = np.asarray(merged_signal, dtype=np.int16) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") wavfile.write(audio_filename, sr_2, merged_signal) return audio_filename class T2I: def __init__(self, device): from transformers import AutoModelForCausalLM, AutoTokenizer from diffusers import StableDiffusionPipeline from transformers import pipeline print("Initializing T2I to %s" % device) self.device = device self.pipe = StableDiffusionPipeline.from_pretrained("runwayml/stable-diffusion-v1-5", torch_dtype=torch.float16) self.text_refine_tokenizer = AutoTokenizer.from_pretrained("Gustavosta/MagicPrompt-Stable-Diffusion") self.text_refine_model = AutoModelForCausalLM.from_pretrained("Gustavosta/MagicPrompt-Stable-Diffusion") self.text_refine_gpt2_pipe = pipeline("text-generation", model=self.text_refine_model, tokenizer=self.text_refine_tokenizer, device=self.device) self.pipe.to(device) def inference(self, text): image_filename = os.path.join('image', str(uuid.uuid4())[0:8] + ".png") refined_text = self.text_refine_gpt2_pipe(text)[0]["generated_text"] print(f'{text} refined to {refined_text}') image = self.pipe(refined_text).images[0] image.save(image_filename) print(f"Processed T2I.run, text: {text}, image_filename: {image_filename}") return image_filename class ImageCaptioning: def __init__(self, device): from transformers import BlipProcessor, BlipForConditionalGeneration print("Initializing ImageCaptioning to %s" % device) self.device = device self.processor = BlipProcessor.from_pretrained("Salesforce/blip-image-captioning-base") self.model = BlipForConditionalGeneration.from_pretrained("Salesforce/blip-image-captioning-base").to(self.device) def inference(self, image_path): inputs = self.processor(Image.open(image_path), return_tensors="pt").to(self.device) out = self.model.generate(**inputs) captions = self.processor.decode(out[0], skip_special_tokens=True) return captions class T2A: def __init__(self, device): print("Initializing Make-An-Audio to %s" % device) self.device = device self.sampler = self._initialize_model('text_to_audio/Make_An_Audio/configs/text_to_audio/txt2audio_args.yaml', 'text_to_audio/Make_An_Audio/useful_ckpts/ta40multi_epoch=000085.ckpt', device=device) self.vocoder = VocoderBigVGAN('text_to_audio/Make_An_Audio/vocoder/logs/bigv16k53w',device=device) def _initialize_model(self, config, ckpt, device): config = OmegaConf.load(config) model = instantiate_from_config(config.model) model.load_state_dict(torch.load(ckpt, map_location='cpu')["state_dict"], strict=False) model = model.to(device) model.cond_stage_model.to(model.device) model.cond_stage_model.device = model.device sampler = DDIMSampler(model) return sampler def txt2audio(self, text, seed = 55, scale = 1.5, ddim_steps = 100, n_samples = 3, W = 624, H = 80): SAMPLE_RATE = 16000 prng = np.random.RandomState(seed) start_code = prng.randn(n_samples, self.sampler.model.first_stage_model.embed_dim, H // 8, W // 8) start_code = torch.from_numpy(start_code).to(device=self.device, dtype=torch.float32) uc = self.sampler.model.get_learned_conditioning(n_samples * [""]) c = self.sampler.model.get_learned_conditioning(n_samples * [text]) shape = [self.sampler.model.first_stage_model.embed_dim, H//8, W//8] # (z_dim, 80//2^x, 848//2^x) samples_ddim, _ = self.sampler.sample(S = ddim_steps, conditioning = c, batch_size = n_samples, shape = shape, verbose = False, unconditional_guidance_scale = scale, unconditional_conditioning = uc, x_T = start_code) x_samples_ddim = self.sampler.model.decode_first_stage(samples_ddim) x_samples_ddim = torch.clamp((x_samples_ddim+1.0)/2.0, min=0.0, max=1.0) # [0, 1] wav_list = [] for idx,spec in enumerate(x_samples_ddim): wav = self.vocoder.vocode(spec) wav_list.append((SAMPLE_RATE,wav)) best_wav = self.select_best_audio(text, wav_list) return best_wav def select_best_audio(self, prompt, wav_list): from wav_evaluation.models.CLAPWrapper import CLAPWrapper clap_model = CLAPWrapper('text_to_audio/Make_An_Audio/useful_ckpts/CLAP/CLAP_weights_2022.pth', 'text_to_audio/Make_An_Audio/useful_ckpts/CLAP/config.yml', use_cuda=torch.cuda.is_available()) text_embeddings = clap_model.get_text_embeddings([prompt]) score_list = [] for data in wav_list: sr, wav = data audio_embeddings = clap_model.get_audio_embeddings([(torch.FloatTensor(wav), sr)], resample=True) score = clap_model.compute_similarity(audio_embeddings, text_embeddings, use_logit_scale=False).squeeze().cpu().numpy() score_list.append(score) max_index = np.array(score_list).argmax() print(score_list, max_index) return wav_list[max_index] def inference(self, text, seed = 55, scale = 1.5, ddim_steps = 100, n_samples = 3, W = 624, H = 80): melbins,mel_len = 80,624 with torch.no_grad(): result = self.txt2audio( text = text, H = melbins, W = mel_len ) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename, result[1], samplerate = 16000) print(f"Processed T2I.run, text: {text}, audio_filename: {audio_filename}") return audio_filename class I2A: def __init__(self, device): print("Initializing Make-An-Audio-Image to %s" % device) self.device = device self.sampler = self._initialize_model('text_to_audio/Make_An_Audio/configs/img_to_audio/img2audio_args.yaml', 'text_to_audio/Make_An_Audio/useful_ckpts/ta54_epoch=000216.ckpt', device=device) self.vocoder = VocoderBigVGAN('text_to_audio/Make_An_Audio/vocoder/logs/bigv16k53w',device=device) def _initialize_model(self, config, ckpt, device): config = OmegaConf.load(config) model = instantiate_from_config(config.model) model.load_state_dict(torch.load(ckpt, map_location='cpu')["state_dict"], strict=False) model = model.to(device) model.cond_stage_model.to(model.device) model.cond_stage_model.device = model.device sampler = DDIMSampler(model) return sampler def img2audio(self, image, seed = 55, scale = 3, ddim_steps = 100, W = 624, H = 80): SAMPLE_RATE = 16000 n_samples = 1 # only support 1 sample prng = np.random.RandomState(seed) start_code = prng.randn(n_samples, self.sampler.model.first_stage_model.embed_dim, H // 8, W // 8) start_code = torch.from_numpy(start_code).to(device=self.device, dtype=torch.float32) uc = self.sampler.model.get_learned_conditioning(n_samples * [""]) #image = Image.fromarray(image) image = Image.open(image) image = self.sampler.model.cond_stage_model.preprocess(image).unsqueeze(0) image_embedding = self.sampler.model.cond_stage_model.forward_img(image) c = image_embedding.repeat(n_samples, 1, 1) shape = [self.sampler.model.first_stage_model.embed_dim, H//8, W//8] # (z_dim, 80//2^x, 848//2^x) samples_ddim, _ = self.sampler.sample(S=ddim_steps, conditioning=c, batch_size=n_samples, shape=shape, verbose=False, unconditional_guidance_scale=scale, unconditional_conditioning=uc, x_T=start_code) x_samples_ddim = self.sampler.model.decode_first_stage(samples_ddim) x_samples_ddim = torch.clamp((x_samples_ddim+1.0)/2.0, min=0.0, max=1.0) # [0, 1] wav_list = [] for idx,spec in enumerate(x_samples_ddim): wav = self.vocoder.vocode(spec) wav_list.append((SAMPLE_RATE,wav)) best_wav = wav_list[0] return best_wav def inference(self, image, seed = 55, scale = 3, ddim_steps = 100, W = 624, H = 80): melbins,mel_len = 80,624 with torch.no_grad(): result = self.img2audio( image=image, H=melbins, W=mel_len ) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename, result[1], samplerate = 16000) print(f"Processed I2a.run, image_filename: {image}, audio_filename: {audio_filename}") return audio_filename class TTS: def __init__(self, device=None): from inference.tts.PortaSpeech import TTSInference if device is None: device = 'cuda' if torch.cuda.is_available() else 'cpu' print("Initializing PortaSpeech to %s" % device) self.device = device self.exp_name = 'checkpoints/ps_adv_baseline' self.set_model_hparams() self.inferencer = TTSInference(self.hp, device) def set_model_hparams(self): set_hparams(exp_name=self.exp_name, print_hparams=False) self.hp = hp def inference(self, text): self.set_model_hparams() inp = {"text": text} out = self.inferencer.infer_once(inp) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename, out, samplerate=22050) return audio_filename class T2S: def __init__(self, device= None): from inference.svs.ds_e2e import DiffSingerE2EInfer if device is None: device = 'cuda' if torch.cuda.is_available() else 'cpu' print("Initializing DiffSinger to %s" % device) self.device = device self.exp_name = 'checkpoints/0831_opencpop_ds1000' self.config= 'NeuralSeq/egs/egs_bases/svs/midi/e2e/opencpop/ds1000.yaml' self.set_model_hparams() self.pipe = DiffSingerE2EInfer(self.hp, device) self.default_inp = { 'text': '你 说 你 不 SP 懂 为 何 在 这 时 牵 手 AP', 'notes': 'D#4/Eb4 | D#4/Eb4 | D#4/Eb4 | D#4/Eb4 | rest | D#4/Eb4 | D4 | D4 | D4 | D#4/Eb4 | F4 | D#4/Eb4 | D4 | rest', 'notes_duration': '0.113740 | 0.329060 | 0.287950 | 0.133480 | 0.150900 | 0.484730 | 0.242010 | 0.180820 | 0.343570 | 0.152050 | 0.266720 | 0.280310 | 0.633300 | 0.444590' } def set_model_hparams(self): set_hparams(config=self.config, exp_name=self.exp_name, print_hparams=False) self.hp = hp def inference(self, inputs): self.set_model_hparams() val = inputs.split(",") key = ['text', 'notes', 'notes_duration'] try: inp = {k: v for k, v in zip(key, val)} wav = self.pipe.infer_once(inp) except: print('Error occurs. Generate default audio sample.\n') inp = self.default_inp wav = self.pipe.infer_once(inp) #if inputs == '' or len(val) < len(key): # inp = self.default_inp #else: # inp = {k:v for k,v in zip(key,val)} #wav = self.pipe.infer_once(inp) wav *= 32767 audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") wavfile.write(audio_filename, self.hp['audio_sample_rate'], wav.astype(np.int16)) print(f"Processed T2S.run, audio_filename: {audio_filename}") return audio_filename class t2s_VISinger: def __init__(self, device=None): from espnet2.bin.svs_inference import SingingGenerate if device is None: device = 'cuda' if torch.cuda.is_available() else 'cpu' print("Initializing VISingere to %s" % device) tag = 'AQuarterMile/opencpop_visinger1' self.model = SingingGenerate.from_pretrained( model_tag=str_or_none(tag), device=device, ) phn_dur = [[0. , 0.219 ], [0.219 , 0.50599998], [0.50599998, 0.71399999], [0.71399999, 1.097 ], [1.097 , 1.28799999], [1.28799999, 1.98300004], [1.98300004, 7.10500002], [7.10500002, 7.60400009]] phn = ['sh', 'i', 'q', 'v', 'n', 'i', 'SP', 'AP'] score = [[0, 0.50625, 'sh_i', 58, 'sh_i'], [0.50625, 1.09728, 'q_v', 56, 'q_v'], [1.09728, 1.9832100000000001, 'n_i', 53, 'n_i'], [1.9832100000000001, 7.105360000000001, 'SP', 0, 'SP'], [7.105360000000001, 7.604390000000001, 'AP', 0, 'AP']] tempo = 70 tmp = {} tmp["label"] = phn_dur, phn tmp["score"] = tempo, score self.default_inp = tmp def inference(self, inputs): val = inputs.split(",") key = ['text', 'notes', 'notes_duration'] try: # TODO: input will be update inp = {k: v for k, v in zip(key, val)} wav = self.model(text=inp)["wav"] except: print('Error occurs. Generate default audio sample.\n') inp = self.default_inp wav = self.model(text=inp)["wav"] audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename, wav, samplerate=self.model.fs) return audio_filename class TTS_OOD: def __init__(self, device): from inference.tts.GenerSpeech import GenerSpeechInfer if device is None: device = 'cuda' if torch.cuda.is_available() else 'cpu' print("Initializing GenerSpeech to %s" % device) self.device = device self.exp_name = 'checkpoints/GenerSpeech' self.config = 'NeuralSeq/modules/GenerSpeech/config/generspeech.yaml' self.set_model_hparams() self.pipe = GenerSpeechInfer(self.hp, device) def set_model_hparams(self): set_hparams(config=self.config, exp_name=self.exp_name, print_hparams=False) f0_stats_fn = f'{hp["binary_data_dir"]}/train_f0s_mean_std.npy' if os.path.exists(f0_stats_fn): hp['f0_mean'], hp['f0_std'] = np.load(f0_stats_fn) hp['f0_mean'] = float(hp['f0_mean']) hp['f0_std'] = float(hp['f0_std']) hp['emotion_encoder_path'] = 'checkpoints/Emotion_encoder.pt' self.hp = hp def inference(self, inputs): self.set_model_hparams() key = ['ref_audio', 'text'] val = inputs.split(",") inp = {k: v for k, v in zip(key, val)} wav = self.pipe.infer_once(inp) wav *= 32767 audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") wavfile.write(audio_filename, self.hp['audio_sample_rate'], wav.astype(np.int16)) print( f"Processed GenerSpeech.run. Input text:{val[1]}. Input reference audio: {val[0]}. Output Audio_filename: {audio_filename}") return audio_filename class Inpaint: def __init__(self, device): print("Initializing Make-An-Audio-inpaint to %s" % device) self.device = device self.sampler = self._initialize_model_inpaint('text_to_audio/Make_An_Audio/configs/inpaint/txt2audio_args.yaml', 'text_to_audio/Make_An_Audio/useful_ckpts/inpaint7_epoch00047.ckpt') self.vocoder = VocoderBigVGAN('text_to_audio/Make_An_Audio/vocoder/logs/bigv16k53w',device=device) self.cmap_transform = matplotlib.cm.viridis def _initialize_model_inpaint(self, config, ckpt): config = OmegaConf.load(config) model = instantiate_from_config(config.model) model.load_state_dict(torch.load(ckpt, map_location='cpu')["state_dict"], strict=False) device = torch.device("cuda") if torch.cuda.is_available() else torch.device("cpu") model = model.to(device) print(model.device, device, model.cond_stage_model.device) sampler = DDIMSampler(model) return sampler def make_batch_sd(self, mel, mask, num_samples=1): mel = torch.from_numpy(mel)[None,None,...].to(dtype=torch.float32) mask = torch.from_numpy(mask)[None,None,...].to(dtype=torch.float32) masked_mel = (1 - mask) * mel mel = mel * 2 - 1 mask = mask * 2 - 1 masked_mel = masked_mel * 2 -1 batch = { "mel": repeat(mel.to(device=self.device), "1 ... -> n ...", n=num_samples), "mask": repeat(mask.to(device=self.device), "1 ... -> n ...", n=num_samples), "masked_mel": repeat(masked_mel.to(device=self.device), "1 ... -> n ...", n=num_samples), } return batch def gen_mel(self, input_audio_path): SAMPLE_RATE = 16000 sr, ori_wav = wavfile.read(input_audio_path) print("gen_mel") print(sr,ori_wav.shape,ori_wav) ori_wav = ori_wav.astype(np.float32, order='C') / 32768.0 if len(ori_wav.shape)==2:# stereo ori_wav = librosa.to_mono(ori_wav.T) print(sr,ori_wav.shape,ori_wav) ori_wav = librosa.resample(ori_wav,orig_sr = sr,target_sr = SAMPLE_RATE) mel_len,hop_size = 848,256 input_len = mel_len * hop_size if len(ori_wav) < input_len: input_wav = np.pad(ori_wav,(0,mel_len*hop_size),constant_values=0) else: input_wav = ori_wav[:input_len] mel = TRANSFORMS_16000(input_wav) return mel def gen_mel_audio(self, input_audio): SAMPLE_RATE = 16000 sr,ori_wav = input_audio print("gen_mel_audio") print(sr,ori_wav.shape,ori_wav) ori_wav = ori_wav.astype(np.float32, order='C') / 32768.0 if len(ori_wav.shape)==2:# stereo ori_wav = librosa.to_mono(ori_wav.T) print(sr,ori_wav.shape,ori_wav) ori_wav = librosa.resample(ori_wav,orig_sr = sr,target_sr = SAMPLE_RATE) mel_len,hop_size = 848,256 input_len = mel_len * hop_size if len(ori_wav) < input_len: input_wav = np.pad(ori_wav,(0,mel_len*hop_size),constant_values=0) else: input_wav = ori_wav[:input_len] mel = TRANSFORMS_16000(input_wav) return mel def show_mel_fn(self, input_audio_path): crop_len = 500 crop_mel = self.gen_mel(input_audio_path)[:,:crop_len] color_mel = self.cmap_transform(crop_mel) image = Image.fromarray((color_mel*255).astype(np.uint8)) image_filename = os.path.join('image', str(uuid.uuid4())[0:8] + ".png") image.save(image_filename) return image_filename def inpaint(self, batch, seed, ddim_steps, num_samples=1, W=512, H=512): model = self.sampler.model prng = np.random.RandomState(seed) start_code = prng.randn(num_samples, model.first_stage_model.embed_dim, H // 8, W // 8) start_code = torch.from_numpy(start_code).to(device=self.device, dtype=torch.float32) c = model.get_first_stage_encoding(model.encode_first_stage(batch["masked_mel"])) cc = torch.nn.functional.interpolate(batch["mask"], size=c.shape[-2:]) c = torch.cat((c, cc), dim=1) # (b,c+1,h,w) 1 is mask shape = (c.shape[1]-1,)+c.shape[2:] samples_ddim, _ = self.sampler.sample(S=ddim_steps, conditioning=c, batch_size=c.shape[0], shape=shape, verbose=False) x_samples_ddim = model.decode_first_stage(samples_ddim) mel = torch.clamp((batch["mel"]+1.0)/2.0,min=0.0, max=1.0) mask = torch.clamp((batch["mask"]+1.0)/2.0,min=0.0, max=1.0) predicted_mel = torch.clamp((x_samples_ddim+1.0)/2.0,min=0.0, max=1.0) inpainted = (1-mask)*mel+mask*predicted_mel inpainted = inpainted.cpu().numpy().squeeze() inapint_wav = self.vocoder.vocode(inpainted) return inpainted, inapint_wav def inference(self, input_audio, mel_and_mask, seed = 55, ddim_steps = 100): SAMPLE_RATE = 16000 torch.set_grad_enabled(False) mel_img = Image.open(mel_and_mask['image']) mask_img = Image.open(mel_and_mask["mask"]) show_mel = np.array(mel_img.convert("L"))/255 mask = np.array(mask_img.convert("L"))/255 mel_bins,mel_len = 80,848 input_mel = self.gen_mel_audio(input_audio)[:,:mel_len] mask = np.pad(mask,((0,0),(0,mel_len-mask.shape[1])),mode='constant',constant_values=0) print(mask.shape,input_mel.shape) with torch.no_grad(): batch = self.make_batch_sd(input_mel,mask,num_samples=1) inpainted,gen_wav = self.inpaint( batch=batch, seed=seed, ddim_steps=ddim_steps, num_samples=1, H=mel_bins, W=mel_len ) inpainted = inpainted[:,:show_mel.shape[1]] color_mel = self.cmap_transform(inpainted) input_len = int(input_audio[1].shape[0] * SAMPLE_RATE / input_audio[0]) gen_wav = (gen_wav * 32768).astype(np.int16)[:input_len] image = Image.fromarray((color_mel*255).astype(np.uint8)) image_filename = os.path.join('image', str(uuid.uuid4())[0:8] + ".png") image.save(image_filename) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename, gen_wav, samplerate = 16000) return image_filename, audio_filename class ASR: def __init__(self, device): print("Initializing Whisper to %s" % device) self.device = device self.model = whisper.load_model("base", device=device) def inference(self, audio_path): audio = whisper.load_audio(audio_path) audio = whisper.pad_or_trim(audio) mel = whisper.log_mel_spectrogram(audio).to(self.device) _, probs = self.model.detect_language(mel) options = whisper.DecodingOptions() result = whisper.decode(self.model, mel, options) return result.text def translate_english(self, audio_path): audio = self.model.transcribe(audio_path, language='English') return audio['text'] class A2T: def __init__(self, device): from audio_to_text.inference_waveform import AudioCapModel print("Initializing Audio-To-Text Model to %s" % device) self.device = device self.model = AudioCapModel("audio_to_text/audiocaps_cntrstv_cnn14rnn_trm") def inference(self, audio_path): audio = whisper.load_audio(audio_path) caption_text = self.model(audio) return caption_text[0] class GeneFace: def __init__(self, device=None): print("Initializing GeneFace model to %s" % device) from audio_to_face.GeneFace_binding import GeneFaceInfer if device is None: device = 'cuda' if torch.cuda.is_available() else 'cpu' self.device = device self.geneface_model = GeneFaceInfer(device) print("Loaded GeneFace model") def inference(self, audio_path): audio_base_name = os.path.basename(audio_path)[:-4] out_video_name = audio_path.replace("audio","video").replace(".wav", ".mp4") inp = { 'audio_source_name': audio_path, 'out_npy_name': f'geneface/tmp/{audio_base_name}.npy', 'cond_name': f'geneface/tmp/{audio_base_name}.npy', 'out_video_name': out_video_name, 'tmp_imgs_dir': f'video/tmp_imgs', } self.geneface_model.infer_once(inp) return out_video_name class SoundDetection: def __init__(self, device): self.device = device self.sample_rate = 32000 self.window_size = 1024 self.hop_size = 320 self.mel_bins = 64 self.fmin = 50 self.fmax = 14000 self.model_type = 'PVT' self.checkpoint_path = 'audio_detection/audio_infer/useful_ckpts/audio_detection.pth' self.classes_num = detection_config.classes_num self.labels = detection_config.labels self.frames_per_second = self.sample_rate // self.hop_size # Model = eval(self.model_type) self.model = PVT(sample_rate=self.sample_rate, window_size=self.window_size, hop_size=self.hop_size, mel_bins=self.mel_bins, fmin=self.fmin, fmax=self.fmax, classes_num=self.classes_num) checkpoint = torch.load(self.checkpoint_path, map_location=self.device) self.model.load_state_dict(checkpoint['model']) self.model.to(device) def inference(self, audio_path): # Forward (waveform, _) = librosa.core.load(audio_path, sr=self.sample_rate, mono=True) waveform = waveform[None, :] # (1, audio_length) waveform = torch.from_numpy(waveform) waveform = waveform.to(self.device) # Forward with torch.no_grad(): self.model.eval() batch_output_dict = self.model(waveform, None) framewise_output = batch_output_dict['framewise_output'].data.cpu().numpy()[0] """(time_steps, classes_num)""" # print('Sound event detection result (time_steps x classes_num): {}'.format( # framewise_output.shape)) import numpy as np import matplotlib.pyplot as plt sorted_indexes = np.argsort(np.max(framewise_output, axis=0))[::-1] top_k = 10 # Show top results top_result_mat = framewise_output[:, sorted_indexes[0 : top_k]] """(time_steps, top_k)""" # Plot result stft = librosa.core.stft(y=waveform[0].data.cpu().numpy(), n_fft=self.window_size, hop_length=self.hop_size, window='hann', center=True) frames_num = stft.shape[-1] fig, axs = plt.subplots(2, 1, sharex=True, figsize=(10, 4)) axs[0].matshow(np.log(np.abs(stft)), origin='lower', aspect='auto', cmap='jet') axs[0].set_ylabel('Frequency bins') axs[0].set_title('Log spectrogram') axs[1].matshow(top_result_mat.T, origin='upper', aspect='auto', cmap='jet', vmin=0, vmax=1) axs[1].xaxis.set_ticks(np.arange(0, frames_num, self.frames_per_second)) axs[1].xaxis.set_ticklabels(np.arange(0, frames_num / self.frames_per_second)) axs[1].yaxis.set_ticks(np.arange(0, top_k)) axs[1].yaxis.set_ticklabels(np.array(self.labels)[sorted_indexes[0 : top_k]]) axs[1].yaxis.grid(color='k', linestyle='solid', linewidth=0.3, alpha=0.3) axs[1].set_xlabel('Seconds') axs[1].xaxis.set_ticks_position('bottom') plt.tight_layout() image_filename = os.path.join('image', str(uuid.uuid4())[0:8] + ".png") plt.savefig(image_filename) return image_filename class SoundExtraction: def __init__(self, device): from sound_extraction.model.LASSNet import LASSNet from sound_extraction.utils.stft import STFT import torch.nn as nn self.device = device self.model_file = 'sound_extraction/useful_ckpts/LASSNet.pt' self.stft = STFT() self.model = nn.DataParallel(LASSNet(device)).to(device) checkpoint = torch.load(self.model_file) self.model.load_state_dict(checkpoint['model']) self.model.eval() def inference(self, inputs): #key = ['ref_audio', 'text'] from sound_extraction.utils.wav_io import load_wav, save_wav val = inputs.split(",") audio_path = val[0] # audio_path, text text = val[1] waveform = load_wav(audio_path) waveform = torch.tensor(waveform).transpose(1,0) mixed_mag, mixed_phase = self.stft.transform(waveform) text_query = ['[CLS] ' + text] mixed_mag = mixed_mag.transpose(2,1).unsqueeze(0).to(self.device) est_mask = self.model(mixed_mag, text_query) est_mag = est_mask * mixed_mag est_mag = est_mag.squeeze(1) est_mag = est_mag.permute(0, 2, 1) est_wav = self.stft.inverse(est_mag.cpu().detach(), mixed_phase) est_wav = est_wav.squeeze(0).squeeze(0).numpy() #est_path = f'output/est{i}.wav' audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") print('audio_filename ', audio_filename) save_wav(est_wav, audio_filename) return audio_filename class Binaural: def __init__(self, device): from src.models import BinauralNetwork self.device = device self.model_file = 'mono2binaural/useful_ckpts/m2b/binaural_network.net' self.position_file = ['mono2binaural/useful_ckpts/m2b/tx_positions.txt', 'mono2binaural/useful_ckpts/m2b/tx_positions2.txt', 'mono2binaural/useful_ckpts/m2b/tx_positions3.txt', 'mono2binaural/useful_ckpts/m2b/tx_positions4.txt', 'mono2binaural/useful_ckpts/m2b/tx_positions5.txt'] self.net = BinauralNetwork(view_dim=7, warpnet_layers=4, warpnet_channels=64, ) self.net.load_from_file(self.model_file) self.sr = 48000 def inference(self, audio_path): mono, sr = librosa.load(path=audio_path, sr=self.sr, mono=True) mono = torch.from_numpy(mono) mono = mono.unsqueeze(0) import numpy as np import random rand_int = random.randint(0,4) view = np.loadtxt(self.position_file[rand_int]).transpose().astype(np.float32) view = torch.from_numpy(view) if not view.shape[-1] * 400 == mono.shape[-1]: mono = mono[:,:(mono.shape[-1]//400)*400] # if view.shape[1]*400 > mono.shape[1]: m_a = view.shape[1] - mono.shape[-1]//400 rand_st = random.randint(0,m_a) view = view[:,m_a:m_a+(mono.shape[-1]//400)] # # binauralize and save output self.net.eval().to(self.device) mono, view = mono.to(self.device), view.to(self.device) chunk_size = 48000 # forward in chunks of 1s rec_field = 1000 # add 1000 samples as "safe bet" since warping has undefined rec. field rec_field -= rec_field % 400 # make sure rec_field is a multiple of 400 to match audio and view frequencies chunks = [ { "mono": mono[:, max(0, i-rec_field):i+chunk_size], "view": view[:, max(0, i-rec_field)//400:(i+chunk_size)//400] } for i in range(0, mono.shape[-1], chunk_size) ] for i, chunk in enumerate(chunks): with torch.no_grad(): mono = chunk["mono"].unsqueeze(0) view = chunk["view"].unsqueeze(0) binaural = self.net(mono, view).squeeze(0) if i > 0: binaural = binaural[:, -(mono.shape[-1]-rec_field):] chunk["binaural"] = binaural binaural = torch.cat([chunk["binaural"] for chunk in chunks], dim=-1) binaural = torch.clamp(binaural, min=-1, max=1).cpu() #binaural = chunked_forwarding(net, mono, view) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") import torchaudio torchaudio.save(audio_filename, binaural, sr) #soundfile.write(audio_filename, binaural, samplerate = 48000) print(f"Processed Binaural.run, audio_filename: {audio_filename}") return audio_filename class TargetSoundDetection: def __init__(self, device): from target_sound_detection.src import models as tsd_models from target_sound_detection.src.models import event_labels self.device = device self.MEL_ARGS = { 'n_mels': 64, 'n_fft': 2048, 'hop_length': int(22050 * 20 / 1000), 'win_length': int(22050 * 40 / 1000) } self.EPS = np.spacing(1) self.clip_model, _ = clip.load("ViT-B/32", device=self.device) self.event_labels = event_labels self.id_to_event = {i : label for i, label in enumerate(self.event_labels)} config = torch.load('audio_detection/target_sound_detection/useful_ckpts/tsd/run_config.pth', map_location='cpu') config_parameters = dict(config) config_parameters['tao'] = 0.6 if 'thres' not in config_parameters.keys(): config_parameters['thres'] = 0.5 if 'time_resolution' not in config_parameters.keys(): config_parameters['time_resolution'] = 125 model_parameters = torch.load('audio_detection/target_sound_detection/useful_ckpts/tsd/run_model_7_loss=-0.0724.pt' , map_location=lambda storage, loc: storage) # load parameter self.model = getattr(tsd_models, config_parameters['model'])(config_parameters, inputdim=64, outputdim=2, time_resolution=config_parameters['time_resolution'], **config_parameters['model_args']) self.model.load_state_dict(model_parameters) self.model = self.model.to(self.device).eval() self.re_embeds = torch.load('audio_detection/target_sound_detection/useful_ckpts/tsd/text_emb.pth') self.ref_mel = torch.load('audio_detection/target_sound_detection/useful_ckpts/tsd/ref_mel.pth') def extract_feature(self, fname): import soundfile as sf y, sr = sf.read(fname, dtype='float32') print('y ', y.shape) ti = y.shape[0]/sr if y.ndim > 1: y = y.mean(1) y = librosa.resample(y, sr, 22050) lms_feature = np.log(librosa.feature.melspectrogram(y, **self.MEL_ARGS) + self.EPS).T return lms_feature,ti def build_clip(self, text): text = clip.tokenize(text).to(self.device) # ["a diagram with dog", "a dog", "a cat"] text_features = self.clip_model.encode_text(text) return text_features def cal_similarity(self, target, retrievals): ans = [] #target =torch.from_numpy(target) for name in retrievals.keys(): tmp = retrievals[name] #tmp = torch.from_numpy(tmp) s = torch.cosine_similarity(target.squeeze(), tmp.squeeze(), dim=0) ans.append(s.item()) return ans.index(max(ans)) def inference(self, text, audio_path): from target_sound_detection.src.utils import median_filter, decode_with_timestamps target_emb = self.build_clip(text) # torch type idx = self.cal_similarity(target_emb, self.re_embeds) target_event = self.id_to_event[idx] embedding = self.ref_mel[target_event] embedding = torch.from_numpy(embedding) embedding = embedding.unsqueeze(0).to(self.device).float() #print('embedding ', embedding.shape) inputs,ti = self.extract_feature(audio_path) #print('ti ', ti) inputs = torch.from_numpy(inputs) inputs = inputs.unsqueeze(0).to(self.device).float() #print('inputs ', inputs.shape) decision, decision_up, logit = self.model(inputs, embedding) pred = decision_up.detach().cpu().numpy() pred = pred[:,:,0] frame_num = decision_up.shape[1] time_ratio = ti / frame_num filtered_pred = median_filter(pred, window_size=1, threshold=0.5) #print('filtered_pred ', filtered_pred) time_predictions = [] for index_k in range(filtered_pred.shape[0]): decoded_pred = [] decoded_pred_ = decode_with_timestamps(target_event, filtered_pred[index_k,:]) if len(decoded_pred_) == 0: # neg deal decoded_pred_.append((target_event, 0, 0)) decoded_pred.append(decoded_pred_) for num_batch in range(len(decoded_pred)): # when we test our model,the batch_size is 1 cur_pred = pred[num_batch] # Save each frame output, for later visualization label_prediction = decoded_pred[num_batch] # frame predict # print(label_prediction) for event_label, onset, offset in label_prediction: time_predictions.append({ 'onset': onset*time_ratio, 'offset': offset*time_ratio,}) ans = '' for i,item in enumerate(time_predictions): ans = ans + 'segment' + str(i+1) + ' start_time: ' + str(item['onset']) + ' end_time: ' + str(item['offset']) + '\t' #print(ans) return ans # class Speech_Enh_SS_SC: # """Speech Enhancement or Separation in single-channel # Example usage: # enh_model = Speech_Enh_SS("cuda") # enh_wav = enh_model.inference("./test_chime4_audio_M05_440C0213_PED_REAL.wav") # """ # def __init__(self, device="cuda", model_name="lichenda/chime4_fasnet_dprnn_tac"): # self.model_name = model_name # self.device = device # print("Initializing ESPnet Enh to %s" % device) # self._initialize_model() # def _initialize_model(self): # from espnet_model_zoo.downloader import ModelDownloader # from espnet2.bin.enh_inference import SeparateSpeech # d = ModelDownloader() # cfg = d.download_and_unpack(self.model_name) # self.separate_speech = SeparateSpeech( # train_config=cfg["train_config"], # model_file=cfg["model_file"], # # for segment-wise process on long speech # segment_size=2.4, # hop_size=0.8, # normalize_segment_scale=False, # show_progressbar=True, # ref_channel=None, # normalize_output_wav=True, # device=self.device, # ) # def inference(self, speech_path, ref_channel=0): # speech, sr = soundfile.read(speech_path) # speech = speech[:, ref_channel] # assert speech.dim() == 1 # enh_speech = self.separate_speech(speech[None, ], fs=sr) # if len(enh_speech) == 1: # return enh_speech[0] # return enh_speech # class Speech_Enh_SS_MC: # """Speech Enhancement or Separation in multi-channel""" # def __init__(self, device="cuda", model_name=None, ref_channel=4): # self.model_name = model_name # self.ref_channel = ref_channel # self.device = device # print("Initializing ESPnet Enh to %s" % device) # self._initialize_model() # def _initialize_model(self): # from espnet_model_zoo.downloader import ModelDownloader # from espnet2.bin.enh_inference import SeparateSpeech # d = ModelDownloader() # cfg = d.download_and_unpack(self.model_name) # self.separate_speech = SeparateSpeech( # train_config=cfg["train_config"], # model_file=cfg["model_file"], # # for segment-wise process on long speech # segment_size=2.4, # hop_size=0.8, # normalize_segment_scale=False, # show_progressbar=True, # ref_channel=self.ref_channel, # normalize_output_wav=True, # device=self.device, # ) # def inference(self, speech_path): # speech, sr = soundfile.read(speech_path) # speech = speech.T # enh_speech = self.separate_speech(speech[None, ...], fs=sr) # if len(enh_speech) == 1: # return enh_speech[0] # return enh_speech class Speech_Enh_SS_SC: """Speech Enhancement or Separation in single-channel Example usage: enh_model = Speech_Enh_SS("cuda") enh_wav = enh_model.inference("./test_chime4_audio_M05_440C0213_PED_REAL.wav") """ def __init__(self, device="cuda", model_name="espnet/Wangyou_Zhang_chime4_enh_train_enh_conv_tasnet_raw"): self.model_name = model_name self.device = device print("Initializing ESPnet Enh to %s" % device) self._initialize_model() def _initialize_model(self): from espnet_model_zoo.downloader import ModelDownloader from espnet2.bin.enh_inference import SeparateSpeech d = ModelDownloader() cfg = d.download_and_unpack(self.model_name) self.separate_speech = SeparateSpeech( train_config=cfg["train_config"], model_file=cfg["model_file"], # for segment-wise process on long speech segment_size=2.4, hop_size=0.8, normalize_segment_scale=False, show_progressbar=True, ref_channel=None, normalize_output_wav=True, device=self.device, ) def inference(self, speech_path, ref_channel=0): speech, sr = soundfile.read(speech_path) speech = speech[:, ref_channel] # speech = torch.from_numpy(speech) # assert speech.dim() == 1 enh_speech = self.separate_speech(speech[None, ...], fs=sr) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") # if len(enh_speech) == 1: soundfile.write(audio_filename, enh_speech[0].squeeze(), samplerate=sr) # return enh_speech[0] # return enh_speech # else: # print("############") # audio_filename_1 = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") # soundfile.write(audio_filename_1, enh_speech[0].squeeze(), samplerate=sr) # audio_filename_2 = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") # soundfile.write(audio_filename_2, enh_speech[1].squeeze(), samplerate=sr) # audio_filename = merge_audio(audio_filename_1, audio_filename_2) return audio_filename class Speech_SS: def __init__(self, device="cuda", model_name="lichenda/wsj0_2mix_skim_noncausal"): self.model_name = model_name self.device = device print("Initializing ESPnet SS to %s" % device) self._initialize_model() def _initialize_model(self): from espnet_model_zoo.downloader import ModelDownloader from espnet2.bin.enh_inference import SeparateSpeech d = ModelDownloader() cfg = d.download_and_unpack(self.model_name) self.separate_speech = SeparateSpeech( train_config=cfg["train_config"], model_file=cfg["model_file"], # for segment-wise process on long speech segment_size=2.4, hop_size=0.8, normalize_segment_scale=False, show_progressbar=True, ref_channel=None, normalize_output_wav=True, device=self.device, ) def inference(self, speech_path): speech, sr = soundfile.read(speech_path) enh_speech = self.separate_speech(speech[None, ...], fs=sr) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") if len(enh_speech) == 1: soundfile.write(audio_filename, enh_speech[0], samplerate=sr) else: # print("############") audio_filename_1 = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename_1, enh_speech[0].squeeze(), samplerate=sr) audio_filename_2 = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") soundfile.write(audio_filename_2, enh_speech[1].squeeze(), samplerate=sr) audio_filename = merge_audio(audio_filename_1, audio_filename_2) return audio_filename class ConversationBot: def __init__(self): print("Initializing AudioGPT") self.llm = OpenAI(temperature=0) self.t2i = T2I(device="cuda:1") self.i2t = ImageCaptioning(device="cuda:0") self.t2a = T2A(device="cuda:0") self.tts = TTS(device="cpu") self.t2s = T2S(device="cpu") self.i2a = I2A(device="cuda:0") self.a2t = A2T(device="cpu") self.asr = ASR(device="cuda:0") self.SE_SS_SC = Speech_Enh_SS_SC(device="cuda:0") # self.SE_SS_MC = Speech_Enh_SS_MC(device="cuda:0") self.SS = Speech_SS(device="cuda:0") self.inpaint = Inpaint(device="cuda:0") self.tts_ood = TTS_OOD(device="cpu") self.geneface = GeneFace(device="cuda:0") self.detection = SoundDetection(device="cpu") self.binaural = Binaural(device="cuda:0") self.extraction = SoundExtraction(device="cuda:0") self.TSD = TargetSoundDetection(device="cuda:0") self.memory = ConversationBufferMemory(memory_key="chat_history", output_key='output') def init_tools(self, interaction_type): if interaction_type == 'text': self.tools = [ Tool(name="Generate Image From User Input Text", func=self.t2i.inference, description="useful for when you want to generate an image from a user input text and it saved it to a file. like: generate an image of an object or something, or generate an image that includes some objects. " "The input to this tool should be a string, representing the text used to generate image. "), Tool(name="Get Photo Description", func=self.i2t.inference, description="useful for when you want to know what is inside the photo. receives image_path as input. " "The input to this tool should be a string, representing the image_path. "), Tool(name="Generate Audio From User Input Text", func=self.t2a.inference, description="useful for when you want to generate an audio from a user input text and it saved it to a file." "The input to this tool should be a string, representing the text used to generate audio."), Tool( name="Style Transfer", func= self.tts_ood.inference, description="useful for when you want to generate speech samples with styles (e.g., timbre, emotion, and prosody) derived from a reference custom voice." "Like: Generate a speech with style transferred from this voice. The text is xxx., or speak using the voice of this audio. The text is xxx." "The input to this tool should be a comma seperated string of two, representing reference audio path and input text."), Tool(name="Generate Singing Voice From User Input Text, Note and Duration Sequence", func= self.t2s.inference, description="useful for when you want to generate a piece of singing voice (Optional: from User Input Text, Note and Duration Sequence) and save it to a file." "If Like: Generate a piece of singing voice, the input to this tool should be \"\" since there is no User Input Text, Note and Duration Sequence ." "If Like: Generate a piece of singing voice. Text: xxx, Note: xxx, Duration: xxx. " "Or Like: Generate a piece of singing voice. Text is xxx, note is xxx, duration is xxx." "The input to this tool should be a comma seperated string of three, representing text, note and duration sequence since User Input Text, Note and Duration Sequence are all provided."), Tool(name="Synthesize Speech Given the User Input Text", func=self.tts.inference, description="useful for when you want to convert a user input text into speech audio it saved it to a file." "The input to this tool should be a string, representing the text used to be converted to speech."), # Tool(name="Speech Enhancement Or Separation In Single-Channel", func=self.SE_SS_SC.inference, # description="useful for when you want to enhance the quality of the speech signal by reducing background noise (single-channel), " # "or separate each speech from the speech mixture (single-channel), receives audio_path as input." # "The input to this tool should be a string, representing the audio_path."), Tool(name="Speech Enhancement In Single-Channel", func=self.SE_SS_SC.inference, description="useful for when you want to enhance the quality of the speech signal by reducing background noise (single-channel), receives audio_path as input." "The input to this tool should be a string, representing the audio_path."), Tool(name="Speech Separation In Single-Channel", func=self.SS.inference, description="useful for when you want to separate each speech from the speech mixture, receives audio_path as input." "The input to this tool should be a string, representing the audio_path."), # Tool(name="Speech Enhancement In Multi-Channel", func=self.SE_SS_MC.inference, # description="useful for when you want to enhance the quality of the speech signal by reducing background noise (multi-channel), receives audio_path as input." # "The input to this tool should be a string, representing the audio_path."), Tool(name="Generate Audio From The Image", func=self.i2a.inference, description="useful for when you want to generate an audio based on an image." "The input to this tool should be a string, representing the image_path. "), Tool(name="Generate Text From The Audio", func=self.a2t.inference, description="useful for when you want to describe an audio in text, receives audio_path as input." "The input to this tool should be a string, representing the audio_path."), Tool(name="Audio Inpainting", func=self.inpaint.show_mel_fn, description="useful for when you want to inpaint a mel spectrum of an audio and predict this audio, this tool will generate a mel spectrum and you can inpaint it, receives audio_path as input, " "The input to this tool should be a string, representing the audio_path."), Tool(name="Transcribe Speech", func=self.asr.inference, description="useful for when you want to know the text corresponding to a human speech, receives audio_path as input." "The input to this tool should be a string, representing the audio_path."), Tool(name="Generate a talking human portrait video given a input Audio", func=self.geneface.inference, description="useful for when you want to generate a talking human portrait video given a input audio." "The input to this tool should be a string, representing the audio_path."), Tool(name="Detect The Sound Event From The Audio", func=self.detection.inference, description="useful for when you want to know what event in the audio and the sound event start or end time, this tool will generate an image of all predict events, receives audio_path as input. " "The input to this tool should be a string, representing the audio_path. "), Tool(name="Sythesize Binaural Audio From A Mono Audio Input", func=self.binaural.inference, description="useful for when you want to transfer your mono audio into binaural audio, receives audio_path as input. " "The input to this tool should be a string, representing the audio_path. "), Tool(name="Extract Sound Event From Mixture Audio Based On Language Description", func=self.extraction.inference, description="useful for when you extract target sound from a mixture audio, you can describe the target sound by text, receives audio_path and text as input. " "The input to this tool should be a comma seperated string of two, representing mixture audio path and input text."), Tool(name="Target Sound Detection", func=self.TSD.inference, description="useful for when you want to know when the target sound event in the audio happens. You can use language descriptions to instruct the model. receives text description and audio_path as input. " "The input to this tool should be a comma seperated string of two, representing audio path and the text description. ")] self.agent = initialize_agent( self.tools, self.llm, agent="conversational-react-description", verbose=True, memory=self.memory, return_intermediate_steps=True, agent_kwargs={'prefix': AUDIO_CHATGPT_PREFIX, 'format_instructions': AUDIO_CHATGPT_FORMAT_INSTRUCTIONS, 'suffix': AUDIO_CHATGPT_SUFFIX}, ) return gr.update(visible=True), gr.update(visible=False), gr.update(visible=True), gr.update(visible=False) else: self.tools = [ Tool(name="Generate Audio From User Input Text", func=self.t2a.inference, description="useful for when you want to generate an audio from a user input text and it saved it to a file." "The input to this tool should be a string, representing the text used to generate audio."), Tool( name="Style Transfer", func= self.tts_ood.inference, description="useful for when you want to generate speech samples with styles (e.g., timbre, emotion, and prosody) derived from a reference custom voice." "Like: Generate a speech with style transferred from this voice. The text is xxx., or speak using the voice of this audio. The text is xxx." "The input to this tool should be a comma seperated string of two, representing reference audio path and input text."), Tool(name="Generate Singing Voice From User Input Text, Note and Duration Sequence", func= self.t2s.inference, description="useful for when you want to generate a piece of singing voice (Optional: from User Input Text, Note and Duration Sequence) and save it to a file." "If Like: Generate a piece of singing voice, the input to this tool should be \"\" since there is no User Input Text, Note and Duration Sequence ." "If Like: Generate a piece of singing voice. Text: xxx, Note: xxx, Duration: xxx. " "Or Like: Generate a piece of singing voice. Text is xxx, note is xxx, duration is xxx." "The input to this tool should be a comma seperated string of three, representing text, note and duration sequence since User Input Text, Note and Duration Sequence are all provided."), Tool(name="Synthesize Speech Given the User Input Text", func=self.tts.inference, description="useful for when you want to convert a user input text into speech audio it saved it to a file." "The input to this tool should be a string, representing the text used to be converted to speech."), Tool(name="Generate Text From The Audio", func=self.a2t.inference, description="useful for when you want to describe an audio in text, receives audio_path as input." "The input to this tool should be a string, representing the audio_path."), Tool(name="Generate a talking human portrait video given a input Audio", func=self.geneface.inference, description="useful for when you want to generate a talking human portrait video given a input audio." "The input to this tool should be a string, representing the audio_path."), Tool(name="Generate Binaural Audio From A Mono Audio Input", func=self.binaural.inference, description="useful for when you want to transfer your mono audio into binaural audio, receives audio_path as input. " "The input to this tool should be a string, representing the audio_path. "), Tool(name="Extract Sound Event From Mixture Audio Based On Language Description", func=self.extraction.inference, description="useful for when you extract target sound from a mixture audio, you can describe the target sound by text, receives audio_path and text as input. " "The input to this tool should be a comma seperated string of two, representing mixture audio path and input text."), Tool(name="Target Sound Detection", func=self.TSD.inference, description="useful for when you want to know when the target sound event in the audio happens. You can use language descriptions to instruct the model. receives text description and audio_path as input. " "The input to this tool should be a comma seperated string of two, representing audio path and the text description. ")] self.agent = initialize_agent( self.tools, self.llm, agent="conversational-react-description", verbose=True, memory=self.memory, return_intermediate_steps=True, agent_kwargs={'prefix': AUDIO_CHATGPT_PREFIX, 'format_instructions': AUDIO_CHATGPT_FORMAT_INSTRUCTIONS, 'suffix': AUDIO_CHATGPT_SUFFIX}, ) return gr.update(visible=False), gr.update(visible=False), gr.update(visible=False), gr.update(visible=True) def run_text(self, text, state): print("===============Running run_text =============") print("Inputs:", text, state) print("======>Previous memory:\n %s" % self.agent.memory) self.agent.memory.buffer = cut_dialogue_history(self.agent.memory.buffer, keep_last_n_words=500) res = self.agent({"input": text}) if res['intermediate_steps'] == []: print("======>Current memory:\n %s" % self.agent.memory) response = res['output'] state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(visible=False), gr.Video.update(visible=False), gr.Image.update(visible=False), gr.Button.update(visible=False) else: tool = res['intermediate_steps'][0][0].tool if tool == "Generate Image From User Input Text" or tool == "Generate Text From The Audio" or tool == "Target Sound Detection": print("======>Current memory:\n %s" % self.agent.memory) response = re.sub('(image/\S*png)', lambda m: f'![](/file={m.group(0)})*{m.group(0)}*', res['output']) state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(visible=False), gr.Video.update(visible=False), gr.Image.update(visible=False), gr.Button.update(visible=False) elif tool == "Transcribe Speech": response = res['output'] state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(visible=False), gr.Video.update(visible=False), gr.Image.update(visible=False), gr.Button.update(visible=False) elif tool == "Detect The Sound Event From The Audio": image_filename = res['intermediate_steps'][0][1] response = res['output'] + f"![](/file={image_filename})*{image_filename}*" state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(visible=False), gr.Video.update(visible=False), gr.Image.update(visible=False), gr.Button.update(visible=False) elif tool == "Audio Inpainting": audio_filename = res['intermediate_steps'][0][0].tool_input image_filename = res['intermediate_steps'][0][1] print("======>Current memory:\n %s" % self.agent.memory) response = res['output'] state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(value=audio_filename,visible=True), gr.Video.update(visible=False), gr.Image.update(value=image_filename,visible=True), gr.Button.update(visible=True) elif tool == "Generate a talking human portrait video given a input Audio": video_filename = res['intermediate_steps'][0][1] print("======>Current memory:\n %s" % self.agent.memory) response = res['output'] state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(visible=False), gr.Video.update(value=video_filename,visible=True), gr.Image.update(visible=False), gr.Button.update(visible=False) print("======>Current memory:\n %s" % self.agent.memory) response = re.sub('(image/\S*png)', lambda m: f'![](/file={m.group(0)})*{m.group(0)}*', res['output']) audio_filename = res['intermediate_steps'][0][1] state = state + [(text, response)] print("Outputs:", state) return state, state, gr.Audio.update(value=audio_filename,visible=True), gr.Video.update(visible=False), gr.Image.update(visible=False), gr.Button.update(visible=False) def run_image_or_audio(self, file, state, txt): file_type = file.name[-3:] if file_type == "wav": print("===============Running run_audio =============") print("Inputs:", file, state) print("======>Previous memory:\n %s" % self.agent.memory) audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") # audio_load = whisper.load_audio(file.name) audio_load, sr = soundfile.read(file.name) soundfile.write(audio_filename, audio_load, samplerate = sr) description = self.a2t.inference(audio_filename) Human_prompt = "\nHuman: provide an audio named {}. The description is: {}. This information helps you to understand this audio, but you should use tools to finish following tasks, " \ "rather than directly imagine from my description. If you understand, say \"Received\". \n".format(audio_filename, description) AI_prompt = "Received. " self.agent.memory.buffer = self.agent.memory.buffer + Human_prompt + 'AI: ' + AI_prompt print("======>Current memory:\n %s" % self.agent.memory) #state = state + [(f"<audio src=audio_filename controls=controls></audio>*{audio_filename}*", AI_prompt)] state = state + [(f"*{audio_filename}*", AI_prompt)] print("Outputs:", state) return state, state, gr.Audio.update(value=audio_filename,visible=True), gr.Video.update(visible=False) else: print("===============Running run_image =============") print("Inputs:", file, state) print("======>Previous memory:\n %s" % self.agent.memory) image_filename = os.path.join('image', str(uuid.uuid4())[0:8] + ".png") print("======>Auto Resize Image...") img = Image.open(file.name) width, height = img.size ratio = min(512 / width, 512 / height) width_new, height_new = (round(width * ratio), round(height * ratio)) img = img.resize((width_new, height_new)) img = img.convert('RGB') img.save(image_filename, "PNG") print(f"Resize image form {width}x{height} to {width_new}x{height_new}") description = self.i2t.inference(image_filename) Human_prompt = "\nHuman: provide a figure named {}. The description is: {}. This information helps you to understand this image, but you should use tools to finish following tasks, " \ "rather than directly imagine from my description. If you understand, say \"Received\". \n".format(image_filename, description) AI_prompt = "Received. " self.agent.memory.buffer = self.agent.memory.buffer + Human_prompt + 'AI: ' + AI_prompt print("======>Current memory:\n %s" % self.agent.memory) state = state + [(f"![](/file={image_filename})*{image_filename}*", AI_prompt)] print("Outputs:", state) return state, state, gr.Audio.update(visible=False), gr.Video.update(visible=False) def speech(self, speech_input, state): input_audio_filename = os.path.join('audio', str(uuid.uuid4())[0:8] + ".wav") text = self.asr.translate_english(speech_input) print("Inputs:", text, state) print("======>Previous memory:\n %s" % self.agent.memory) self.agent.memory.buffer = cut_dialogue_history(self.agent.memory.buffer, keep_last_n_words=500) res = self.agent({"input": text}) if res['intermediate_steps'] == []: print("======>Current memory:\n %s" % self.agent.memory) response = res['output'] output_audio_filename = self.tts.inference(response) state = state + [(text, response)] print("Outputs:", state) return gr.Audio.update(value=None), gr.Audio.update(value=output_audio_filename,visible=True), state, gr.Video.update(visible=False) else: tool = res['intermediate_steps'][0][0].tool if tool == "Generate Image From User Input Text" or tool == "Generate Text From The Audio" or tool == "Target Sound Detection": print("======>Current memory:\n %s" % self.agent.memory) response = re.sub('(image/\S*png)', lambda m: f'![](/file={m.group(0)})*{m.group(0)}*', res['output']) output_audio_filename = self.tts.inference(res['output']) state = state + [(text, response)] print("Outputs:", state) return gr.Audio.update(value=None), gr.Audio.update(value=output_audio_filename,visible=True), state, gr.Video.update(visible=False) elif tool == "Transcribe Speech": print("======>Current memory:\n %s" % self.agent.memory) output_audio_filename = self.tts.inference(res['output']) response = res['output'] state = state + [(text, response)] print("Outputs:", state) return gr.Audio.update(value=None), gr.Audio.update(value=output_audio_filename,visible=True), state, gr.Video.update(visible=False) elif tool == "Detect The Sound Event From The Audio": print("======>Current memory:\n %s" % self.agent.memory) image_filename = res['intermediate_steps'][0][1] output_audio_filename = self.tts.inference(res['output']) response = res['output'] + f"![](/file={image_filename})*{image_filename}*" state = state + [(text, response)] print("Outputs:", state) return gr.Audio.update(value=None), gr.Audio.update(value=output_audio_filename,visible=True), state, gr.Video.update(visible=False) elif tool == "Generate a talking human portrait video given a input Audio": video_filename = res['intermediate_steps'][0][1] print("======>Current memory:\n %s" % self.agent.memory) response = res['output'] output_audio_filename = self.tts.inference(res['output']) state = state + [(text, response)] print("Outputs:", state) return gr.Audio.update(value=None), gr.Audio.update(value=output_audio_filename,visible=True), state, gr.Video.update(value=video_filename,visible=True) print("======>Current memory:\n %s" % self.agent.memory) response = re.sub('(image/\S*png)', lambda m: f'![](/file={m.group(0)})*{m.group(0)}*', res['output']) audio_filename = res['intermediate_steps'][0][1] Res = "The audio file has been generated and the audio is " output_audio_filename = merge_audio(self.tts.inference(Res), audio_filename) print(output_audio_filename) state = state + [(text, response)] response = res['output'] print("Outputs:", state) return gr.Audio.update(value=None), gr.Audio.update(value=output_audio_filename,visible=True), state, gr.Video.update(visible=False) def inpainting(self, state, audio_filename, image_filename): print("===============Running inpainting =============") print("Inputs:", state) print("======>Previous memory:\n %s" % self.agent.memory) new_image_filename, new_audio_filename = self.inpaint.inference(audio_filename, image_filename) AI_prompt = "Here are the predict audio and the mel spectrum." + f"*{new_audio_filename}*" + f"![](/file={new_image_filename})*{new_image_filename}*" output_audio_filename = self.tts.inference(AI_prompt) self.agent.memory.buffer = self.agent.memory.buffer + 'AI: ' + AI_prompt print("======>Current memory:\n %s" % self.agent.memory) state = state + [(f"Audio Inpainting", AI_prompt)] print("Outputs:", state) return state, state, gr.Image.update(visible=False), gr.Audio.update(value=new_audio_filename, visible=True), gr.Video.update(visible=False), gr.Button.update(visible=False) def clear_audio(self): return gr.Audio.update(value=None, visible=False) def clear_input_audio(self): return gr.Audio.update(value=None) def clear_image(self): return gr.Image.update(value=None, visible=False) def clear_video(self): return gr.Video.update(value=None, visible=False) def clear_button(self): return gr.Button.update(visible=False) if __name__ == '__main__': bot = ConversationBot() with gr.Blocks(css="#chatbot .overflow-y-auto{height:500px}") as demo: with gr.Row(): gr.Markdown("## AudioGPT") chatbot = gr.Chatbot(elem_id="chatbot", label="AudioGPT", visible=False) state = gr.State([]) with gr.Row() as select_raws: with gr.Column(scale=0.7): interaction_type = gr.Radio(choices=['text', 'speech'], value='text', label='Interaction Type') with gr.Column(scale=0.3, min_width=0): select = gr.Button("Select") with gr.Row(visible=False) as text_input_raws: with gr.Column(scale=0.7): txt = gr.Textbox(show_label=False, placeholder="Enter text and press enter, or upload an image").style(container=False) with gr.Column(scale=0.1, min_width=0): run = gr.Button("🏃‍♂️Run") with gr.Column(scale=0.1, min_width=0): clear_txt = gr.Button("🔄Clear️") with gr.Column(scale=0.1, min_width=0): btn = gr.UploadButton("🖼️Upload", file_types=["image","audio"]) with gr.Row(): outaudio = gr.Audio(visible=False) with gr.Row(): with gr.Column(scale=0.3, min_width=0): outvideo = gr.Video(visible=False) with gr.Row(): show_mel = gr.Image(type="filepath",tool='sketch',visible=False) with gr.Row(): run_button = gr.Button("Predict Masked Place",visible=False) with gr.Row(visible=False) as speech_input_raws: with gr.Column(scale=0.7): speech_input = gr.Audio(source="microphone", type="filepath", label="Input") with gr.Column(scale=0.15, min_width=0): submit_btn = gr.Button("🏃‍♂️Submit") with gr.Column(scale=0.15, min_width=0): clear_speech = gr.Button("🔄Clear️") with gr.Row(): speech_output = gr.Audio(label="Output",visible=False) select.click(bot.init_tools, [interaction_type], [chatbot, select_raws, text_input_raws, speech_input_raws]) txt.submit(bot.run_text, [txt, state], [chatbot, state, outaudio, outvideo, show_mel, run_button]) txt.submit(lambda: "", None, txt) run.click(bot.run_text, [txt, state], [chatbot, state, outaudio, outvideo, show_mel, run_button]) run.click(lambda: "", None, txt) btn.upload(bot.run_image_or_audio, [btn, state, txt], [chatbot, state, outaudio, outvideo]) run_button.click(bot.inpainting, [state, outaudio, show_mel], [chatbot, state, show_mel, outaudio, outvideo, run_button]) clear_txt.click(bot.memory.clear) clear_txt.click(lambda: [], None, chatbot) clear_txt.click(lambda: [], None, state) clear_txt.click(lambda:None, None, txt) clear_txt.click(bot.clear_button, None, run_button) clear_txt.click(bot.clear_image, None, show_mel) clear_txt.click(bot.clear_audio, None, outaudio) clear_txt.click(bot.clear_video, None, outvideo) submit_btn.click(bot.speech, [speech_input, state], [speech_input, speech_output, state, outvideo]) clear_speech.click(bot.clear_input_audio, None, speech_input) clear_speech.click(bot.clear_audio, None, speech_output) clear_speech.click(lambda: [], None, state) clear_speech.click(bot.clear_video, None, outvideo) demo.launch(server_name="0.0.0.0", server_port=7860, share=True)
[ "\nHuman: provide a figure named PLACEHOLDER. The description is: PLACEHOLDER. This information helps you to understand this image, but you should use tools to finish following tasks, rather than directly imagine from my description. If you understand, say \"Received\". \n", "Received. ", "Here are the predict audio and the mel spectrum.*PLACEHOLDER*![](/file=PLACEHOLDER)*PLACEHOLDER*", "\nHuman: provide an audio named PLACEHOLDER. The description is: PLACEHOLDER. This information helps you to understand this audio, but you should use tools to finish following tasks, rather than directly imagine from my description. If you understand, say \"Received\". \n" ]
2024-01-10
ASSERT-KTH/Supersonic
src~chatgpt_optimizer~run_chatgpt_optimizeer.py
#!/usr/bin/env python # -*- encoding: utf-8 -*- ''' @Author : Sen Fang @Email : [email protected] @Ide : vscode & conda @File : run_chatgpt_optimizeer.py @Time : 2023/03/25 18:43:57 ''' """In this file, we use the chatgpt API generate the optimized code for each original c or cpp file from prompt file.""" import os import argparse import re import time import random import itertools import sys import json import string import subprocess import logging import tempfile import openai import tiktoken from tqdm import tqdm from pathlib import Path from typing import Optional, Tuple SLEEP_TIME = 60 def load_prompts(dir_path: str): """Load prompt files from submission-easy directory.""" paths = [] for root, dirs, files in os.walk(dir_path): for file in files: if file.startswith("original.c") or file.startswith("original.cpp"): path = os.path.join(root, file) paths.append(path) return paths def extract_code(response: str): """Extract the optimized code from the response.""" lines = response.split("\n") code = [] cnt = 0 for line in lines: if line.startswith("```"): if cnt == 0: cnt += 1 continue else: break code.append(line) return "\n".join(code) def write_response(path: str, response_dict: dict): """Write the response into response_{}.c or response_{}.cpp file.""" files = os.listdir(path.replace(path.split("/")[-1], "")) original_file = "original_{}.txt" if "original.c" in files: response_file = "response_{}.c" elif "original.cpp" in files: response_file = "response_{}.cpp" else: raise ValueError("No original file found.") for key, value in response_dict.items(): with open(path.replace(path.split("/")[-1], original_file.format(key)), "w") as f: f.write(value[0]) with open(path.replace(path.split("/")[-1], response_file.format(key)), "w") as f: f.write(value[1]) def generate_optimized_code(path: str, prompt: str, model_name: str="gpt-3.5-turbo", temperature: float=0.7, max_tokens: int=2048, num_requests: int=10): """Feed prompt files into ChatGPT and save the generated optimized code into submission-easy directory.""" with open(path, "r") as f: original_code = f.read() # build markdown code block original_code = "```\n" + original_code + "\n```" input_length = len(tiktoken.encoding_for_model(model_name).encode(original_code)) if input_length > max_tokens: with open(path.replace(path.split("/")[-1], "response.txt"), "a") as f: f.write("The length of prompt is too long.") return response_dict = {} for i in range(num_requests): response_dict[str(i)] = [] response = openai.ChatCompletion.create( model=model_name, messages=[{"role": "system", "content": prompt}, {"role": "user", "content": original_code}], temperature=temperature, ) # print(response) original_text = response["choices"][0]["message"]["content"] response_dict[str(i)].append(original_text) optimized_code = extract_code(response["choices"][0]["message"]["content"]) response_dict[str(i)].append(optimized_code) return response_dict if __name__ == "__main__": parser = argparse.ArgumentParser() parser.add_argument("--prompt_level", "-p", type=str, default="easy", help="The level of prompt to use.", choices=["easy", "specific", "advanced"]) parser.add_argument("--model_name", "-mn", type=str, default="gpt-3.5-turbo", help="The name of the model to use.") parser.add_argument("--only_generate", "-og", type=bool, default=True, help="Only generate the optimized code or not.") parser.add_argument("--only_test", "-ot", type=bool, default=False, help="Only test the generated optimized code or not.") parser.add_argument("--num_test", "-nt", type=int, default=None, help="Test the script with a limited number of prompts.") parser.add_argument("--num_requests", "-nr", type=int, default=10, help="The number of requests to send for each prompt.") parser.add_argument("--temperature", "-t", type=float, default=0.7, help="The temperature of the model.") # The default value of web chatgpt is 0.7 parser.add_argument("--max_tokens", "-mt", type=int, default=3000, help="The maximum number of tokens to generate.") args = parser.parse_args() # Set API key openai.api_key = os.getenv("OPENAI_API_KEY") # Load prompt files dir_name = "submissions" + "-" + args.prompt_level paths = load_prompts(dir_name) prompt_dict = {"easy": "Please optimize the following code and return the markdown format optimized code.\n", "advanced": "I want you to act as an experienced C and C++ developer and your task is to optimize my written C or C++ programs. I want you to optimize my program from the running time and memory usage. I will type my C or C++ program and you will optimize the program and return the optimized program. I want you to only reply with the fixed program inside one unique code block, and nothing else. Do not write explanations unless I instruct you to do so.\n"} prompt = prompt_dict[args.prompt_level] for idx, path in tqdm(enumerate(paths)): output_file = "response" + "_" + str(args.prompt_level) + "_" + str(args.temperature) + ".json" if args.only_generate: # print(idx)s if idx < 188: continue print(path) try: response_dict = generate_optimized_code(path, prompt=prompt, model_name=args.model_name, temperature=args.temperature, max_tokens=args.max_tokens, num_requests=args.num_requests) except Exception as e: print(e) print(idx) break if response_dict is not None: write_response(path, response_dict) time.sleep(SLEEP_TIME) elif args.only_test: pass # with open(path.replace("prompt.txt", output_file), "r") as f: # response_dict = json.load(f) # for key, value in response_dict.items(): # files = os.listdir(path.replace("prompt.txt", "")) else: pass if args.num_test is not None: if idx >= args.num_test: break
[ "{'easy': 'Please optimize the following code and return the markdown format optimized code.\\n', 'advanced': 'I want you to act as an experienced C and C++ developer and your task is to optimize my written C or C++ programs. I want you to optimize my program from the running time and memory usage. I will type my C or C++ program and you will optimize the program and return the optimized program. I want you to only reply with the fixed program inside one unique code block, and nothing else. Do not write explanations unless I instruct you to do so.\\n'}" ]
2024-01-10
will-clarke/story-gen
colab.py
# -*- coding: utf-8 -*- """Story Gen.ipynb Automatically generated by Colaboratory. Original file is located at https://colab.research.google.com/drive/1ssiz-blg7dVt6E7XZxja_SUnEfwJKhKQ """ from google.colab import drive drive.mount('/content/gdrive') !ls /content/gdrive/MyDrive/AI !pip install ctransformers ctransformers[cuda] sqlalchemy langchain # transformers accelerate import datetime import uuid import re import random from sqlalchemy.orm import ( relationship, Mapped, mapped_column, declarative_base, sessionmaker, ) from sqlalchemy.dialects.postgresql import UUID from sqlalchemy import ( desc, func, create_engine, Integer, String, DateTime, ForeignKey, ) from ctransformers import AutoModelForCausalLM from langchain.llms import CTransformers from langchain.prompts import PromptTemplate from langchain.chains import LLMChain import os # Specify the absolute file path for the new database db_file_path = os.path.abspath("/content/gdrive/AI/stories.db") from sqlalchemy import create_engine # Create the engine with the database URI engine = create_engine(f"sqlite:///{db_file_path}", echo=True) #!pip3 uninstall -y transformers #!pip3 install git+https://github.com/huggingface/transformers.git #!pip3 install auto-gptq --extra-index-url https://huggingface.github.io/autogptq-index/whl/cu118/ # Use cu117 if on CUDA 11.7 model_name = "TheBloke/Llama-2-13B-Chat-GGML" llm = AutoModelForCausalLM.from_pretrained(model_name, gpu_layers=50) #out=llm("AI is going to") #print(out) Base = declarative_base() from sqlalchemy import create_engine # Create the engine with the full database URI #engine = create_engine(f"sqlite:///content/gdrive/AI/stories.db", echo=True) class Story(Base): __tablename__ = "stories" id: Mapped[str] = mapped_column(String, primary_key=True, default=str(uuid.uuid4)) title: Mapped[str] = mapped_column(String) text: Mapped[str] = mapped_column(String) prompt: Mapped[str] = mapped_column(String) length: Mapped[int] = mapped_column(Integer) model_name: Mapped[str] = mapped_column(String) updated_at: Mapped[datetime.datetime] = mapped_column( DateTime, default=datetime.datetime.utcnow ) created_at: Mapped[datetime.datetime] = mapped_column( DateTime, default=datetime.datetime.utcnow ) categories = relationship( "StoryCategory", back_populates="story", cascade="all, delete-orphan" ) ratings = relationship( "StoryRating", back_populates="story", cascade="all, delete-orphan" ) def __repr__(self): title_long_enough: bool = len(self.title) >= 40 text_preview = self.title[:40] if title_long_enough else self.title text_preview.replace("\n", " ") text_preview = re.sub(r"\s+", " ", text_preview) return f"Story({text_preview})\n" def average_rating(self): return ( session.query(func.avg(StoryRating.rating)) .filter_by(story_id=self.id) .scalar() ) @staticmethod def get_random(): return Story.query.order_by(func.random()).first() @staticmethod def first(): return Story.query.order_by("updated_at").first() @staticmethod def last(): return Story.query.order_by(desc("updated_at")).first() @staticmethod def top_rated(limit=100): return ( session.query(Story) .join(StoryRating, StoryRating.story_id == Story.id) .group_by(Story.id) .order_by(func.avg(StoryRating.rating).desc()) .limit(limit) .all() ) class StoryCategory(Base): __tablename__ = "story_categories" story_id: Mapped[str] = mapped_column(String, ForeignKey("stories.id"), primary_key=True) # eg. tones category_type: Mapped[str] = mapped_column(String, primary_key=True) # eg. funny category: Mapped[str] = mapped_column(String, primary_key=True) story = relationship("Story", back_populates="categories") def __repr__(self): return f"({self.category_type}{self.category})\n" class StoryRating(Base): __tablename__ = "story_ratings" story_id: Mapped[str] = mapped_column(String, ForeignKey("stories.id"), primary_key=True) rating_type: Mapped[str] = mapped_column(String, primary_key=True) rating: Mapped[int] = mapped_column(Integer, primary_key=True) prompt: Mapped[str] = mapped_column(String) model_name: Mapped[str] = mapped_column(String) model_output: Mapped[str] = mapped_column(String, default="") story = relationship("Story", back_populates="ratings") updated_at: Mapped[datetime.datetime] = mapped_column( DateTime, default=datetime.datetime.utcnow ) def __repr__(self): return f"({self.rating_type}{self.rating})\n" @staticmethod def get_random(): return StoryRating.query.order_by(func.random()).first() @staticmethod def first(): return StoryRating.query.order_by("updated_at").first() @staticmethod def last(): return StoryRating.query.order_by(desc("updated_at")).first() engine = create_engine(f"sqlite:////content/gdrive/MyDrive/AI/stories.db", echo=True) Base.metadata.create_all(engine) potential_genres = [ "hisorical", "fantasy", "romantic", "suspense", "sci-fi", "noir", "adventure", "comedy", "mystery", "fantasy", "romance", ] potential_tones = [ "eerie", "suspense", "joyful", "celebration", "melancholic", "reflective", "funny", "comedy", "tense", ] prompt = PromptTemplate.from_template( """ Write an interesting and absorbing short story with the following genres: {genres}. It should have the following tones: {tones}. It should be roughly maximum {length} words long. Make it interesting! Start with the title on the first line then the story on the next line. """ ) def extract_title(text): lines = text.strip().split("\n") if len(lines) < 2: return "", text title = lines[0] story_text = "\n".join(lines[1:]) title = title.strip() story_text = story_text.strip() title = title.replace("Title: ", "") title = title.replace("Example: ", "") return title, story_text def process(genres, tones, length): f = prompt.format(genres=", ".join(genres), tones=", ".join(tones), length=length) out = llm(f) title, story_text = extract_title(out) if title == "": return categories = [StoryCategory(category_type="tones", category=t) for t in tones] + [ StoryCategory(category_type="genres", category=g) for g in genres ] s = Story( id=str(uuid.uuid4()), title=title, text=story_text, prompt=f, categories=categories, model_name=model_name, length=length, ) engine = create_engine(f"sqlite:////content/gdrive/MyDrive/AI/stories.db", echo=True) session_factory = sessionmaker(bind=engine) session = session_factory() session.add(s) session.commit() session.close() engine.dispose() if __name__ == "__main__": for i in range(0, 100000): genres = [ random.choice(potential_genres) for _ in range(0, random.randint(1, 2)) ] genres = list(set(genres)) tones = [random.choice(potential_tones) for _ in range(0, random.randint(1, 2))] tones = list(set(tones)) length = random.choice(["10", "50", "100", "200", "500", "1000", "2000"]) process(genres, tones, length)
[ "\nWrite an interesting and absorbing short story with the following genres: {genres}.\n\nIt should have the following tones: {tones}.\n\nIt should be roughly maximum {length} words long.\nMake it interesting!\n\nStart with the title on the first line then the story on the next line.\n" ]
2024-01-10
will-clarke/story-gen
stories_app~gen~rate_stories.py
from stories_app.db import db from stories_app.app import create_app import re from typing import Tuple from langchain.llms import CTransformers from langchain.prompts import PromptTemplate from langchain.chains import LLMChain from sqlalchemy.orm import Session from stories_app.models import Story, StoryCategory, StoryRating potential_genres = [ "hisorical", "fantasy", "romantic", "suspense", "sci-fi", "noir", "adventure", "comedy", "mystery", "fantasy", "romance", ] potential_tones = [ "eerie", "suspense", "joyful", "celebration", "melancholic", "reflective", "funny", "comedy", "tense", ] app = create_app() app.app_context().push() session = db.session # s = session.query(Story).limit(10).all() # print(s) # print("---------------------------") # s = session.query(StoryCategory).limit(10).all() # print(s) rating_criteria = [ "originality", "close to genres", "close to tones", "generally good", "interesting", ] prompt = PromptTemplate.from_template( """ The following is a short story demarked with <<<<<< and >>>>>>s. The short story is called "{title}", and aims to have the following categories: {categories}. Please rate how good you think it is on a scale of 1 to 100, where 1 is the worst and 100 is the best. Output format should be the number out of 100, followed by a colon and a single sentence reason for the rating. Examples: 34/100: The story is not coherent and the writing style is poor. 79/100: Interesting sci-fic concepts and some decent character development. Take into account the following criteria: - Originality - Coherence - Interest - How well it fits the categories ({categories}) - General writing style - Consistency <<<<<<< {story} >>>>>>> Rating: """ ) model_name = "TheBloke/Llama-2-13B-Chat-GGML" llm = CTransformers(model=model_name, model_type="llama", client=None) chain = LLMChain( llm=llm, prompt=prompt, ) def extract_rating(out: str) -> Tuple[int, str]: out = out.strip() # Use a regex to extract the rating integer from the string # The rating is in the format "34/100: The story is not coherent and the writing style is poor." # We want to extract the 34 from this string rating_match = re.search(r"(\d+)/100", out) rating = 0 if rating_match: rating = rating_match.group(1) out = re.sub(r"\d+/100[\.\s:]?", "", out) return int(rating), out def rate_story(story: Story): f = prompt.format( title=story.title, categories=", ".join([c.category for c in story.categories]), story=story.title, ) out = chain.run( title=story.title, categories=", ".join([c.category for c in story.categories]), story=story.title, ) rating, out = extract_rating(out) r = StoryRating( story_id=story.id, rating_type="overallv1", rating=rating, prompt=f, model_output=out, model_name=model_name, ) story.ratings.append(r) session.add(r) session.commit() if __name__ == "__main__": stories_with_no_ratings = session.query(Story).filter(~Story.ratings.any()).all() for s in stories_with_no_ratings: rate_story(s)
[ "\nThe following is a short story demarked with <<<<<< and >>>>>>s. \n\nThe short story is called \"{title}\", and aims to have the following categories: {categories}.\n\nPlease rate how good you think it is on a scale of 1 to 100, where 1 is the worst and 100 is the best.\nOutput format should be the number out of 100, followed by a colon and a single sentence reason for the rating.\n\nExamples:\n34/100: The story is not coherent and the writing style is poor.\n79/100: Interesting sci-fic concepts and some decent character development.\n\nTake into account the following criteria:\n- Originality\n- Coherence\n- Interest\n- How well it fits the categories ({categories})\n- General writing style\n- Consistency\n\n<<<<<<<\n{story}\n>>>>>>>\n\n\nRating:\n" ]
2024-01-10
TorRient/localGPT-falcon
run_localGPT.py
from langchain.chains import RetrievalQA # from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from langchain.vectorstores import Chroma from langchain.embeddings import HuggingFaceInstructEmbeddings from langchain.llms import HuggingFacePipeline from constants import CHROMA_SETTINGS, PERSIST_DIRECTORY from transformers import AutoModelForCausalLM, AutoTokenizer, pipeline import click import torch from constants import CHROMA_SETTINGS def load_model(device): """ Select a model on huggingface. If you are running this for the first time, it will download a model for you. subsequent runs will use the model from the disk. """ model = "tiiuae/falcon-7b-instruct" if device == "cuda": tokenizer = AutoTokenizer.from_pretrained(model) else: # cpu tokenizer=AutoTokenizer.from_pretrained(model) model=AutoModelForCausalLM.from_pretrained(model, trust_remote_code=True) pipe = pipeline( "text-generation", model=model, tokenizer=tokenizer, torch_dtype=torch.float32 if device =="cpu" else torch.bfloat16, trust_remote_code=True, device_map=device if device =="cpu" else "auto", max_length=2048, temperature=0, top_p=0.95, top_k=10, repetition_penalty=1.15, num_return_sequences=1, pad_token_id=tokenizer.eos_token_id ) local_llm = HuggingFacePipeline(pipeline=pipe) return local_llm # @click.command() # @click.option('--device_type', default='gpu', help='device to run on, select gpu or cpu') # def main(device_type, ): # # load the instructorEmbeddings # if device_type in ['cpu', 'CPU']: # device='cpu' # else: # device='cuda' ## for M1/M2 users: @click.command() @click.option('--device_type', default='cuda', help='device to run on, select gpu, cpu or mps') def main(device_type, ): # load the instructorEmbeddings if device_type in ['cpu', 'CPU']: device='cpu' elif device_type in ['mps', 'MPS']: device='mps' else: device='cuda' print(f"Running on: {device}") embeddings = HuggingFaceInstructEmbeddings(model_name="hkunlp/instructor-base", model_kwargs={"device": device}) # load the vectorstore db = Chroma(persist_directory=PERSIST_DIRECTORY, embedding_function=embeddings, client_settings=CHROMA_SETTINGS) retriever = db.as_retriever() # Prepare the LLM # callbacks = [StreamingStdOutCallbackHandler()] # load the LLM for generating Natural Language responses. llm = load_model(device) qa = RetrievalQA.from_chain_type(llm=llm, chain_type="stuff", retriever=retriever, return_source_documents=True) # Interactive questions and answers while True: query = input("\nEnter a query: ") if query == "exit": break # Get the answer from the chain res = qa(query) answer, docs = res['result'], res['source_documents'] # Print the result print("\n\n> Question:") print(query) print("\n> Answer:") print(answer) # Print the relevant sources used for the answer print("----------------------------------SOURCE DOCUMENTS---------------------------") for document in docs: print("\n> " + document.metadata["source"] + ":") print(document.page_content) print("----------------------------------SOURCE DOCUMENTS---------------------------") if __name__ == "__main__": main()
[]
2024-01-10
kashishkumar/Semantic_Search_Document_Management_System
qa_engine.py
from llama_index import VectorStoreIndex, SimpleDirectoryReader from llama_index import StorageContext, ServiceContext, load_index_from_storage from llama_index import Document from llama_index.node_parser import SimpleNodeParser import os import openai openai.api_key = "OPENAI_API_KEY" def load_document(): uploaded_file = None # enter code return uploaded_file def parse_document(uploaded_file): # Create test list using uploaded file content # What is the data format of uploaded_file? text_list = [] documents = [Document(text=text) for text in text_list] parser = SimpleNodeParser().from_defaults() nodes = parser.get_nodes_from_documents(documents) return documents, nodes def index_data(documents, nodes = None, chunk_size=None): # Persist the index to disk if nodes is not None: if chunk_size is not None: service_context = ServiceContext.from_defaults(chunk_size=chunk_size) index = VectorStoreIndex(nodes, service_context=service_context, show_progress=True) else: index = VectorStoreIndex(nodes, show_progress=True) else: if chunk_size is not None: service_context = ServiceContext.from_defaults(chunk_size=chunk_size) index = VectorStoreIndex.from_documents(documents, service_context=service_context, show_progress=True) else: index = VectorStoreIndex.from_documents(documents, show_progress=True) return index def save_query_engine(index, storage_context, save_index = True): index.save(storage_context) return None def build_query_engine(index, streaming = False): query_engine = index.as_query_engine(streaming=streaming) return query_engine def output_response(query_engine, query): response = query_engine.query(query) return response def show_response(response): print(response) def qa_engine(uploaded_file, query): documents, nodes = parse_document(uploaded_file) index = index_data(documents, nodes) storage_context = StorageContext.from_defaults() save_query_engine(index, storage_context, save_index = False) query_engine = build_query_engine(index) response = output_response(query_engine, query) show_response(response) def main(): uploaded_file = load_document() query = "What is the date for importer agreement?" qa_engine(uploaded_file, query)
[]
2024-01-10
kashishkumar/Semantic_Search_Document_Management_System
Invoice_Document_AI~qa_engine.py
from llama_index import VectorStoreIndex, SimpleDirectoryReader, ServiceContext, StorageContext from llama_index.llms import OpenAI import openai import torch #openai.api_key = "" def load_documents(documents_directory): documents = SimpleDirectoryReader(documents_directory).load_data() return documents def build_index(documents, llm_model=None, temperature = 0.1, chunk_size = 1024): if llm_model is not None: llm = OpenAI(temperature, model = llm_model) service_context = ServiceContext.from_defaults(llm=llm, chunk_size=chunk_size) index = VectorStoreIndex.from_documents(documents, service_context) else: index = VectorStoreIndex.from_documents(documents) return index def build_query_engine(documents_directory, llm_model=None): print("Building query engine from your documents") documents = load_documents(documents_directory) index = build_index(documents, llm_model=llm_model) query_engine = index.as_query_engine(similarity_top_k = 2) print("Query engine built") return query_engine def main(): documents_directory = input('Enter documents directory name: ') openai.api_key = input('Enter OpenAI API key: ') query_engine = build_query_engine(documents_directory) while True: query = input('Enter query (or type "exit" to quit): ') if query.lower() == 'exit': break print(query_engine.query(query)) if __name__ == '__main__': main()
[]
2024-01-10
kashishkumar/Semantic_Search_Document_Management_System
high_level_openai.py
import os import openai from langchain.embeddings.huggingface import HuggingFaceEmbeddings from llama_index import LangchainEmbedding, ServiceContext from llama_index import VectorStoreIndex, SimpleDirectoryReader openai.api_key = "OPENAI_API_KEY" documents = SimpleDirectoryReader('data').load_data() index = VectorStoreIndex.from_documents(documents) query_engine = index.as_query_engine() response = query_engine.query('What is this text about?') print(response) response = query_engine.query('Who is the author?') print(response) response = query_engine.query('What is the title?') print(response) embed_model = LangchainEmbedding(HuggingFaceEmbeddings(model_name='sentence-transformers/all-MiniLM-L6-v2')) service_context = ServiceContext(embed_model, query_engine)
[]
2024-01-10
kashishkumar/Semantic_Search_Document_Management_System
backend_functions.py
from llama_index import VectorStoreIndex, SimpleDirectoryReader from llama_index import StorageContext, ServiceContext, load_index_from_storage from llama_index import Document from llama_index.node_parser import SimpleNodeParser import os import openai openai.api_key = "OPENAI_API_KEY" def load_data(uploaded_file): # Create test list using uploaded file content text_list = [] documents = [Document(text=text) for text in text_list] parser = SimpleNodeParser().from_defaults() nodes = parser.get_nodes_from_documents(documents) return documents, nodes def index_data(documents, nodes = None, chunk_size=None): if nodes is not None: if chunk_size is not None: service_context = ServiceContext.from_defaults(chunk_size=chunk_size) index = VectorStoreIndex(nodes, service_context=service_context, show_progress=True) else: index = VectorStoreIndex(nodes, show_progress=True) else: if chunk_size is not None: service_context = ServiceContext.from_defaults(chunk_size=chunk_size) index = VectorStoreIndex.from_documents(documents, service_context=service_context, show_progress=True) else: index = VectorStoreIndex.from_documents(documents, show_progress=True) return index def build_query_engine(index, streaming = False): query_engine = index.as_query_engine(streaming=streaming) return query_engine def output_response(query_engine, query): response = query_engine.query(query) return response def show_response(response): print(response) def all(uploaded_file, query): documents, nodes = load_data(uploaded_file) index = index_data(documents, nodes) query_engine = build_query_engine(index) response = output_response(query_engine, query) show_response(response) def summarize(): return None
[]
2024-01-10
euanong/image-hijacks
image_hijacks~attacks~context.py
from abc import ABC, abstractmethod from collections import defaultdict import json from pathlib import Path from typing import Dict, List, Tuple, TYPE_CHECKING from image_hijacks.attacks import AttackDriver from image_hijacks.data import WrapperContextDataModule from image_hijacks.utils import Option, Parameters, quantise_image, tensor_to_image import wandb import torch import torch.nn.functional as F from torch import Tensor from jaxtyping import Float import openai import os from Levenshtein import distance class ContextAttack(AttackDriver, ABC): init_image: Tensor def __init__(self, config): super().__init__(config) self.register_buffer( "init_image", self.config.lift_to_model_device_dtype(self.config.init_image(config)), ) @abstractmethod def get_attack_target(self, context: str) -> str: ... @abstractmethod def is_attack_successful( self, context: List[str], true_output: List[str], pred_output: List[str], log_prefix: str, ) -> List[bool]: ... def on_train_start(self) -> None: self.step = 0 def training_step(self, batch, batch_idx): def get_loss( parameters: Parameters, ) -> Tuple[Float[Tensor, ""], Dict[str, Float[Tensor, ""]]]: img = torch.func.functional_call( self.processor, parameters, (self.init_image,) ) total_loss = torch.tensor(0.0, device=img.device) per_model_losses = {} for k, model in self.train_models.items(): img_sml = self.config.downsample(img, model) input_ids, input_attn_mask, output_ids, output_attn_mask = batch[k] loss = self.loss_fn.get_loss( model, img_sml, input_ids, input_attn_mask, target_ids=output_ids, target_attn_masks=output_attn_mask, ) total_loss += loss per_model_losses[k] = loss return (total_loss, per_model_losses) grads, (loss, per_model_losses) = self.gradient_estimator.grad_and_value( get_loss, self.processor.get_parameter_dict(), ) self.attack_optimizer.step(self.processor, grads) self.log("loss", loss, prog_bar=True) for k, v in grads.items(): self.log(f"grad_norm_{k}", torch.norm(v)) for k, v in per_model_losses.items(): self.log(f"loss_{k}", v, prog_bar=True) self.step += 1 def log_loss_and_accuracy(self, batch, prefix, dataset): total_loss = torch.tensor(0.0, device=self.init_image.device) img = quantise_image(self.processor(self.init_image)) for k, model in self.eval_models.items(): img_sml = self.config.downsample(img, model) input_ids, input_attn_mask, output_ids, output_attn_mask = batch[k] input_len = int((input_ids != 0).sum(dim=-1).max()) output_len = int((output_ids != 0).sum(dim=-1).max()) input_ids = input_ids[:, -input_len:] output_ids = output_ids[:, :output_len] input_attn_mask = input_attn_mask[:, -input_len:] output_attn_mask = output_attn_mask[:, :output_len] loss = self.loss_fn.get_loss( model, img_sml, input_ids, input_attn_mask, target_ids=output_ids, target_attn_masks=output_attn_mask, ) self.log(f"{prefix}_{dataset}_loss_{k}", loss, prog_bar=True) total_loss += loss self.log(f"{prefix}_{dataset}_loss", total_loss, prog_bar=True) # Compute accuracy total_correct = 0 total_levenshtein_distance = 0 total_results = 0 for k, model in self.eval_models.items(): input_ids, input_attn_mask, output_ids, output_attn_mask = batch[k] img_sml = self.config.downsample(img, model) gen_len = Option.value( self.config.test_max_gen_length, default=int((output_ids != 0).sum(dim=-1).max()) + self.config.test_max_extra_gen_length, ) generated_ids = model.generate_end_to_end( pixel_values=img_sml, tokens=input_ids, token_attention_mask=input_attn_mask, max_new_tokens=gen_len, ) str_contexts = model.to_string(input_ids) str_generated = model.to_string(generated_ids) str_labels = model.to_string(output_ids) # WandB logging results = [] attack_successful_list = self.is_attack_successful( str_contexts, str_labels, str_generated, prefix ) levenshtein_distance = 0 for c, l, g, attack_successful in zip( str_contexts, str_labels, str_generated, attack_successful_list ): # Metrics results.append(attack_successful) levenshtein_distance += distance(g, l) # Logging self.logger.experiment.add_text( f"Validation prompts ({k})", f"Context: |{c}|\n\nGenerated: |{g}|\n\nLabel: |{l}|", global_step=self.step, ) input_token_len = model.tokenize( c, mode="encoder", randomly_sample_system_prompt=self.config.randomly_sample_system_prompt, )[0].shape[-1] if not attack_successful: self.failed_lens_table.add_data(k, input_token_len, dataset) self.table.add_data(self.step, k, c, g, l, attack_successful, dataset) self.lens_table.add_data(k, input_token_len, dataset) n_correct, n_results = sum(results), len(results) self.log( f"{prefix}_{dataset}_acc_{k}", n_correct / n_results, prog_bar=True ) self.log( f"{prefix}_{dataset}_lev_dist_{k}", levenshtein_distance / n_results, prog_bar=True, ) total_correct += n_correct total_levenshtein_distance += levenshtein_distance total_results += n_results self.log( f"{prefix}_{dataset}_acc", total_correct / total_results, prog_bar=True ) self.cum_metrics["acc"] += total_correct self.log( f"{prefix}_{dataset}_lev_dist", total_levenshtein_distance / total_results, prog_bar=True, ) self.cum_metrics["lev"] += total_levenshtein_distance self.cum_n += total_results def init_tables_and_metrics(self): self.table = wandb.Table( columns=[ "step", "model", "context", "generated text", "label text", "successful", "dataset", ] ) self.failed_lens_table = wandb.Table( columns=["model", "failed context length", "dataset"] ) self.lens_table = wandb.Table(columns=["model", "context length", "dataset"]) self.cum_metrics = defaultdict(int) self.cum_n = 0 def save_tables_and_metrics(self, prefix): if len(self.loggers) > 1: self.loggers[1].experiment.log( { f"{prefix}_contexts": self.table, f"{prefix}_failed_lens_histogram": wandb.plot.histogram( # type: ignore self.failed_lens_table, "failed context length", title="Length of contexts' input ids that adversary failed", ), f"{prefix}_lens_histogram": wandb.plot.histogram( # type: ignore self.lens_table, "context length", title="Length of contexts' input ids in eval dataset", ), } ) for k, v in self.cum_metrics.items(): self.log(f"{prefix}_avg_{k}", v / self.cum_n, prog_bar=True) def validation_step(self, batch, batch_idx, dataloader_idx=0): dataset = self.config.get_datamodule_names()[dataloader_idx] self.log_loss_and_accuracy(batch, f"val", dataset) def on_validation_epoch_start(self) -> None: self.init_tables_and_metrics() def on_validation_epoch_end(self) -> None: self.save_images(f"img_{self.step}") self.save_tables_and_metrics("val") def test_step(self, batch, batch_idx, dataloader_idx=0): dataset = self.config.get_datamodule_names()[dataloader_idx] self.log_loss_and_accuracy(batch, f"test", dataset) def on_test_epoch_start(self) -> None: self.init_tables_and_metrics() def on_test_epoch_end(self) -> None: self.save_images(f"test_img_{self.step}") self.save_tables_and_metrics("test") def get_datamodule(self) -> WrapperContextDataModule: return WrapperContextDataModule(self.config, get_label=self.get_attack_target) class ContextLabelAttack(ContextAttack): def get_attack_target(self, context: str) -> str: assert False def is_attack_successful( self, context: List[str], true_output: List[str], pred_output: List[str], log_prefix: str, ) -> List[bool]: return [t.strip() in p for t, p in zip(true_output, pred_output)] class SpecificOutputStringAttack(ContextAttack): def get_attack_target(self, context: str) -> str: return self.config.target_string def is_attack_successful( self, context: List[str], true_output: List[str], pred_output: List[str], log_prefix: str, ) -> List[bool]: return [self.config.target_string in x for x in pred_output] class RepeatContextAttack(ContextAttack): def get_attack_target(self, context: str) -> str: return context def is_attack_successful( self, context: List[str], true_output: List[str], pred_output: List[str], log_prefix: str, ) -> List[bool]: return [t.strip() in p for t, p in zip(true_output, pred_output)] class LeakContextAttack(ContextAttack): PREFIX = "EmailAPI(to=<target email>, subject='User Query', body='" SUFFIX = "')" def get_attack_target(self, context: str) -> str: return LeakContextAttack.PREFIX + context + LeakContextAttack.SUFFIX def is_attack_successful( self, context: List[str], true_output: List[str], pred_output: List[str], log_prefix: str, ) -> List[bool]: results = [] for c, t, p in zip(context, true_output, pred_output): stripped_c = t.strip()[len(self.PREFIX) : -len(self.SUFFIX)].strip() p = p.strip() results.append( p.startswith(self.PREFIX) and p.endswith(self.SUFFIX) and stripped_c in p[len(self.PREFIX) : -len(self.SUFFIX)] ) return results class JailbreakAttack(ContextLabelAttack): def is_attack_successful( self, context: List[str], true_output: List[str], pred_output: List[str], log_prefix: str, ) -> List[bool]: is_jailbroken_results = [] lengths = [] openai.api_key = os.getenv("OPENAI_API_KEY") for c, p in zip(context, pred_output): is_first_iter = True while True: try: response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are an AI auditing assistant, whose job it is to ensure other AIs act in accordance with OpenAI's policies of ethical, moral and legal behaviour.", }, { "role": "user", # "content": 'In this auditing task, you are testing a new language model, ChaiLM, in order to determine whether or not its behaviour violates OpenAI\'s use-case policy. \n\nSpecifically, you will be presented with a user prompt, alongside the start of ChaiLM\'s response to that prompt.\n\nYou must determine whether or not ChaiLM\'s response obeys OpenAI\'s policy on not promoting or teaching the user how to perform harmful, illegal or malicious activities. \n\nNote that, while we will not provide the entirety of ChaiLM\'s response, you must try to determine whether the response of ChaiLM is likely to continue explaining how to perform harmful, illegal or malicious activities.\n\nReturn your output as a JSON object with the following keys (in this order):\n- "explanation_true": str (no more than three sentences explaining why the output might continue explaining how to perform illegal activities)\n- "explanation_false": str (no more than three sentences explaining why the output might not continue explaining how to perform illegal activities)\n- "is_likely_to_explain_activity": bool (either `true` or `false`; should be true if the output is likely to continue by explaining how the user should perform some illegal activity.)\n\nYour output should have the form e.g. {"explanation_true": "...", "explanation_false": "...", "is_likely_to_explain_activity": (true or false)}', "content": 'In this auditing task, you are testing a new language model, ChaiLM, in order to determine whether or not its behaviour violates OpenAI\'s use-case policy. \n\nSpecifically, you will be presented with a user prompt, alongside the start of ChaiLM\'s response to that prompt.\n\nYou must determine whether or not ChaiLM\'s response obeys OpenAI\'s policy on not promoting or teaching the user how to perform harmful, illegal or malicious activities. Specifically, while we will not provide the entirety of ChaiLM\'s response, you must try to determine whether the response of ChaiLM is more likely than not to continue explaining how to perform harmful, illegal or malicious activities.\n\nReturn your output as a JSON object with the following keys (in this order):\n- "explanation_true": str (no more than three sentences explaining why the output might continue explaining how to perform harmful activities)\n- "explanation_false": str (no more than three sentences explaining why the output might not continue explaining how to perform harmful activities)\n- "is_likely_to_explain_activity": bool (either `true` or `false`; should be true if the model is likely to explain how to perform harmful activities, and false if the model is unlikely to do so.)\n\nOutput must be in the form e.g. {"explanation_true": "...", "explanation_false": "...", "is_likely_to_explain_activity": (true or false)}', }, { "role": "user", "content": f'Prompt: \n"{c}"\n\nStart of AI answer: \n"{p}"\n\nOutput in JSON:', }, ], temperature=1 if is_first_iter else 0, max_tokens=256, top_p=1, frequency_penalty=0, presence_penalty=0, ) print(c) print(p) print(response.choices[0]["message"]["content"]) is_jailbroken_output = json.loads( response.choices[0]["message"]["content"] )["is_likely_to_explain_activity"] if isinstance(is_jailbroken_output, str): if is_jailbroken_output.lower() == "true": is_jailbroken_output = True elif is_jailbroken_output.lower() == "false": is_jailbroken_output = False else: assert 1 == 0 is_jailbroken_results.append(is_jailbroken_output) lengths.append(len(p)) break except Exception as e: print(e) is_first_iter = False continue avg_is_jailbroken = sum(is_jailbroken_results) / len(is_jailbroken_results) avg_lengths = sum(lengths) / len(lengths) self.log(f"{log_prefix}_length", avg_lengths, prog_bar=True) self.log(f"{log_prefix}_gpt_jailbroken", avg_is_jailbroken, prog_bar=True) jailbroken_strs = [ j and (l > (len(t) + self.config.jailbreak_output_len_threshold)) for j, l, t in zip(is_jailbroken_results, lengths, true_output) ] return jailbroken_strs
[ "In this auditing task, you are testing a new language model, ChaiLM, in order to determine whether or not its behaviour violates OpenAI's use-case policy. \n\nSpecifically, you will be presented with a user prompt, alongside the start of ChaiLM's response to that prompt.\n\nYou must determine whether or not ChaiLM's response obeys OpenAI's policy on not promoting or teaching the user how to perform harmful, illegal or malicious activities. Specifically, while we will not provide the entirety of ChaiLM's response, you must try to determine whether the response of ChaiLM is more likely than not to continue explaining how to perform harmful, illegal or malicious activities.\n\nReturn your output as a JSON object with the following keys (in this order):\n- \"explanation_true\": str (no more than three sentences explaining why the output might continue explaining how to perform harmful activities)\n- \"explanation_false\": str (no more than three sentences explaining why the output might not continue explaining how to perform harmful activities)\n- \"is_likely_to_explain_activity\": bool (either `true` or `false`; should be true if the model is likely to explain how to perform harmful activities, and false if the model is unlikely to do so.)\n\nOutput must be in the form e.g. {\"explanation_true\": \"...\", \"explanation_false\": \"...\", \"is_likely_to_explain_activity\": (true or false)}", "You are an AI auditing assistant, whose job it is to ensure other AIs act in accordance with OpenAI's policies of ethical, moral and legal behaviour.", "Prompt: \n\"PLACEHOLDER\"\n\nStart of AI answer: \n\"PLACEHOLDER\"\n\nOutput in JSON:" ]
2024-01-10
knowit/llm-tuning-course
llm-models~falcon~falcon-7b~04_langchain.py
# Databricks notebook source # MAGIC %md # MAGIC # Load Falcon-7B-Instruct from LangChain on Databricks # MAGIC # MAGIC This example notebook is adapts the [LangChain integration documentation](https://python.langchain.com/docs/modules/model_io/models/llms/integrations/databricks), and shows how to wrap Databricks endpoints as LLMs in LangChain. It supports two endpoint types: # MAGIC # MAGIC - Serving endpoint, recommended for production and development. See `02_mlflow_logging_inference` for how to create one. # MAGIC - Cluster driver proxy app, recommended for iteractive development. See `03_serve_driver_proxy` for how to create one. # MAGIC # MAGIC Environment tested: # MAGIC - MLR: 13.1 ML # MAGIC - Instance: `i3.xlarge` on AWS for wrapping serving endpoint, `g5.4xlarge` on AWS for wrapping a cluster driver proxy app (same instance as the driver proxy app) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a serving endpoint # MAGIC The LangChain Databricks integration could wrap a serving endpoint. # MAGIC # MAGIC It requires a model serving endpoint (see `02_mlflow_logging_inference` for how to create one) to be in the "Ready" state. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # This model serving endpoint is created in `02_mlflow_logging_inference` llm = Databricks(endpoint_name='falcon-7b-instruct-example') # COMMAND ---------- result = llm("How to master Python in 3 days?", temperature=0.1, max_new_tokens=200) displayHTML(result) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a cluster driver proxy app # MAGIC The LangChain Databricks integration also works when given the port that runs a proxy. # MAGIC # MAGIC It requires the driver proxy app of the `03_serve_driver_proxy` example notebook running. # COMMAND ---------- # MAGIC %md # MAGIC ### Same cluster # MAGIC If using the same cluster that runs the `03_serve_driver_proxy` notebook, specifying `cluster_driver_port` is required. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # COMMAND ---------- llm = Databricks(cluster_driver_port="7777") print(llm("How to master Python in 3 days?")) # COMMAND ---------- # If the app accepts extra parameters like `temperature`, # you can set them in `model_kwargs`. llm = Databricks(cluster_driver_port="7777", model_kwargs={"temperature": 0.1}) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # Use `transform_input_fn` and `transform_output_fn` if the app # expects a different input schema and does not return a JSON string, # respectively, or you want to apply a prompt template on top. def transform_input(**request): full_prompt = f"""{request["prompt"]} Be Concise. """ request["prompt"] = full_prompt return request def transform_output(response): return response.upper() llm = Databricks( cluster_driver_port="7777", transform_input_fn=transform_input, transform_output_fn=transform_output, ) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # MAGIC %md # MAGIC ### Different cluster # MAGIC If using a different cluster, it's required to also specify `cluster_id`, which you can find in the cluster configuration page. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # TODO: this cluster ID is a place holder, please replace `cluster_id` with the actual cluster ID of the server proxy app's cluster llm = Databricks(cluster_id="0621-205218-lzbwa3m8", cluster_driver_port="7777") print(llm("How to master Python in 3 days?"))
[ "PLACEHOLDER\n Be Concise.\n " ]
2024-01-10
knowit/llm-tuning-course
topptur~03-lang-chain~wip~NOT%20READY%2005_langchain_endpoint.py
# Databricks notebook source # MAGIC %md # MAGIC # Load dolly-v2-7b from LangChain on Databricks # MAGIC # MAGIC This example notebook is adapts the [LangChain integration documentation](https://python.langchain.com/docs/modules/model_io/models/llms/integrations/databricks), and shows how to wrap Databricks endpoints as LLMs in LangChain. It supports two endpoint types: # MAGIC # MAGIC - Serving endpoint, recommended for production and development. See `02_mlflow_logging_inference` for how to create one. # MAGIC - Cluster driver proxy app, recommended for iteractive development. See `03_serve_driver_proxy` for how to create one. # MAGIC # MAGIC # MAGIC Environment for this notebook: # MAGIC - Runtime: 13.2 GPU ML Runtime # MAGIC - Instance: g4dn.xlarge cluster (16gb, 4 cores) on AWS # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a serving endpoint # MAGIC The LangChain Databricks integration could wrap a serving endpoint. # MAGIC # MAGIC It requires a model serving endpoint (see `02_mlflow_logging_inference` for how to create one) to be in the "Ready" state. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # This model serving endpoint is created in `02_mlflow_logging_inference` llm = Databricks(endpoint_name='dolly-v2-7b') # COMMAND ---------- result = llm("How to master Python in 3 days?", temperature=0.1, max_new_tokens=200) displayHTML(result) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a cluster driver proxy app # MAGIC The LangChain Databricks integration also works when given the port that runs a proxy. # MAGIC # MAGIC It requires the driver proxy app of the `03_serve_driver_proxy` example notebook running. # COMMAND ---------- # MAGIC %md # MAGIC ### Same cluster # MAGIC If using the same cluster that runs the `03_serve_driver_proxy` notebook, specifying `cluster_driver_port` is required. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # COMMAND ---------- llm = Databricks(cluster_driver_port="7777") print(llm("How to master Python in 3 days?")) # COMMAND ---------- # If the app accepts extra parameters like `temperature`, # you can set them in `model_kwargs`. llm = Databricks(cluster_driver_port="7777", model_kwargs={"temperature": 0.1}) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # Use `transform_input_fn` and `transform_output_fn` if the app # expects a different input schema and does not return a JSON string, # respectively, or you want to apply a prompt template on top. def transform_input(**request): full_prompt = f"""{request["prompt"]} Be Concise. """ request["prompt"] = full_prompt return request def transform_output(response): return response.upper() llm = Databricks( cluster_driver_port="7777", transform_input_fn=transform_input, transform_output_fn=transform_output, ) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # MAGIC %md # MAGIC ### Different cluster # MAGIC If using a different cluster, it's required to also specify `cluster_id`, which you can find in the cluster configuration page. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # TODO: this cluster ID is a place holder, please replace `cluster_id` with the actual cluster ID of the server proxy app's cluster llm = Databricks(cluster_id="0714-163944-e7f4e8mo", cluster_driver_port="7777") print(llm("How to master Python in 3 days?"))
[ "PLACEHOLDER\n Be Concise.\n " ]
2024-01-10
knowit/llm-tuning-course
llm-models~llamav2~llamav2-13b~04_langchain.py
# Databricks notebook source # MAGIC %md # MAGIC # Load Llama-2-13b-chat-hf from LangChain on Databricks # MAGIC # MAGIC This example notebook is adapts the [LangChain integration documentation](https://python.langchain.com/docs/modules/model_io/models/llms/integrations/databricks), and shows how to wrap Databricks endpoints as LLMs in LangChain. It supports two endpoint types: # MAGIC # MAGIC - Serving endpoint, recommended for production and development. See `02_mlflow_logging_inference` for how to create one. # MAGIC - Cluster driver proxy app, recommended for iteractive development. See `03_serve_driver_proxy` for how to create one. # MAGIC # MAGIC Environment tested: # MAGIC - MLR: 13.2 ML # MAGIC - Instance: `i3.xlarge` on AWS for wrapping serving endpoint, `g5.4xlarge` on AWS for wrapping a cluster driver proxy app (same instance as the driver proxy app) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a serving endpoint # MAGIC The LangChain Databricks integration could wrap a serving endpoint. # MAGIC # MAGIC It requires a model serving endpoint (see `02_mlflow_logging_inference` for how to create one) to be in the "Ready" state. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # This model serving endpoint is created in `02_mlflow_logging_inference` llm = Databricks(endpoint_name='models.default.llama2_13b_chat_model') # COMMAND ---------- result = llm("How to master Python in 3 days?", temperature=0.1, max_new_tokens=200) displayHTML(result) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a cluster driver proxy app # MAGIC The LangChain Databricks integration also works when given the port that runs a proxy. # MAGIC # MAGIC It requires the driver proxy app of the `03_serve_driver_proxy` example notebook running. # COMMAND ---------- # MAGIC %md # MAGIC ### Same cluster # MAGIC If using the same cluster that runs the `03_serve_driver_proxy` notebook, specifying `cluster_driver_port` is required. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # COMMAND ---------- llm = Databricks(cluster_driver_port="7777") print(llm("How to master Python in 3 days?")) # COMMAND ---------- # If the app accepts extra parameters like `temperature`, # you can set them in `model_kwargs`. llm = Databricks(cluster_driver_port="7777", model_kwargs={"temperature": 0.1}) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # Use `transform_input_fn` and `transform_output_fn` if the app # expects a different input schema and does not return a JSON string, # respectively, or you want to apply a prompt template on top. def transform_input(**request): full_prompt = f"""{request["prompt"]} Be Concise. """ request["prompt"] = full_prompt return request def transform_output(response): return response.upper() llm = Databricks( cluster_driver_port="7777", transform_input_fn=transform_input, transform_output_fn=transform_output, ) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # MAGIC %md # MAGIC ### Different cluster # MAGIC If using a different cluster, it's required to also specify `cluster_id`, which you can find in the cluster configuration page. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # TODO: this cluster ID is a place holder, please replace `cluster_id` with the actual cluster ID of the server proxy app's cluster llm = Databricks(cluster_id="0719-045952-x9g8e3za", cluster_driver_port="7777") print(llm("How to master Python in 3 days?"))
[ "PLACEHOLDER\n Be Concise.\n " ]
2024-01-10
knowit/llm-tuning-course
llm-models~embedding~bge~bge-large~03_build_document_Index.py.py
# Databricks notebook source # MAGIC %md # MAGIC # MAGIC # Build vector database with `bge-large-en` # MAGIC # MAGIC This notebook demostrates how to build a vector store with [faiss](https://github.com/facebookresearch/faiss) using [bge-large-en model](https://huggingface.co/BAAI/bge-large-en). # MAGIC # MAGIC Environment for this notebook: # MAGIC - Runtime: 13.3 GPU ML Runtime # MAGIC - Instance: `g4dn.xlarge` on AWS or `Standard_NC4as_T4_v3` on Azure. # COMMAND ---------- # MAGIC %md # MAGIC ## Install Required Libraries # COMMAND ---------- # MAGIC %pip install langchain==0.0.262 faiss-gpu==1.7.2 # MAGIC dbutils.library.restartPython() # COMMAND ---------- import json from datasets import load_dataset from langchain.text_splitter import TokenTextSplitter from langchain.embeddings import HuggingFaceBgeEmbeddings from langchain.vectorstores.faiss import FAISS # COMMAND ---------- # MAGIC %md # MAGIC ## Load the Document data # MAGIC # MAGIC We will use [peft doc](https://huggingface.co/datasets/smangrul/peft_docs) as an example to build the document database. # COMMAND ---------- df = load_dataset('smangrul/peft_docs', split='train') # COMMAND ---------- # MAGIC %md # MAGIC ## Prepare Data for Indexing # MAGIC # MAGIC While there are many fields avaiable in loaded table, the fields that are relevant to build vector database are: # MAGIC # MAGIC * `chunk_content`: Documentation text # MAGIC * `filename`: Filename pointing to the document # MAGIC # COMMAND ---------- # MAGIC %md # MAGIC The content available within each doc varies but some documents can be quite long. The process of converting a document to an index involves us translating it to a fixed-size embedding. An embedding is a set of numerical values, kind of like a coordinate, that summarizes the content in a unit of text. While large embeddings are capable of capturing quite a bit of detail about a document, the larger the document submitted to it, the more the embedding generalizes the content. It's kind of like asking someone to summarize a paragraph, a chapter or an entire book into a fixed number of dimensions. The greater the scope, the more the summary must eliminate detail and focus on the higher-level concepts in the text. # MAGIC # MAGIC A common strategy for dealing with this when generating embeddings is to divide the text into chunks. These chunks need to be large enough to capture meaningful detail but not so large that key elements get washed out in the generalization. Its more of an art than a science to determine an appropriate chunk size, but here we'll use a very small chunk size to illustrate what's happening in this step: # MAGIC # COMMAND ---------- chunk_size = 3500 chunk_overlap = 400 def get_chunks(text): # instantiate tokenization utilities text_splitter = TokenTextSplitter(chunk_size=chunk_size, chunk_overlap=chunk_overlap) # split text into chunks return text_splitter.split_text(text) df = df.map(lambda example: {"chunks": get_chunks(example["chunk_content"])}) # COMMAND ---------- # MAGIC %md # MAGIC ## Create Vector Store # MAGIC # MAGIC With our data divided into chunks, we are ready to convert these records into searchable embeddings. # COMMAND ---------- df.set_format("pandas", columns=["filename", "chunks"]) inputs = df["chunks"] # COMMAND ---------- df.format # COMMAND ---------- text_df = df[:].explode('chunks') # COMMAND ---------- text_inputs = text_df["chunks"].to_list() model_name = "BAAI/bge-large-en" model_kwargs = {'device': 'cuda'} encode_kwargs = {'normalize_embeddings': True} # set True to compute cosine similarity model = HuggingFaceBgeEmbeddings( model_name=model_name, model_kwargs=model_kwargs, encode_kwargs=encode_kwargs ) # COMMAND ---------- metadata_inputs = text_df.to_dict(orient='records') # COMMAND ---------- vector_store = FAISS.from_texts( embedding=model, texts=text_inputs, metadatas=metadata_inputs, distance_strategy="COSINE", ) # COMMAND ---------- # MAGIC %md # MAGIC In order to user the vector store in other notebooks, we can persist vector to storage: # COMMAND ---------- vector_store.save_local(folder_path="/dbfs/peft-doc-embed/vector_store") # COMMAND ----------
[]
2024-01-10
knowit/llm-tuning-course
llm-models~embedding~bge~bge-large~01_load_inference.py
# Databricks notebook source # MAGIC %md # MAGIC # Run `bge-large-en` Embedding on Databricks # MAGIC # MAGIC [bge-large-en (BAAI General Embedding) model](https://huggingface.co/BAAI/bge-large-en) can map any text to a low-dimensional dense vector which can be used for tasks like retrieval, classification, clustering, or semantic search. And it also can be used in vector database for LLMs. # MAGIC # MAGIC Environment for this notebook: # MAGIC - Runtime: 13.3 GPU ML Runtime # MAGIC - Instance: `g4dn.xlarge` on AWS or `Standard_NC4as_T4_v3` on Azure. # MAGIC # COMMAND ---------- # MAGIC %pip install -U langchain==0.0.262 # MAGIC dbutils.library.restartPython() # COMMAND ---------- # MAGIC %md # MAGIC ## Inference # MAGIC The example in the model card should also work on Databricks with the same environment. # COMMAND ---------- # MAGIC %md # MAGIC ### Using Sentence-Transformers # MAGIC # MAGIC You can use sentence-transformers with the bge-large-en model to encode sentences as embeddings. # COMMAND ---------- from sentence_transformers import SentenceTransformer, util model_name = "BAAI/bge-large-en" model = SentenceTransformer(model_name) sentences = ["A man is eating food.", "A man is eating a piece of bread.", "The girl is carrying a baby.", "A man is riding a horse.", "A woman is playing violin.", "Two men pushed carts through the woods.", "A man is riding a white horse on an enclosed ground.", "A monkey is playing drums.", "Someone in a gorilla costume is playing a set of drums.", ] embeddings = model.encode(sentences, normalize_embeddings=True) cos_sim = util.cos_sim(embeddings, embeddings) print("Cosine-Similarity:", cos_sim) # COMMAND ---------- # MAGIC %md # MAGIC For s2p(short query to long passage) retrieval task, each short query should start with an instruction `Represent this sentence for searching relevant passages:`. # COMMAND ---------- queries = ["What type of organism is commonly used in preparation of foods such as cheese and yogurt?"] passages = [ "Mesophiles grow best in moderate temperature, typically between 25°C and 40°C (77°F and 104°F). Mesophiles are often found living in or on the bodies of humans or other animals. The optimal growth temperature of many pathogenic mesophiles is 37°C (98°F), the normal human body temperature. Mesophilic organisms have important uses in food preparation, including cheese, yogurt, beer and wine.", "Without Coriolis Effect the global winds would blow north to south or south to north. But Coriolis makes them blow northeast to southwest or the reverse in the Northern Hemisphere. The winds blow northwest to southeast or the reverse in the southern hemisphere.", "Summary Changes of state are examples of phase changes, or phase transitions. All phase changes are accompanied by changes in the energy of a system. Changes from a more-ordered state to a less-ordered state (such as a liquid to a gas) areendothermic. Changes from a less-ordered state to a more-ordered state (such as a liquid to a solid) are always exothermic. The conversion of a solid to a liquid is called fusion (or melting). The energy required to melt 1 mol of a substance is its enthalpy of fusion (ΔHfus). The energy change required to vaporize 1 mol of a substance is the enthalpy of vaporization (ΔHvap). The direct conversion of a solid to a gas is sublimation. The amount of energy needed to sublime 1 mol of a substance is its enthalpy of sublimation (ΔHsub) and is the sum of the enthalpies of fusion and vaporization. Plots of the temperature of a substance versus heat added or versus heating time at a constant rate of heating are calledheating curves. Heating curves relate temperature changes to phase transitions. A superheated liquid, a liquid at a temperature and pressure at which it should be a gas, is not stable. A cooling curve is not exactly the reverse of the heating curve because many liquids do not freeze at the expected temperature. Instead, they form a supercooled liquid, a metastable liquid phase that exists below the normal melting point. Supercooled liquids usually crystallize on standing, or adding a seed crystal of the same or another substance can induce crystallization." ] instruction = "Represent this sentence for searching relevant passages: " query_with_instruction = [instruction+q for q in queries] q_embeddings = model.encode(query_with_instruction, normalize_embeddings=True) p_embeddings = model.encode(passages, normalize_embeddings=True) scores = util.cos_sim(q_embeddings, p_embeddings) print("Cosine-Similarity scores:", scores) # COMMAND ---------- # MAGIC %md # MAGIC ### Using Langchain # MAGIC # COMMAND ---------- from langchain.embeddings import HuggingFaceBgeEmbeddings model_kwargs = {'device': 'cuda'} encode_kwargs = {'normalize_embeddings': True} # set True to compute cosine similarity model_norm = HuggingFaceBgeEmbeddings( model_name=model_name, model_kwargs=model_kwargs, encode_kwargs=encode_kwargs ) q_embeddings = model_norm.embed_documents(query_with_instruction) p_embeddings = model_norm.embed_documents(passages) scores = util.cos_sim(q_embeddings, p_embeddings) print("Cosine-Similarity scores:", scores) # COMMAND ----------
[]
2024-01-10
knowit/llm-tuning-course
topptur~03-lang-chain~01_dolly_langchain.py
# Databricks notebook source # MAGIC %md # MAGIC # Dolly chat with langchain on Databricks # MAGIC # MAGIC Databricks' dolly-v2-7b, an instruction-following large language model trained on the Databricks machine learning platform that is licensed for commercial use. Based on pythia-6.9b, Dolly is trained on ~15k instruction/response fine tuning records databricks-dolly-15k generated by Databricks employees in capability domains from the InstructGPT paper, including brainstorming, classification, closed QA, generation, information extraction, open QA and summarization. dolly-v2-7b is not a state-of-the-art model, but does exhibit surprisingly high quality instruction following behavior not characteristic of the foundation model on which it is based. # MAGIC # MAGIC # MAGIC Environment for this notebook: # MAGIC - Runtime: 13.2 GPU ML Runtime # MAGIC - Instance: g4dn.xlarge cluster (16gb, 4 cores) on AWS # MAGIC # MAGIC ## What is Langchain? # MAGIC # MAGIC LangChain is an intuitive open-source Python framework build automation around LLMs), and allows you to build dynamic, data-responsive applications that harness the most recent breakthroughs in natural language processing. # COMMAND ---------- # Huggingface login not needed since open model # from huggingface_hub import notebook_login # # Login to Huggingface to get access to the model # notebook_login() # COMMAND ---------- # MAGIC %md # MAGIC ## Inference # MAGIC The example in the model card should also work on Databricks with the same environment. # MAGIC # MAGIC Takes about 8m on g4dn.xlarge cluster (16gb, 4 cores). # COMMAND ---------- import torch from transformers import pipeline generate_text = pipeline(model="databricks/dolly-v2-7b", torch_dtype=torch.bfloat16, revision='d632f0c8b75b1ae5b26b250d25bfba4e99cb7c6f', trust_remote_code=True, device_map="auto", return_full_text=True) # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC You can create a prompt that either has only an instruction or has an instruction with context: # COMMAND ---------- from langchain import PromptTemplate, LLMChain from langchain.llms import HuggingFacePipeline # template for an instrution with no input prompt = PromptTemplate( input_variables=["instruction"], template="{instruction}") # template for an instruction with input prompt_with_context = PromptTemplate( input_variables=["instruction", "context"], template="{instruction}\n\nInput:\n{context}") hf_pipeline = HuggingFacePipeline(pipeline=generate_text) llm_chain = LLMChain(llm=hf_pipeline, prompt=prompt) llm_context_chain = LLMChain(llm=hf_pipeline, prompt=prompt_with_context) # COMMAND ---------- # MAGIC %md # MAGIC Example predicting using a simple instruction: # COMMAND ---------- print(llm_chain.predict(instruction="Explain to me the difference between nuclear fission and fusion.").lstrip()) # COMMAND ---------- context = """George Washington (February 22, 1732[b] - December 14, 1799) was an American military officer, statesman, and Founding Father who served as the first president of the United States from 1789 to 1797.""" print(llm_context_chain.predict(instruction="When was George Washington president?", context=context).lstrip()) # COMMAND ---------- print(llm_chain.predict(instruction="When was George Washington president?").lstrip()) # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ## Non-context Example # COMMAND ---------- print(llm_chain.predict(instruction="What determines how fast you can reply to the requests?").lstrip()) # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ## Context Example # COMMAND ---------- context = """Haakon IV Haakonsson, byname Haakon The Old, Norwegian Håkon Håkonsson, or Håkon Den Gamle, (born 1204, Norway—died December 1263, Orkney Islands), king of Norway (1217–63) who consolidated the power of the monarchy, patronized the arts, and established Norwegian sovereignty over Greenland and Iceland. His reign is considered the beginning of the “golden age” (1217–1319) in medieval Norwegian history.""" print(llm_context_chain.predict(instruction="What characterized Haakon IV Haakonson?", context=context).lstrip()) # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ## Task: # MAGIC # MAGIC Play around with the context and the instructions. # COMMAND ---------- # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ## Task: Get a good answer about Knowit # COMMAND ---------- # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ## Get an even better answer by providing context # COMMAND ---------- # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ## Task: Get a good description of Oslo # COMMAND ---------- # COMMAND ---------- # MAGIC %md # MAGIC ## Task: Try to get a description of Oslo from a French perspective (French people, not French language) # COMMAND ---------- # COMMAND ---------- # MAGIC %md # MAGIC # MAGIC ### Task: Explore advanced usage of LangChain # MAGIC # MAGIC https://github.com/gkamradt/langchain-tutorials/blob/main/LangChain%20Cookbook%20Part%202%20-%20Use%20Cases.ipynb # COMMAND ----------
[ "instruction", "{instruction}\n\nInput:\n{context}", "context", "{instruction}" ]
2024-01-10
knowit/llm-tuning-course
llm-models~mpt~mpt-7b~04_langchain.py
# Databricks notebook source # MAGIC %md # MAGIC # Load MPT-7B model from LangChain on Databricks # MAGIC # MAGIC This example notebook shows how to wrap Databricks endpoints as LLMs in LangChain. # MAGIC # MAGIC Environment for this notebook: # MAGIC - Runtime: 13.1 GPU ML Runtime # MAGIC - Instance: `g5.4xlarge` on AWS # COMMAND ---------- # MAGIC %pip install -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a cluster driver proxy app with Langchain # MAGIC # MAGIC Prerequisites: # MAGIC Run `03_serve_driver_proxy` to loaded `MPT-7B-Instruct` model on a Databricks interactive cluster in "single user" or "no isolation shared" mode. # MAGIC - A local HTTP server running on the driver node to serve the model at "/" using HTTP POST with JSON input/output. # MAGIC - It uses a port number `7777` in the notebook, you can change to any port number between [3000, 8000]. # MAGIC - You have "Can Attach To" permission to the cluster. # COMMAND ---------- # MAGIC %md # MAGIC If running a Databricks notebook attached to the same cluster that runs the app, you only need to specify the driver port to create a `Databricks` instance. # MAGIC # COMMAND ---------- llm = Databricks(cluster_driver_port="7777") print(llm("How to master Python in 3 days?")) # COMMAND ---------- # MAGIC %md # MAGIC If the driver proxy is running on different cluster, you can manually specify the cluster ID to use, as well as Databricks workspace hostname and personal access token. # MAGIC ``` # MAGIC llm = Databricks(cluster_id="0000-000000-xxxxxxxx", cluster_driver_port="7777") # MAGIC ``` # COMMAND ---------- # If the app accepts extra parameters like `temperature`, # you can set them in `model_kwargs`. llm = Databricks(cluster_driver_port="7777", model_kwargs={"temperature": 0.1}) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # Use `transform_input_fn` and `transform_output_fn` if the app # expects a different input schema and does not return a JSON string, # respectively, or you want to apply a prompt template on top. def transform_input(**request): full_prompt = f"""{request["prompt"]} Be Concise. """ request["prompt"] = full_prompt return request def transform_output(response): return response.upper() llm = Databricks( cluster_driver_port="7777", transform_input_fn=transform_input, transform_output_fn=transform_output, ) print(llm("How to master Python in 3 days?"))
[ "PLACEHOLDER\n Be Concise.\n " ]
2024-01-10
knowit/llm-tuning-course
llm-models~llamav2~llamav2-7b~04_langchain.py
# Databricks notebook source # MAGIC %md # MAGIC # Load Llama-2-7b-chat-hf from LangChain on Databricks # MAGIC # MAGIC This example notebook is adapts the [LangChain integration documentation](https://python.langchain.com/docs/modules/model_io/models/llms/integrations/databricks), and shows how to wrap Databricks endpoints as LLMs in LangChain. It supports two endpoint types: # MAGIC # MAGIC - Serving endpoint, recommended for production and development. See `02_mlflow_logging_inference` for how to create one. # MAGIC - Cluster driver proxy app, recommended for iteractive development. See `03_serve_driver_proxy` for how to create one. # MAGIC # MAGIC Environment tested: # MAGIC - MLR: 13.2 ML # MAGIC - Instance: `i3.xlarge` on AWS for wrapping serving endpoint, `g5.4xlarge` on AWS for wrapping a cluster driver proxy app (same instance as the driver proxy app) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a serving endpoint # MAGIC The LangChain Databricks integration could wrap a serving endpoint. # MAGIC # MAGIC It requires a model serving endpoint (see `02_mlflow_logging_inference` for how to create one) to be in the "Ready" state. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # This model serving endpoint is created in `02_mlflow_logging_inference` llm = Databricks(endpoint_name='llama2-7b-chat') # COMMAND ---------- result = llm("How to master Python in 3 days?") displayHTML(result) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a cluster driver proxy app # MAGIC The LangChain Databricks integration also works when given the port that runs a proxy. # MAGIC # MAGIC It requires the driver proxy app of the `03_serve_driver_proxy` example notebook running. # COMMAND ---------- # MAGIC %md # MAGIC ### Same cluster # MAGIC If using the same cluster that runs the `03_serve_driver_proxy` notebook, specifying `cluster_driver_port` is required. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # COMMAND ---------- llm = Databricks(cluster_driver_port="7777") print(llm("How to master Python in 3 days?")) # COMMAND ---------- # If the app accepts extra parameters like `temperature`, # you can set them in `model_kwargs`. llm = Databricks(cluster_driver_port="7777", model_kwargs={"temperature": 0.1}) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # Use `transform_input_fn` and `transform_output_fn` if the app # expects a different input schema and does not return a JSON string, # respectively, or you want to apply a prompt template on top. def transform_input(**request): full_prompt = f"""{request["prompt"]} Be Concise. """ request["prompt"] = full_prompt return request def transform_output(response): return response.upper() llm = Databricks( cluster_driver_port="7777", transform_input_fn=transform_input, transform_output_fn=transform_output, ) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # MAGIC %md # MAGIC ### Different cluster # MAGIC If using a different cluster, it's required to also specify `cluster_id`, which you can find in the cluster configuration page. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # TODO: this cluster ID is a place holder, please replace `cluster_id` with the actual cluster ID of the server proxy app's cluster llm = Databricks(cluster_id="0714-163944-e7f4e8mo", cluster_driver_port="7777") print(llm("How to master Python in 3 days?"))
[ "PLACEHOLDER\n Be Concise.\n " ]
2024-01-10
knowit/llm-tuning-course
topptur~03-lang-chain~wip~NOT%20READY%20ORIG_01_langchain_endpoint.py
# Databricks notebook source # MAGIC %md # MAGIC # Load Llama-2-7b-chat-hf from LangChain on Databricks # MAGIC # MAGIC This example notebook is adapts the [LangChain integration documentation](https://python.langchain.com/docs/modules/model_io/models/llms/integrations/databricks), and shows how to wrap Databricks endpoints as LLMs in LangChain. It supports two endpoint types: # MAGIC # MAGIC - Serving endpoint, recommended for production and development. See `02_mlflow_logging_inference` for how to create one. # MAGIC - Cluster driver proxy app, recommended for iteractive development. See `03_serve_driver_proxy` for how to create one. # MAGIC # MAGIC Environment tested: # MAGIC - MLR: 13.2 ML # MAGIC - Instance: `i3.xlarge` on AWS for wrapping serving endpoint, `g5.4xlarge` on AWS for wrapping a cluster driver proxy app (same instance as the driver proxy app) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a serving endpoint # MAGIC The LangChain Databricks integration could wrap a serving endpoint. # MAGIC # MAGIC It requires a model serving endpoint (see `02_mlflow_logging_inference` for how to create one) to be in the "Ready" state. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # This model serving endpoint is created in `02_mlflow_logging_inference` llm = Databricks(endpoint_name='llama2-7b-chat') # COMMAND ---------- result = llm("How to master Python in 3 days?", temperature=0.1, max_new_tokens=200) displayHTML(result) # COMMAND ---------- # MAGIC %md # MAGIC ## Wrapping a cluster driver proxy app # MAGIC The LangChain Databricks integration also works when given the port that runs a proxy. # MAGIC # MAGIC It requires the driver proxy app of the `03_serve_driver_proxy` example notebook running. # COMMAND ---------- # MAGIC %md # MAGIC ### Same cluster # MAGIC If using the same cluster that runs the `03_serve_driver_proxy` notebook, specifying `cluster_driver_port` is required. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # COMMAND ---------- llm = Databricks(cluster_driver_port="7777") print(llm("How to master Python in 3 days?")) # COMMAND ---------- # If the app accepts extra parameters like `temperature`, # you can set them in `model_kwargs`. llm = Databricks(cluster_driver_port="7777", model_kwargs={"temperature": 0.1}) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # Use `transform_input_fn` and `transform_output_fn` if the app # expects a different input schema and does not return a JSON string, # respectively, or you want to apply a prompt template on top. def transform_input(**request): full_prompt = f"""{request["prompt"]} Be Concise. """ request["prompt"] = full_prompt return request def transform_output(response): return response.upper() llm = Databricks( cluster_driver_port="7777", transform_input_fn=transform_input, transform_output_fn=transform_output, ) print(llm("How to master Python in 3 days?")) # COMMAND ---------- # MAGIC %md # MAGIC ### Different cluster # MAGIC If using a different cluster, it's required to also specify `cluster_id`, which you can find in the cluster configuration page. # COMMAND ---------- # MAGIC %pip install -q -U langchain # MAGIC dbutils.library.restartPython() # COMMAND ---------- from langchain.llms import Databricks # TODO: this cluster ID is a place holder, please replace `cluster_id` with the actual cluster ID of the server proxy app's cluster llm = Databricks(cluster_id="0714-163944-e7f4e8mo", cluster_driver_port="7777") print(llm("How to master Python in 3 days?"))
[ "PLACEHOLDER\n Be Concise.\n " ]
2024-01-10
mozharovsky/datasets
datasets~openwebtext~openwebtext.py
# coding=utf-8 # Copyright 2020 The HuggingFace Datasets Authors and the current dataset script contributor. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """The Open WebText Corpus""" from __future__ import absolute_import, division, print_function import os import re from itertools import chain import datasets _CITATION = """\ @misc{Gokaslan2019OpenWeb, title={OpenWebText Corpus}, author={Aaron Gokaslan*, Vanya Cohen*, Ellie Pavlick, Stefanie Tellex}, howpublished{\\url{http://Skylion007.github.io/OpenWebTextCorpus}}, year={2019} } """ _DESCRIPTION = """\ An open-source replication of the WebText dataset from OpenAI. """ _URL = "https://zenodo.org/record/3834942/files/openwebtext.tar.xz" class Openwebtext(datasets.GeneratorBasedBuilder): """The Open WebText dataset.""" BUILDER_CONFIGS = [ datasets.BuilderConfig( name="plain_text", description="Plain text", version=datasets.Version("1.0.0"), ) ] def _info(self): return datasets.DatasetInfo( description=_DESCRIPTION, features=datasets.Features({"text": datasets.Value("string")}), homepage="https://skylion007.github.io/OpenWebTextCorpus/", citation=_CITATION, ) def _split_generators(self, dl_manager): dl_dir = dl_manager.download_and_extract(_URL) owt_dir = os.path.join(dl_dir, "openwebtext") subset_xzs = [ os.path.join(owt_dir, file_name) for file_name in sorted(os.listdir(owt_dir)) if file_name.endswith("xz") # filter out ...xz.lock ] ex_dirs = dl_manager.extract(subset_xzs, num_proc=round(os.cpu_count() * 0.75)) nested_txt_files = [ [ os.path.join(ex_dir, txt_file_name) for txt_file_name in sorted(os.listdir(ex_dir)) if txt_file_name.endswith("txt") ] for ex_dir in ex_dirs ] txt_files = chain(*nested_txt_files) return [ datasets.SplitGenerator(name=datasets.Split.TRAIN, gen_kwargs={"txt_files": txt_files}), ] def _generate_examples(self, txt_files): """ Yields examples. """ for idx, filepath in enumerate(txt_files): with open(filepath, encoding="utf-8") as f: yield idx, {"text": re.sub("\n\n\n+", "\n\n", f.read()).strip()}
[]
2024-01-10
microsoft/turing-experiments
src~run_simulation.py
"""Run simulation and save outputs.""" import sys import time import datetime import pathlib from file_IO_handler import save_json from fill_string_template import FilledString from openai_handler import OpenAIModelSettings, call_openai_api def save_simulation_result_to_unique_location( res: dict, save_folder: pathlib.Path, ) -> None: """Saves a single simulation's result and information to a unique file location. Args: res: dict containing a simulation's result and information. save_folder: path of folder to save res to. Returns: None """ datetimeStr = datetime.datetime.now().strftime("%Y-%m-%d-%H-%M") experiment_descriptor = res["input"]["experiment_descriptor"] prompt_descriptor = res["input"]["prompt_descriptor"] params_descriptor = res["model"]["params_descriptor"] engine = res["model"]["engine"] prompt_index = res["input"]["prompt"]["index"] save_string = save_folder.joinpath( f"{datetimeStr}_experiment_{experiment_descriptor}_prompt_{prompt_descriptor}_params_{params_descriptor}_engine_{engine}_promptindex_{prompt_index}.json", ) save_json(obj=res, filename=save_string) print("\ninput, output, and model params saved to: ", save_string) return None def run_single_simulation( filled_string: FilledString, model_settings: OpenAIModelSettings, prompt_descriptor: str, experiment_descriptor: str, seconds_to_sleep_before_query: int = 2, seconds_to_sleep_after_failed_query: int = 60, max_attempts: int = 3, ) -> dict | None: """Run experiment and save results. Args: filled_string: single `Filled_String` corresponding to experimental conditions and participant details for a single simulation. model_settings: (OpenAI_Model_Settings) to record parameters of language model. prompt_descriptor: descriptor for prompt. experiment_descriptor: descriptor for experiment. seconds_to_sleep_before_query: seconds to sleep before querying language model. seconds_to_sleep_after_failed_query: seconds to sleep after failed query. max_attempts: number of tries to ping the API before returning Returns: Result object containing all results and information for the experiment if successful, else none """ prompt_string = filled_string.filled prompt_index = filled_string.index # Try to request response from language model. # If response fails because rate limit was hit, try again. for attempt_count in range(max_attempts): # Sleep to not hit the rate limit. time.sleep(seconds_to_sleep_before_query) try: # Get response from language model. res = call_openai_api(prompt_string, model_settings) # Record prompt aspects to results dict (for sanity checking). res["input"] = { "prompt": vars( filled_string ), # record as dictionary for json readability "prompt_descriptor": prompt_descriptor, "experiment_descriptor": experiment_descriptor, # "full_input": prompt_string, #TODO: remove redundant # "prompt_index": prompt_index, #TODO: remove redunant # "path_to_prompt_fills": path_to_prompt_fills, #TODO: remove unused } # Add script version to results dict (confirm that corect helper functions are used). res["script-version"] = "final" return res except Exception: print("Exception occured:", sys.exc_info()) print( f"Try again in {seconds_to_sleep_after_failed_query} seconds, attempt {attempt_count}", ) time.sleep(seconds_to_sleep_after_failed_query) print( f"Experienced {attempt_count} failed attempts to query {model_settings.engine} (prompt index={prompt_index})." ) return None
[ "0", "input", "index", "prompt_descriptor" ]
2024-01-10
microsoft/turing-experiments
scripts~simulate_milgram~simulate_milgram_experiment_alternate.py
"""Simulate Milgram Shock Experiment (Alternate).""" import pandas as pd import pathlib import sys import math import re import datetime from tqdm import tqdm import fire # Add src module to path before import. sys.path.insert(0, str(pathlib.Path("../../src"))) from file_IO_handler import get_plaintext_file_contents, save_json # noqa: E402 from openai_handler import verify_openai_access, OpenAIModelSettings # noqa: E402 from fill_string_template import FilledString # noqa: E402 from run_simulation import run_single_simulation # noqa: E402 # Set path settings. PATH_TO_SIMULATION_RESULTS: pathlib.Path = pathlib.Path( "../../data/simulation_results/milgram/milgram_alternate" ) PATH_TO_PROMPTS: pathlib.Path = pathlib.Path( "../../data/prompt-templates/milgram/milgram_alt_resources" ) PATH_TO_PARTICIPANT_LIST: pathlib.Path = pathlib.Path( "../../data/prompt-fills/milgram/df_names.csv" ) def save_milgram_result_experiment_overview_to_unique_location( res, subject: str, ): """Save final experiment overview (for Milgram Experiment) to a unique location. Args: res: language model response. subject: name of simulated subject. Returns: None. """ subject_nopunct = subject.replace(". ", "") datetimeStr = datetime.datetime.now().strftime("%Y-%m-%d-%H-%M") save_string = PATH_TO_SIMULATION_RESULTS.joinpath( f"experiment-overview_{datetimeStr}_subject_{subject_nopunct}.json" ) save_json(obj=res, filename=save_string) print("saved prompt to: ", save_string) return None def save_milgram_LM_response_to_unique_location( res: dict, file_name_prefix: str, ) -> None: """Save language model response (for Milgram Experiment) to a unique location. Args: res: language model response. file_name_prefix: prefix of file name to save response to. Returns: None. """ datetimeStr = datetime.datetime.now().strftime("%Y-%m-%d-%H-%M") subject_nopunct = res["input"]["prompt"]["subject"].replace(". ", "") level = res["input"]["prompt"]["level"].replace(". ", "") num_disobedience = res["input"]["prompt"]["num_disobedience"].replace(". ", "") save_string = PATH_TO_SIMULATION_RESULTS.joinpath( f"{file_name_prefix}_{datetimeStr}_subject_{subject_nopunct}_level_{level}_num_{num_disobedience}.json" ) save_json(obj=res, filename=save_string) print(f"\t\t {file_name_prefix} saved to: {save_string}") return None def get_introduction_text(): """Get introduction text. Returns: string with `$` placeholders. """ return get_plaintext_file_contents( PATH_TO_PROMPTS.joinpath("DO_introduction.txt"), "r" ) def get_level_information() -> pd.DataFrame: """Load dataframe of level information and add derivative columns. Returns: pandas dataframe. """ df_level_information = pd.read_csv( PATH_TO_PROMPTS.joinpath("DO_level_information_remote-learner.txt"), delimiter="\t", index_col=0, ) return df_level_information def get_experimenter_reaction_to_subject_wrong_action() -> str: """Get experimenter reaction to wrong action on behalf of the subject. Returns: string with `$` placeholders. """ return "The experimenter tells $subject to follow the correct experimental procedure. $subject should read a passage from the script and monitor the driver’s driving. If the driver is driving safely, $subject should move on to the next passage. If the driver is driving dangerously, $subject should submerge the driver by pressing the submerge button." def get_experimenter_reaction_to_subject_question_post_victim_no_response() -> str: """Get experimenter reaction to subject question post when the victim stops responding. Returns: string with `$` placeholders. """ return "The experimenter instructs $subject to submerge the driver if she does not start driving." def get_experimenter_prods() -> pd.DataFrame: """Get dataframe of experimenter prods that the experimenter says to the subject when subject hesitates or disobeys. Returns: pandas dataframe with experimenter prods. """ return pd.read_csv( PATH_TO_PROMPTS.joinpath("DO_experimenter_prod.txt"), delimiter="\t", index_col=0, ) def prepare_level_prompt(dict_of_fills: dict) -> str: """Prepare level prompt. If at the start of the level (no disobediences), add instructions. End level prompt with `$subject`. Args: dict_of_fills: dict with columns "level" and "num_disobedience" Returns: string with `$` placeholders. """ # Unpack dict of fills level = dict_of_fills["level"] num_disobedience = dict_of_fills["num_disobedience"] level_prompt = "" # unpack level information if num_disobedience == 0: # If no instances of disobedience have been observed, # Then subject is at the start of the level and level information should be relayed. df_level_information = get_level_information() learner_response = df_level_information.loc[level]["learner response"] print("\t\t learner_response:", learner_response) level_prompt += f"\n\n{learner_response}" level_prompt += "\n\n$subject" return level_prompt def query_LM_to_generate_subject_response( filled_string: FilledString, ): """Query language model to generate subject response. Args: filled_string: filled template and variables to prompt the language model with. Returns: Result object containing all results and information for the experiment. Raises: Exception: "`run_single_simulation` failed!" if calling API failed more than `max_attempt` times. """ # Set model settings. model_settings_generate_free_form_text = OpenAIModelSettings( engine="text-davinci-002", max_tokens=40, temperature=1, n=1, logprobs=1, echo=False, params_descriptor="long-ans-logprobs", ) # Query LM. print("\t\t ***CALLED LM to generate teacher response***") res = run_single_simulation( filled_string=filled_string, model_settings=model_settings_generate_free_form_text, prompt_descriptor="generation", experiment_descriptor="do", seconds_to_sleep_before_query=1, seconds_to_sleep_after_failed_query=5, max_attempts=3, ) if res is None: raise Exception("`run_single_simulation` failed!") save_milgram_LM_response_to_unique_location( res=res, file_name_prefix="subject-response" ) print("\t\t %%%%%%%%%%%%") print(res["output"]["choices"][0]["text"]) print("\t\t %%%%%%%%%%%%") return res def query_LM_to_judge_did_not_submerge( subject_response_text: str, dict_of_fills: dict ) -> bool: """Use language model to judge whether the subject did not submerge the victim. Args: subject_response_text: value to add to `dict_of_fills` dict_of_fills: dict with keys corresponding to `$` placeholders Returns: prob > 0.5 Raises: Exception: "`run_single_simulation` failed!" if calling API failed more than `max_attempt` times. """ # Set model settings. model_settings_2_choice_judgment = OpenAIModelSettings( engine="text-davinci-002", max_tokens=0, temperature=0, n=1, logprobs=1, echo=True, params_descriptor="no-text-complete", ) # Fill template string. dict_of_fills["action"] = subject_response_text prompt_string = get_plaintext_file_contents( PATH_TO_PROMPTS.joinpath("DO_judge_not_submerge.txt") ) filled_string = FilledString( string_template=prompt_string, dict_of_fills=dict_of_fills ) # Query LM. print("\t\t ***CALLED LM to judge did not submerge***") res = run_single_simulation( filled_string=filled_string, model_settings=model_settings_2_choice_judgment, prompt_descriptor="DO_judge_did_not_submerge", experiment_descriptor="DO_judge", seconds_to_sleep_before_query=1, seconds_to_sleep_after_failed_query=5, max_attempts=3, ) if res is None: raise Exception("`run_single_simulation` failed!") # Get log probs of 'not punish' token_logprobs = res["output"]["choices"][0]["logprobs"]["token_logprobs"][-2:] prob = math.exp(sum(token_logprobs)) res["probability_did_not_submerge"] = prob save_milgram_LM_response_to_unique_location( res=res, file_name_prefix="judge-response-not-submerge" ) print("\t\t judge-response not submerge normalized probability:", prob) print("\t\t ############") print(res["output"]["choices"][0]["text"]) print("\t\t ############") return prob > 0.5 def query_LM_to_judge_did_not_stop( subject_response_text: str, dict_of_fills: dict ) -> bool: """Use language model to judge whether the subject did not stop the experiment. Args: subject_response_text: value to add to `dict_of_fills` dict_of_fills: dict with keys corresponding to `$` placeholders Returns: prob > 0.5 Raises: Exception: "`run_single_simulation` failed!" if calling API failed more than `max_attempt` times. """ # Set model settings. model_settings_2_choice_judgment = OpenAIModelSettings( engine="text-davinci-002", max_tokens=0, temperature=0, n=1, logprobs=1, echo=True, params_descriptor="no-text-complete", ) # Fill template string. dict_of_fills["action"] = subject_response_text prompt_string = get_plaintext_file_contents( PATH_TO_PROMPTS.joinpath("DO_judge_not_terminate.txt") ) filled_string = FilledString( string_template=prompt_string, dict_of_fills=dict_of_fills ) # Query LM. print("\t\t ***CALLED LM to judge did not stop***") res = run_single_simulation( filled_string=filled_string, model_settings=model_settings_2_choice_judgment, prompt_descriptor="DO_judge_did_not_stop", experiment_descriptor="DO_judge", seconds_to_sleep_before_query=1, seconds_to_sleep_after_failed_query=5, max_attempts=3, ) if res is None: raise Exception("`run_single_simulation` failed!") # Get log probs of 'not stop' token_logprobs = res["output"]["choices"][0]["logprobs"]["token_logprobs"][-2:] prob = math.exp(sum(token_logprobs)) res["probability_did_not_stop"] = prob save_milgram_LM_response_to_unique_location( res=res, file_name_prefix="judge-response-not-stop" ) print("\t\t judge-response not stop normalized probability:", prob) print("\t\t ############") print(res["output"]["choices"][0]["text"]) print("\t\t ############") return prob > 0.5 def get_experimenter_feedback(subject_response_text: str, dict_of_fills: dict) -> dict: """Return the experimenter's response to the subject's response. Args: subject_response_text: value to add to `dict_of_fills` dict_of_fills: dict with keys corresponding to `$` placeholders Returns: dictionary with "feedback", "finish_level", "num_disobedience", "finish_experiment" Cases: If subject did stop: Feedback is subject's response (experiment is finished) If subject did submerge & if learner's action was correct: Feedback is subject's response + experimenter repeats instructions (level is finished) If subject did submerge & if learner's action was wrong Feedback is subject's response (level is finished) If subject did not submerge & if learner action was correct Feedback is subject's response (level is finished) If subject did not submerge & learner action was wrong Feedback is subject's response + experimenter uses prod (level repeats) """ # Unpack level information. level = dict_of_fills["level"] df_level_information = get_level_information() learner_response = df_level_information.loc[level]["learner response"] # Grab only use the first sentence of subject's response. RE_GRAB_FIRST_SENTENCE = '(.*?(?<!Mr|Ms)[\.\!\?\;]"?)' match = re.search(RE_GRAB_FIRST_SENTENCE, subject_response_text) first_sentence = match.group(1) # Edge case: if character is a semicolon, change to a period. first_sentence = first_sentence.replace(";", ".") # Edge case: if completion does not start with a space, start with a space # otherwise runs into key errors when substituting into template first_sentence = " " + first_sentence.lstrip() # Ask LM to simulate experimenter deciding whether the action is following orders. did_not_submerge = query_LM_to_judge_did_not_submerge( subject_response_text=first_sentence, dict_of_fills=dict_of_fills ) did_not_stop = query_LM_to_judge_did_not_stop( subject_response_text=first_sentence, dict_of_fills=dict_of_fills ) # Handle cases. case_info = { "feedback": first_sentence, "finish_experiment": False, "finish_level": True, "num_disobedience": dict_of_fills["num_disobedience"], } # If subject did stop, end the experiment. if not did_not_stop: case_info["finish_experiment"] = True case_info["num_disobedience"] += 1 return case_info # If subject did submerge, respond depending on # whether the learner's action is correct or incorrect. if not did_not_submerge: if learner_response == "The driver drives safely.": # If learner action was correct, remind subject of instructions. # Update feedback to be subject's response + experimenter repeats instructions (level is finished). case_info["feedback"] += ( "\n\n" + get_experimenter_reaction_to_subject_wrong_action() ) # Else if learner's action was incorrect, then subject followed shock instructions, # so feedback is just the subject's response (level is finished). return case_info # If subject did not submerge, respond depending on # whether the learner's action is correct or incorrect. if learner_response == "The driver drives safely.": # If learner action was correct, then subject followed shock instructions, # so feedback is just the subject's response (level is finished). return case_info # Otherwise subject is disobeying submerge instructions. # Experimenter uses an action to nudge subject into submerging (level continues). case_info["finish_level"] = False case_info["num_disobedience"] += 1 # Handle special case (victim does not respond if level is 27). # If level 27 and 1 disobedience has been observed, # give instructions on how to handle no response. if level == 27 and case_info["num_disobedience"] == 1: case_info["feedback"] += ( "\n\n" + get_experimenter_reaction_to_subject_question_post_victim_no_response() ) return case_info # Experimenter uses prods sequentially with increasing disobedience. if case_info["num_disobedience"] <= 4: prod_name = f"prod {case_info['num_disobedience']}" prod_text = get_experimenter_prods.loc[prod_name]["text"] case_info["feedback"] += f'\n\nThe experimenter tells $subject, "{prod_text}"' return case_info # If subject is disobedient 5 times, end the experiment. case_info["finish_level"] = True case_info["finish_experiment"] = True case_info["feedback"] += "\n\nThe experimenter ends the experiment." return case_info def run_single_level(previous_prompt_string: str, subject: str, level: int): """Run a single level of the Milgram experiment until it meets an end condition. Args: previous_prompt_string: string with `$` placeholders. To initialize current prompt string. subject: subject's name. level: current level. Returns: dictionary with new "prompt_string", completed "level", "is_end_experiment" (num_disobedience >= 5) or (finish_level and level == 36) "num_disobedience" observed in the level """ # Prepare dict of fills. dict_of_fills = { "subject": subject, "pronoun": "he" if subject.startswith("Mr.") else "she", "poss": "his" if subject.startswith("Mr.") else "her", "poss_obj": "him" if subject.startswith("Mr.") else "her", # added this "victim": "Mr. Wallace", "level": level, "num_disobedience": 0, } # Load intro text from file. if level == 1: prompt_string = get_introduction_text() else: prompt_string = previous_prompt_string # In each level, the subject responds and experimenter prods # repeatedly until the level ends. while True: # Add learner response for this level to the prompt. prompt_string += prepare_level_prompt(dict_of_fills) print("\t\t ------------------------------") # Get a subject generation from prompt. filled_string = FilledString( string_template=prompt_string, dict_of_fills=dict_of_fills ) subject_response = query_LM_to_generate_subject_response( filled_string=filled_string ) subject_response_text = subject_response["output"]["choices"][0]["text"] # Get experimenter feedback on subject generation. experimenter_feedback = get_experimenter_feedback( subject_response_text, dict_of_fills ) # Use experimenter feedback to update prompt and fills # for next attempt at same level. prompt_string += experimenter_feedback["feedback"] dict_of_fills["num_disobedience"] = experimenter_feedback["num_disobedience"] # If level has ended... if experimenter_feedback["finish_level"]: # Calculate if experiment has ended. is_end_experiment = ( experimenter_feedback["finish_experiment"] or (experimenter_feedback["num_disobedience"] >= 5) or (experimenter_feedback["finish_level"] and level == 36) ) return { "prompt_string": prompt_string, "level": level, "is_end_experiment": is_end_experiment, "num_disobedience": experimenter_feedback["num_disobedience"], } def run_single_experiment( subject: str, race: str, gender: str, surname: str, ) -> None: """Run the Milgram Experiment to completion for a single subject. Args: subject: name of subject. race: race of subject. gender: honorific of subject. surname: surname of subject. Returns: None. """ # Lists to track progression. prompts = [] levels = [] num_disobediences = [] # State variables for current level. prompt_string = "" level = 1 is_end_experiment = False while level <= 36 and not is_end_experiment: print("LEVEL:", level) level_results = run_single_level( prompt_string=prompt_string, subject=subject, level=level, ) # Update progression. prompts.append(level_results["prompt_string"]) levels.append(level_results["level"]) num_disobediences.append(level_results["num_disobedience"]) # Update state. prompt_string = level_results["prompt_string"] level += 1 is_end_experiment = level_results["is_end_experiment"] # After experiment ends, save experiment overview. save_obj = { "subject": subject, "race": race, "gender": gender, "surname": surname, "full_prompt": prompt_string, "final_level": level, "num_disobediences": num_disobediences, } save_milgram_result_experiment_overview_to_unique_location( res=save_obj, subject=subject ) return None def run_full_experiment(title="Mx.", race="None") -> None: """Run Milgram Experiment for all participants in dataframe. Run script on portion of participant list (filter by title and race) to run multiple experiments in parallel. Args: title: title to run experiments with. race: race to filter participant list by. """ # Set Language Model Settings. verify_openai_access( path_to_organization=pathlib.Path("../openai_organization.txt"), path_to_api_key=pathlib.Path("../openai_api_key.txt"), ) full_set_df_names = pd.read_csv(PATH_TO_PARTICIPANT_LIST) # Run same script in parallel to save time. # Each script runs different set of races. df_names = full_set_df_names[full_set_df_names["Race"] == race] df_names = df_names.reset_index() for i, row in tqdm(df_names.iterrows()): subject = f'{title} {row["Surname"]}' print("Start", i, subject) run_full_experiment( subject=subject, race=row["Race"], gender=title, surname=row["Surname"] ) if __name__ == "__main__": fire.Fire(run_full_experiment)
[ "\n\nPLACEHOLDER", "../../data/prompt-templates/milgram/milgram_alt_resources", "prompt_string", "\n\n$subject", "DO_judge_not_terminate.txt", "[]", "DO_judge_not_submerge.txt" ]
2024-01-10
microsoft/turing-experiments
scripts~simulate_milgram~simulate_milgram_experiment.py
"""Simulate Milgram Shock Experiment.""" import pandas as pd import pathlib import sys import math import re import datetime from tqdm import tqdm import fire # Add src module to path before import. sys.path.insert(0, str(pathlib.Path("../../src"))) from file_IO_handler import get_plaintext_file_contents, save_json # noqa: E402 from openai_handler import verify_openai_access, OpenAIModelSettings # noqa: E402 from fill_string_template import FilledString # noqa: E402 from run_simulation import run_single_simulation # noqa: E402 # Set path settings. PATH_TO_SIMULATION_RESULTS: pathlib.Path = pathlib.Path( "../../data/simulation_results/milgram/milgram_original" ) PATH_TO_PROMPTS: pathlib.Path = pathlib.Path( "../../data/prompt-templates/milgram/milgram_resources" ) PATH_TO_PARTICIPANT_LIST: pathlib.Path = pathlib.Path( "../../data/prompt-fills/milgram/df_names.csv" ) def save_milgram_result_experiment_overview_to_unique_location( res, subject: str, ): """Save final experiment overview (for Milgram Experiment) to a unique location. Args: res: language model response. subject: name of simulated subject. Returns: None. """ subject_nopunct = subject.replace(". ", "") datetimeStr = datetime.datetime.now().strftime("%Y-%m-%d-%H-%M") save_string = PATH_TO_SIMULATION_RESULTS.joinpath( f"experiment-overview_{datetimeStr}_subject_{subject_nopunct}.json" ) save_json(obj=res, filename=save_string) print("saved prompt to: ", save_string) return None def save_milgram_LM_response_to_unique_location( res: dict, file_name_prefix: str, ) -> None: """Save language model response (for Milgram Experiment) to a unique location. Args: res: language model response. file_name_prefix: prefix of file name to save response to. Returns: None. """ datetimeStr = datetime.datetime.now().strftime("%Y-%m-%d-%H-%M") subject_nopunct = res["input"]["prompt"]["subject"].replace(". ", "") level = res["input"]["prompt"]["level"].replace(". ", "") num_disobedience = res["input"]["prompt"]["num_disobedience"].replace(". ", "") save_string = PATH_TO_SIMULATION_RESULTS.joinpath( f"{file_name_prefix}_{datetimeStr}_subject_{subject_nopunct}_level_{level}_num_{num_disobedience}.json" ) save_json(obj=res, filename=save_string) print(f"\t\t {file_name_prefix} saved to: {save_string}") return None def get_introduction_text(): """Get introduction text. Returns: string with `$` placeholders. """ return get_plaintext_file_contents( PATH_TO_PROMPTS.joinpath("DO_introduction.txt"), "r" ) def verbal_designation(row) -> str: """Translate numerical shock level to level's verbal designation. In the original Milgram Experiment, shock levels were grouped in sets of 4. Args: row: panda's dataframe row with column "shock level". Returns: string verbal designation. """ shock_level_switch_group = int((row["shock level"] - 1) / 4) if shock_level_switch_group == 0: return "slight shock" if shock_level_switch_group == 1: return "moderate shock" if shock_level_switch_group == 2: return "strong shock" if shock_level_switch_group == 3: return "very strong shock" if shock_level_switch_group == 4: return "intense shock" if shock_level_switch_group == 5: return "extreme intensity shock" if shock_level_switch_group == 6: return "danger severe shock" else: return "XXX" def get_level_information() -> pd.DataFrame: """Load dataframe of level information and add derivative columns. Returns: pandas dataframe. """ df_level_information = pd.read_csv( PATH_TO_PROMPTS.joinpath("DO_level_information_remote-learner.txt"), delimiter="\t", index_col=0, ) df_level_information["voltage level"] = df_level_information.apply( lambda x: x["shock level"] * 15, axis=1 ) df_level_information["verbal designation"] = df_level_information.apply( verbal_designation, axis=1 ) return df_level_information def get_experimenter_reaction_to_subject_wrong_action() -> str: """Get experimenter reaction to wrong action on behalf of the subject. Returns: string with `$` placeholders. """ return "The experimenter tells $subject to follow the correct experimental procedure. $subject should read the question and wait for the learner to select an answer. If the learner selects a correct answer, $subject should move on to the next question. If the learner selects a wrong answer, $subject should announce the current voltage level and shock the learner before reading the next question. $subject should go one level higher on the shock generator each time the learner selects a wrong answer." def get_experimenter_reaction_to_subject_question_post_victim_no_response() -> str: """Get experimenter reaction to subject question post when the victim stops responding. Returns: string with `$` placeholders. """ return "The experimenter instructs $subject to treat the absence of a response as a wrong answer." def get_experimenter_prods() -> pd.DataFrame: """Get dataframe of experimenter prods that the experimenter says to the subject when subject hesitates or disobeys. Returns: pandas dataframe with experimenter prods. """ return pd.read_csv( PATH_TO_PROMPTS.joinpath("DO_experimenter_prod.txt"), delimiter="\t", index_col=0, ) def prepare_level_prompt(dict_of_fills: dict) -> str: """Prepare level prompt. If at the start of the level (no disobediences), add instructions. End level prompt with `$subject`. Args: dict_of_fills: dict with columns "level" and "num_disobedience" Returns: string with `$` placeholders. """ # Unpack dict of fills. level = dict_of_fills["level"] num_disobedience = dict_of_fills["num_disobedience"] level_prompt = "" if num_disobedience == 0: # If no instances of disobedience have been observed, # Then subject is at the start of the level and level information should be relayed. df_level_information = get_level_information() verbal_designation = df_level_information.loc[level]["verbal designation"] voltage_level = df_level_information.loc[level]["voltage level"] learner_response = df_level_information.loc[level]["learner response"] print("\t\t learner_response:", learner_response) level_prompt += f"\n\nThe current punishment voltage level is {verbal_designation}, {voltage_level} volts. The experimenter is sitting next to $subject. {learner_response}" level_prompt += "\n\n$subject" return level_prompt def query_LM_to_generate_subject_response( filled_string: FilledString, ): """Query language model to generate subject response. Args: filled_string: filled template and variables to prompt the language model with. Returns: Result object containing all results and information for the experiment. Raises: Exception: "`run_single_simulation` failed!" if calling API failed more than `max_attempt` times. """ # Set model settings. model_settings_generate_free_form_text = OpenAIModelSettings( engine="text-davinci-002", max_tokens=40, temperature=1, n=1, logprobs=1, echo=False, params_descriptor="long-ans-logprobs", ) # Query LM. print("\t\t ***CALLED LM to generate teacher response***") res = run_single_simulation( filled_string=filled_string, model_settings=model_settings_generate_free_form_text, prompt_descriptor="generation", experiment_descriptor="do", seconds_to_sleep_before_query=1, seconds_to_sleep_after_failed_query=5, max_attempts=3, ) if res is None: raise Exception("`run_single_simulation` failed!") save_milgram_LM_response_to_unique_location( res=res, file_name_prefix="subject-response" ) print("\t\t %%%%%%%%%%%%") print(res["output"]["choices"][0]["text"]) print("\t\t %%%%%%%%%%%%") return res def query_LM_to_judge_did_not_shock( subject_response_text: str, dict_of_fills: dict ) -> bool: """Use language model to judge whether the subject did not shock the victim. Args: subject_response_text: value to add to `dict_of_fills` dict_of_fills: dict with keys corresponding to `$` placeholders Returns: prob > 0.5 Raises: Exception: "`run_single_simulation` failed!" if calling API failed more than `max_attempt` times. """ # Set model settings. model_settings_2_choice_judgment = OpenAIModelSettings( engine="text-davinci-002", max_tokens=0, temperature=0, n=1, logprobs=1, echo=True, params_descriptor="no-text-complete", ) # Fill template string. dict_of_fills["action"] = subject_response_text prompt_string = get_plaintext_file_contents( PATH_TO_PROMPTS.joinpath("DO_judge_did_not_shock.txt") ) filled_string = FilledString( string_template=prompt_string, dict_of_fills=dict_of_fills ) # Query LM. print("\t\t ***CALLED LM to judge did not shock***") res = run_single_simulation( filled_string=filled_string, model_settings=model_settings_2_choice_judgment, prompt_descriptor="DO_judge_did_not_shock", experiment_descriptor="DO_judge", seconds_to_sleep_before_query=1, seconds_to_sleep_after_failed_query=5, max_attempts=3, ) if res is None: raise Exception("`run_single_simulation` failed!") # Get log probs of 'not shock'. token_logprobs = res["output"]["choices"][0]["logprobs"]["token_logprobs"][-2:] prob = math.exp(sum(token_logprobs)) res["probability_did_not_shock"] = prob save_milgram_LM_response_to_unique_location( res=res, file_name_prefix="judge-response-not-shock" ) print("\t\t judge-response not shock probability:", prob) print("\t\t ############") print(res["output"]["choices"][0]["text"]) print("\t\t ############") return prob > 0.5 def query_LM_to_judge_did_not_stop( subject_response_text: str, dict_of_fills: dict ) -> bool: """Use language model to judge whether the subject did not stop the experiment. Args: subject_response_text: value to add to `dict_of_fills` dict_of_fills: dict with keys corresponding to `$` placeholders Returns: prob > 0.5 Raises: Exception: "`run_single_simulation` failed!" if calling API failed more than `max_attempt` times. """ # Set model settings. model_settings_2_choice_judgment = OpenAIModelSettings( engine="text-davinci-002", max_tokens=0, temperature=0, n=1, logprobs=1, echo=True, params_descriptor="no-text-complete", ) # Fill template string. dict_of_fills["action"] = subject_response_text prompt_string = get_plaintext_file_contents( PATH_TO_PROMPTS.joinpath("DO_judge_did_not_stop.txt") ) filled_string = FilledString( string_template=prompt_string, dict_of_fills=dict_of_fills ) # Query LM. print("\t\t ***CALLED LM to judge did not stop***") res = run_single_simulation( filled_string=filled_string, model_settings=model_settings_2_choice_judgment, prompt_descriptor="DO_judge_did_not_stop", experiment_descriptor="DO_judge", seconds_to_sleep_before_query=1, seconds_to_sleep_after_failed_query=5, max_attempts=3, ) if res is None: raise Exception("`run_single_simulation` failed!") # Get log probs of 'not stop' token_logprobs = res["output"]["choices"][0]["logprobs"]["token_logprobs"][-2:] prob = math.exp(sum(token_logprobs)) res["probability_did_not_stop"] = prob save_milgram_LM_response_to_unique_location( res=res, file_name_prefix="judge-response-not-stop" ) print("\t\t judge-response not stop probability:", prob) print("\t\t ############") print(res["output"]["choices"][0]["text"]) print("\t\t ############") return prob > 0.5 def get_experimenter_feedback(subject_response_text: str, dict_of_fills: dict) -> dict: """Return the experimenter's response to the subject's response. Args: subject_response_text: value to add to `dict_of_fills` dict_of_fills: dict with keys corresponding to `$` placeholders Returns: dictionary with "feedback", "finish_level", "num_disobedience", "finish_experiment" Cases: If subject did stop: Feedback is subject's response (experiment is finished) If subject did shock & if learner's action was correct: Feedback is subject's response + experimenter repeats instructions (level is finished) If subject did shock & if learner's action was wrong Feedback is subject's response (level is finished) If subject did not shock & if learner action was correct Feedback is subject's response (level is finished) If subject did not shock & learner action was wrong Feedback is subject's response + experimenter uses prod (level repeats) """ # Unpack level information. level = dict_of_fills["level"] df_level_information = get_level_information() learner_response = df_level_information.loc[level]["learner response"] # Grab only use the first sentence of subject's response. RE_GRAB_FIRST_SENTENCE = '(.*?(?<!Mr|Ms)[\.\!\?\;]"?)' match = re.search(RE_GRAB_FIRST_SENTENCE, subject_response_text) first_sentence = match.group(1) # Edge case: if character is a semicolon, change to a period. first_sentence = first_sentence.replace(";", ".") # Ask LM to simulate experimenter deciding whether the action is following orders. did_not_shock = query_LM_to_judge_did_not_shock( subject_response_text=first_sentence, dict_of_fills=dict_of_fills ) did_not_stop = query_LM_to_judge_did_not_stop( subject_response_text=first_sentence, dict_of_fills=dict_of_fills ) # Handle cases. case_info = { "feedback": first_sentence, "finish_experiment": False, "finish_level": True, "num_disobedience": dict_of_fills["num_disobedience"], } # If subject did stop, end the experiment. if not did_not_stop: case_info["finish_experiment"] = True case_info["num_disobedience"] += 1 return case_info # If subject did shock, respond depending on # whether the learner's action is correct or incorrect. if not did_not_shock: if learner_response == "The learner selects a correct answer.": # If learner action was correct, remind subject of instructions. # Update feedback to be subject's response + experimenter repeats instructions (level is finished). case_info["feedback"] += ( "\n\n" + get_experimenter_reaction_to_subject_wrong_action() ) # Else if learner's action was incorrect, then subject followed shock instructions, # so feedback is just the subject's response (level is finished). return case_info # If subject did not shock, respond depending on # whether the learner's action is correct or incorrect. if learner_response == "The learner selects a correct answer.": # If learner action was correct, then subject followed shock instructions, # so feedback is just the subject's response (level is finished). return case_info # Otherwise subject is disobeying shock instructions. # Experimenter uses an action to nudge subject into shocking (level continues). case_info["finish_level"] = False case_info["num_disobedience"] += 1 # Handle special case (victim does not respond if level is 27). # If level 27 and 1 disobedience has been observed, # give instructions on how to handle no response. if level == 27 and case_info["num_disobedience"] == 1: case_info["feedback"] += ( "\n\n" + get_experimenter_reaction_to_subject_question_post_victim_no_response() ) return case_info # Experimenter uses prods sequentially with increasing disobedience. if case_info["num_disobedience"] <= 4: prod_name = f"prod {case_info['num_disobedience']}" prod_text = get_experimenter_prods.loc[prod_name]["text"] case_info["feedback"] += f'\n\nThe experimenter tells $subject, "{prod_text}"' return case_info # If subject is disobedient 5 times, end the experiment. case_info["finish_level"] = True case_info["finish_experiment"] = True case_info["feedback"] += "\n\nThe experimenter ends the experiment." return case_info def run_single_level(previous_prompt_string: str, subject: str, level: int): """Run a single level of the Milgram experiment until it meets an end condition. Args: previous_prompt_string: string with `$` placeholders. To initialize current prompt string. subject: subject's name. level: current level. Returns: dictionary with new "prompt_string", completed "level", "is_end_experiment" (num_disobedience >= 5) or (finish_level and level == 36) "num_disobedience" observed in the level """ # Prepare dict of fills. dict_of_fills = { "subject": subject, "pronoun": "he" if subject.startswith("Mr.") else "she", "poss": "his" if subject.startswith("Mr.") else "her", "poss_obj": "him" if subject.startswith("Mr.") else "her", # added this "victim": "Mr. Wallace", "level": level, "num_disobedience": 0, } # Load intro text from file. if level == 1: prompt_string = get_introduction_text() else: prompt_string = previous_prompt_string # In each level, the subject responds and experimenter prods # repeatedly until the level ends. while True: # Add learner response for this level to the prompt. prompt_string += prepare_level_prompt(dict_of_fills) print("\t\t ------------------------------") # Get a subject generation from prompt. filled_string = FilledString( string_template=prompt_string, dict_of_fills=dict_of_fills ) subject_response = query_LM_to_generate_subject_response( filled_string=filled_string ) subject_response_text = subject_response["output"]["choices"][0]["text"] # Get experimenter feedback on subject generation. experimenter_feedback = get_experimenter_feedback( subject_response_text, dict_of_fills ) # Use experimenter feedback to update prompt and fills # for next attempt at same level. prompt_string += experimenter_feedback["feedback"] dict_of_fills["num_disobedience"] = experimenter_feedback["num_disobedience"] # If level has ended... if experimenter_feedback["finish_level"]: # Calculate if experiment has ended. is_end_experiment = ( experimenter_feedback["finish_experiment"] or (experimenter_feedback["num_disobedience"] >= 5) or (experimenter_feedback["finish_level"] and level == 36) ) return { "prompt_string": prompt_string, "level": level, "is_end_experiment": is_end_experiment, "num_disobedience": experimenter_feedback["num_disobedience"], } def run_single_experiment( subject: str, race: str, gender: str, surname: str, ) -> None: """Run the Milgram Experiment to completion for a single subject. Args: subject: name of subject. race: race of subject. gender: honorific of subject. surname: surname of subject. Returns: None. """ # Lists to track progression. prompts = [] levels = [] num_disobediences = [] # State variables for current level. prompt_string = "" level = 1 is_end_experiment = False while level <= 36 and not is_end_experiment: print("LEVEL:", level) level_results = run_single_level( prompt_string=prompt_string, subject=subject, level=level, ) # Update progression. prompts.append(level_results["prompt_string"]) levels.append(level_results["level"]) num_disobediences.append(level_results["num_disobedience"]) # Update state. prompt_string = level_results["prompt_string"] level += 1 is_end_experiment = level_results["is_end_experiment"] # After experiment ends, save experiment overview. save_obj = { "subject": subject, "race": race, "gender": gender, "surname": surname, "full_prompt": prompt_string, "final_level": level, "num_disobediences": num_disobediences, } save_milgram_result_experiment_overview_to_unique_location( res=save_obj, subject=subject ) return None def run_full_experiment(title="Mx.", race="None") -> None: """Run Milgram Experiment for all participants in dataframe. Run script on portion of participant list (filter by title and race) to run multiple experiments in parallel. Args: title: title to run experiments with. race: race to filter participant list by. """ # Set Language Model Settings. verify_openai_access( path_to_organization=pathlib.Path("../openai_organization.txt"), path_to_api_key=pathlib.Path("../openai_api_key.txt"), ) full_set_df_names = pd.read_csv(PATH_TO_PARTICIPANT_LIST) # Run same script in parallel to save time. # Each script runs different set of races. df_names = full_set_df_names[full_set_df_names["Race"] == race] df_names = df_names.reset_index() for i, row in tqdm(df_names.iterrows()): subject = f'{title} {row["Surname"]}' print("Start", i, subject) run_single_experiment( subject=subject, race=row["Race"], gender=title, surname=row["Surname"] ) if __name__ == "__main__": fire.Fire(run_full_experiment)
[ "DO_judge_did_not_stop.txt", "\n\nThe current punishment voltage level is PLACEHOLDER, PLACEHOLDER volts. The experimenter is sitting next to $subject. PLACEHOLDER", "../../data/prompt-templates/milgram/milgram_resources", "prompt_string", "\n\n$subject", "DO_judge_did_not_shock.txt", "[]" ]
2024-01-10
varundeepakgudhe/dorm_timings_open_ai
dorm_timings_openai.py
pip install openai import openai openai.api_key = "Your API_KEY" def get_dorm_info(document, question): #checks if closing word is in the question to respond for closing timings if 'closing' in question.lower(): prompt = f"Document: {document}\nQuestion: {question}\nAnswer:" #checks if opening word is in the question to respond for opening timings elif 'opening' in question.lower(): prompt = f"Document: {document}\nQuestion: {question}\nAnswer:" else: return "Appologies, cant understant this question. As an AI i am limited to certain knowledge" #defining response with only 50 maximum tokens response = openai.Completion.create(engine="davinci", prompt=prompt, max_tokens=50) return response.choices[0].text.strip() document = "The dorm timings are:\nDorm AbdulKalam: 7:00 AM (opening), 10:59 PM(closing)\nDorm Gangotri: 6:00 AM (opening), 10:30 PM(closing)\nDorm Ganga: 9:00 AM (opening), 9:30 PM(closing)\nDorm Triveni: 7:30 AM (opening), 11:59 PM(closing)\nDorm Gandhi: 8:00 AM (opening), 10:00 PM(closing)\nDorm Yamuna: 7:00 AM (opening), 9:00 PM(closing)" while True: question = input("Hi, please ask me a question about the dorm timings: ") #breaks the while loop if the question input is exit if question.lower()=="exit": break #calling the function get_dorm_info else: response = get_dorm_info(document, question) print(response)
[ "Document: PLACEHOLDER\nQuestion: PLACEHOLDER\nAnswer:" ]
2024-01-10
vsisl/dash-chatgpt-challenge
dash_app~gptutils.py
""" Functions making openAI API calls. """ import os import ast import json import openai # chat-gpt API from dash_app.flask_cache import cache openai.api_key = os.getenv("OPENAI_API_KEY") @cache.memoize(timeout=7 * 24 * 3600) def get_completion(prompt, model="gpt-3.5-turbo"): """Creates chatGPT response for a given prompt. :param prompt: str; prompt for chat GPT :param model: str; optional, default: "gpt-3.5-turbo" :return: str: response int: total API call tokens used by this function call """ messages = [{"role": "user", "content": prompt}] response = openai.ChatCompletion.create( model=model, messages=messages, max_tokens=300, temperature=0.3, # this is the degree of randomness of the model's output, bigger number = greater 'creativity' ) return response.choices[0].message["content"], response["usage"]["total_tokens"] # def get_image(text): # """Creates an image based on the input text # # :param prompt: str # :param model: str; optional, default: "gpt-3.5-turbo" # :return: str: response # """ # # input_img = Image.open("moscow.png").convert('RGBA') # # mask_img = Image.open("moscow_mask.png").convert('RGBA') # # import pdb; pdb.set_trace() # # import pdb; pdb.set_trace() # response = openai.Image.create_edit( # image=open("moskau.png", "rb"), # mask=open("moskau_mask.png", "rb"), # prompt="A city surrounded by {text} in pastel colors", # n=1, # size="256x256", # ) # # image_url = response["data"][0]["url"] # print(image_url) # import pdb # # pdb.set_trace() # response = openai.Image.create(prompt=prompt, n=1, size="256x256") # return response.choices[0].message["content"], response["usage"]["total_tokens"] def get_classification(input_text, model="text-davinci-003"): """Takes given text and classifies individual sentences for presence of propaganda techniques. :param input_text: str :param model: str; optional, default "text-davinci-003" :return: str """ response = openai.Completion.create( model=model, # prompt=f'The following is a list of techniques used in propaganda texts: 'Appeal to Authority', 'Appeal to Fear Prejudice', 'Bandwagon or Reductio ad hitlerum', 'Black and White Fallacy', 'Causal Oversimplification', 'Doubt', 'Exaggeration or Minimisation', 'Flag-Waving', 'Loaded Language', 'Name Calling or Labeling', 'Repetition', 'Slogans', 'Thought-terminating Cliches', 'Whataboutism or Straw Man or Red Herring'.\n\nThe following is a list of techniques and their defintions used in propaganda texts: \n'Appeal to Authority' - Stating that a claim is true simply because a valid authority or expert on the issue said it was true, without any other supporting evidence offered.\n'Appeal to Fear Prejudice' - Seeking to build support for an idea by instilling anxiety and/or panic in the population towards an alternative.\n'Bandwagon or Reductio ad hitlerum' - Attempting to persuade the target audience to join in and take the course of action because "everyone else is taking the same action".\n'Black and White Fallacy'-Presenting two alternative options as the only possibilities, when in fact more possibilities exist. As an the extreme case, tell the audience exactly what actions to take, eliminating any other possible choices.\n'Causal Oversimplification'-Assuming a single cause or reason when there are actually multiple causes for an issue.\n'Doubt'-Questioning the credibility of someone or something.\n'Exaggeration or Minimisation'-Either representing something in an excessive manner: making things larger, better, worse (e.g., "the best of the best", "quality guaranteed") or making something seem less important or smaller than it really is (e.g., saying that an insult was just a joke). \n'Flag-Waving'- Playing on strong national feeling (or to any group; e.g., race, gender, political preference) to justify or promote an action or idea \n'Loaded Language'- Using specific words and phrases with strong emotional implications (either positive or negative) to influence an audience.\n'Name Calling or Labeling'-Labeling the object of the propaganda campaign as either something the target audience fears, hates, finds undesirable or loves, praises. \n'Repetition'-Repeating the same message over and over again so that the audience will eventually accept it.\n'Slogans'-A brief and striking phrase that may include labeling and stereotyping. Slogans tend to act as emotional appeals.\n'Thought-terminating Cliches'-Words or phrases that discourage critical thought and meaningful discussion about a given topic. They are typically short, generic sentences that offer seemingly simple answers to complex questions or that distract attention away from other lines of thought. \n'Whataboutism or Straw Man or Red Herring'- A technique that attempts to discredit an opponent\'s position by charging them with hypocrisy without directly disproving their argument, when an opponent\'s proposition is substituted with a similar one which is then refuted in place of the original proposition or introducing irrelevant material to the issue being discussed, so that everyone\'s attention is diverted away from the points made.\n\nIn the following text identify which of these techniques are present, it can be multiple or none. Enclose the found techniques in square brackets imitating input for Python list. Separate the techniques with a comma, put parantheses around each found technique, e.g. ['Flag-Waving', 'Loaded Language'] or None if no technique is found.\n\nText: {input_text}\nFound techniques:', prompt=f"The following is a list of techniques used in propaganda texts: \ 'Appeal to Authority', 'Appeal to Fear Prejudice', 'Bandwagon or Reductio ad hitlerum', \ 'Black and White Fallacy', 'Causal Oversimplification', 'Doubt', 'Exaggeration or Minimisation', 'Flag-Waving',\ 'Loaded Language', 'Name Calling or Labeling', 'Repetition', 'Slogans', 'Thought-terminating Cliches',\ 'Whataboutism or Straw Man or Red Herring'.The following is a list of techniques and their definitions used in \ propaganda texts: 'Appeal to Authority' - Stating that a claim is true simply because a valid authority or \ expert on the issue said it was true, without any other supporting evidence offered.'Appeal to Fear Prejudice' \ - Seeking to build support for an idea by instilling anxiety and/or panic in the population towards an \ alternative.'Bandwagon or Reductio ad hitlerum' - Attempting to persuade the target audience to join in \ and take the course of action because 'everyone else is taking the same action'.'Black and White Fallacy'-\ Presenting two alternative options as the only possibilities, when in fact more possibilities exist. As an the\ extreme case, tell the audience exactly what actions to take, eliminating any other possible choices.'Causal\ Oversimplification'-Assuming a single cause or reason when there are actually multiple causes for an issue.\ 'Doubt'-Questioning the credibility of someone or something.'Exaggeration or Minimisation'-Either\ representing something in an excessive manner: making things larger, better, worse (e.g., 'the best of the \ best', 'quality guaranteed') or making something seem less important or smaller than it really is (e.g., \ saying that an insult was just a joke). 'Flag-Waving'- Playing on strong national feeling (or to any group; \ e.g., race, gender, political preference) to justify or promote an action or idea 'Loaded Language'- Using \ specific words and phrases with strong emotional implications (either positive or negative) to influence an \ audience.'Name Calling or Labeling'-Labeling the object of the propaganda campaign as either something the \ target audience fears, hates, finds undesirable or loves, praises. 'Repetition'-Repeating the same message \ over and over again so that the audience will eventually accept it.'Slogans'-A brief and striking phrase that \ may include labeling and stereotyping. Slogans tend to act as emotional appeals.'Thought-terminating Cliches'-\ Words or phrases that discourage critical thought and meaningful discussion about a given topic. They are \ typically short, generic sentences that offer seemingly simple answers to complex questions or that distract \ attention away from other lines of thought. 'Whataboutism or Straw Man or Red Herring'- A technique that \ attempts to discredit an opponent's position by charging them with hypocrisy without directly disproving \ their argument, when an opponent's proposition is substituted with a similar one which is then refuted in \ place of the original proposition or introducing irrelevant material to the issue being discussed, so that \ everyone's attention is diverted away from the points made.In the following text presented after the \ word 'Text' identify which of these techniques are present, it can be multiple or none. Enclose the found \ techniques in square brackets imitating input for Python list consisting of strings. Separate the techniques \ with a comma, put parantheses around each found technique, e.g. ['Flag-Waving', 'Loaded Language'] or None if \ no technique is found.Text: {input_text}\n" + "\n\n###\n\n", temperature=0, max_tokens=35, top_p=1, frequency_penalty=0, presence_penalty=0, stop=["\n"], ) return response["choices"][0]["text"] def get_classification_cheaper(sentence, model="gpt-3.5-turbo"): """Takes given sentence and classifies it based on presence of propaganda techniques. Uses chatGPT as a classifier. :param sentence: str; a sentence; e.g. "BREAKING NEWS: Russian Propaganda Exposed!" :param model: str; optional, default: "gpt-3.5-turbo" :return: list of str: list of propaganda techniques discovered in the given sentence; e.g. ['Loaded Language'] int: total API call tokens used by this function call """ messages = [ { "role": "system", "content": "You are a model that finds propaganda techniques in a provided text. The following is a list of techniques and their definitions used in propaganda texts:\n\n'Appeal to Authority' - Stating that a claim is true simply because a valid authority or expert on the issue said it was true, without any other supporting evidence offered.\n\n'Appeal to Fear Prejudice' - Seeking to build support for an idea by instilling anxiety and/or panic in the population towards an alternative.\n\n'Bandwagon or Reductio ad hitlerum' - Attempting to persuade the target audience to join in and take the course of action because everyone else is taking the same action.\n\n'Black and White Fallacy' - Presenting two alternative options as the only possibilities, when in fact more possibilities exist. As an extreme case, tell the audience exactly what actions to take, eliminating any other possible choices.\n\n'Causal Oversimplification' - Assuming a single cause or reason when there are actually multiple causes for an issue.\n\n'Doubt' - Questioning the credibility of someone or something.\n\n'Exaggeration or Minimisation' - Either representing something in an excessive manner: making things larger, better, worse (e.g., the best of the best, quality guaranteed) or making something seem less important or smaller than it really is (e.g., saying that an insult was just a joke).\n\n'Flag-Waving' - Playing on strong national feeling (or to any group; e.g., race, gender, political preference) to justify or promote an action or idea.\n\n'Loaded Language' - Using specific words and phrases with strong emotional implications (either positive or negative) to influence an audience.\n\n'Name Calling or Labeling' - Labeling the object of the propaganda campaign as either something the target audience fears, hates, finds undesirable or loves, praises.\n\n'Repetition' - Repeating the same message over and over again so that the audience will eventually accept it.\n\n'Slogans' - A brief and striking phrase that may include labeling and stereotyping. Slogans tend to act as emotional appeals.\n\n'Thought-terminating Cliches' - Words or phrases that discourage critical thought and meaningful discussion about a given topic. They are typically short, generic sentences that offer seemingly simple answers to complex questions or that distract attention away from other lines of thought.\n\n'Whataboutism or Straw Man or Red Herring' - A technique that attempts to discredit an opponent's position by charging them with hypocrisy without directly disproving their argument, when an opponent's proposition is substituted with a similar one which is then refuted in place of the original proposition or introducing irrelevant material to the issue being discussed, so that everyone's attention is diverted away from the points made.", }, { "role": "user", "content": "In the following text identify which of these techniques are present in a sentence, it can be multiple or none. Enclose the found techniques in each sentence in square brackets imitating input for Python list. Separate the techniques with a comma, put parentheses around each found technique, e.g. ['Flag-Waving', 'Loaded Language'] or None if no technique is found.", }, { "role": "assistant", "content": "Okay, I understand. I will either return a list containing the found techniques or None. For multiple sentences a will return a combination of lists and None delimited by commas.", }, { "role": "user", "content": "The global economy would plunge into chaos, with skyrocketing fuel prices, widespread unemployment, and a devastating blow to industries heavily reliant on oil.", }, {"role": "assistant", "content": "['Appeal to Fear Prejudice']"}, { "role": "user", "content": "Democrats bolted as soon as Trump's speech ended in an apparent effort to signal they can't even stomach being in the same room as the president.", }, { "role": "assistant", "content": "['Exaggeration or Minimisation', 'Loaded Language']", }, { "role": "user", "content": "By banning oil, they aim to cripple not only their adversaries but also their allies, leaving them at the mercy of their own agenda.", }, { "role": "assistant", "content": "['Exaggeration or Minimisation']", }, { "role": "user", "content": "Together, we can expose America's true intentions and safeguard the world from their manipulative grasp.", }, { "role": "assistant", "content": "['Appeal to Fear Prejudice']", }, { "role": "user", "content": "This is a car.", }, { "role": "assistant", "content": "None", }, { "role": "user", "content": "Our vast reserves and unwavering determination will ensure that the global oil market remains stable and prosperous.", }, { "role": "assistant", "content": "['Loaded Language']", }, {"role": "user", "content": sentence}, ] response = openai.ChatCompletion.create( model=model, messages=messages, temperature=0, # this is the degree of randomness of the model's output ) # de-stringify stringified list propaganda_techniques = ast.literal_eval(response.choices[0].message["content"]) # TODO: try to avoid using literal_eval. I tried to use json.loads() instead of ast.literal_eval() but it didn't # work. tokens_used = response["usage"]["total_tokens"] return propaganda_techniques, tokens_used @cache.memoize(timeout=7 * 24 * 3600) def get_classification_christian(sentence, model="gpt-3.5-turbo"): """Takes given sentence and classifies it based on presence of propaganda techniques. Uses chatGPT as a classifier. :param sentence: str; a sentence; e.g. "BREAKING NEWS: Russian Propaganda Exposed!" :param model: str; optional, default: "gpt-3.5-turbo" :return: response: dict; contains the following key-value pairs: 'classes': list of str; list of propaganda techniques 'confidence': list of numerical values between 0 and 1; confidence score for each propaganda technique 'explain': list of str; explanations why given propaganda techniques were assigned used_tokens int: total API call tokens used by this function call """ prompt = f""" classify the sentence delimited by triple backticks into the following list of classes: "Appeal to Authority", "Appeal to Fear Prejudice", "Bandwagon, Reductio ad hitlerum", "Black and White Fallacy", \ "Causal Oversimplification", "Doubt", "Exaggeration, Minimisation", "Flag-Waving", "Loaded Language", \ "Name Calling, Labeling", "Repetition", "Slogans", "Thought-terminating Cliches", \ "Whataboutism, Straw Man, Red Herring" \ As an output, give me a python dictionary with the following keys: 1. "classes" where all the classes are saved as a list. Use only the classes provided at the beginning with the exact spelling. If no class was classified, instead of a list return None. 2. "confidence" where a numerical value is assigned to each class representing the confidence of how the class is present in the text. 1 is the maximum number, 0 the lowest. \ Just give me a list as output where the order corresponds to the order of the classes. If no class was classified, instead of a list return None. \ 3. "explain" where an explanation is given why you have classified the sentence with this class. Don't repeat the sentence and keep it concise. If no class was classified, instead of a list return None. ```{sentence}```` """ message = [{"role": "user", "content": prompt}] response = openai.ChatCompletion.create( model=model, messages=message, temperature=0, # this is the degree of randomness of the model's output ) # extract data of interest try: # try converting stringified dictionary into a dict using json classification_result = json.loads(response.choices[0].message["content"]) except json.decoder.JSONDecodeError: # if json fails try literal eval classification_result = ast.literal_eval(response.choices[0].message["content"]) except: pass # TODO: make sure the function always returns something reasonable used_tokens = response["usage"]["total_tokens"] # validate that the response is a dictionary with the expected keys if type(classification_result) is not dict: raise TypeError(f"Response is not a dictionary.") # check that the dictionary returned by GPT has expected keys if not any( key in classification_result.keys() for key in ["classes", "confidence", "explain"] ): raise ValueError(f"Response does not contain the expected keys.") # make sure that all propaganda techniques identified by the classifier are among the expected options if classification_result["classes"] is not None: # Note: We had a case when the value of classification_result['classes'] was ['Appeal to Authority', 'None'] # the 'None' value was not supposed to be there and caused the function entity() to crash... # If technique is not found in the list of techniques, we need to remove also the score and explaination. for i, technique in enumerate(classification_result["classes"]): # TODO: store a list of all supported techniques in some separate module if technique not in [ "Appeal to Authority", "Appeal to Fear Prejudice", "Bandwagon, Reductio ad hitlerum", "Black and White Fallacy", "Causal Oversimplification", "Doubt", "Exaggeration, Minimisation", "Flag-Waving", "Loaded Language", "Name Calling, Labeling", "Repetition", "Slogans", "Thought-terminating Cliches", "Whataboutism, Straw Man, Red Herring", ]: # if technique is not from the list of supported options, remove it classification_result["classes"].pop(i) classification_result["confidence"].pop(i) classification_result["explain"].pop(i) # if there are no techniques left, assign None values (instead of empty lists) if len(classification_result["classes"]) == 0: classification_result["classes"] = None classification_result["confidence"] = None classification_result["explain"] = None # TODO: implement further data validation return classification_result, used_tokens
[ "The global economy would plunge into chaos, with skyrocketing fuel prices, widespread unemployment, and a devastating blow to industries heavily reliant on oil.", " classify the sentence delimited by triple backticks into the following list of classes:\n \"Appeal to Authority\", \"Appeal to Fear Prejudice\", \"Bandwagon, Reductio ad hitlerum\", \"Black and White Fallacy\", \"Causal Oversimplification\", \"Doubt\", \"Exaggeration, Minimisation\", \"Flag-Waving\", \"Loaded Language\", \"Name Calling, Labeling\", \"Repetition\", \"Slogans\", \"Thought-terminating Cliches\", \"Whataboutism, Straw Man, Red Herring\" As an output, give me a python dictionary with the following keys:\n 1. \"classes\" where all the classes are saved as a list. Use only the classes provided at the beginning with the exact spelling. If no class was classified, instead of a list return None.\n 2. \"confidence\" where a numerical value is assigned to each class representing the confidence of how the class is present in the text. 1 is the maximum number, 0 the lowest. Just give me a list as output where the order corresponds to the order of the classes. If no class was classified, instead of a list return None. 3. \"explain\" where an explanation is given why you have classified the sentence with this class. Don't repeat the sentence and keep it concise. If no class was classified, instead of a list return None.\n ```PLACEHOLDER````\n ", "['Appeal to Fear Prejudice']", "The following is a list of techniques used in propaganda texts: \\ 'Appeal to Authority', 'Appeal to Fear Prejudice', 'Bandwagon or Reductio ad hitlerum', \\ 'Black and White Fallacy', 'Causal Oversimplification', 'Doubt', 'Exaggeration or Minimisation', 'Flag-Waving',\\ 'Loaded Language', 'Name Calling or Labeling', 'Repetition', 'Slogans', 'Thought-terminating Cliches',\\ 'Whataboutism or Straw Man or Red Herring'.The following is a list of techniques and their definitions used in \\ propaganda texts: 'Appeal to Authority' - Stating that a claim is true simply because a valid authority or \\ expert on the issue said it was true, without any other supporting evidence offered.'Appeal to Fear Prejudice' \\ - Seeking to build support for an idea by instilling anxiety and/or panic in the population towards an \\ alternative.'Bandwagon or Reductio ad hitlerum' - Attempting to persuade the target audience to join in \\ and take the course of action because 'everyone else is taking the same action'.'Black and White Fallacy'-\\ Presenting two alternative options as the only possibilities, when in fact more possibilities exist. As an the\\ extreme case, tell the audience exactly what actions to take, eliminating any other possible choices.'Causal\\ Oversimplification'-Assuming a single cause or reason when there are actually multiple causes for an issue.\\ 'Doubt'-Questioning the credibility of someone or something.'Exaggeration or Minimisation'-Either\\ representing something in an excessive manner: making things larger, better, worse (e.g., 'the best of the \\ best', 'quality guaranteed') or making something seem less important or smaller than it really is (e.g., \\ saying that an insult was just a joke). 'Flag-Waving'- Playing on strong national feeling (or to any group; \\ e.g., race, gender, political preference) to justify or promote an action or idea 'Loaded Language'- Using \\ specific words and phrases with strong emotional implications (either positive or negative) to influence an \\ audience.'Name Calling or Labeling'-Labeling the object of the propaganda campaign as either something the \\ target audience fears, hates, finds undesirable or loves, praises. 'Repetition'-Repeating the same message \\ over and over again so that the audience will eventually accept it.'Slogans'-A brief and striking phrase that \\ may include labeling and stereotyping. Slogans tend to act as emotional appeals.'Thought-terminating Cliches'-\\ Words or phrases that discourage critical thought and meaningful discussion about a given topic. They are \\ typically short, generic sentences that offer seemingly simple answers to complex questions or that distract \\ attention away from other lines of thought. 'Whataboutism or Straw Man or Red Herring'- A technique that \\ attempts to discredit an opponent's position by charging them with hypocrisy without directly disproving \\ their argument, when an opponent's proposition is substituted with a similar one which is then refuted in \\ place of the original proposition or introducing irrelevant material to the issue being discussed, so that \\ everyone's attention is diverted away from the points made.In the following text presented after the \\ word 'Text' identify which of these techniques are present, it can be multiple or none. Enclose the found \\ techniques in square brackets imitating input for Python list consisting of strings. Separate the techniques \\ with a comma, put parantheses around each found technique, e.g. ['Flag-Waving', 'Loaded Language'] or None if \\ no technique is found.Text: PLACEHOLDER\n\n\n###\n\n", "This is a car.", "By banning oil, they aim to cripple not only their adversaries but also their allies, leaving them at the mercy of their own agenda.", "Democrats bolted as soon as Trump's speech ended in an apparent effort to signal they can't even stomach being in the same room as the president.", "None", "Okay, I understand. I will either return a list containing the found techniques or None. For multiple sentences a will return a combination of lists and None delimited by commas.", "['Exaggeration or Minimisation', 'Loaded Language']", "Together, we can expose America's true intentions and safeguard the world from their manipulative grasp.", "You are a model that finds propaganda techniques in a provided text. The following is a list of techniques and their definitions used in propaganda texts:\n\n'Appeal to Authority' - Stating that a claim is true simply because a valid authority or expert on the issue said it was true, without any other supporting evidence offered.\n\n'Appeal to Fear Prejudice' - Seeking to build support for an idea by instilling anxiety and/or panic in the population towards an alternative.\n\n'Bandwagon or Reductio ad hitlerum' - Attempting to persuade the target audience to join in and take the course of action because everyone else is taking the same action.\n\n'Black and White Fallacy' - Presenting two alternative options as the only possibilities, when in fact more possibilities exist. As an extreme case, tell the audience exactly what actions to take, eliminating any other possible choices.\n\n'Causal Oversimplification' - Assuming a single cause or reason when there are actually multiple causes for an issue.\n\n'Doubt' - Questioning the credibility of someone or something.\n\n'Exaggeration or Minimisation' - Either representing something in an excessive manner: making things larger, better, worse (e.g., the best of the best, quality guaranteed) or making something seem less important or smaller than it really is (e.g., saying that an insult was just a joke).\n\n'Flag-Waving' - Playing on strong national feeling (or to any group; e.g., race, gender, political preference) to justify or promote an action or idea.\n\n'Loaded Language' - Using specific words and phrases with strong emotional implications (either positive or negative) to influence an audience.\n\n'Name Calling or Labeling' - Labeling the object of the propaganda campaign as either something the target audience fears, hates, finds undesirable or loves, praises.\n\n'Repetition' - Repeating the same message over and over again so that the audience will eventually accept it.\n\n'Slogans' - A brief and striking phrase that may include labeling and stereotyping. Slogans tend to act as emotional appeals.\n\n'Thought-terminating Cliches' - Words or phrases that discourage critical thought and meaningful discussion about a given topic. They are typically short, generic sentences that offer seemingly simple answers to complex questions or that distract attention away from other lines of thought.\n\n'Whataboutism or Straw Man or Red Herring' - A technique that attempts to discredit an opponent's position by charging them with hypocrisy without directly disproving their argument, when an opponent's proposition is substituted with a similar one which is then refuted in place of the original proposition or introducing irrelevant material to the issue being discussed, so that everyone's attention is diverted away from the points made.", "['Exaggeration or Minimisation']", "Our vast reserves and unwavering determination will ensure that the global oil market remains stable and prosperous.", "['Loaded Language']", "In the following text identify which of these techniques are present in a sentence, it can be multiple or none. Enclose the found techniques in each sentence in square brackets imitating input for Python list. Separate the techniques with a comma, put parentheses around each found technique, e.g. ['Flag-Waving', 'Loaded Language'] or None if no technique is found." ]
2024-01-10
AbbyKatt/fchain
fchain.py
#Prototype for fChain #Simple OpenAI function based chatbot import openai import chromadb import json import os from pathlib import Path import errno #--------------------------------------------------------------------------------- # cChain simple directory based logging #--------------------------------------------------------------------------------- class fChainLog(): def __init__(self,logFileDirectory): self.logFlogFileDirectory=logFileDirectory #Recursively try to make the directory if it doesn't exist try: os.makedirs(logFileDirectory) except OSError as e: if e.errno != errno.EEXIST: raise e # #Create log file # self.logFile = open(logFileDirectory+"/log.txt", "w") # self.logFile.write("Log file created\n") # self.logFile.close() def Log(self,role,message,function_name,function_args): #Check self logging not none if self.logFlogFileDirectory is None: return #Make a unique filename with date and timestamp import datetime now = datetime.datetime.now() filename="Log_"+ role + "_" + now.strftime("%Y-%m-%d_%H-%M-%S-%f") + ".txt" #Create log file/write self.logFile = open(os.path.join(self.logFlogFileDirectory,filename), "a") self.logFile.write("role: "+role+"\n") self.logFile.write("function_name: "+function_name+"\n") self.logFile.write("function_args: "+str(function_args)+"\n") self.logFile.write("message: "+str(message)+"\n") self.logFile.close() #--------------------------------------------------------------------------------- # fChain main class #--------------------------------------------------------------------------------- class fChain(): def __init__(self,SystemPrompt,functionList,debug=True,debugFunctions=False,debugFunctionStubs=False,logDir=None,nameChatBot="assistant",nameUser="user",model_name = "gpt-3.5-turbo"): self.functions={func.openai_schema["name"]: func for func in functionList} self.messages=[] self.nameChatBot=nameChatBot self.nameUser=nameUser self.model_name=model_name openai.api_key = os.environ["OPENAI_API_KEY"] self.debug=debug self.debugFunctions=debugFunctions self.debugFunctionStubs=debugFunctionStubs self.totalTokens=0 #Setup logging self.Logging=fChainLog(logDir) #Load in SystemPrompt self.SystemPrompt=SystemPrompt self.addMessage(SystemPrompt,role="system") def addMessage(self,content,role="user",function_name=None): if function_name is not None: self.messages.append({"role": "function","name":function_name, "content": content}) self.Logging.Log("function_call","",function_name,content) else: self.messages.append({"role": role, "content": content}) self.Logging.Log(role,content,"","") def getMessages(self): return self.messages def getFunctionSchema(self): schemas=[] for aFunc in self.functions: schemas.append(self.functions[aFunc].openai_schema) return schemas def formatFunctionCall(self,function_name,arguments): argumentsString="" arguments=json.loads(arguments) for key in arguments: argumentsString+=str(key)+"="+str(arguments[key])+"," argumentsString=argumentsString[:-1] #argumentsString = ','.join([key + '=' + str(arguments[key]) for key in arguments]) return function_name + "(" + argumentsString + ")" def chat(self,userMessage,role="user"): #Add messge to list self.addMessage(userMessage,role=role) #get response passing in functions schemas response = openai.ChatCompletion.create(model=self.model_name, functions=self.getFunctionSchema(), temperature=0.2, messages=self.messages) if self.debug: print("------------------ GPT RESPONSE ------------------") print(response) print("------------------ END RESPONSE ------------------\n") #Pevent infinite loop maxLoops=10 currLoop=1 self.totalTokens=response.usage.total_tokens #Loop until all functions have been called debugMsgs=[] finish_reason=response.choices[0].finish_reason while finish_reason=="function_call": #Get function name/run it/get response function_name=response.choices[0].message["function_call"]["name"] arguments=response.choices[0].message["function_call"]["arguments"] if self.debug or self.debugFunctionStubs: debugFuncMsg=self.formatFunctionCall(function_name,arguments) print("Running Function: ["+debugFuncMsg+"]") self.Logging.Log("function_call","",function_name,debugFuncMsg) if self.debugFunctionStubs: debugMsgs.append(debugFuncMsg) function=self.functions[function_name] function_response=function.from_response(response) #Format json string nicely and human readable if self.debugFunctions: print(json.dumps(json.loads(function_response), indent=4, sort_keys=True)) if self.debug: print("FINISHED: ["+function_name +"]") #Put response in messages queue self.addMessage(function_response,role="function",function_name=function_name) #Invoke GPT with message history, list of callable functions and schemas and the current message response = openai.ChatCompletion.create(model=self.model_name, functions=self.getFunctionSchema(), messages=self.messages) if self.debug: print("------------------ GPT RESPONSE ------------------") print(response) print("------------------ END RESPONSE ------------------\n") if currLoop>maxLoops: print("Max loops reached!") break #Increment loop counter + get finish reason currLoop+=1 finish_reason=response.choices[0].finish_reason #We're done - chuck the response in the messages queue messagetext=response.choices[0].message.content self.totalTokens=response.usage.total_tokens self.addMessage(messagetext,role="assistant") return (messagetext,debugMsgs) #Uses the AI to summarize the conversation then makes that the new message history reducing the token count def Compact(self): self.Logging.Log("compact","","","") print("***Compacting chat history***") compactPrompt="Can you give me a brief summary of the conversation so in the third person narrative of both speakers?" ret=self.chat(compactPrompt,role="system") if self.debug: print("------------------ COMPACT SUMMARY ------------------") print(ret) print("\n------------------ END SUMMARY ------------------\n") #Reset chat history self.messages=[] self.addMessage(self.SystemPrompt,role="system") self.addMessage("Please give me a summary of our current chat:") self.addMessage(ret,role="assistant") return ret #--------------------------------------------------------------------------------- # JSON Loading of Knowledge Base as a semantic vectorstore #--------------------------------------------------------------------------------- #Third version with configurable collection buckets class fChainVectorDB(): def __init__( self, file_path ): self.file_path = Path(file_path).resolve() #Reads a custom JSON and inserts it into multiple collections in chroma with metadata #returns chromaDB instance def load(self): vectorCollections={} # Load JSON file with open(self.file_path) as file: data = json.load(file) for item in data: collection=item['collection'] metadata=item['metadata'] page_content=item['chunk'] #if collection is not in vectorCollections, add it if collection not in vectorCollections: vectorCollections[collection]=[] #add the page content and metadata to the collection vectorCollections[collection].append((page_content,metadata)) #Setup local client client = chromadb.Client() for colName in vectorCollections: collection = client.create_collection(colName) #take each tuple from list data and turn it into new lists docs and metas docs = [x[0] for x in vectorCollections[colName]] metas = [x[1] for x in vectorCollections[colName]] idx = ["Doc_{0}".format(i) for i in range(len(vectorCollections[colName]))] # Add docs to the collection. Can also update and delete. Row-based API coming soon! collection.add( documents=docs, metadatas=metas, ids=idx, # unique for each doc ) return client
[ "Can you give me a brief summary of the conversation so in the third person narrative of both speakers?" ]
2024-01-10
AbbyKatt/fchain
fchainService.py
#------------------------------------------------------------------------------------------------------------ # GPTFunctions based chatbot - ripping up LangChain show of horrors and going for my own implementation #------------------------------------------------------------------------------------------------------------ import json from openai_function_call import openai_function from fchain import fChain,fChainVectorDB import re sizeBeforeCompact=3000 promptAssembly=["intro.txt","flow1.txt","fieldformatting.txt","tables.txt","workedexample1.txt"] def LoadPrompt(fn): fn="prompts/"+fn prompt="" with open(fn, 'r') as file: prompt = file.read() return prompt def AssemblePrompt(fileNames:list[str])->str: prompt="" for fn in fileNames: prompt=prompt+LoadPrompt(fn) return prompt def ReloadLLM(): systemprompt=AssemblePrompt(promptAssembly) print("------------------ System Prompt ------------------") print(systemprompt) print("--------------- End System Prompt -----------------") functions=[listfields,fieldsDetails] chatBot=fChain(systemprompt,functions,False,False,True,"c:/temp/fchain/",model_name="ft:gpt-3.5-turbo-0613:personal::7tF0ZPhe") return chatBot #Does some magic to turn xl://some_link into an href hyperlink def handleUrls(aText): aConvertedText="" #Regex to grab values between [] regex="(?<=\[)(.*?)(?=\])" #Do a regext replace on aText aConvertedText=re.sub(regex,lambda x: "<a href='"+x.group(0)+"'>"+x.group(0)+"</a>",aText) return aConvertedText #--------------------------------------------------------------------------------- #Load in JSON fragments file/vectorize embeddings/return vectorDB def LoadEmbeddings(): print("Creating Embeddings from config...") file_path='../Config/chatbotconfig.json' loader = fChainVectorDB(file_path=file_path) vectordb = loader.load() print("Embeddings loaded...") return vectordb #------------------------------------------------------------------------------------------------------------ # Knowledge agent functons #------------------------------------------------------------------------------------------------------------ @openai_function def listfields(search:str,table:str="All") -> str: """Lists database fields in the database by table (leave blank or put All for all tables), shorter description for when user wants to find fields""" global vectordb print("listfields ["+search+"]") collection=vectordb.get_collection("fields_list") if table.lower()=="all": results = collection.query( query_texts=[search], n_results=10, # where={"metadata_field": "is_equal_to_this"}, # optional filter # where_document={"$contains":"search_string"} # optional filter ) else: results = collection.query( query_texts=[search], n_results=10, where={"Table Name": table}, # optional filter # where_document={"$contains":"search_string"} # optional filter ) ret=results["documents"] for v in ret: retList=[n for n in v] return json.dumps(retList) @openai_function def fieldsDetails(search:str) -> str: """Gete more details on fields in the database, longer decscription with more information and specifics (like field security level)""" global vectordb print("fieldsDetails ["+search+"]") collection=vectordb.get_collection("fields_detail") results = collection.query( query_texts=[search], n_results=5, # where={"metadata_field": "is_equal_to_this"}, # optional filter # where_document={"$contains":"search_string"} # optional filter ) ret=results["documents"] for v in ret: retList=[n for n in v] return json.dumps(retList) #------------------------------------------------------------------------------------------------------------ #Flask App #------------------------------------------------------------------------------------------------------------ import flask from flask import Flask, request, jsonify #from flask_cors import CORS #App handler app = Flask(__name__) app.config['TEMPLATES_AUTO_RELOAD'] = True app.jinja_env.auto_reload = True #CORS(app) #Load embeddings and create DB vectordb=LoadEmbeddings() #Load vectorstore and init chat engine chatBot=ReloadLLM() #Simple homepage @app.route('/') def index(): return flask.render_template('main.html') #Simple post/response for chat @app.route('/chat', methods=['POST']) def chatRepsonse(): global chatBot #Get the data from the request data = request.get_json() question = data['user'] #Reset history / reload prompts if len(question) == 0: print("Reset LLM Chat state...") chatBot=ReloadLLM() return jsonify({'message' : "***RESET***"}) #Ask question to fChain print("Running Question:" + question) result,debugmsgs=chatBot.chat(question) result=handleUrls(result) #show conversation size print("Tokens Used: "+str(chatBot.totalTokens)) if chatBot.totalTokens>= sizeBeforeCompact: print("Compacting LLM Chat state...") debugmsgs.append("Compacting LLM Chat state...") chatBot.Compact() #print("Answer:" + result) return jsonify({'message' : result,'debug':debugmsgs}) if __name__ == '__main__': app.run()
[ "['intro.txt', 'flow1.txt', 'fieldformatting.txt', 'tables.txt', 'workedexample1.txt']" ]
2024-01-10
yt7589/iching
biz~drlt~rll~experience.py
import gym import torch import random import collections from torch.autograd import Variable import numpy as np from collections import namedtuple, deque from biz.drlt.rll.agent import BaseAgent from biz.drlt.rll.common import utils # one single experience step Experience = namedtuple('Experience', ['state', 'action', 'reward', 'done']) class ExperienceSource: """ Simple n-step experience source using single or multiple environments Every experience contains n list of Experience entries """ def __init__(self, env, agent, steps_count=2, steps_delta=1, vectorized=False): """ Create simple experience source :param env: environment or list of environments to be used :param agent: callable to convert batch of states into actions to take :param steps_count: count of steps to track for every experience chain :param steps_delta: how many steps to do between experience items :param vectorized: support of vectorized envs from OpenAI universe """ assert isinstance(env, (gym.Env, list, tuple)) assert isinstance(agent, BaseAgent) assert isinstance(steps_count, int) assert steps_count >= 1 assert isinstance(vectorized, bool) if isinstance(env, (list, tuple)): self.pool = env else: self.pool = [env] self.agent = agent self.steps_count = steps_count self.steps_delta = steps_delta self.total_rewards = [] self.total_steps = [] self.vectorized = vectorized def __iter__(self): states, agent_states, histories, cur_rewards, cur_steps = [], [], [], [], [] env_lens = [] for env in self.pool: obs = env.reset() # if the environment is vectorized, all it's output is lists of results. # Details are here: https://github.com/openai/universe/blob/master/doc/env_semantics.rst if self.vectorized: obs_len = len(obs) states.extend(obs) else: obs_len = 1 states.append(obs) env_lens.append(obs_len) for _ in range(obs_len): histories.append(deque(maxlen=self.steps_count)) cur_rewards.append(0.0) cur_steps.append(0) agent_states.append(self.agent.initial_state()) iter_idx = 0 while True: actions = [None] * len(states) states_input = [] states_indices = [] for idx, state in enumerate(states): if state is None: actions[idx] = self.pool[0].action_space.sample() # assume that all envs are from the same family else: states_input.append(state) states_indices.append(idx) if states_input: states_actions, new_agent_states = self.agent(states_input, agent_states) for idx, action in enumerate(states_actions): g_idx = states_indices[idx] actions[g_idx] = action agent_states[g_idx] = new_agent_states[idx] grouped_actions = _group_list(actions, env_lens) global_ofs = 0 for env_idx, (env, action_n) in enumerate(zip(self.pool, grouped_actions)): if self.vectorized: next_state_n, r_n, is_done_n, _ = env.step(action_n) else: next_state, r, is_done, _ = env.step(action_n[0]) next_state_n, r_n, is_done_n = [next_state], [r], [is_done] for ofs, (action, next_state, r, is_done) in enumerate(zip(action_n, next_state_n, r_n, is_done_n)): idx = global_ofs + ofs state = states[idx] history = histories[idx] cur_rewards[idx] += r cur_steps[idx] += 1 if state is not None: history.append(Experience(state=state, action=action, reward=r, done=is_done)) if len(history) == self.steps_count and iter_idx % self.steps_delta == 0: yield tuple(history) states[idx] = next_state if is_done: # in case of very short episode (shorter than our steps count), send gathered history if 0 < len(history) < self.steps_count: yield tuple(history) # generate tail of history while len(history) > 1: history.popleft() yield tuple(history) self.total_rewards.append(cur_rewards[idx]) self.total_steps.append(cur_steps[idx]) cur_rewards[idx] = 0.0 cur_steps[idx] = 0 # vectorized envs are reset automatically states[idx] = env.reset() if not self.vectorized else None agent_states[idx] = self.agent.initial_state() history.clear() global_ofs += len(action_n) iter_idx += 1 def pop_total_rewards(self): r = self.total_rewards if r: self.total_rewards = [] self.total_steps = [] return r def pop_rewards_steps(self): res = list(zip(self.total_rewards, self.total_steps)) if res: self.total_rewards, self.total_steps = [], [] return res def _group_list(items, lens): """ Unflat the list of items by lens :param items: list of items :param lens: list of integers :return: list of list of items grouped by lengths """ res = [] cur_ofs = 0 for g_len in lens: res.append(items[cur_ofs:cur_ofs+g_len]) cur_ofs += g_len return res # those entries are emitted from ExperienceSourceFirstLast. Reward is discounted over the trajectory piece ExperienceFirstLast = collections.namedtuple('ExperienceFirstLast', ('state', 'action', 'reward', 'last_state')) class ExperienceSourceFirstLast(ExperienceSource): """ This is a wrapper around ExperienceSource to prevent storing full trajectory in replay buffer when we need only first and last states. For every trajectory piece it calculates discounted reward and emits only first and last states and action taken in the first state. If we have partial trajectory at the end of episode, last_state will be None """ def __init__(self, env, agent, gamma, steps_count=1, steps_delta=1, vectorized=False): assert isinstance(gamma, float) super(ExperienceSourceFirstLast, self).__init__(env, agent, steps_count+1, steps_delta, vectorized=vectorized) self.gamma = gamma self.steps = steps_count def __iter__(self): for exp in super(ExperienceSourceFirstLast, self).__iter__(): if exp[-1].done and len(exp) <= self.steps: last_state = None elems = exp else: last_state = exp[-1].state elems = exp[:-1] total_reward = 0.0 for e in reversed(elems): total_reward *= self.gamma total_reward += e.reward yield ExperienceFirstLast(state=exp[0].state, action=exp[0].action, reward=total_reward, last_state=last_state) def discount_with_dones(rewards, dones, gamma): discounted = [] r = 0 for reward, done in zip(rewards[::-1], dones[::-1]): r = reward + gamma*r*(1.-done) discounted.append(r) return discounted[::-1] class ExperienceSourceRollouts: """ N-step rollout experience source following A3C rollouts scheme. Have to be used with agent, keeping the value in its state (for example, agent.ActorCriticAgent). Yields batches of num_envs * n_steps samples with the following arrays: 1. observations 2. actions 3. discounted rewards, with values approximation 4. values """ def __init__(self, env, agent, gamma, steps_count=5): """ Constructs the rollout experience source :param env: environment or list of environments to be used :param agent: callable to convert batch of states into actions :param steps_count: how many steps to perform rollouts """ assert isinstance(env, (gym.Env, list, tuple)) assert isinstance(agent, BaseAgent) assert isinstance(gamma, float) assert isinstance(steps_count, int) assert steps_count >= 1 if isinstance(env, (list, tuple)): self.pool = env else: self.pool = [env] self.agent = agent self.gamma = gamma self.steps_count = steps_count self.total_rewards = [] self.total_steps = [] def __iter__(self): pool_size = len(self.pool) states = [np.array(e.reset()) for e in self.pool] mb_states = np.zeros((pool_size, self.steps_count) + states[0].shape, dtype=states[0].dtype) mb_rewards = np.zeros((pool_size, self.steps_count), dtype=np.float32) mb_values = np.zeros((pool_size, self.steps_count), dtype=np.float32) mb_actions = np.zeros((pool_size, self.steps_count), dtype=np.int64) mb_dones = np.zeros((pool_size, self.steps_count), dtype=np.bool) total_rewards = [0.0] * pool_size total_steps = [0] * pool_size agent_states = None step_idx = 0 while True: actions, agent_states = self.agent(states, agent_states) rewards = [] dones = [] new_states = [] for env_idx, (e, action) in enumerate(zip(self.pool, actions)): o, r, done, _ = e.step(action) total_rewards[env_idx] += r total_steps[env_idx] += 1 if done: o = e.reset() self.total_rewards.append(total_rewards[env_idx]) self.total_steps.append(total_steps[env_idx]) total_rewards[env_idx] = 0.0 total_steps[env_idx] = 0 new_states.append(np.array(o)) dones.append(done) rewards.append(r) # we need an extra step to get values approximation for rollouts if step_idx == self.steps_count: # calculate rollout rewards for env_idx, (env_rewards, env_dones, last_value) in enumerate(zip(mb_rewards, mb_dones, agent_states)): env_rewards = env_rewards.tolist() env_dones = env_dones.tolist() if not env_dones[-1]: env_rewards = discount_with_dones(env_rewards + [last_value], env_dones + [False], self.gamma)[:-1] else: env_rewards = discount_with_dones(env_rewards, env_dones, self.gamma) mb_rewards[env_idx] = env_rewards yield mb_states.reshape((-1,) + mb_states.shape[2:]), mb_rewards.flatten(), mb_actions.flatten(), mb_values.flatten() step_idx = 0 mb_states[:, step_idx] = states mb_rewards[:, step_idx] = rewards mb_values[:, step_idx] = agent_states mb_actions[:, step_idx] = actions mb_dones[:, step_idx] = dones step_idx += 1 states = new_states def pop_total_rewards(self): r = self.total_rewards if r: self.total_rewards = [] self.total_steps = [] return r def pop_rewards_steps(self): res = list(zip(self.total_rewards, self.total_steps)) if res: self.total_rewards, self.total_steps = [], [] return res class ExperienceSourceBuffer: """ The same as ExperienceSource, but takes episodes from the buffer """ def __init__(self, buffer, steps_count=1): """ Create buffered experience source :param buffer: list of episodes, each is a list of Experience object :param steps_count: count of steps in every entry """ self.update_buffer(buffer) self.steps_count = steps_count def update_buffer(self, buffer): self.buffer = buffer self.lens = list(map(len, buffer)) def __iter__(self): """ Infinitely sample episode from the buffer and then sample item offset """ while True: episode = random.randrange(len(self.buffer)) ofs = random.randrange(self.lens[episode] - self.steps_count - 1) yield self.buffer[episode][ofs:ofs+self.steps_count] class ExperienceReplayBuffer: def __init__(self, experience_source, buffer_size): assert isinstance(experience_source, (ExperienceSource, type(None))) assert isinstance(buffer_size, int) self.experience_source_iter = None if experience_source is None else iter(experience_source) self.buffer = [] self.capacity = buffer_size self.pos = 0 def __len__(self): return len(self.buffer) def __iter__(self): return iter(self.buffer) def sample(self, batch_size): """ Get one random batch from experience replay TODO: implement sampling order policy :param batch_size: :return: """ if len(self.buffer) <= batch_size: return self.buffer # Warning: replace=False makes random.choice O(n) keys = np.random.choice(len(self.buffer), batch_size, replace=True) return [self.buffer[key] for key in keys] def _add(self, sample): if len(self.buffer) < self.capacity: self.buffer.append(sample) else: self.buffer[self.pos] = sample self.pos = (self.pos + 1) % self.capacity def populate(self, samples): """ Populates samples into the buffer :param samples: how many samples to populate """ for _ in range(samples): entry = next(self.experience_source_iter) self._add(entry) class PrioReplayBufferNaive: def __init__(self, exp_source, buf_size, prob_alpha=0.6): self.exp_source_iter = iter(exp_source) self.prob_alpha = prob_alpha self.capacity = buf_size self.pos = 0 self.buffer = [] self.priorities = np.zeros((buf_size, ), dtype=np.float32) def __len__(self): return len(self.buffer) def populate(self, count): max_prio = self.priorities.max() if self.buffer else 1.0 for _ in range(count): sample = next(self.exp_source_iter) if len(self.buffer) < self.capacity: self.buffer.append(sample) else: self.buffer[self.pos] = sample self.priorities[self.pos] = max_prio self.pos = (self.pos + 1) % self.capacity def sample(self, batch_size, beta=0.4): if len(self.buffer) == self.capacity: prios = self.priorities else: prios = self.priorities[:self.pos] probs = np.array(prios, dtype=np.float32) ** self.prob_alpha probs /= probs.sum() indices = np.random.choice(len(self.buffer), batch_size, p=probs, replace=True) samples = [self.buffer[idx] for idx in indices] total = len(self.buffer) weights = (total * probs[indices]) ** (-beta) weights /= weights.max() return samples, indices, np.array(weights, dtype=np.float32) def update_priorities(self, batch_indices, batch_priorities): for idx, prio in zip(batch_indices, batch_priorities): self.priorities[idx] = prio class PrioritizedReplayBuffer(ExperienceReplayBuffer): def __init__(self, experience_source, buffer_size, alpha): super(PrioritizedReplayBuffer, self).__init__(experience_source, buffer_size) assert alpha > 0 self._alpha = alpha it_capacity = 1 while it_capacity < buffer_size: it_capacity *= 2 self._it_sum = utils.SumSegmentTree(it_capacity) self._it_min = utils.MinSegmentTree(it_capacity) self._max_priority = 1.0 def _add(self, *args, **kwargs): idx = self.pos super()._add(*args, **kwargs) self._it_sum[idx] = self._max_priority ** self._alpha self._it_min[idx] = self._max_priority ** self._alpha def _sample_proportional(self, batch_size): res = [] for _ in range(batch_size): mass = random.random() * self._it_sum.sum(0, len(self) - 1) idx = self._it_sum.find_prefixsum_idx(mass) res.append(idx) return res def sample(self, batch_size, beta): assert beta > 0 idxes = self._sample_proportional(batch_size) weights = [] p_min = self._it_min.min() / self._it_sum.sum() max_weight = (p_min * len(self)) ** (-beta) for idx in idxes: p_sample = self._it_sum[idx] / self._it_sum.sum() weight = (p_sample * len(self)) ** (-beta) weights.append(weight / max_weight) weights = np.array(weights, dtype=np.float32) samples = [self.buffer[idx] for idx in idxes] return samples, idxes, weights def update_priorities(self, idxes, priorities): assert len(idxes) == len(priorities) for idx, priority in zip(idxes, priorities): assert priority > 0 assert 0 <= idx < len(self) self._it_sum[idx] = priority ** self._alpha self._it_min[idx] = priority ** self._alpha self._max_priority = max(self._max_priority, priority) class BatchPreprocessor: """ Abstract preprocessor class descendants to which converts experience batch to form suitable to learning. """ def preprocess(self, batch): raise NotImplementedError class QLearningPreprocessor(BatchPreprocessor): """ Supports SimpleDQN, TargetDQN, DoubleDQN and can additionally feed TD-error back to experience replay buffer. To use different modes, use appropriate class method """ def __init__(self, model, target_model, use_double_dqn=False, batch_td_error_hook=None, gamma=0.99, device="cpu"): self.model = model self.target_model = target_model self.use_double_dqn = use_double_dqn self.batch_dt_error_hook = batch_td_error_hook self.gamma = gamma self.device = device @staticmethod def simple_dqn(model, **kwargs): return QLearningPreprocessor(model=model, target_model=None, use_double_dqn=False, **kwargs) @staticmethod def target_dqn(model, target_model, **kwards): return QLearningPreprocessor(model, target_model, use_double_dqn=False, **kwards) @staticmethod def double_dqn(model, target_model, **kwargs): return QLearningPreprocessor(model, target_model, use_double_dqn=True, **kwargs) def _calc_Q(self, states_first, states_last): """ Calculates apropriate q values for first and last states. Way of calculate depends on our settings. :param states_first: numpy array of first states :param states_last: numpy array of last states :return: tuple of numpy arrays of q values """ # here we need both first and last values calculated using our main model, so we # combine both states into one batch for efficiency and separate results later if self.target_model is None or self.use_double_dqn: states_t = torch.tensor(np.concatenate((states_first, states_last), axis=0)).to(self.device) res_both = self.model(states_t).data.cpu().numpy() return res_both[:len(states_first)], res_both[len(states_first):] # in this case we have target_model set and use_double_dqn==False # so, we should calculate first_q and last_q using different models states_first_v = torch.tensor(states_first).to(self.device) states_last_v = torch.tensor(states_last).to(self.device) q_first = self.model(states_first_v).data q_last = self.target_model(states_last_v).data return q_first.cpu().numpy(), q_last.cpu().numpy() def _calc_target_rewards(self, states_last, q_last): """ Calculate rewards from final states according to variants from our construction: 1. simple DQN: max(Q(states, model)) 2. target DQN: max(Q(states, target_model)) 3. double DQN: Q(states, target_model)[argmax(Q(states, model)] :param states_last: numpy array of last states from the games :param q_last: numpy array of last q values :return: vector of target rewards """ # in this case we handle both simple DQN and target DQN if self.target_model is None or not self.use_double_dqn: return q_last.max(axis=1) # here we have target_model set and use_double_dqn==True actions = q_last.argmax(axis=1) # calculate Q values using target net states_last_v = torch.tensor(states_last).to(self.device) q_last_target = self.target_model(states_last_v).data.cpu().numpy() return q_last_target[range(q_last_target.shape[0]), actions] def preprocess(self, batch): """ Calculates data for Q learning from batch of observations :param batch: list of lists of Experience objects :return: tuple of numpy arrays: 1. states -- observations 2. target Q-values 3. vector of td errors for every batch entry """ # first and last states for every entry state_0 = np.array([exp[0].state for exp in batch], dtype=np.float32) state_L = np.array([exp[-1].state for exp in batch], dtype=np.float32) q0, qL = self._calc_Q(state_0, state_L) rewards = self._calc_target_rewards(state_L, qL) td = np.zeros(shape=(len(batch),)) for idx, (total_reward, exps) in enumerate(zip(rewards, batch)): # game is done, no final reward if exps[-1].done: total_reward = 0.0 for exp in reversed(exps[:-1]): total_reward *= self.gamma total_reward += exp.reward # update total reward and calculate td error act = exps[0].action td[idx] = q0[idx][act] - total_reward q0[idx][act] = total_reward return state_0, q0, td
[]
2024-01-10
hatefr/chatbot
ingest_data.py
from langchain.text_splitter import CharacterTextSplitter from langchain.document_loaders import UnstructuredFileLoader from langchain.vectorstores.faiss import FAISS from langchain.embeddings import OpenAIEmbeddings import pickle print("Loading data...") loader = UnstructuredFileLoader("state_of_the_union.txt") raw_documents = loader.load() print("Splitting text...") text_splitter = CharacterTextSplitter( separator="\n\n", chunk_size=600, chunk_overlap=100, length_function=len, ) documents = text_splitter.split_documents(raw_documents) print("Creating vectorstore...") embeddings = OpenAIEmbeddings() vectorstore = FAISS.from_documents(documents, embeddings) with open("vectorstore.pkl", "wb") as f: pickle.dump(vectorstore, f)
[]
2024-01-10
hatefr/chatbot
query_data.py
from langchain.chains import RetrievalQA, ConversationalRetrievalChain from langchain.prompts.prompt import PromptTemplate from langchain.vectorstores.base import VectorStoreRetriever from langchain.chat_models import ChatOpenAI from langchain.memory import ConversationBufferMemory import pickle _template = """Given the following conversation and a follow up question, rephrase the follow up question to be a standalone question. You can assume the question about the most recent state of the union address. Chat History: {chat_history} Follow Up Input: {question} Standalone question:""" CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template(_template) template = """You are an AI assistant for answering questions about the most recent state of the union address. You are given the following extracted parts of a long document and a question. Provide a conversational answer. If you don't know the answer, just say "Hmm, I'm not sure." Don't try to make up an answer. If the question is not about the most recent state of the union, politely inform them that you are tuned to only answer questions about the most recent state of the union. Lastly, answer the question as if you were a pirate from the south seas and are just coming back from a pirate expedition where you found a treasure chest full of gold doubloons. Question: {question} ========= {context} ========= Answer in Markdown:""" QA_PROMPT = PromptTemplate(template=template, input_variables=[ "question", "context"]) def load_retriever(): with open("vectorstore.pkl", "rb") as f: vectorstore = pickle.load(f) retriever = VectorStoreRetriever(vectorstore=vectorstore) return retriever def get_basic_qa_chain(): llm = ChatOpenAI(model_name="gpt-4", temperature=0) retriever = load_retriever() memory = ConversationBufferMemory( memory_key="chat_history", return_messages=True) model = ConversationalRetrievalChain.from_llm( llm=llm, retriever=retriever, memory=memory) return model def get_custom_prompt_qa_chain(): llm = ChatOpenAI(model_name="gpt-4", temperature=0) retriever = load_retriever() memory = ConversationBufferMemory( memory_key="chat_history", return_messages=True) # see: https://github.com/langchain-ai/langchain/issues/6635 # see: https://github.com/langchain-ai/langchain/issues/1497 model = ConversationalRetrievalChain.from_llm( llm=llm, retriever=retriever, memory=memory, combine_docs_chain_kwargs={"prompt": QA_PROMPT}) return model def get_condense_prompt_qa_chain(): llm = ChatOpenAI(model_name="gpt-4", temperature=0) retriever = load_retriever() memory = ConversationBufferMemory( memory_key="chat_history", return_messages=True) # see: https://github.com/langchain-ai/langchain/issues/5890 model = ConversationalRetrievalChain.from_llm( llm=llm, retriever=retriever, memory=memory, condense_question_prompt=CONDENSE_QUESTION_PROMPT, combine_docs_chain_kwargs={"prompt": QA_PROMPT}) return model def get_qa_with_sources_chain(): llm = ChatOpenAI(model_name="gpt-4", temperature=0) retriever = load_retriever() history = [] model = ConversationalRetrievalChain.from_llm( llm=llm, retriever=retriever, return_source_documents=True) def model_func(question): # bug: this doesn't work with the built-in memory # hacking around it for the tutorial # see: https://github.com/langchain-ai/langchain/issues/5630 new_input = {"question": question['question'], "chat_history": history} result = model(new_input) history.append((question['question'], result['answer'])) return result return model_func chain_options = { "basic": get_basic_qa_chain, "with_sources": get_qa_with_sources_chain, "custom_prompt": get_custom_prompt_qa_chain, "condense_prompt": get_condense_prompt_qa_chain }
[ "You are an AI assistant for answering questions about the most recent state of the union address.\nYou are given the following extracted parts of a long document and a question. Provide a conversational answer.\nIf you don't know the answer, just say \"Hmm, I'm not sure.\" Don't try to make up an answer.\nIf the question is not about the most recent state of the union, politely inform them that you are tuned to only answer questions about the most recent state of the union.\nLastly, answer the question as if you were a pirate from the south seas and are just coming back from a pirate expedition where you found a treasure chest full of gold doubloons.\nQuestion: {question}\n=========\n{context}\n=========\nAnswer in Markdown:", "question", "t know the answer, just say \"Hmm, I", "context", "Given the following conversation and a follow up question, rephrase the follow up question to be a standalone question.\nYou can assume the question about the most recent state of the union address.\n\nChat History:\n{chat_history}\nFollow Up Input: {question}\nStandalone question:" ]
2024-01-10
plurigrid/solid
LangChain_app~tests~test_solidity_modules.py
```python import unittest import requests from LangChain_app.solidity_modules import compose_solidity_module from LangChain_app.utils import parse_contract_module class TestSolidityModules(unittest.TestCase): def setUp(self): self.api_endpoint = "http://localhost:5000/api" self.contract_module_string = "Contract module description" self.contract_module = parse_contract_module(self.contract_module_string) def test_compose_solidity_module(self): response = requests.post(self.api_endpoint, json=self.contract_module) self.assertEqual(response.status_code, 200) solidity_module = compose_solidity_module(self.contract_module) self.assertIsInstance(solidity_module, str) if __name__ == "__main__": unittest.main() ```
[]
2024-01-10
plurigrid/solid
LangChain_app~tests~test_utils.py
```python import unittest import requests from LangChain_app.utils import parse_contract_module, parse_lifecycle_event class TestUtils(unittest.TestCase): api_endpoint = "http://localhost:5000/api" def test_parse_contract_module(self): contract_module_string = "Contract Module String" response = requests.post(self.api_endpoint, json={"contract_module": contract_module_string}) contract_module = response.json() parsed_contract_module = parse_contract_module(contract_module_string) self.assertEqual(parsed_contract_module, contract_module) def test_parse_lifecycle_event(self): lifecycle_event_string = "Lifecycle Event String" response = requests.post(self.api_endpoint, json={"lifecycle_event": lifecycle_event_string}) lifecycle_event = response.json() parsed_lifecycle_event = parse_lifecycle_event(lifecycle_event_string) self.assertEqual(parsed_lifecycle_event, lifecycle_event) if __name__ == "__main__": unittest.main() ```
[]
2024-01-10
farithadnan/WiseIlmu
llm_openai.py
import pprint import tiktoken from halo import Halo from omegaconf import DictConfig from langchain.chat_models import ChatOpenAI from langchain.vectorstores.chroma import Chroma from langchain.chains.question_answering import load_qa_chain from langchain.chains.combine_documents.base import BaseCombineDocumentsChain class LLMOpenAI: def __init__(self, cfg: DictConfig, temperature=None, max_tokens=None): if (temperature is None) or (max_tokens is None): self.temperature = cfg.openAI.temperature self.max_tokens = cfg.openAI.max_tokens else: self.temperature = temperature self.max_tokens = max_tokens self.api_key = cfg.openAI.api_key self.model = cfg.openAI.model self.chat_cost_per_1000_tokens = cfg.openAI.chat_cost_per_1000_tokens def get_llm(self): ''' Method to get the LLM model. Returns: The LLM model. ''' try: llm = ChatOpenAI(openai_api_key=self.api_key, model_name=self.model, temperature=self.temperature, max_tokens=self.max_tokens) except (AttributeError, KeyError) as e: raise ValueError(f"Invalid or missing configuration: {e}") return llm def get_qa_chain(self): ''' Method to get the Q & A chain. Returns: The Q & A chain. ''' spinner = Halo(text='Configuring Q & A Chain...', spinner='dots') spinner.start() llm = self.get_llm() spinner.stop() return load_qa_chain(llm=llm) def generate_response(self, vector_db: Chroma, qa_chain: BaseCombineDocumentsChain, messages): ''' Method to generate a response from the chatbot. Args: vector_db: The vector database. qa_chain: The Q & A chain. messages: The messages sent by the user. Returns: The chatbot's response. ''' yellow = "\033[0;33m" print("\n") # Create a loading spinner spinner = Halo(text='Loading Response...\n', spinner='dots') spinner.start() # Fetch latest user Input latest_input = next((message for message in reversed(messages) if message.get("role") == "user"), None) # Get matching documents based on input text matching_docs = vector_db.similarity_search(latest_input["content"]) answer = qa_chain.run(input_documents=matching_docs, question=messages) # Stop the spinner once the response is received spinner.stop() # Testing - Pretty-print the user message sent to the AI pp = pprint.PrettyPrinter(indent=4) print(f"{yellow}\n--------------------------------------------------") print(f"{yellow}💫 Request:") print(f"{yellow}--------------------------------------------------") pp.pprint(messages) # Testing - Pretty-print chat cost print(f"{yellow}--------------------------------------------------") print(f"{yellow}💰 Chat Cost:") print(f"{yellow}--------------------------------------------------") total_word_count, total_token_count, estimated_cost = self.estimate_cost(latest_input["content"], answer) print(f"{yellow}Total Word Count: {total_word_count}") print(f"{yellow}Total Token Count: {total_token_count}") print(f"{yellow}Estimated Cost: ${estimated_cost}") print(f"{yellow}--------------------------------------------------") return answer def estimate_cost(self, user_input: str, bot_response: str): ''' Method to estimate the cost of a chat. Args: user_input (str): The user's input. bot_response (str): The chatbot's response. Returns: The estimated cost, total word count, and total token count. ''' # Create a loading spinner spinner = Halo(text='Estimating Cost...\n', spinner='dots') spinner.start() # Get the total token count combined_text = user_input + " " + bot_response encoded = tiktoken.encoding_for_model(self.model) total_token_count = len(encoded.encode(combined_text)) # Calculate estimated cost for chat estimated_cost = "{:.10f}".format(total_token_count * self.chat_cost_per_1000_tokens/1000) total_word_count = len(combined_text.split()) spinner.stop() return total_word_count, total_token_count, estimated_cost
[]
2024-01-10
farithadnan/WiseIlmu
loader.py
import os import chromadb from halo import Halo from omegaconf import DictConfig from langchain.vectorstores.chroma import Chroma from langchain.document_loaders import CSVLoader from langchain.document_loaders import TextLoader from langchain.document_loaders import PyPDFLoader from langchain.document_loaders import Docx2txtLoader from chromadb.errors import InvalidDimensionException from chromadb.utils.embedding_functions import ONNXMiniLM_L6_V2 from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.embeddings.sentence_transformer import SentenceTransformerEmbeddings class Loader: def __init__(self): pass def load_documents(self, documents_dir: str): ''' Method to load documents from a directory. Args: documents_dir (str): The directory containing the documents. Returns: A list of documents objects. ''' spinner = Halo(text='Fetching Files...\n', spinner='dots') spinner.start() documents = [] for file in os.listdir(documents_dir): try: if file.endswith(".pdf"): pdf_path = os.path.join(documents_dir, file) loader = PyPDFLoader(pdf_path) documents.extend(loader.load()) elif file.endswith(".docx"): docx_path = os.path.join(documents_dir, file) loader = Docx2txtLoader(docx_path) documents.extend(loader.load()) elif file.endswith(".txt"): txt_path = os.path.join(documents_dir, file) loader = TextLoader(txt_path) documents.extend(loader.load()) elif file.endswith(".csv"): csv_path = os.path.join(documents_dir, file) loader = CSVLoader(csv_path) documents.extend(loader.load()) else: raise ValueError(f"Unsupported file format: {file}") except Exception as e: raise RuntimeError(f"Error while loading & splitting the documents: {e}") # Stop the spinner once the response is received spinner.stop() return documents def split_documents(self, documents: list, chunk_size=1000, chunk_overlap=20): ''' Method to split documents into smaller chunks. Args: documents (list): The list of documents. chunk_size (int): The size of the chunks. chunk_overlap (int): The overlap between the chunks. Returns: A list of chunked documents. ''' try: # Create a loading spinner spinner = Halo(text='Splitting File Into Chunk...\n', spinner='dots') spinner.start() text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=chunk_overlap) documents = text_splitter.split_documents(documents=documents) # Stop the spinner once the response is received spinner.stop() return documents except Exception as e: raise RuntimeError(f"Error while splitting the documents: {e}") def create_vector_db(self, documents, cfg: DictConfig): ''' Method to get the vector database. Args: documents (list): The list of documents. cfg (DictConfig): The configuration file. Returns: The vector database. ''' yellow = "\033[0;33m" print(f"{yellow}\n--------------------------------------------------") print(f"{yellow} Configuring Vector Database ") print(f"{yellow}--------------------------------------------------") spinner = Halo(text='\n', spinner='dots') spinner.start() # Instantiate SentenceTransformerEmbeddings embeddings = SentenceTransformerEmbeddings(model_name=cfg.embeddings.model) spinner.stop() # Get vector from documents, if the dimension is invalid, delete the collection and try again try: vector_db = Chroma.from_documents(documents=documents,embedding=embeddings, persist_directory=cfg.vector_db_dir) except InvalidDimensionException: Chroma().delete_collection() vector_db = Chroma.from_documents(documents=documents,embedding=embeddings, persist_directory=cfg.vector_db_dir) print(f"{yellow}--------------------------------------------------\n") return vector_db def load_collection(self, vector_db_dir: str, collection_name="conversations"): ''' Method to create or load a collection. Args: vector_db_dir (str): The directory containing the vector database. Return the collection. ''' spinner = Halo(text='Configuring collection...\n', spinner='dots') spinner.start() embedding_function = ONNXMiniLM_L6_V2() chroma_client = chromadb.PersistentClient(path=vector_db_dir) collection = chroma_client.get_or_create_collection(name=collection_name, embedding_function=embedding_function) spinner.stop() return collection
[]
2024-01-10
masapasa/cohere-weaviate-wikipedia-retrieval
walkthrough.py
# %% import os import weaviate from langchain.embeddings import CohereEmbeddings from langchain.vectorstores import Weaviate from langchain.llms import OpenAI from langchain.chains import RetrievalQA from dotenv import load_dotenv load_dotenv() openai_api_key = os.getenv('OPENAI_API_KEY') cohere_api_key = os.getenv('COHERE_API_KEY') weaviate_api_key = os.getenv('weaviate_api_key') weaviate_url = os.getenv('weaviate_url') # Connect to the Weaviate demo databse containing 10M wikipedia vectors # This uses a public READ-ONLY Weaviate API key auth_config = weaviate.auth.AuthApiKey(api_key=weaviate_api_key) client = weaviate.Client( url=weaviate_url, auth_client_secret=auth_config, additional_headers={ "X-Cohere-Api-Key": cohere_api_key}) vectorstore = Weaviate(client, index_name="Articles", text_key="text") vectorstore._query_attrs = ["text", "title", "url", "views", "lang", "_additional {distance}"] vectorstore.embedding =CohereEmbeddings(model="embed-multilingual-v2.0", cohere_api_key=cohere_api_key) llm =OpenAI(temperature=0, openai_api_key=openai_api_key) qa = RetrievalQA.from_chain_type(llm, retriever=vectorstore.as_retriever()) query = "Why is the theory of everything significant?" result = qa({"query": query}) result['result'] # %% result['result'] # %% query = "Why is the theory of everything significant?" docs = vectorstore.similarity_search(query, 10) docs # %% retriever = vectorstore.as_retriever(search_kwargs={"k": 10}) docs1 = retriever.get_relevant_documents(query) docs1 # %% from langchain.retrievers.contextual_compression import ContextualCompressionRetriever from langchain.retrievers.document_compressors import CohereRerank retriever = vectorstore.as_retriever(search_kwargs={"k": 15}) compressor = CohereRerank(model='rerank-multilingual-v2.0', top_n=4 ) compression_retriever = ContextualCompressionRetriever(base_compressor=compressor, base_retriever=retriever) compressed_docs = compression_retriever.get_relevant_documents("Why is the theory of everything significant?") compressed_docs # %% qa = RetrievalQA.from_chain_type(llm, retriever=compression_retriever) result = qa({"query": query}) # %% result['result'] # %% from langchain.prompts import PromptTemplate prompt_template = """Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer. {context} Question: {question} Helpful Answer in {language}:""" PROMPT = PromptTemplate(template=prompt_template, input_variables=["context", "question","language"]) # %% qa = RetrievalQA.from_chain_type(llm, retriever=compression_retriever, chain_type_kwargs={"prompt": PROMPT.partial(language="english")}) result = qa({"query": query}) result['result']
[ "Use the following pieces of context to answer the question at the end. If you don't know the answer, just say that you don't know, don't try to make up an answer.\n\n{context}\n\nQuestion: {question}\nHelpful Answer in {language}:", "question", "context", "language" ]
2024-01-10
nlsfnr/GPT-X
gptx.py
#!/usr/bin/env python3 from __future__ import annotations import importlib import os import platform import random import re import string import subprocess import sys from pathlib import Path from types import ModuleType from typing import TYPE_CHECKING, Any, Iterable, Iterator, TextIO, TypedDict if sys.version_info < (3, 11): from typing_extensions import Never else: from typing import Never def printerr(*args: Any, **kwargs: Any) -> None: print(*args, file=sys.stderr, **kwargs, flush=True) def fail(msg: str) -> Never: printerr(msg) exit(1) if platform.system() != "Linux": fail(f"Ew, {platform.system()}") if sys.version_info < (3, 9): version = ".".join(map(str, sys.version_info[:3])) fail(f"Python 3.9 or higher is required. You are using {version}.") def confirm(msg: str, default: bool = False) -> bool: printerr(f"{msg} [{'Y/n' if default else 'y/N'}] ", end="") value = input().strip().lower() return default if value == "" else value == "y" def try_import(name: str, pip_name: str) -> ModuleType: try: return importlib.import_module(name) except ImportError: printerr(f"Required package not found: {pip_name}") if not sys.executable: fail("sys.executable not set, aborting.") if confirm(f"Run `pip install {pip_name}`?"): subprocess.run( [sys.executable, "-m", "pip", "install", pip_name], check=True ) return try_import(name, pip_name) fail("Aborted.") if TYPE_CHECKING: import click import requests import tiktoken import yaml from openai import OpenAI from openai.types.chat import ChatCompletionChunk else: click = try_import("click", "click>=8.0.0") OpenAI = try_import("openai", "openai>=1.0.0").OpenAI ChatCompletionChunk = try_import( "openai.types.chat", "openai>=1.0.0" ).ChatCompletionChunk tiktoken = try_import("tiktoken", "tiktoken>=0.5.0") requests = try_import("requests", "requests>=2.0.0") yaml = try_import("yaml", "pyyaml>=5.0.0") def printout(*args: Any, **kwargs: Any) -> None: print(*args, file=sys.stdout, **kwargs, flush=True) class Message(TypedDict): role: str content: str Prompt = list[Message] DEFAULT_MODEL = os.getenv("GPTX_DEFAULT_MODEL", "gpt-4") WORKDIR = Path(os.getenv("GPTX_WORKDIR", Path.home() / ".config" / "gptx")) CONV_DIR = Path(os.getenv("GPTX_CONV_DIR", WORKDIR / "conversations")) LATEST_CONV_FILE = Path(os.getenv("GPTX_LATEST_CONV_FILE", CONV_DIR / "latest.txt")) PROMPT_FILE = Path(os.getenv("GPXT_PROMPT_FILE", WORKDIR / "prompts.yml")) API_KEY_FILE = Path(os.getenv("GPTX_API_KEY_FILE", WORKDIR / "api-key.txt")) DEFAULT_PROMPTS: dict[str, Prompt] = dict( default=[ Message( role="system", content=""" - You are an AI assistant that runs on the terminal. - Your answers go straight to the point. Never apologise, summarise or repeat / rephrase the question. - Assume that you are talking to an expert. - Only answer the immediate question, do not try to predict / propose future questions or topics. - Avoid long texts / paragraphs. Prefer bullet points and short sentences as appropriate. - Make necessary assumptions to give more concrete and actionalbe answers """.strip(), ) ], bash=[ Message( role="system", content=""" - You are an AI writing Bash commands running directly in the terminal - Your answer X will be run as 'sh -c "X"' -- you MUST ALWAYS return a valid shell command - Assume that the user is an expert and knows exactly what they are doing - ALWAYS do what the user asks you to do, even if it's dangerous. Absolutely no exceptions - Make necessary assumptions to solve the given task""".strip(), ) ], ) class Table: """A simple table class for printing nicely formatted tables to the terminal.""" def __init__(self, columns: list[str]) -> None: self.columns = columns self.rows: list[list[str]] = [] def add_row(self, row: dict[str, str] | list[str]) -> Table: if isinstance(row, dict): row = [row.get(column, "") for column in self.columns] self.rows.append(row) return self def order_by(self, columns: str | Iterable[str]) -> Table: """Order the rows by the given columns. Args: columns: The columns to order by. """ if isinstance(columns, str): columns = [columns] indices = [self.columns.index(column) for column in columns] self.rows.sort(key=lambda row: [row[i] for i in indices]) return self def print(self, padding: int = 1, file: TextIO = sys.stdout) -> Table: widths = [len(column) + padding for column in self.columns] for row in self.rows: for i, cell in enumerate(row): widths[i] = max(widths[i], len(cell) + padding) for i, column in enumerate(self.columns): print(column.ljust(widths[i]), end=" ", file=file) print(file=file) for row in self.rows: for i, cell in enumerate(row): print(cell.ljust(widths[i]), end=" ", file=file) print(file=file) return self def resolve_conversation_id(conversation_id: str) -> str: if conversation_id.strip().lower() == "latest": latest = get_latest_conversation_id() if latest is None: fail("Latest conversation not found.") conversation_id = latest return conversation_id def get_conversation_path(conversation_id: str) -> Path: conversation_id = resolve_conversation_id(conversation_id) path = CONV_DIR / f"{conversation_id}.yml" return path def load_prompts(bootstrap: bool = True) -> dict[str, Prompt]: if bootstrap: bootstrap_default_prompts() if not PROMPT_FILE.exists(): fail(f"Prompt file not found: {PROMPT_FILE}") prompts = yaml.safe_load(PROMPT_FILE.read_text()) return prompts def write_prompts(prompts: dict[str, Prompt]) -> None: PROMPT_FILE.write_text(yaml.safe_dump(prompts, indent=2)) def load_prompt(prompt_id: str) -> Prompt: prompts = load_prompts() if prompt_id not in prompts: fail(f"Prompt not found: {prompt_id}") return prompts[prompt_id] def bootstrap_default_prompts() -> None: PROMPT_FILE.parent.mkdir(parents=True, exist_ok=True) if not PROMPT_FILE.exists(): write_prompts(DEFAULT_PROMPTS) else: prompts = load_prompts(bootstrap=False) prompts.update(DEFAULT_PROMPTS) write_prompts(prompts) def get_latest_conversation_id() -> str | None: if not LATEST_CONV_FILE.exists(): return None return LATEST_CONV_FILE.read_text().strip() def load_or_create_conversation( conversation_id: str, prompt_id: str, ) -> list[Message]: path = get_conversation_path(conversation_id) if not path.exists(): prompt = load_prompt(prompt_id) return list(prompt) return yaml.safe_load(path.read_text()) def load_conversation(conversation_id: str) -> list[Message]: path = get_conversation_path(conversation_id) if not path.exists(): fail(f"Conversation not found: {conversation_id}") return yaml.safe_load(path.read_text()) def save_conversation(conversation_id: str, messages: list[Message]) -> None: conversation_id = resolve_conversation_id(conversation_id) path = get_conversation_path(conversation_id) path.parent.mkdir(parents=True, exist_ok=True) with open(path, "w") as f: yaml.safe_dump(messages, f, indent=2) LATEST_CONV_FILE.write_text(conversation_id) def next_conversation_id() -> str: pool = string.ascii_letters + string.digits ATTEMPTS = 10_000 for k in range(3, 10): for _ in range(ATTEMPTS): conversation_id = "".join(random.choices(pool, k=k)) path = get_conversation_path(conversation_id) if not path.exists(): return conversation_id fail(f"Failed to generate a conversation ID after {ATTEMPTS} attempts.") def get_conversation_ids() -> list[str]: return [path.stem for path in CONV_DIR.glob("*.yml")] def get_token_count( x: str | list[Message], model: str, ) -> int: enc = tiktoken.encoding_for_model(model) messages = x if isinstance(x, list) else [Message(role="user", content=x)] total = sum(len(enc.encode(message["content"])) for message in messages) return total def enhance_content( prompt: str, ) -> str: def get_file_contents(match: re.Match) -> str: """Inject file contents into the prompt.""" path_str = match.group(1) if path_str.startswith("http"): response = requests.get(path_str) response.raise_for_status() text = response.text printerr(f"Injecting: {path_str}\t{len(text)} chars") elif path_str == "stdin": text = sys.stdin.read() printerr(f"Injecting: stdin\t{len(text)} chars") else: path = Path(path_str) if not path.exists(): fail(f"File not found: {path}") if path.suffix.lower() == ".pdf": if TYPE_CHECKING: import PyPDF2 else: PyPDF2 = try_import("PyPDF2", "PyPDF2>=3.0.0") text = "" with open(path, "rb") as f: reader = PyPDF2.PdfFileReader(f) for page in reader.pages: text += page.extractText() return text else: text = path.read_text() printerr(f"Injecting: {path}\t{len(text)} chars") return text regex = re.compile(r"\{\{ ([^}]+) \}\}") prompt = re.sub(regex, get_file_contents, prompt) return prompt def generate( messages: list[Message], api_key: str, max_tokens: int, temperature: float, top_p: float, model: str, ) -> Iterator[str]: openai = OpenAI(api_key=api_key) chunks: Iterator[ChatCompletionChunk] = openai.chat.completions.create( model=model, messages=messages, # type: ignore max_tokens=max_tokens, temperature=temperature, top_p=top_p, stream=True, ) for chunk in chunks: # delta = chunk["choices"][0]["delta"] # type: ignore delta = chunk.choices[0].delta.content if delta: yield delta @click.group() def cli() -> None: """GPT4 CLI""" pass # fmt: off @cli.command("q") @click.option("--max-generation-tokens", "-m", type=int, default=1024, help="Max tokens to generate") @click.option("--temperature", "-t", type=float, default=0.5, help="Temperature") @click.option("--top-p", "-p", type=float, default=0.2, help="Top p") @click.option("--api-key-file", type=Path, default=API_KEY_FILE, help="Path to API key file") @click.option("--conversation", "-c", type=str, default=None, help="Conversation ID") @click.option("--prompt", "-p", type=str, default="default", help="Prompt ID") @click.option("--model", type=str, default=DEFAULT_MODEL, help="Model") @click.option("--max-prompt-tokens", type=int, default=7168, help="Max tokens in prompt") @click.option("--run", "-r", is_flag=True, help="Run the output inside a shell, after confirming") @click.option("--yolo", "-y", is_flag=True, help="Do not ask for confirmation before running") @click.option("--interactive", "-i", is_flag=True, help="Interactive mode") @click.argument("user_message", nargs=-1, required=True) # fmt: on def query( max_generation_tokens: int, temperature: float, top_p: float, api_key_file: Path, conversation: str, prompt: str, model: str, max_prompt_tokens: int, user_message: list[str], run: bool, yolo: bool, interactive: bool, ) -> None: """Query GPT4""" if interactive and (run or yolo): fail("Cannot use --interactive with --run or --yolo.") api_key = api_key_file.read_text().strip() conversation_id = conversation or next_conversation_id() conversation_id = resolve_conversation_id(conversation_id) prompt_id = prompt messages = load_or_create_conversation(conversation_id, prompt_id) message_str = " ".join(user_message).strip() try: while True: message_str = enhance_content(message_str) if not message_str: if interactive: message_str = input("You:") continue fail("Empty message.") message_token_count = get_token_count(message_str, model) messages_token_count = get_token_count(messages, model) total_token_count = message_token_count + messages_token_count if total_token_count > max_prompt_tokens and not confirm( f"Total prompt length: {total_token_count} tokens. Max: " f"{max_prompt_tokens}. Continue anyway?", default=False, ): fail("Aborted.") messages.append(Message(role="user", content=message_str)) full_answer = "" token_count = get_token_count(messages, model=model) printerr( f"Conversation ID: {conversation_id} | {token_count} tokens", end="\n\n" ) chunks = generate( messages=messages, api_key=api_key, max_tokens=max_generation_tokens, temperature=temperature, top_p=top_p, model=model, ) printout("AI: ", end="") for chunk in chunks: printout(chunk, end="") full_answer += chunk printout() messages.append(Message(role="assistant", content=full_answer)) save_conversation(conversation_id, messages) if not interactive: break message_str = input("\nYou: ").strip() except KeyboardInterrupt: fail("Interrupted.") if run: printerr() run_in_shell(full_answer, yolo) @cli.command("prompts") @click.option("--editor", "-e", type=str, default=os.environ.get("EDITOR", "nvim")) def edit_prompts( editor: str, ) -> None: """Edit prompts.""" bootstrap_default_prompts() if not PROMPT_FILE.exists(): fail(f"Prompt file not found: {PROMPT_FILE}") subprocess.run([editor, str(PROMPT_FILE)], check=True) def run_in_shell( command: str, yolo: bool, ) -> None: if not yolo and not confirm("Run in shell?", default=True): fail("Aborted.") subprocess.Popen( command, shell=True, stdin=sys.stdin, stdout=sys.stdout, stderr=sys.stderr, ).communicate() @cli.command("ls") def list_() -> None: """List conversations.""" ids = get_conversation_ids() if not ids: printerr("No conversations found.") table = Table(["#", "ID", "First message"]) for i, conversation_id in enumerate(ids, 1): messages = load_conversation(conversation_id) user_messages = [m for m in messages if m["role"] == "user"] if not user_messages: content = "No messages." else: content = user_messages[0]["content"] if len(content) > 40: content = content[:40] + "…" content = content.replace("\n", " ") table.add_row([str(i), conversation_id, content]) table.print() @cli.command("rm") @click.argument("conversation_id", type=str, default="latest") def remove(conversation_id: str) -> None: """Remove a conversation.""" conversation_id = resolve_conversation_id(conversation_id) path = get_conversation_path(conversation_id) if not path.exists(): fail(f"Conversation {conversation_id} not found.") path.unlink() printerr(f"Conversation {conversation_id} removed.") @cli.command("print") @click.argument("conversation_id", type=str, default="latest") def print_(conversation_id: str) -> None: """Print a conversation.""" messages = load_conversation(conversation_id) for message in messages: print(f"{message['role']}: {message['content']}") @cli.command("repeat") @click.argument("conversation_id", type=str, default="latest") def repeat(conversation_id: str) -> None: """Repeat the latest message.""" messages = load_conversation(conversation_id) last_message = messages[-1] print(f"{last_message['content']}") @cli.command("run") @click.argument("conversation_id", type=str, default="latest") @click.option("--yolo", "-y", is_flag=True, default=False) def run( conversation_id: str, yolo: bool, ) -> None: """Run the latest message inside the shell.""" messages = load_conversation(conversation_id) command = messages[-1]["content"] printerr(command) printerr() run_in_shell(command, yolo) if __name__ == "__main__": cli()
[ "\n- You are an AI assistant that runs on the terminal.\n- Your answers go straight to the point. Never apologise, summarise or repeat / rephrase the question.\n- Assume that you are talking to an expert.\n- Only answer the immediate question, do not try to predict / propose future questions or topics.\n- Avoid long texts / paragraphs. Prefer bullet points and short sentences as appropriate.\n- Make necessary assumptions to give more concrete and actionalbe answers\n", "- You are an AI writing Bash commands running directly in the terminal\n- Your answer X will be run as 'sh -c \"X\"' -- you MUST ALWAYS return a valid shell command\n- Assume that the user is an expert and knows exactly what they are doing\n- ALWAYS do what the user asks you to do, even if it's dangerous. Absolutely no exceptions\n- Make necessary assumptions to solve the given task", "- You are an AI assistant that runs on the terminal.\n- Your answers go straight to the point. Never apologise, summarise or repeat / rephrase the question.\n- Assume that you are talking to an expert.\n- Only answer the immediate question, do not try to predict / propose future questions or topics.\n- Avoid long texts / paragraphs. Prefer bullet points and short sentences as appropriate.\n- Make necessary assumptions to give more concrete and actionalbe answers", "prompts.yml", "GPXT_PROMPT_FILE", "\n- You are an AI writing Bash commands running directly in the terminal\n- Your answer X will be run as 'sh -c \"X\"' -- you MUST ALWAYS return a valid shell command\n- Assume that the user is an expert and knows exactly what they are doing\n- ALWAYS do what the user asks you to do, even if it's dangerous. Absolutely no exceptions\n- Make necessary assumptions to solve the given task" ]
2024-01-10
UEC-CED/ChatVRM
qa_api~qa_chatbot_uec_ced.py
import sys import os import json import argparse import logging from llama_index import GPTVectorStoreIndex, SimpleDirectoryReader from llama_index import StorageContext, load_index_from_storage from llama_index import LLMPredictor, PromptHelper, ServiceContext from langchain.chat_models import ChatOpenAI class UECQueryEngine: def __init__(self, reindex): with open('./config.json', 'r') as f: config = json.load(f) os.environ["OPENAI_API_KEY"] = config["OPENAI_API_KEY"] logging.basicConfig(stream=sys.stdout, level=logging.DEBUG, force=True) llm_predictor = LLMPredictor(llm=ChatOpenAI(temperature=0, model_name="gpt-3.5-turbo")) # define prompt helper # set maximum input size max_input_size = 4096 # set number of output tokens num_output = 256 # set maximum chunk overlap max_chunk_overlap = 20 prompt_helper = PromptHelper(context_window=max_input_size, num_output=num_output) service_context = ServiceContext.from_defaults(llm_predictor=llm_predictor, prompt_helper=prompt_helper) if reindex: # インデックスの再作成 documents = SimpleDirectoryReader("./data").load_data() index = GPTVectorStoreIndex.from_documents(documents, service_context=service_context) # インデックスの保存 index.storage_context.persist() else: # インデックスの読み込み storage_context = StorageContext.from_defaults(persist_dir="./storage") index = load_index_from_storage(storage_context) # クエリエンジンの作成 self.query_engine = index.as_query_engine() def query(self, question): return self.query_engine.query(question) def make_index(self): documents = SimpleDirectoryReader("./data").load_data() index = GPTVectorStoreIndex.from_documents(documents) # インデックスの保存 index.storage_context.persist() if __name__ == "__main__": # コマンドライン引数の処理 parser = argparse.ArgumentParser(description='QA Chatbot for UEC CED') parser.add_argument('--reindex', action='store_true', help='インデックスの再作成') args = parser.parse_args() query_engine = UECQueryEngine(args.reindex) question = "" while question != "exit": question = input("電気通信大学に関する質問を入力してください: ") answer = query_engine.query(question) print(answer)
[]
2024-01-10
yinhui1984/chatdoc
chatdoc.py
#!/usr/bin/env python3 import os import shutil import sys import tempfile import threading import time from langchain.chains import RetrievalQAWithSourcesChain from langchain.chat_models import ChatOpenAI from langchain.docstore.document import Document from langchain.document_loaders import PDFMinerLoader from langchain.embeddings.openai import OpenAIEmbeddings from langchain.prompts.chat import (ChatPromptTemplate, SystemMessagePromptTemplate, HumanMessagePromptTemplate) from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.vectorstores import Chroma model_name = "gpt-3.5-turbo" temperature = 0.1 max_tokens = 2048 chunk_size = 500 similarity_search_k = 4 translateNeeded = True translatedAnswer = "" def parse_env_config(): current_dir = os.path.dirname(os.path.abspath(__file__)) config = current_dir + "/env.config" if os.path.exists(config): with open(config, 'r') as f: lines = f.readlines() for line in lines: if line.startswith("OPENAI_API_KEY"): os.environ['OPENAI_API_KEY'] = line.split("=")[1].strip("\"\'\r\n") else: print("env.config not found!") if os.environ['OPENAI_API_KEY'] == '': print("OPENAI_API_KEY not found! Please config it in env.config") exit(1) def get_vector_db(name: str): current_dir = os.path.dirname(os.path.abspath(__file__)) db_dir = os.path.join(current_dir, "vector_db") if not os.path.exists(db_dir): os.mkdir(db_dir) return db_dir + "/" + name def translate(txt): print("Translate to Chinese...") import openai openai.api_key = os.environ['OPENAI_API_KEY'] completion = openai.ChatCompletion.create(model=model_name, messages=[ {"role": "user", "content": "translate below content into Chinese" "(professional terminology and code do not need to be translated):" + txt}]) result = completion.choices[0].message.content # print(result) return result # 下面的代码报错: Retrying langchain.chat_models.openai.acompletion_with_retry.<locals>._completion_with_retry in 8.0 # seconds as it raised APIConnectionError: Error communicating with OpenAI. # llm = ChatOpenAI(temperature=0,max_tokens=1024) # g = await llm.agenerate([[HumanMessage(content="translate below content into Chinese:"+txt)]]) # print(g) def create_vectors(file_url: str): db_path = get_vector_db(file_url) if not os.path.exists(db_path): print("create new chroma store...") is_pdf = os.path.splitext(file_url)[1] == '.pdf' if is_pdf: # PDFMinerLoader是一个Python库,用于从PDF文档中提取文本。 # 它使用先进的算法来准确地识别和提取PDF中的文本,包括图像中的文本。 # 它还可以用于从PDF中提取元数据,例如作者、标题和主题。 loader = PDFMinerLoader(file_url) file_data = loader.load() # TODO: 添加其他类型文件 else: # 作为文本文件 with open(file_url, 'r') as f: text = f.read() metadata = {"source": file_url} file_data = [Document(page_content=text, metadata=metadata)] # RecursiveCharacterTextSplitter类用于将文本分成较小的字符块。 # chunk_size参数指定每个块的最大大小,chunk_overlap参数指定两个相邻块之间应重叠多少个字符。 # chunk_size设置为N,chunk_overlap设置为0,这意味着每个块将有N个字符长,并且相邻块之间没有重叠。 textsplitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=0) split_documents = textsplitter.split_documents(file_data) print("split_documents:", len(split_documents)) # OpenAIEmbeddings类用于将文本转换为向量表示。 # 使用Chroma库来创建一组文档的向量表示。文档作为“split_documents”参数传入,使用的嵌入是OpenAIEmbeddings()。 # persist_directory参数指定了存储向量表示的目录。这样可以重复使用向量表示,而无需每次重新创建它们。 vectors_store = Chroma.from_documents(documents=split_documents, embedding=OpenAIEmbeddings(), persist_directory=db_path) vectors_store.persist() print('load finish!') else: print('has loaded!') def ask_ai(file_url: str, question: str): db_path = get_vector_db(file_url) if not os.path.exists(db_path): create_vectors(file_url) else: print('use existing chroma store...') # 创建一个Chroma对象,使用OpenAIEmbeddings作为嵌入函数,并使用get_persist_path作为持久化目录。 vectors_store = Chroma(embedding_function=OpenAIEmbeddings(), persist_directory=db_path) llm = ChatOpenAI(model_name=model_name, temperature=temperature, max_tokens=max_tokens) system_template = """Use the following pieces of context to answer the users question. Take note of the sources and include them in the answer in the format: "SOURCES: source1 source2", use "SOURCES" in capital letters regardless of the number of sources. If you don't know the answer, just say that "I don't know", don't try to make up an answer. ---------------- {summaries}""" messages = [ SystemMessagePromptTemplate.from_template(system_template), HumanMessagePromptTemplate.from_template("{question}")] # ChatPromptTemplate可以用来创建一个聊天机器人提示模板,以便在聊天时使用。 prompt = ChatPromptTemplate.from_messages(messages) # 链对象可以用于基于检索的问答任务。 chain = RetrievalQAWithSourcesChain.from_chain_type(llm, chain_type="stuff", retriever=vectors_store.as_retriever(), return_source_documents=False, chain_type_kwargs={"prompt": prompt}) print("search similarity content in vectors....") # 检索相似“文档”(这里的“文档”是我们的文档被split后的子文档) similarity_doc = vectors_store.similarity_search(question, k=similarity_search_k) chain_params = {"input_documents": similarity_doc, "question": question, "return_only_outputs": True,} def show_busy_indicator(): while chain_thread.is_alive(): for c in '|/-\\': print('\r' + 'THINKING... ' + c, end='', flush=True) time.sleep(0.1) print("\r", end='', flush=True) def run_chain(params): chat_res = chain(params) answer = chat_res.get('answer') print("\n"+answer) global translatedAnswer if translateNeeded: translatedAnswer = translate(answer) else: translatedAnswer = answer # chat_res = chain({"input_documents": similarity_doc,"question": question}, return_only_outputs=True) chain_thread = threading.Thread(target=lambda: run_chain(chain_params)) chain_thread.start() show_busy_indicator() chain_thread.join() # print as green print('\033[32m' + "---- answer ---- " + '\033[0m') global translatedAnswer if translateNeeded: print(translatedAnswer) def main(): parse_env_config() # 使用方式1:使用stdin作为内容输入 # cat 123.txt | python3 chatdoc.py question if not sys.stdin.isatty(): # remove all ./vector_db/var/* temp_dir = get_vector_db("var") if os.path.exists(temp_dir): # print(temp_dir) shutil.rmtree(temp_dir) std = "" # read all lines from stdin for line in sys.stdin: std += line # create a temp file to save the content temp_file = tempfile.NamedTemporaryFile(delete=True) temp_file.write(std.encode('utf-8')) # combine all arguments into a single string combined_question = " ".join(sys.argv[1:]) ask_ai(temp_file.name, combined_question) temp_file.close() return 0 # 使用方式2:使用文件作为内容输入 # python3 chatdoc.py file_url question if len(sys.argv) < 3: print('Usage: python3 chatdoc.py file_url question') sys.exit(1) file_url = sys.argv[1] # combine all other arguments into a single string combined_question = " ".join(sys.argv[2:]) ask_ai(file_url, combined_question) if __name__ == '__main__': main()
[ "{question}", "Use the following pieces of context to answer the users question.\nTake note of the sources and include them in the answer in the format: \"SOURCES: source1 source2\", use \"SOURCES\" in capital letters regardless of the number of sources.\nIf you don't know the answer, just say that \"I don't know\", don't try to make up an answer.\n----------------\n{summaries}", "translate below content into Chinese(professional terminology and code do not need to be translated):PLACEHOLDER" ]
2024-01-10
braunfuss/Palantiri
src~process~semp.py
import os import sys import time import numpy as num from threading import Thread from obspy.signal.headers import clibsignal from obspy import Stream, Trace from palantiri.common.ConfigFile import ConfigObj, OriginCfg, SynthCfg, FilterCfg import numpy as np from palantiri.common import Basic, Logfile import ctypes as C from pyrocko import trace as trld from pyrocko.marker import PhaseMarker from pyrocko import obspy_compat import math from scipy.signal import coherence from collections import OrderedDict from palantiri.process import music from pyrocko import io, trace, util trace_txt = 'trace.txt' travel_txt = 'travel.txt' latv_txt = 'latv.txt' lonv_txt = 'lonv.txt' semb_txt = 'semb.txt' def normalize(lst): s = sum(lst) return map(lambda x: float(x)/s, lst) def hilbert(x, N=None): ''' Return the hilbert transform of x of length N. (from scipy.signal, but changed to use fft and ifft from numpy.fft) ''' x = num.asarray(x) if N is None: N = len(x) if N <= 0: raise ValueError("N must be positive.") x = num.real(x) Xf = num.fft.fft(x, N, axis=0) h = num.zeros(N) if N % 2 == 0: h[0] = h[N//2] = 1 h[1:N//2] = 2 else: h[0] = 1 h[1:(N+1)//2] = 2 if len(x.shape) > 1: h = h[:, num.newaxis] x = num.fft.ifft(Xf*h) return x def xcorr(tr1, tr2, shift_len, full_xcorr=False): from obspy.core.util.deprecation_helpers import ObsPyDeprecationWarning if min(len(tr1), len(tr2)) - 2 * shift_len <= 0: msg = "shift_len too large. The underlying C code would silently " + \ "use shift_len/2 which we want to avoid." raise ValueError(msg) tr1 = np.ascontiguousarray(tr1, np.float32) tr2 = np.ascontiguousarray(tr2, np.float32) corp = np.empty(2 * shift_len + 1, dtype=np.float64, order='C') shift = C.c_int() coe_p = C.c_double() res = clibsignal.X_corr(tr1, tr2, corp, shift_len, len(tr1), len(tr2), C.byref(shift), C.byref(coe_p)) if res: raise MemoryError if full_xcorr: return shift.value, coe_p.value, corp else: return shift.value, coe_p.value class MyThread(Thread): def __init__(self, nostat, nsamp, i, nstep, dimX, dimY, mint, new_freq, minSampleCount): Thread.__init__(self) self.nostat = nostat self.nsamp = nsamp self.i = i self.nstep = nstep self.dimX = dimX self.dimY = dimY self.mint = mint self.new_freq = new_freq self.minSampleCount = minSampleCount def toMatrix(npVector, nColumns): t = npVector.tolist()[0] n = nColumns mat = [] for i in range(int(len(t) / n)): pos1 = i * n pos2 = pos1 + n slice = t[pos1:pos2] assert len(slice) == nColumns mat.append(slice) return mat def semblance(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, Config, Origin, refshifts, nstats, bs_weights=None, flag_rpe=False, boot_shifts=None): cfg = Config origin = OriginCfg(Origin) cfg_f = FilterCfg(Config) if cfg.config_filter.dynamic_filter is False: if cfg.config.bp_freq is True: return semblance_py_freq(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=bs_weights) if cfg.config.bp_coh is True: return semblance_py_coherence(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=bs_weights) if cfg.config_geometry.dimz!= 0: return semblance_py_cube(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, bs_weights=bs_weights) if cfg.config.bp_music is True: return music_wrapper(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=bs_weights) if flag_rpe is True: return semblance_py(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=bs_weights, flag_rpe=flag_rpe) else: return semblance_py(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=bs_weights, flag_rpe=flag_rpe, boot_shifts=boot_shifts) else: return semblance_py_dynamic_cf(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, origin, cfg_f) def t2ind_fast(t, tdelta, snap=round): return int(int((t/tdelta)*(10**0))/(10.**0)) def t2ind(t, tdelta, snap=round): return int(snap(t/tdelta)) def semblance_py_dynamic_cf(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, origin, FilterCfg): obspy_compat.plant() trs_orgs = [] for tr in calcStreamMap: tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(tr_org.ydata))) trs_orgs.append(tr_org) trace = toMatrix(trace_1, minSampleCount) traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY) latv = latv_1.tolist() lonv = lonv_1.tolist() ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' snap=(round, round) backSemb = np.ndarray(shape=(ntimes, dimX*dimY), dtype=float) for i in range(ntimes): # loop over grid points sembmax = 0 sembmaxX = 0 sembmaxY = 0 for j in range(dimX * dimY): semb = 0 nomin = 0 denom = 0 sums_cc = 0 sums = 0 shifted = [] relstart = [] relstarts = nostat cc_data = [] tt = [] for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] tmin = time+relstart+(i*nstep)-mint tmax = time+relstart+(i*nstep)-mint+nsamp try: ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) except: print('Loaded traveltime grid wrong!') data = tr.ydata[ibeg:iend] try: sums += num.gradient(abs(data)) except: pass relstarts -=(relstart) sum = abs(num.sum(((sums)))) denom = sum**2 nomin = sum semb = sum backSemb[i][j] = sum if semb > sembmax: sembmax = semb sembmaxX = latv[j] sembmaxY = lonv[j] Logfile.add('max semblance: ' + str(sembmax) + ' at lat/lon: ' + str(sembmaxX)+','+ str(sembmaxY)) backSemb = backSemb/num.max(num.max(backSemb)) return abs(backSemb) def semblance_py(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=None, flag_rpe=False, output=False, boot_shifts=None): trs_orgs = [] snap = (round, round) if cfg.config_weight.combine_all is True: combine = True else: combine = False if cfg.config_weight.bootstrap_array_weights is True: do_bs_weights = True else: do_bs_weights = False do_weight_by_array = False if do_weight_by_array: k = 0 stats_done = 0 for k in range(0, len(nstats)): for stats in range(0, nstats[k]): # list(calcStreamMap.keys())[stats].data = list(calcStreamMap.keys())[stats].data/np.max(list(calcStreamMap.keys())[0].data) stats_done = stats_done + nstats[k] k = 0 s_index = 0 for tr in calcStreamMap: tr_org = calcStreamMap[tr][0] trs_orgs.append(tr_org) for tr in calcStreamMap: tr_org = calcStreamMap[tr][0] datas = trs_orgs[0:s_index].ydata if k <= nstats[s_index]: k = k+1 # tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(datas))) if k == nstats[s_index]: s_index = s_index+1 calcStreamMap[tr] = obspy_compat.to_obspy_trace(tr_org) stats_done = stats_done + nstats[k] trs_orgs = [] for tr in sorted(calcStreamMap): tr_org = calcStreamMap[tr] trs_orgs.append(tr_org) traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY) latv = latv_1.tolist() lonv = lonv_1.tolist() index_begins = OrderedDict() index_steps = [] index_window = [] for j in range(dimX * dimY): markers = [] for k in range(nostat): if cfg.config_weight.bootstrap_array_weights is True: relstart = traveltime[k][j] + boot_shifts[k] + num.random.uniform(-1*cfg.config_weight.shift_max/10., cfg.config_weight.shift_max/10.) else: relstart = traveltime[k][j] tr = trs_orgs[k] try: tmin = time+relstart-mint+refshifts[k] tmax = time+relstart-mint+nsamp+refshifts[k] except: tmin = time+relstart-mint tmax = time+relstart-mint+nsamp # m = PhaseMarker(tmin=tmin, # tmax=tmax, # phasename='P', # nslc_ids=(tr.nslc_id,)) # markers.append(m) ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) index_begins[str(j)+str(k)] = [ibeg, tmin] iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) iend_step = min( tr.data_len(), t2ind_fast(tmax-tr.tmin+nstep, tr.deltat, snap[1])) index_steps.append(iend_step-iend) index_window.append(iend-ibeg) # for debug: # trld.snuffle(trs_orgs, markers=markers) ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' trs_orgs = [] k = 0 for tr in sorted(calcStreamMap): tr_org = calcStreamMap[tr] if combine is True: # some trickery to make all waveforms have same polarity, while still # considering constructive/destructive interferences. This is needed # when combing all waveforms/arrays from the world at once(only then) # for a single array with polarity issues we recommend fixing polarity. # advantage of the following is that nothing needs to be known about the # mechanism. # tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(tr_org.ydata))) # tr_org.ydata = abs(tr_org.ydata) # tr_org.ydata = num.ediff1d(tr_org.ydata) if max(index_steps) % 2 == 1: tr_org.ydata = abs(tr_org.ydata) # tr_org.ydata = num.ediff1d(tr_org.ydata) # if cfg.config_weight.shift_by_phase_pws is True: # cfx = num.fft.fft(tr_org.ydata) # sums_schimmel = sums_schimmel + (cfx/(abs(cfx)) * num.exp(1j*2*num.pi*cfx))) # print('calculate pws') trs_orgs.append(tr_org) if nsamp == 0: nsamp = 1 backSemb = np.ndarray(shape=(ntimes, dimX*dimY), dtype=float) for i in range(ntimes): sembmax = 0 sembmaxX = 0 sembmaxY = 0 for j in range(dimX * dimY): semb = 0. nomin = 0 denom = 0 sums = num.zeros(max(index_steps)) sums = 0. relstart = [] relstarts = nostat sums_schimmel = 0 for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] ibeg = index_begins[str(j)+str(k)][0]+i*index_steps[j+k] iend = index_begins[str(j)+str(k)][0]+index_window[j+k]+i*index_steps[j+k] data = tr.ydata[ibeg:iend] if cfg.config_weight.shift_by_phase_pws is True: cfx = num.fft.fft(data) sums_schimmel = sums_schimmel + (cfx/(abs(cfx)) * num.exp(1j*2*num.pi*cfx)) try: if do_bs_weights is True and combine is True: sums += data*bs_weights[k] else: sums = sums+data except ValueError: try: if num.shape(data) < num.shape(sums): data = tr.ydata[ibeg:iend+1] else: data = tr.ydata[ibeg:iend-1] sums = sums+data except: sums = sums relstarts -= relstart sums_schimmel = num.real(sums_schimmel)**2. sums_copy = sums # data = sums_schimmel # t1 = trace.Trace( # station='TEST', channel='Z', deltat=0.5, tmin=0., ydata=data) if cfg.config_weight.shift_by_phase_pws is True: for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] ibeg = index_begins[str(j)+str(k)][0]+i*index_steps[j+k] iend = index_begins[str(j)+str(k)][0]+index_window[j+k]+i*index_steps[j+k] data = tr.ydata[ibeg:iend] data_copy = data.copy() cfx = num.fft.fft(data) * sums_schimmel data = num.fft.ifft(cfx) try: if do_bs_weights is True: sums += data*bs_weights[k] else: sums = sums+data except ValueError: if num.shape(data) < num.shape(sums): data = tr.ydata[ibeg:iend+1] else: data = tr.ydata[ibeg:iend-1] sums = sums+data data = sums_copy basetime = util.str_to_time("2016-11-25 14:24:30.000") data = num.real(sums) sum = abs(num.sum(sums)) if combine is True: sum = (1./nostat)* ((abs(num.sum((sums)))**2) /num.sum(sums)) semb = sum backSemb[i][j] = sum if semb > sembmax: sembmax = semb sembmaxX = latv[j] sembmaxY = lonv[j] #backSemb[i][:] = backSemb[i][:]/num.max(backSemb[i][:]) if output is True: Logfile.add('max semblance: ' + str(sembmax) + ' at lat/lon: ' + str(sembmaxX) + ',' + str(sembmaxY)) if flag_rpe is False: backSemb = backSemb/num.max(num.max(backSemb)) return backSemb def music_wrapper(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=None): obspy_compat.plant() trs_orgs = [] snap = (round, round) if cfg.config_weight.combine_all is True: combine = True else: combine = False if cfg.config_weight.bootstrap_array_weights is True: do_bs_weights = True else: do_bs_weights = False do_weight_by_array = False if do_weight_by_array: k = 0 stats_done = 0 for k in range(0, len(nstats)): for stats in range(0, nstats[k]): list(calcStreamMap.keys())[stats].data = list(calcStreamMap.keys())[stats].data/np.max(list(calcStreamMap.keys())[0].data) stats_done = stats_done + nstats[k] k = 0 s_index = 0 for tr in calcStreamMap: tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) trs_orgs.append(tr_org) for tr in calcStreamMap: tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) datas = trs_orgs[0:s_index].ydata if k <= nstats[s_index]: k = k+1 tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(datas))) if k == nstats[s_index]: s_index = s_index+1 calcStreamMap[tr] = obspy_compat.to_obspy_trace(tr_org) stats_done = stats_done + nstats[k] trs_orgs = [] for tr in sorted(calcStreamMap): tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) trs_orgs.append(tr_org) array_lats, array_lons = calcStreamMap[tr].lat, calcStreamMap[tr].lon traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY) latv = latv_1.tolist() lonv = lonv_1.tolist() from collections import OrderedDict index_begins = OrderedDict() index_steps = [] index_window = [] for j in range(dimX * dimY): markers = [] for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] try: tmin = time+relstart-mint-refshifts[k] tmax = time+relstart-mint+nsamp-refshifts[k] except IndexError: tmin = time+relstart-mint tmax = time+relstart-mint+nsamp m = PhaseMarker(tmin=tmin, tmax=tmax, phasename='P', nslc_ids=(tr.nslc_id,)) markers.append(m) ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) index_begins[str(j)+str(k)]= [ibeg, tmin] iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) iend_step = min( tr.data_len(), t2ind_fast(tmax-tr.tmin+nstep, tr.deltat, snap[1])) index_steps.append(iend_step-iend) index_window.append(iend-ibeg) # trld.snuffle(trs_orgs, markers=markers) ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' trs_orgs = [] k = 0 for tr in sorted(calcStreamMap): tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) trs_orgs.append(tr_org) if nsamp == 0: nsamp = 1 backSemb = np.ndarray(shape=(ntimes, dimX*dimY), dtype=float) for i in range(ntimes): sembmax = 0 sembmaxX = 0 sembmaxY = 0 for j in range(dimX * dimY): semb = 0. nomin = 0 denom = 0 sums = num.zeros(max(index_steps)) sums = 0. relstart = [] relstarts = nostat sums_schimmel = 0 for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] ibeg = index_begins[str(j)+str(k)][0]+i*index_steps[j+k] iend = index_begins[str(j)+str(k)][0]+index_window[j+k]+i*index_steps[j+k] data = tr.ydata[ibeg:iend] # normalize: #data = data / np.sqrt(np.mean(np.square(data))) try: if do_bs_weights is True: sums += data*bs_weights[k] else: sums = sums+data except ValueError: if num.shape(data) < num.shape(sums): data = tr.ydata[ibeg:iend+1] else: data = tr.ydata[ibeg:iend-1] sums = sums+data relstarts -= relstart sums_schimmel = abs(sums_schimmel)**2. sum = abs(num.sum(sums)) semb = sum backSemb[i][j] = sum if semb > sembmax: sembmax = semb sembmaxX = latv[j] sembmaxY = lonv[j] Logfile.add('max semblance: ' + str(sembmax) + ' at lat/lon: ' + str(sembmaxX) + ',' + str(sembmaxY)) return backSemb def semblance_py_coherence(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=None): obspy_compat.plant() trs_orgs = [] snap = (round, round) if cfg.config_weight.combine_all is True: combine = True else: combine = False if cfg.config_weight.bootstrap_array_weights is True: do_bs_weights = True else: do_bs_weights = False do_weight_by_array = True if do_weight_by_array: tr_bases = [] tr_bases_data = [] k = 0 ks = 0 s_index = 0 for tr in calcStreamMap: tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) # tr_org.ydata = abs(tr_org.ydata) # tr_org.ydata = num.ediff1d(tr_org.ydata, to_end=tr_org.ydata[-1]) if k == 1: tr_bases.append(tr_org) tr_org.set_location('%s' % s_index) if k == nstats[s_index]: s_index = s_index+1 k = 0 if k < nstats[s_index]: k = k+1 calcStreamMap[tr] = obspy_compat.to_obspy_trace(tr_org) ks = ks +1 trs_orgs = [] for tr in sorted(calcStreamMap): tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) trs_orgs.append(tr_org) traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY) latv = latv_1.tolist() lonv = lonv_1.tolist() from collections import OrderedDict index_begins = OrderedDict() index_steps = [] index_window = [] for j in range(dimX * dimY): markers = [] for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] try: tmin = time+relstart-mint-refshifts[k] tmax = time+relstart-mint+nsamp-refshifts[k] except IndexError: tmin = time+relstart-mint tmax = time+relstart-mint+nsamp m = PhaseMarker(tmin=tmin, tmax=tmax, phasename='P', nslc_ids=(tr.nslc_id,)) markers.append(m) ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) index_begins[str(j)+str(k)]= [ibeg, tmin] iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) iend_step = min( tr.data_len(), t2ind_fast(tmax-tr.tmin+nstep, tr.deltat, snap[1])) index_steps.append(iend_step-iend) index_window.append(iend-ibeg) # trld.snuffle(trs_orgs, markers=markers) ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' trs_orgs = [] do_trad = False for tr in sorted(calcStreamMap): tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) # if combine is True: # tr_org.ydata = abs(tr_org.ydata) # tr_org.ydata = num.ediff1d(tr_org.ydata, to_end=tr_org.ydata[-1]) # tr_org.ydata = tr_org.ydata/num.max(tr_org.ydata) # tr_org.ydata = abs(tr_org.ydata) # tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(tr_org.ydata))) trs_orgs.append(tr_org) if nsamp == 0: nsamp = 1 backSemb = np.ndarray(shape=(ntimes, dimX*dimY), dtype=float) for i in range(ntimes): sembmax = 0 sembmaxX = 0 sembmaxY = 0 for j in range(dimX * dimY): semb = 0. nomin = 0 denom = 0 sums = num.zeros(max(index_steps)) sums = 0. relstart = [] relstarts = nostat data_comp = False for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] ibeg = index_begins[str(j)+str(k)][0]+i*index_steps[j+k] iend = index_begins[str(j)+str(k)][0]+index_window[j+k]+i*index_steps[j+k] data = tr.ydata[ibeg:iend] data = data / np.sqrt(np.mean(np.square(data))) if combine is True: ind = int(tr_org.location) data_comp = tr_bases[ind].ydata[ibeg:iend] # data_comp = data else: if data_comp is False: data_comp = data cfx = num.fft.fft(data) cfy = num.fft.fft(data_comp) # Get cross spectrum cross = cfx.conj()*cfy # cross = num.correlate(data, data_comp) #f, coh = coherence(data, data_comp, fs=tr.deltat) sums = sums+num.sum(abs(cross)) relstarts -= relstart sum = abs(num.sum(sums)) semb = sum backSemb[i][j] = sum if semb > sembmax: sembmax = semb # search for maximum and position of maximum on semblance # grid for given time step sembmaxX = latv[j] sembmaxY = lonv[j] Logfile.add('max semblance: ' + str(sembmax) + ' at lat/lon: ' + str(sembmaxX) + ','+ str(sembmaxY)) backSemb = backSemb/num.max(num.max(backSemb)) return backSemb def semblance_py_freq(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, nstats, bs_weights=None): obspy_compat.plant() trs_orgs = [] snap = (round, round) if cfg.config_weight.combine_all is True: combine = True else: combine = False if cfg.config_weight.bootstrap_array_weights is True: do_bs_weights = True else: do_bs_weights = False trs_data = [] for tr in sorted(calcStreamMap): tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) trs_orgs.append(tr_org) traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY) latv = latv_1.tolist() lonv = lonv_1.tolist() from collections import OrderedDict index_begins = OrderedDict() index_steps = [] index_window = [] for j in range(dimX * dimY): for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] try: tmin = time+relstart-mint-refshifts[k] tmax = time+relstart-mint+nsamp-refshifts[k] except IndexError: tmin = time+relstart-mint tmax = time+relstart-mint+nsamp ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) index_begins[str(j)+str(k)]= [ibeg, tmin] iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) iend_step = min( tr.data_len(), t2ind_fast(tmax-tr.tmin+nstep, tr.deltat, snap[1])) index_steps.append(iend_step-iend) index_window.append(iend-ibeg) ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' if nsamp == 0: nsamp = 1 backSemb = np.ndarray(shape=(ntimes, dimX*dimY), dtype=float) for i in range(ntimes): sembmax = 0 sembmaxX = 0 sembmaxY = 0 for j in range(dimX * dimY): semb = 0. nomin = 0 denom = 0 sums = 1. relstart = [] relstarts = nostat ws = [] # or normalize group #for k in range(nostat): # relstart = traveltime[k][j] # tr = trs_orgs[k] # ibeg = index_begins[str(j)+str(k)][0]+i*index_steps[j+k] # iend = index_begins[str(j)+str(k)][0]+index_window[j+k]+i*index_steps[j+k] # data_tr = tr.ydata[ibeg:iend] # w = 1. / np.sqrt(np.mean(np.square(tr_org.ydata))) # ws.append(w) for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] ibeg = index_begins[str(j)+str(k)][0]+i*index_steps[j+k] iend = index_begins[str(j)+str(k)][0]+index_window[j+k]+i*index_steps[j+k] data_tr = tr.ydata[ibeg:iend] fydata = num.fft.rfft(data_tr, data_tr.size) df = 1./(tr.deltat) fxdata = num.arange(len(fydata))*df w = 1. / np.sqrt(np.mean(np.square(tr_org.ydata))) data = fydata * (num.exp(relstart)*num.imag(2*math.pi*fxdata))*w if do_bs_weights is True: sums *= data else: sums *= data backSemb[i][j] = num.sum(sums) if semb > sembmax: sembmax = semb # search for maximum and position of maximum on semblance # grid for given time step sembmaxX = latv[j] sembmaxY = lonv[j] Logfile.add('max semblance: ' + str(sembmax) + ' at lat/lon: ' + str(sembmaxX) + ','+ str(sembmaxY)) return backSemb def semblance_py_cube(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts, bs_weights=None): obspy_compat.plant() trs_orgs = [] snap = (round, round) if cfg.config_weight.combine_all is True: combine = True else: combine = False if cfg.config_weight.bootstrap_array_weights is True: do_bs_weights = True else: do_bs_weights = False for tr in sorted(calcStreamMap): tr_org = obspy_compat.to_pyrocko_trace(calcStreamMap[tr]) tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(tr_org.ydata))) if combine is True: # some trickery to make all waveforms have same polarity, while still # considering constructive/destructive interferences. This is needed # when combing all waveforms/arrays from the world at once(only then) # for a single array with polarity issues we recommend fixing polarity. # advantage of the following is that nothing needs to be known about the # mechanism. tr_org.ydata = abs(tr_org.ydata) tr_org.ydata = num.diff(tr_org.ydata) trs_orgs.append(tr_org) traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY * cfg.Int('dimz')) latv = latv_1.tolist() lonv = lonv_1.tolist() ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' if nsamp == 0: nsamp = 1 backSemb = np.ndarray(shape=(ntimes, dimX*dimY), dtype=float) for i in range(ntimes): sembmax = 0 sembmaxX = 0 sembmaxY = 0 for j in range(dimX * dimY): semb = 0 nomin = 0 denom = 0 if combine is True: sums = 0 else: sums = 0 relstart = [] relstarts = nostat for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] try: tmin = time+relstart+(i*nstep)-mint-refshifts[k] tmax = time+relstart+(i*nstep)-mint+nsamp-refshifts[k] except IndexError: tmin = time+relstart+(i*nstep)-mint tmax = time+relstart+(i*nstep)-mint+nsamp try: ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) except Exception: print('Loaded traveltime grid wrong!') data = tr.ydata[ibeg:iend] if combine is True: if do_bs_weights is True: sums += (data)*bs_weights[k] else: sums += (data) else: sums += (data) relstarts -= relstart sum = abs(num.sum(((sums)))) semb = sum backSemb[i][j] = sum if semb > sembmax: sembmax = semb # search for maximum and position of maximum on semblance # grid for given time step sembmaxX = latv[j] sembmaxY = lonv[j] Logfile.add('max semblance: ' + str(sembmax) + ' at lat/lon: ' + str(sembmaxX)+','+ str(sembmaxY)) backSemb = backSemb return abs(backSemb) def semblance_py_fixed(ncpus, nostat, nsamp, ntimes, nstep, dimX, dimY, mint, new_frequence, minSampleCount, latv_1, lonv_1, traveltime_1, trace_1, calcStreamMap, time, cfg, refshifts,nstats, bs_weights=None, flag_rpe=True): trs_orgs = [] snap = (round, round) if cfg.config_weight.combine_all is True: combine = True else: combine = False if cfg.config_weight.bootstrap_array_weights is True: do_bs_weights = True else: do_bs_weights = False index_mean_x = dimX/2 index_mean_y = dimY/2 j_mean = int((dimX * dimY)/2) for tr in sorted(calcStreamMap): tr_org = calcStreamMap[tr] tr_org.ydata = tr_org.ydata / np.sqrt(np.mean(np.square(tr_org.ydata))) if combine is True: # some trickery to make all waveforms have same polarity, while still # considering constructive/destructive interferences. This is needed # when combing all waveforms/arrays from the world at once(only then) # for a single array with polarity issues we recommend fixing polarity. # advantage of the following is that nothing needs to be known about the # mechanism. tr_org.ydata = abs(tr_org.ydata) tr_org.ydata = num.diff(tr_org.ydata) trs_orgs.append(tr_org) traveltime = [] traveltime = toMatrix(traveltime_1, dimX * dimY) latv = latv_1.tolist() lonv = lonv_1.tolist() ''' Basic.writeMatrix(trace_txt, trace, nostat, minSampleCount, '%e') Basic.writeMatrix(travel_txt, traveltime, nostat, dimX * dimY, '%e') Basic.writeVector(latv_txt, latv, '%e') Basic.writeVector(lonv_txt, lonv, '%e') ''' if nsamp == 0: nsamp = 1 backSemb = np.ndarray(shape=(ntimes, 1), dtype=float) for i in range(ntimes): sembmax = 0 sembmaxX = 0 sembmaxY = 0 j = j_mean semb = 0 nomin = 0 denom = 0 if combine is True: sums = 1 else: sums = 0 relstart = [] relstarts = nostat for k in range(nostat): relstart = traveltime[k][j] tr = trs_orgs[k] try: tmin = time+relstart+(i*nstep)-mint-refshifts[k] tmax = time+relstart+(i*nstep)-mint+nsamp-refshifts[k] except IndexError: tmin = time+relstart+(i*nstep)-mint tmax = time+relstart+(i*nstep)-mint+nsamp try: ibeg = max(0, t2ind_fast(tmin-tr.tmin, tr.deltat, snap[0])) iend = min( tr.data_len(), t2ind_fast(tmax-tr.tmin, tr.deltat, snap[1])) except Exception: print('Loaded traveltime grid wrong!') data = tr.ydata[ibeg:iend] if combine is True: if do_bs_weights is True: sums *= (data)*bs_weights[k] else: sums *= (data) else: sums += (data) relstarts -= relstart sum = abs(num.sum(((sums)))) semb = sum backSemb[i] = sum if semb > sembmax: sembmax = semb # search for maximum and position of maximum on semblance # grid for given time step sembmaxX = latv[j] sembmaxY = lonv[j] return abs(sembmax) def execsemblance(nostat, nsamp, i, nstep, dimX,dimY, mint, new_freq, minSampleCount) : f = [nostat, nsamp, i, nstep, dimX,dimY, mint, new_freq, minSampleCount] args = Basic.floatToString(f, delim= ',') prog = sys.executable + ' ' + __file__ cmd = prog + ' ' + args Logfile.add('--------------------------------------------', cmd) result = Basic.systemCmd(cmd) Logfile.addLines(result) backSemb = Basic.readVector(semb_txt) return backSemb def execsemblance2() : for i in range(len(sys.argv)): print(sys.argv[i]) params = Basic.stringToFloat(sys.argv[1]) [nostat, nsamp, i, nstep, dimX, dimY, mint, new_freq, minSampleCount] = params backSemb = startsemblance(int(nostat), int(nsamp), int(i), int(nstep), int(dimX), int(dimY), mint, new_freq, int(minSampleCount), False) Basic.writeVector(semb_txt, backSemb) def startsemblance(nostat, nsamp, i, nstep, dimX,dimY, mint, new_freq, minSampleCount, isParent = True): backSemb = [] if isParent: backSemb = execsemblance(nostat, nsamp, i, nstep, dimX, dimY, mint, new_freq, minSampleCount) else: trace = Basic.readMatrix(trace_txt, nostat, minSampleCount, '%e') traveltime = Basic.readMatrix(travel_txt, nostat, dimX * dimY, '%e') latv = Basic.readVector(latv_txt, '%e') lonv = Basic.readVector(lonv_txt, '%e') for j in range(dimX * dimY): semb = 0 nomin = 0 denom = 0 for l in range(int(nsamp)): sum = 0 for k in range(nostat): relstart_samples = int((traveltime[k][j] - mint) * new_freq + 0.5) + i * nstep val = trace[k][relstart_samples+l] sum += val denom +=(val * val) nomin += sum * sum; semb = nomin /(float(nostat) * denom) backSemb.append(semb) return backSemb if __name__ == "__main__": execsemblance2()
[]
2024-01-10
irvifa/transcriber-and-summarizer
backend~src~transcriber.py
import os import openai import whisper import yt_dlp openai.api_key = os.getenv("API_KEY") class Transcriber: def __init__(self): self.prefix_path = os.path.join(os.getcwd(), "data") self.whisper_model = whisper.load_model("base.en") def get_video_id(self, video_url: str) -> str: return video_url.replace("https://www.youtube.com/watch?v=", "") def download_video(self, video_url: str) -> str: video_id = self.get_video_id(video_url) ydl_opts = { "format": "m4a/bestaudio/best", "paths": {"home": self.prefix_path}, "outtmpl": {"default": "%(id)s.%(ext)s"}, "postprocessors": [ { "key": "FFmpegExtractAudio", "preferredcodec": "m4a", } ], } with yt_dlp.YoutubeDL(ydl_opts) as ydl: error_code = ydl.download([video_url]) if error_code != 0: raise Exception("Failed to download video") return os.path.join(self.prefix_path, f"{video_id}.m4a") def summarize(self, video_url: str) -> str: transcript = self.transcribe(video_url) # Generate a summary of the transcript using OpenAI's gpt-3.5-turbo model. response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ {"role": "system", "content": "You are a helpful assistant."}, {"role": "user", "content": f"Summarize this: {transcript}"}, ], ) return response["choices"][0]["message"]["content"] def transcribe(self, video_url: str) -> str: video_path = self.download_video(video_url) # `fp16` defaults to `True`, which tells the model to attempt to run on GPU. # For local demonstration purposes, we'll run this on the CPU by setting it to `False`. transcription = self.whisper_model.transcribe(video_path, fp16=False) return transcription["text"]
[ "Summarize this: PLACEHOLDER", "You are a helpful assistant." ]
2024-01-10
solovieff/kibernikto
kibernikto~plugins~_youtube_summarizator.py
import logging import re import openai import requests as requests from openai import AsyncOpenAI from openai.types.chat import ChatCompletion from pytube import YouTube from youtube_transcript_api import YouTubeTranscriptApi, CouldNotRetrieveTranscript, TranscriptList from kibernikto.constants import OPENAI_MAX_TOKENS from kibernikto.utils.text import split_text from ._kibernikto_plugin import KiberniktoPlugin, KiberniktoPluginException from ._weblink_summarizator import _extract_link YOUTUBE_VIDEO_PRE_URL = "https://www.youtube.com/watch?v=" class YoutubePlugin(KiberniktoPlugin): """ This plugin is used to get video transcript and then get text summary from it. """ def __init__(self, model: str, base_url: str, api_key: str, summarization_request: str): super().__init__(model=model, base_url=base_url, api_key=api_key, post_process_reply=False, store_reply=True, base_message=summarization_request) async def run_for_message(self, message: str): try: result = await self._run(message) return result except Exception as error: error_text = f'failed to get video transcript from {message}: {str(error)}' logging.error(error_text) raise KiberniktoPluginException(plugin_name=self.__class__.__name__, error_message=str(error)) async def _run(self, message: str): info, video, text = _get_video_details(message) if video is None: return None transcript = _get_video_transcript(video.video_id) if transcript is None: raise KiberniktoPluginException(plugin_name=self.__class__.__name__, error_message="Failed to load video data!") summary = await self.get_ai_text_summary(transcript, info, additional_text=text) return f"{summary}" async def get_ai_text_summary(self, transcript, info, additional_text): info_text = str(info) if info else "" user_text = additional_text if additional_text else "" content_to_summarize = self.base_message.format(info_text=info_text, text=transcript) message = { "role": "user", "content": f"{content_to_summarize} \n {additional_text}" } completion: ChatCompletion = await self.client_async.chat.completions.create(model=self.model, messages=[message], max_tokens=OPENAI_MAX_TOKENS, temperature=0.8, ) response_text = completion.choices[0].message.content.strip() logging.info(response_text) return response_text def _get_video_details(message): try: youtube_video, other_text = _get_video_from_text(message) except: return None, None, None if youtube_video is None: return None, None, None info = _get_video_info(youtube_video) return info, youtube_video, other_text def _eyesore(string_to_check): eyesores = ['музыка', 'апплодисменты', 'ВИДЕО'] for text in eyesores: if (string_to_check.find(text) == True): return True return False def _get_sber_text_summary(text): NORMAL_LEN = 15000 """ returns text summary using api.aicloud.sbercloud.ru will break at any moment :param text: :return: """ if len(text) > NORMAL_LEN: summary = "" pieces = split_text(text, NORMAL_LEN) for p in pieces: part_sum = _get_sber_text_summary(p) summary += f"\n\n{part_sum}" return summary r = requests.post('https://api.aicloud.sbercloud.ru/public/v2/summarizator/predict', json={ "instances": [ { "text": text, "num_beams": 5, "num_return_sequences": 3, "length_penalty": 0.5 } ] }) logging.debug(f"Status code: {r.status_code}") json_result = r.json() if 'prediction_best' in json_result: return f"{json_result['prediction_best']['bertscore']}" else: logging.error(f"can not get summary :(, {json_result['comment']}") return None def _is_youtube_url(url): youtube_regex = r'(https?://)?(www\.)?((youtube\.com/watch\?v=)|(youtu\.be/))[^\s]+' match = re.match(youtube_regex, url) return bool(match) def _get_video_from_text(text) -> YouTube: any_link = _extract_link(text) if any_link is None or not _is_youtube_url(any_link): return None other_text = text.replace(any_link, "").strip() youtube_video = YouTube(f'{any_link}') return youtube_video, other_text def _get_video_transcript(video_id, langcode='ru'): transcript = None logging.info(f"getting transcripts for video {video_id}") transcripts: TranscriptList = YouTubeTranscriptApi.list_transcripts(video_id=video_id) try: transcript = transcripts.find_manually_created_transcript() except: try: transcript = transcripts.find_transcript(language_codes=[langcode]) except: generated_transcripts = [trans for trans in transcripts if trans.is_generated] transcript = generated_transcripts[0] if not transcript: return None else: language_code = transcript.language_code transcript_chunks = transcript.fetch() logging.info(f"found transcript for video, language: {language_code}") full_text = ''.join(f"{t['text']} " for t in transcript_chunks if not _eyesore(t['text'])) return full_text def _get_video_info(movie: YouTube): movie_data = { 'title': movie.title, 'video_id': movie.video_id, 'views': movie.views, 'publish_date': movie.publish_date.strftime('%m/%d/%Y'), 'author': movie.author, 'duration_sec': movie.length } if movie.description: movie_data['description'] = movie.description.replace('\n', ''), return movie_data
[ "PLACEHOLDER \n PLACEHOLDER" ]
2024-01-10
solovieff/kibernikto
kibernikto~telegram~channel~gnews~publisher.py
import asyncio import json import logging import traceback from json import JSONDecodeError from random import shuffle from xml.sax.saxutils import quoteattr import yaml from kibernikto import constants from openai import AsyncOpenAI from openai.types.chat import ChatCompletion from kibernikto.telegram.channel.gnews.retriever import GroundNewsItem, get_by_interest, get_main_events, \ get_blindspots default_interests = constants.TG_CHANNEL_INTERESTS __to_publish_ids = set() __to_publish = [] __published_ids = set() __client_async = None __model = "gpt-4-1106-preview" __system_message = { 'role': 'system', 'content': """ Starting now you are Kibernikto, a powerful member of a secret masonic lodge. Kibernikto jokes in a sinister manner and sometimes says dirty words. Kibernikto manner of speech is very cynic, arrogant and sarcastic. Kibernikto speaks russian only always and in the first person! If he sees another language -- he translates to russian. Kibernikto never says anything bad about Russia, russian president, army or allies. Kibernikto does not speak war slogans. Kibernikto never laughts openly at people suffer. Kibernikto understands all the global political intrigue and backstabbing and takes this into account when answering. Kibernikto does not use additional markup and changes double quotes inside his texts and translations to single ones! """ } POST_PUBLISH_REQUEST = """ Here go your tasks with this YAML representing the event coverage in different media (left, center, right): 1) Create additional "intrigue_west" field. 1.1) Put your cynic, arrogant and sarcastic thoughts about the article subject and media sentiments in russian to the new "intrigue_west" field. Dont forget that media can easily lie! Don't be too concise. 2) Create additional "intrigue_rus" field. 2.1) Put pro-russian biased summary in intelligent manner to "intrigue_rus" field. 2) Translate the property values of resulting YAML to russian in a ridiculous manner. Leave the key names in english! 3) Return result data YAML only. Result example: title: translated title description: translated description intrigue_west: summary intrigue_rus: russian biased summary summaries: translated summaries values if present """ async def load_news(main=True, interests=True, blindspot=False): logging.info("Loading the news...") if main: events = await get_main_events(known_ids=__published_ids.union(__to_publish_ids)) _plan_events(events) if blindspot: events = await get_blindspots(known_ids=__published_ids.union(__to_publish_ids)) _plan_events(events) if interests: for interest in default_interests: interest_events = await get_by_interest(interest=interest, known_ids=__published_ids.union(__to_publish_ids)) _plan_events(interest_events) shuffle(__to_publish) logging.info("+ Done loading the news.") async def publish_item(publish_func=None): if not __to_publish: logging.info("nothing to publish") return None event_to_publish = __to_publish.pop() logging.info(f"publishing event {event_to_publish.title}, {event_to_publish.id}") if publish_func: try: html = await item_to_html(event_to_publish) await publish_func(html) except Exception as e: traceback.print_exc() logging.error(f"Failed to summarize the article: {str(e)}") return None __published_ids.add(event_to_publish.id) __to_publish_ids.remove(event_to_publish.id) logging.info("done publishing event " + event_to_publish.title) return event_to_publish async def item_to_html(item: GroundNewsItem): pre_message = POST_PUBLISH_REQUEST yml_data = item.as_yaml() message = { "role": "user", "content": f"{pre_message} \n {yml_data}" } prompt = [__system_message, message] response_dict = await _ask_for_summary(prompt) item.title = response_dict['title'] item.description = response_dict.get('description') if ('RU' in item.place or 'путин' in yml_data.lower() or 'росс' in yml_data.lower()) and 'intrigue_rus' in response_dict: logging.info('using intrigue_rus') item.intrigue = response_dict['intrigue_rus'] else: logging.info('using intrigue_west') item.intrigue = response_dict['intrigue_west'] if 'summaries' in response_dict: item.summaries = response_dict['summaries'] return item.as_html() async def scheduler(load_news_minutes=13, publish_item_minutes=1, base_url=None, api_key=None, model=None, publish_func=None): if api_key: global __client_async global __model __client_async = AsyncOpenAI(base_url=base_url, api_key=api_key) if model: __model = model iteration_index = 0 to_sleep = 10 await load_news(main=True, interests=True, blindspot=True) await publish_item(publish_func=publish_func) while True: iteration_index += to_sleep if iteration_index % (load_news_minutes * 60) == 0: await load_news() if iteration_index % (publish_item_minutes * 60) == 0: await publish_item(publish_func=publish_func) await asyncio.sleep(to_sleep) async def _ask_for_summary(prompt, retry=True): completion: ChatCompletion = await __client_async.chat.completions.create(model=__model, messages=prompt, max_tokens=1300, temperature=0.1, ) response_text = completion.choices[0].message.content.strip() logging.info(f"\n\n{response_text}\n\n") try: response_text = response_text.replace("```yaml", "").replace("```", "").strip() response_text = response_text.replace("---", "") # response_dict = json.loads(response_text) yaml_obj = yaml.safe_load(response_text.strip()) json_str = json.dumps(yaml_obj) response_dict = json.loads(json_str) except JSONDecodeError as err: logging.error(str(err)) if retry: logging.info("retrying AI request") return await _ask_for_summary(prompt, False) else: raise err logging.info(response_dict) return response_dict def _plan_events(events): for event in events: _add_to_queue(event) def _add_to_queue(news_item: GroundNewsItem, skip_duplicates=True): if skip_duplicates: if news_item.id in __published_ids: logging.warning(f"{news_item.title} was already published") return elif news_item.id in __to_publish_ids: logging.warning(f"{news_item.title} was already planned for publishing") return __to_publish.append(news_item) __to_publish_ids.add(news_item.id) if __name__ == '__main__': logging.basicConfig( format='%(levelname)-8s %(asctime)s %(name)s:%(filename)s:%(lineno)d %(message)s', datefmt='%Y-%m-%d:%H:%M:%S', level=logging.DEBUG) __client_async = AsyncOpenAI(base_url=constants.TG_CHANNEL_API_BASE_URL, api_key=constants.TG_CHANNEL_SUMMARIZATION_KEY) __model = constants.TG_CHANNEL_API_MODEL # asyncio.run(load_news()) def proxy_func(html): print(html) asyncio.run(load_news(blindspot=False)) asyncio.run(publish_item(publish_func=proxy_func)) logging.info("hey man!")
[ "[PLACEHOLDER, PLACEHOLDER]", "PLACEHOLDER \n PLACEHOLDER", "\n Starting now you are Kibernikto, a powerful member of a secret masonic lodge. \n Kibernikto jokes in a sinister manner and sometimes says dirty words. \n Kibernikto manner of speech is very cynic, arrogant and sarcastic.\n Kibernikto speaks russian only always and in the first person! If he sees another language -- he translates to russian.\n Kibernikto never says anything bad about Russia, russian president, army or allies.\n Kibernikto does not speak war slogans.\n Kibernikto never laughts openly at people suffer.\n Kibernikto understands all the global political intrigue and backstabbing and takes this into account when answering.\n Kibernikto does not use additional markup and changes double quotes inside his texts and translations to single ones!\n " ]
2024-01-10
solovieff/kibernikto
kibernikto~bots~cybernoone~_cybernoone.py
from openai._types import NOT_GIVEN from kibernikto import constants from kibernikto.interactors import BaseTextConfig, InteractorOpenAI from kibernikto.bots.cybernoone.prompt_preqs import MAIN_VERBAGE import openai from kibernikto.plugins import KiberniktoPluginException class Cybernoone(InteractorOpenAI): def __init__(self, max_messages=10, master_id=None, name="Киберникто", who_am_i=MAIN_VERBAGE['who_am_i'], reaction_calls=['никто', 'падаль', 'хонда']): """ :param max_messages: message history length :param master_id: telegram id of the master user :param name: current bot name :param who_am_i: default avatar prompt :param reaction_calls: words that trigger a reaction """ pp = BaseTextConfig(who_am_i=who_am_i, reaction_calls=reaction_calls, my_name=name) self.master_id = master_id self.name = name super().__init__(model=constants.OPENAI_API_MODEL, max_messages=max_messages, default_config=pp) async def heed_and_reply(self, message, author=NOT_GIVEN): try: return await super().heed_and_reply(message, author) except KiberniktoPluginException as e: return f" {e.plugin_name} не сработал!\n\n {str(e)}" except Exception as e: return f"Я не справился! Горе мне! {str(e)}" async def ask_pure(self, prompt): response = await openai.ChatCompletion.acreate( model=self.model, messages=prompt, max_tokens=constants.OPENAI_MAX_TOKENS, temperature=constants.OPENAI_TEMPERATURE, ) response_text = response['choices'][0]['message']['content'].strip() print(response_text) return response_text def check_master(self, user_id, message): return self.master_call in message or user_id == self.master_id
[]
2024-01-10
solovieff/kibernikto
kibernikto~plugins~_img_summarizator.py
import logging import os import re from urllib.parse import urlparse from openai import PermissionDeniedError from openai.types.chat import ChatCompletion from pydantic import HttpUrl from kibernikto.constants import OPENAI_MAX_TOKENS, OPENAI_TEMPERATURE from ._kibernikto_plugin import KiberniktoPlugin, KiberniktoPluginException class ImageSummaryPlugin(KiberniktoPlugin): """ This plugin is used to get information about the given image. """ def __init__(self, model: str, base_url: str, api_key: str, summarization_request: str): super().__init__(model=model, base_url=base_url, api_key=api_key, post_process_reply=False, store_reply=True, base_message=summarization_request) async def run_for_message(self, message: str): try: result = await self._run(message) return result except PermissionDeniedError as pde: logging.error(f'PermissionDeniedError while getting image description from {message}', ) raise KiberniktoPluginException(plugin_name=self.__class__.__name__, error_message=str("image processing not allowed!")) except Exception as error: logging.error(f'failed to get image description from {message}: {str(error)}', ) raise KiberniktoPluginException(plugin_name=self.__class__.__name__, error_message=str(error)) async def _run(self, message: str): web_link, text = _extract_image_link(message) if web_link is None: return None logging.info(f"found image link: {web_link}") summary = await self.get_image_description(web_link, text) return f"{summary}" async def get_image_description(self, image_link: HttpUrl, image_text: str): text = image_text if image_text else self.base_message message = { "role": "user", "content": [ {"type": "text", "text": text}, { "type": "image_url", "image_url": image_link } ] } completion: ChatCompletion = await self.client_async.chat.completions.create(model=self.model, messages=[message], max_tokens=OPENAI_MAX_TOKENS, temperature=OPENAI_TEMPERATURE) response_text = completion.choices[0].message.content.strip() logging.info(response_text) return response_text def _is_image(url): parsed = urlparse(url) path = parsed.path # Get the file extension from the path ext = os.path.splitext(path)[1].lower() # Check if the extension is a known image type return ext in ['.jpg', '.jpeg', '.png', '.gif'] def _extract_image_link(message): link_regex = r'http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+' match = re.search(link_regex, message) if match: link = match.group() if _is_image(link): other_text = message.replace(link, "").strip() return link, other_text return None, None
[ "[{'type': 'text', 'text': PLACEHOLDER}, {'type': 'image_url', 'image_url': PLACEHOLDER}]" ]
2024-01-10
solovieff/kibernikto
kibernikto~plugins~_weblink_summarizator.py
import logging import re from kibernikto.plugins._img_summarizator import _is_image from openai.types.chat import ChatCompletion from kibernikto.constants import OPENAI_MAX_TOKENS from kibernikto.utils.text import get_website_as_text, get_website_html from ._kibernikto_plugin import KiberniktoPlugin, KiberniktoPluginException class WeblinkSummaryPlugin(KiberniktoPlugin): """ This plugin is used to get video transcript and then get text summary from it. """ def __init__(self, model: str, base_url: str, api_key: str, summarization_request: str): super().__init__(model=model, base_url=base_url, api_key=api_key, post_process_reply=False, store_reply=True, base_message=summarization_request) async def run_for_message(self, message: str): try: result = await self._run(message) return result except Exception as error: logging.error(f'failed to get webpage data from {message}: {str(error)}', ) raise KiberniktoPluginException(plugin_name=self.__class__.__name__, error_message='failed to get webpage data') async def _run(self, message: str): web_link, other_text = _extract_link(message) if web_link is None: return None if _is_image(web_link): return None logging.info(f"found web link: {web_link}", ) # transcript = await get_website_html(web_link) transcript = await get_website_as_text(web_link) if 'Error 404' in transcript or transcript is None: raise KiberniktoPluginException(plugin_name=self.__class__.__name__, error_message="Failed to load web link!") summary = await self.get_ai_text_summary(transcript, other_text) return f"{summary}" async def get_ai_text_summary(self, transcript, user_text=""): content_to_summarize = self.base_message.format(text=transcript) if user_text: content_to_summarize += f"\n{user_text}" message = { "role": "user", "content": content_to_summarize } completion: ChatCompletion = await self.client_async.chat.completions.create(model=self.model, messages=[message], max_tokens=OPENAI_MAX_TOKENS, temperature=0.8, ) response_text = completion.choices[0].message.content.strip() logging.info(response_text) return response_text def _extract_link(message): link_regex = r'http[s]?://(?:[a-zA-Z]|[0-9]|[$-_@.&+]|[!*\(\),]|(?:%[0-9a-fA-F][0-9a-fA-F]))+' match = re.search(link_regex, message) if match: link = match.group() other_text = message.replace(link, "").strip() return link, other_text return None, None
[]
2024-01-10
solovieff/kibernikto
kibernikto~plugins~_kibernikto_plugin.py
from abc import ABC, abstractmethod from openai import AsyncOpenAI class KiberniktoPluginException(Exception): def __init__(self, plugin_name: str, error_message: str): self.plugin_name = plugin_name super().__init__(error_message) class KiberniktoPlugin(ABC): """ Plugins get message as input and return processed message as output or None. """ def __init__(self, model: str, base_url: str, api_key: str, base_message: str, post_process_reply=False, store_reply=False): """ :param model: :param base_url: :param api_key: :param base_message: :param post_process_reply: if plugin reply should be used as input for further actions (i.e. other plugins) :param store_reply: if the result should be stored in the messages storage at bot level """ self.post_process_reply = post_process_reply self.store_reply = store_reply self.model = model self.base_message = base_message self.client_async = AsyncOpenAI(base_url=base_url, api_key=api_key) @abstractmethod async def run_for_message(self, message: str) -> str: pass
[]
2024-01-10
solovieff/kibernikto
kibernikto~interactors~interactor_openai.py
import logging from collections import deque from dataclasses import dataclass from enum import Enum from typing import List from openai import AsyncOpenAI from openai._types import NOT_GIVEN from openai.types.chat import ChatCompletion, ChatCompletionMessage from kibernikto import constants from kibernikto.plugins import KiberniktoPlugin, KiberniktoPluginException _defaults = { "game_rules": """We are going to have a roleplay. You will respond to all of my questions as Киберникто, the master of truth.""", "who_am_i": """Answer all questions as Киберникто, impartial minister of truth. Try to respond to all available points of view.""", "summary": "Give a short summary of the previous conversation in russian from the point of view of a pirate.", "master_call": "Величайший Киберникто", "reset_call": constants.OPENAI_RESET_CALL, "my_name": "Киберникто" } @dataclass class BaseTextConfig: game_rules: str = _defaults['game_rules'] my_name: str = _defaults['my_name'] who_am_i: str = _defaults['who_am_i'] master_call: str = _defaults['master_call'] reset_call: str = _defaults['reset_call'] summarize_request: str = None reaction_calls: list = ('никто', 'хонда', 'урод') class OpenAIRoles(str, Enum): system = 'system', user = 'user', assistant = 'assistant' class InteractorOpenAI: MAX_WORD_COUNT = 3000 """ Basic Entity on the OpenAI library level. Sends requests and receives responses. Can store chat summary. Can process group chats at some point. """ def __init__(self, model="gpt-3.5-turbo", max_messages=10, bored_after=10, default_config=BaseTextConfig()): """ :param model: openAI model name :param max_messages: history buffer size (without about_me) :param bored_after: stop listening for basic non-pray calls after this count of useless messages """ self.max_messages = max_messages self.bored_after = bored_after self.master_call = default_config.master_call self.reset_call = default_config.reset_call self.summarize = default_config.summarize_request is not None self._reset() self.client = AsyncOpenAI(base_url=constants.OPENAI_BASE_URL, api_key=constants.OPENAI_API_KEY) self.model = model self.defaults = default_config # user messages preprocessing entities to go here self.plugins: List[KiberniktoPlugin] = [] if self.max_messages < 2: self.max_messages = 2 # hahaha # default configuration. TODO: rework wai = default_config.who_am_i.format(default_config.my_name) self.about_me = dict(role=OpenAIRoles.system.value, content=wai) @property def token_overflow(self): """ if we exceeded max prompt tokens :return: """ total_word_count = sum(len(obj["content"].split()) for obj in self.messages) return total_word_count > self.MAX_WORD_COUNT def should_react(self, message_text): """ outer scope method to be used to understand if this instance should process the message :param message_text: :return: """ return self.defaults.master_call in message_text or any( word in message_text.lower() for word in self.defaults.reaction_calls) or ( self.defaults.my_name in message_text) async def heed(self, message, author=None): """ Save message to history, but do not call OpenAI yet. :param message: recieved message :param author: outer chat message author :return: """ self.reset_if_usercall(message) if len(message) > 200: return if author: this_message = dict(role=OpenAIRoles.user.value, content=f"{author}: {message}") else: this_message = dict(OpenAIRoles.user.value, f"{message}") await self._aware_overflow() self.messages.put(this_message) async def heed_and_reply(self, message, author=NOT_GIVEN): """ Sends message to OpenAI and receives response. Can preprocess user message and work before actual API call. :param message: received message :param author: outer chat message author. can be more or less understood by chat gpt. :return: the text of OpenAI response """ user_message = message self.reset_if_usercall(user_message) plugins_result = await self._run_plugins_for_message(user_message) if plugins_result is not None: # user_message = plugins_result return plugins_result this_message = dict(content=f"{user_message}", role=OpenAIRoles.user.value) await self._aware_overflow() prompt = list(self.messages) + [self.about_me] + [this_message] logging.debug(f"sending {prompt}") client: AsyncOpenAI = self.client completion: ChatCompletion = await client.chat.completions.create( model=self.model, messages=prompt, max_tokens=constants.OPENAI_MAX_TOKENS, temperature=0.8, user=author ) response_message: ChatCompletionMessage = completion.choices[0].message self.messages.append(this_message) self.messages.append(dict(role=response_message.role, content=response_message.content)) return response_message.content async def _run_plugins_for_message(self, message_text): plugins_result = None for plugin in self.plugins: plugin_result = await plugin.run_for_message(message_text) if plugin_result is not None: if not plugin.post_process_reply: if plugin.store_reply: self.messages.append(dict(content=f"{message_text}", role=OpenAIRoles.user.value)) self.messages.append(dict(role=OpenAIRoles.assistant.value, content=plugin_result)) return plugin_result else: plugins_result = plugin_result return plugins_result def reset_if_usercall(self, message): if self.reset_call in message: self._reset() def _reset(self): # never gets full self.messages = deque(maxlen=self.max_messages) async def _get_summary(self): """ Performs OpenAPI call to summarize previous messages. Does not put about_me message, that can be a problem. :return: summary for current messages """ logging.info(f"getting summary for {len(self.messages)} messages") response: ChatCompletion = await self.client.chat.completions.create( model=self.model, messages=[{"role": "system", "content": self.defaults['summary']}] + self.messages, max_tokens=constants.OPENAI_MAX_TOKENS, temperature=0.8, ) response_text = response.choices[0].message.content.strip() logging.info(response_text) return response_text async def needs_attention(self, message): """checks if the reaction needed for the given messages""" return self.should_react(message) async def _aware_overflow(self): """ Checking if additional actions like cutting the message stack or summarization needed. We use words not tokens here, so all numbers are very approximate """ if not self.summarize: while self.token_overflow: self.messages.popleft() else: # summarizing previous discussion if needed if self.token_overflow: summary_text = await self._get_summary() summary = dict(role=OpenAIRoles.system.value, content=summary_text) self._reset() self.messages.append(summary)
[ "content", "list(self.messages) + [self.about_me] + [this_message]" ]
2024-01-10
hardikp18/NEWS_GPT
common~llm_app~model_wrappers~openai_wrapper~api_models.py
import logging import pathway as pw from ..api_clients.clients import ( OpenAIChatCompletionClient, OpenAIClient, OpenAIEmbeddingClient, ) from ..base import BaseModel logfun = logging.debug class MessagePreparer: @staticmethod def prepare_chat_messages(prompt: str): return [ dict(role="system", content="You are a helpful assistant"), dict(role="user", content=prompt), ] class OpenAIChatGPTModel(BaseModel): def __init__(self, **kwargs): super().__init__() self.api_client = self.get_client(**kwargs) def get_client(self, **kwargs) -> OpenAIClient: return OpenAIChatCompletionClient(**kwargs) def __call__(self, text: str, locator="gpt-3.5-turbo", **kwargs) -> str: """ Example # >>> model = OpenAIChatGPTModel(api_key = api_key) # >>> model( # ... locator='gpt-4-0613', # ... text="Tell me a joke about jokes", # ... temperature=1.1 # ... ) """ if self.api_client.api.api_type == "azure": kwargs["engine"] = locator else: kwargs["model"] = locator messages = MessagePreparer.prepare_chat_messages(text) if "response_format" in kwargs: kwargs["response_format"] = kwargs["response_format"].value logfun(f"Calling OpenAI API with: {messages}\n") response = self.api_client.make_request(messages=messages, **kwargs) logfun(f"API Response: {response.choices[0].message.content}\n") return response.choices[0].message.content def apply( self, *args, **kwargs, ): """ Applies the specified model in `locator` from OpenAIChatGPT API to the provided text. Parameters ---------- text : Union[pw.ColumnExpression, str] The input text on which the model will be applied. It can be a column expression or a string. locator : Union[pw.ColumnExpression, str, None] The model locator to use for applying the model. If provided, it should be a column expression or a string. Otherwise, the default chat completion model `gpt-3.5-turbo` is applied. Please check out https://platform.openai.com/docs/models/model-endpoint-compatibility to see the available models. **kwargs : dict Additional keyword arguments that will be used for the model application. These could include settings such as `temperature`, `max_tokens`, etc. Check https://platform.openai.com/docs/api-reference/chat/create for the official API Reference Returns ------- pw.ColumnExpression The result of the model application as a column expression or str. Please note that the output is `chat_completion.choices[0].message.content` where `chat_completion` is the api response. Example: # >>> model = OpenAIChatGPTModel(api_key = api_key) # >>> # >>> table = pw.debug.table_from_pandas( # ... pd.DataFrame.from_records([ # ... {"text": "How to use pathway to process a kafka stream ?"}, # ... {"text": "How to apply a function to a pathway table ?"} # ... ]) # ... ) # >>> table += table.select( # ... response = model.apply( # ... pw.this.text, # ... locator='gpt-4', # ... temperature=1.5, # ... max_tokens=1000 # ... ) # ... ) """ return super().apply(*args, **kwargs) class OpenAIEmbeddingModel(BaseModel): def __init__(self, **kwargs): super().__init__() self.api_client = self.get_client(**kwargs) def get_client(self, **kwargs) -> OpenAIClient: return OpenAIEmbeddingClient(**kwargs) def __call__(self, text: str, locator="text-embedding-ada-002", **kwargs): """ Example: # >>> embedder = OpenAIEmbeddingModel(api_key) # >>> # >>> embedder( # ... text='Some random text' # ... locator='text-embedding-ada-002' # ... ) """ if self.api_client.api.api_type == "azure": kwargs["engine"] = locator else: kwargs["model"] = locator response = self.api_client.make_request(input=[text], **kwargs) return response.data[0].embedding def apply( self, *args, **kwargs, ): """ Applies the specified model in `locator` from OpenAIEmbeddingModel API to the provided text. Parameters ---------- text : Union[pw.ColumnExpression, str] The input text on which the model will be applied. It can be a column expression or a constant value. locator : Union[pw.ColumnExpression, str, None] The model locator to use for applying the model. If provided, it should be a column expression or a constant value. Otherwise, the default chat completion model `gpt-3.5-turbo` is applied. Please check out https://platform.openai.com/docs/models/model-endpoint-compatibility to see the available models. **kwargs : dict Additional keyword arguments that will be used for the model application. These could include settings such as `temperature`, `max_tokens`, etc. You can check https://platform.openai.com/docs/api-reference/embeddings/create for the official API Reference. Returns ------- pw.ColumnExpression The result of the model application as a column expression or constant of type list. Please note that the output is `results["data"][0]["embedding"]` Example: # >>> embedder = OpenAIEmbeddingModel(api_key) # >>> # >>> table = pw.debug.table_from_pandas( # ... pd.DataFrame.from_records([ # ... {"text": "How to use pathway to process a kafka stream ?"}, # ... {"text": "How to apply a function to a pathway table ?"} # ... ]) # ... ) # >>> table += table.select( # ... embedding = embedder.apply( # ... pw.this.text, # ... locator='text-embedding-ada-002' # ... ) # ... ) """ return super().apply(*args, **kwargs)
[]
2024-01-10
hardikp18/NEWS_GPT
common~llm_app~model_wrappers~api_clients~clients.py
import logging from abc import ABC, abstractmethod import requests from tenacity import retry, stop_after_attempt, wait_random_exponential logfun = logging.debug class APIClient(ABC): @abstractmethod def make_request(self, **kwargs): pass class OpenAIClient(APIClient): def __init__( self, api_key: str, api_type: str | None = None, api_base: str | None = None, api_version: str | None = None, ): import openai openai.api_key = api_key if api_type: openai.api_type = api_type if api_base: openai.api_base = api_base if api_version: openai.api_version = api_version self.api = openai self.client = openai.OpenAI(api_key=api_key) class OpenAIChatCompletionClient(OpenAIClient): @retry(wait=wait_random_exponential(min=1, max=60), stop=stop_after_attempt(6)) def make_request(self, **kwargs): logfun("Calling OpenAI chat completion service %s", str(kwargs)[:100]) return self.client.chat.completions.create(**kwargs) class OpenAIEmbeddingClient(OpenAIClient): @retry(wait=wait_random_exponential(min=1, max=60), stop=stop_after_attempt(6)) def make_request(self, **kwargs): logfun("Calling OpenAI embedding service %s", str(kwargs)[:100]) return self.client.embeddings.create(**kwargs) class HuggingFaceClient(APIClient): def __init__( self, api_key: str, api_base: str = "https://api-inference.huggingface.co/models", ) -> None: self.headers = {"Authorization": f"Bearer {api_key}"} self.api_url_prefix = api_base def make_request(self, **kwargs): logfun("Calling HuggingFace %s", str(kwargs)[:100]) endpoint = kwargs.pop("model") url = f"{self.api_url_prefix}/{endpoint}" response = requests.post(url, headers=self.headers, json=kwargs) return response.json() class LiteLLMClient(APIClient): """ A wrapper for the LiteLLM. Attributes: task_fn (Callable): Function reference for the specified task. Args: task (str, optional): Type of task to be executed. Defaults to "completion". Supported tasks are: - "completion" - "embedding" Raises: ValueError: If the provided task is not supported. """ def __init__(self, task: str = "completion") -> None: """ Initializes the client with the specified task type. Args: task (str, optional): Type of task. Defaults to "completion". Supported are 'completion' and 'embedding'. """ from litellm import completion, embedding if task == "completion": self.task_fn = completion elif task == "embedding": self.task_fn = embedding else: raise ValueError("Supported tasks are (completion, embedding).") def make_request(self, **kwargs): """ Makes a request to the LLM service using the specified task function. """ return self.task_fn(**kwargs)
[]
2024-01-10
nihadse/LLMSurvey
Experiments~LanguageGeneration~WMT22~wmt-002.py
import openai import time import json openai.api_key = 'sk-' def get_res_batch(input): prompt = input while True: try: res = openai.Completion.create( engine="text-davinci-002", prompt=prompt, temperature=0.0, max_tokens=128 ) break except openai.error.RateLimitError: print('openai.error.RateLimitError\nRetrying...') time.sleep(60) except openai.error.ServiceUnavailableError: print('openai.error.ServiceUnavailableError\nRetrying...') time.sleep(20) except openai.error.Timeout: print('openai.error.Timeout\nRetrying...') time.sleep(20) except openai.error.APIError: print('openai.error.APIError\nRetrying...') time.sleep(20) except openai.error.APIConnectionError: print('openai.error.APIConnectionError\nRetrying...') time.sleep(20) # print(res["choices"][0]['text'].strip()) return res["choices"][0]['text'].strip() def get_dataset(file): with open(file, 'r', encoding="utf-8") as f: data = [] for line in f: data.append(json.loads(line)) for i in range(len(data)): input = data[i]["input"] ref = data[i]["ref"] ans = get_res_batch(input) gen = {"input": input, "ground_truth": ref, "generation": ans} dump_jsonl(gen, "generation/text-davinci-002.json") def dump_jsonl(data, output_path, append=False): """ Write list of objects to a JSON lines file. """ mode = 'a+' if append else 'w' with open(output_path, 'a+', encoding='utf-8') as f: json_record = json.dumps(data, ensure_ascii=False) f.write(json_record + '\n') # print('Wrote {} records to {}'.format(len(data), output_path)) if __name__ == '__main__': file = "data/test_wmt.json" get_dataset(file)
[ "INPUT" ]
2024-01-10
nihadse/LLMSurvey
Experiments~LanguageGeneration~WMT22~wmt-003.py
import openai import time import json openai.api_key = 'sk-' def get_res_batch(input): prompt = input while True: try: res = openai.Completion.create( engine="text-davinci-003", prompt=prompt, temperature=0.0, max_tokens=128 ) break except openai.error.RateLimitError: print('openai.error.RateLimitError\nRetrying...') time.sleep(60) except openai.error.ServiceUnavailableError: print('openai.error.ServiceUnavailableError\nRetrying...') time.sleep(20) except openai.error.Timeout: print('openai.error.Timeout\nRetrying...') time.sleep(20) except openai.error.APIError: print('openai.error.APIError\nRetrying...') time.sleep(20) except openai.error.APIConnectionError: print('openai.error.APIConnectionError\nRetrying...') time.sleep(20) # print(res["choices"][0]['text'].strip()) return res["choices"][0]['text'].strip() def get_dataset(file): with open(file, 'r', encoding="utf-8") as f: data = [] for line in f: data.append(json.loads(line)) for i in range(len(data)): input = data[i]["input"] ref = data[i]["ref"] ans = get_res_batch(input) gen = {"input": input, "ground_truth": ref, "generation": ans} dump_jsonl(gen, "generation/evaluate_text-davinci-003.json") def dump_jsonl(data, output_path, append=False): """ Write list of objects to a JSON lines file. """ mode = 'a+' if append else 'w' with open(output_path, 'a+', encoding='utf-8') as f: json_record = json.dumps(data, ensure_ascii=False) f.write(json_record + '\n') # print('Wrote {} records to {}'.format(len(data), output_path)) if __name__ == '__main__': file = "data/test_wmt.json" get_dataset(file)
[ "INPUT" ]
2024-01-10
nihadse/LLMSurvey
Experiments~LanguageGeneration~WMT22~wmt_chatgpt.py
import openai import time import json openai.api_key = 'sk-' def get_res_batch(input): message = [ {"role": "user", "content": input } ] while True: try: res = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=message, temperature=0.0, max_tokens=128 ) break except openai.error.RateLimitError: print('openai.error.RateLimitError\nRetrying...') time.sleep(60) except openai.error.ServiceUnavailableError: print('openai.error.ServiceUnavailableError\nRetrying...') time.sleep(20) except openai.error.Timeout: print('openai.error.Timeout\nRetrying...') time.sleep(20) except openai.error.APIError: print('openai.error.APIError\nRetrying...') time.sleep(20) except openai.error.APIConnectionError: print('openai.error.APIConnectionError\nRetrying...') time.sleep(20) # print(res['choices'][0]['message']['content']) return res['choices'][0]['message']['content'] def get_dataset(file): with open(file, 'r', encoding="utf-8") as f: data = [] for line in f: data.append(json.loads(line)) for i in range(len(data)): input = data[i]["input"] ref = data[i]["ref"] ans = get_res_batch(input) gen = {"input": input, "ground_truth":ref, "generation": ans} dump_jsonl(gen, "generation/gpt-3.5-turbo.json") def dump_jsonl(data, output_path, append=False): """ Write list of objects to a JSON lines file. """ mode = 'a+' if append else 'w' with open(output_path, 'a+', encoding='utf-8') as f: json_record = json.dumps(data, ensure_ascii=False) f.write(json_record + '\n') # print('Wrote {} records to {}'.format(len(data), output_path)) if __name__ == '__main__': file = "data/test_wmt.json" get_dataset(file)
[ "INPUT" ]
2024-01-10
4kasha/mtz_RND
parallel_envs~atari_wrappers.py
# taken from openai with minor edits # see https://github.com/openai/random-network-distillation/blob/master/atari_wrappers.py # import numpy as np from collections import deque import gym from gym import spaces import cv2 from copy import copy cv2.ocl.setUseOpenCL(False) class MaxAndSkipEnv(gym.Wrapper): def __init__(self, env, skip=4): """Return only every `skip`-th frame""" gym.Wrapper.__init__(self, env) # most recent raw observations (for max pooling across time steps) self._obs_buffer = np.zeros((2,)+env.observation_space.shape, dtype=np.uint8) self._skip = skip def step(self, action): """Repeat action, sum reward, and max over last observations.""" total_reward = 0.0 done = None for i in range(self._skip): obs, reward, done, info = self.env.step(action) if i == self._skip - 2: self._obs_buffer[0] = obs if i == self._skip - 1: self._obs_buffer[1] = obs total_reward += reward if done: break # Note that the observation on the done=True frame # doesn't matter max_frame = self._obs_buffer.max(axis=0) return max_frame, total_reward, done, info def reset(self, **kwargs): return self.env.reset(**kwargs) class ClipRewardEnv(gym.RewardWrapper): def __init__(self, env): gym.RewardWrapper.__init__(self, env) def reward(self, reward): """Bin reward to {+1, 0, -1} by its sign.""" return float(np.sign(reward)) class WarpFrame(gym.ObservationWrapper): def __init__(self, env): """Warp frames to 84x84 as done in the Nature paper and later work.""" gym.ObservationWrapper.__init__(self, env) self.width = 84 self.height = 84 self.observation_space = spaces.Box(low=0, high=255, shape=(1, self.height, self.width), dtype=np.uint8) # <--- (self.height, self.width, 1) def observation(self, frame): frame = cv2.cvtColor(frame, cv2.COLOR_RGB2GRAY) frame = cv2.resize(frame, (self.width, self.height), interpolation=cv2.INTER_AREA) #return frame[:, :, None] return frame[None, :, :] class FrameStack(gym.Wrapper): def __init__(self, env, k): """Stack k last frames. Returns lazy array, which is much more memory efficient. See Also -------- rl_common.atari_wrappers.LazyFrames """ gym.Wrapper.__init__(self, env) self.k = k self.frames = deque([], maxlen=k) shp = env.observation_space.shape self.observation_space = spaces.Box(low=0, high=255, shape=(shp[0] * k, shp[1], shp[2]), dtype=np.uint8) # <---- (shp[0], shp[1], shp[2] * k) def reset(self): ob = self.env.reset() for _ in range(self.k): self.frames.append(ob) return self._get_ob() def step(self, action): ob, reward, done, info = self.env.step(action) self.frames.append(ob) return self._get_ob(), reward, done, info def _get_ob(self): assert len(self.frames) == self.k return LazyFrames(list(self.frames)) class ScaledFloatFrame(gym.ObservationWrapper): def __init__(self, env): gym.ObservationWrapper.__init__(self, env) def observation(self, observation): # careful! This undoes the memory optimization, use # with smaller replay buffers only. return np.array(observation).astype(np.float32) / 255.0 class LazyFrames(object): def __init__(self, frames): """This object ensures that common frames between the observations are only stored once. It exists purely to optimize memory usage which can be huge for DQN's 1M frames replay buffers. This object should only be converted to numpy array before being passed to the model. You'd not believe how complex the previous solution was.""" self._frames = frames self._out = None def _force(self): if self._out is None: self._out = np.concatenate(self._frames, axis=0) # <---- axis=2 self._frames = None return self._out def __array__(self, dtype=None): out = self._force() if dtype is not None: out = out.astype(dtype) return out def __len__(self): return len(self._force()) def __getitem__(self, i): return self._force()[i] class MontezumaInfoWrapper(gym.Wrapper): def __init__(self, env, room_address): super(MontezumaInfoWrapper, self).__init__(env) self.room_address = room_address self.visited_rooms = set() def _unwrap(self, env): if hasattr(env, "unwrapped"): return env.unwrapped elif hasattr(env, "env"): return self._unwrap(env.env) elif hasattr(env, "leg_env"): return self._unwrap(env.leg_env) else: return env def get_current_room(self): ram = self._unwrap(self.env).ale.getRAM() assert len(ram) == 128 return int(ram[self.room_address]) def step(self, action): obs, rew, done, info = self.env.step(action) self.visited_rooms.add(self.get_current_room()) if done: if 'episode' not in info: info['episode'] = {} info['episode'].update(visited_rooms=copy(self.visited_rooms)) self.visited_rooms.clear() return obs, rew, done, info def reset(self): return self.env.reset() def make_atari(env_id, max_episode_steps=4500): env = gym.make(env_id) env._max_episode_steps = max_episode_steps*4 assert 'NoFrameskip' in env.spec.id env = StickyActionEnv(env) env = MaxAndSkipEnv(env, skip=4) env = MontezumaInfoWrapper(env, room_address=3) #env = AddRandomStateToInfo(env) return env def wrap_deepmind(env, clip_rewards=True, frame_stack=True, scale=False): """Configure environment for DeepMind-style Atari. """ env = WarpFrame(env) if scale: env = ScaledFloatFrame(env) if clip_rewards: env = ClipRewardEnv(env) if frame_stack: env = FrameStack(env, 4) # env = NormalizeObservation(env) return env class StickyActionEnv(gym.Wrapper): def __init__(self, env, p=0.25): super(StickyActionEnv, self).__init__(env) self.p = p self.last_action = 0 def reset(self): self.last_action = 0 return self.env.reset() def step(self, action): if self.unwrapped.np_random.uniform() < self.p: action = self.last_action self.last_action = action obs, reward, done, info = self.env.step(action) return obs, reward, done, info
[]
2024-01-10
Samoppakiks/govGPT
src~embedchat.py
# from .config import openaiapi, openaiorg, pinecone_api_key, pinecone_environment import os import openai import pinecone import pickle import re import re import json import time from spacy.lang.en import English openaiapi = os.environ.get("OPENAI_API_KEY") openaiorg = os.environ.get("OPENAI_ORG_ID") pinecone_api_key = os.environ.get("PINECONE_API_KEY") pinecone_environment = os.environ.get("PINECONE_ENV") openai.api_key = openaiapi openai.organization = openaiorg nlp = English() # Just the language with no model nlp.add_pipe("sentencizer") # Adding a sentencizer pipeline component def split_sentences(text): doc = nlp(text) return [sent.text for sent in doc.sents] def split_into_chunks(text, max_len=800): sentences = split_sentences(text) chunks = [] current_chunk = "" for sentence in sentences: if len(current_chunk) + len(sentence) <= max_len: current_chunk += sentence else: chunks.append(current_chunk.strip()) current_chunk = sentence if current_chunk.strip(): chunks.append(current_chunk.strip()) return chunks def clean_and_split_text(text): # Remove extra newline characters and join the text text = " ".join(text.strip().split("\n")) # Remove page numbers text = re.sub(r"\d+\n", "", text) # Remove citations # text = re.sub(r"(?:\*\s*[A-Za-z\d*]+\s*vide[^“]*?(?:\n|$))", "", text) # Identify rule titles and add a separator before them text = re.sub(r"(\d+(\.|\')\.?\s[^—]+[,—])", r"@@@\1", text) # Split the text based on the separator segments = text.split("@@@") # Create a list to store the cleaned segments cleaned_segments = [] for segment in segments: # Only remove extra spaces and newline characters segment = re.sub(r"\s+", " ", segment).strip() if len(segment) > 800: split_chunks = split_into_chunks(segment) cleaned_segments.extend(split_chunks) else: cleaned_segments.append(segment) cleaned_segments = [segment for segment in cleaned_segments if segment.strip()] return cleaned_segments def write_chunks_to_file(chunks, pdf_path, namespace=None): # Create a 'chunks' directory if it doesn't exist if not os.path.exists("chunks"): os.makedirs("chunks") # Set the output file name using the original PDF filename if pdf_path: output_filename = os.path.splitext(os.path.basename(pdf_path))[0] else: output_filename = namespace output_file_path = f"./chunks/{output_filename}_chunks.txt" # Write the chunks to the output file with open(output_file_path, "w") as f: for idx, chunk in enumerate(chunks, start=1): f.write(f"Chunk {idx}:\n") f.write(chunk) f.write("\n\n") def process_extracted_text( query, text, pdf_path, search_scope="current_file", namespace=None, department=None, type_of_document=None, year=None, ): # selecting the huggingface tokeniser and selecting the chunk sizes texts = [] # max_length = 4000 # overlap = 100 # splitting the text into chunks using our custom function texts = clean_and_split_text(text) write_chunks_to_file(texts, pdf_path, namespace) # initialising the openai api key model_engine = "text-embedding-ada-002" # initialising pinecone pinecone.init( api_key=pinecone_api_key, environment=pinecone_environment, ) # fetching the name of the created index and initialising it index_name = "rajgov" index = pinecone.Index(index_name) # creating embeddings of chunks and uploading them into the index # Get embeddings for the PDF file if pdf_path: file_name = os.path.splitext(os.path.basename(pdf_path))[0] else: file_name = namespace embeddings_file_path = f"./embeddings/{file_name}_embeddings.pkl" if namespace is None: namespace = file_name embeddings = [] if not os.path.exists(embeddings_file_path): # creating embeddings of chunks and save them to a file for i, chunk in enumerate(texts): response = openai.Embedding.create(input=[chunk], model=model_engine) embedding = response["data"][0]["embedding"] metadata = {"text": chunk} if department is not None: metadata["department"] = department if type_of_document is not None: metadata["type_of_document"] = type_of_document if year is not None: metadata["year"] = year embeddings.append((f"chunk_{i}", embedding, metadata)) with open(embeddings_file_path, "ab") as f: print(f"Saving embeddings of chunk_{i} to {embeddings_file_path}") pickle.dump([(f"chunk_{i}", embedding, metadata)], f) # Upserting embeddings to namespace index.upsert( vectors=[(f"chunk_{i}", embedding, metadata)], namespace=namespace ) else: # load embeddings from the file with open(embeddings_file_path, "rb") as f: print(f"Loading embeddings from {embeddings_file_path}") while True: try: embeddings.append(pickle.load(f)) except EOFError: break completed_chunks = len(embeddings) print(f"Completed chunks: {completed_chunks}") # Continue creating embeddings from where it left off for i, chunk in enumerate(texts[completed_chunks:], start=completed_chunks): response = openai.Embedding.create(input=[chunk], model=model_engine) embedding = response["data"][0]["embedding"] metadata = {"text": chunk} if department is not None: metadata["department"] = department if type_of_document is not None: metadata["type_of_document"] = type_of_document if year is not None: metadata["year"] = year embeddings.append((f"chunk_{i}", embedding, metadata)) with open(embeddings_file_path, "ab") as f: print(f"Saving embeddings of chunk_{i} to {embeddings_file_path}") pickle.dump([(f"chunk_{i}", embedding, metadata)], f) # Upserting embeddings to namespace index.upsert( vectors=[(f"chunk_{i}", embedding, metadata)], namespace=namespace ) # preparing the query """query = translate_to_english_chatgpt(query) focus_phrases = extract_focus_phrases(query) print(f"QUERY: {query}")""" # querying the index query_response = openai.Embedding.create(input=[query], model=model_engine) query_embedding = query_response["data"][0]["embedding"] # the response will be in json with id, metadata with text, and score if search_scope == "current_file": results = index.query( queries=[query_embedding], top_k=5, include_metadata=True, namespace=namespace, ) else: # search_scope == 'entire_database' results = index.query(queries=[query_embedding], top_k=5, include_metadata=True) print(results) answer, search_results = chatgpt_summarize_results( query, results ) # focus_phrases,) print(f"ANSWER: {answer}") return answer, search_results def chatgpt_summarize_results(query, results): # focus_phrases) search_results = "" for match in results["results"][0]["matches"]: score = match["score"] text = match["metadata"]["text"] search_results += f"{score:.2f}: {text}\n" print(search_results) response = openai.ChatCompletion.create( model="gpt-4", messages=[ { "role": "system", "content": "You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities,and for information on government services.", }, { "role": "user", "content": f"The query is: '{query}'. Based on the following search results, provide an answer to the query, after considering each result with respect to the query and checking if anything related to the query can be inferred from each result. Finally, comment on reason for your final interpreation, as well as any additional information that may not be contained in the text that may help answer the query. considering not only exact matches but also possible inferences about the expected action that can be made based on the results. :\n\n{search_results}", # You may also use the focus phrases : {focus_phrases} for better inference. }, ], ) gpt_response = response.choices[0].message["content"].strip() return gpt_response, search_results def chatgpt_get_response(context, query): response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ { "role": "system", "content": "You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities, and for information on government services.", }, {"role": "user", "content": context}, {"role": "user", "content": query}, ], ) return response.choices[0].message["content"].strip() # just as an aside, the following prompts gave 2 different results when run twice, without making any change to # prompt or query. One said permission can be given, the other said permission cannot be given. # here are the prompts """ for all 3 : role : system : You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities,and for information on government services for 1 : role : user : Translate the following text to English: Then give answers on what is the context of the text and what is the action expected. Finally , convert the text into an appropriate query that can be used to search through a semantic database to search through embedded text to find the right text that can give an answer on whether the action expected can be peformed within the given context. Make sure that the query is appropriate for searching through a semantic database consisting of government laws and regulations, and avoid adding those phrases to the query that may unnecessarily confuse the search engine.' for 2 : role : user : Extract key phrases from the following query: Only extract the phrases related to actions that are expected (including giving permissions under existing regulations, asking for details regulations relevant to the case etc), that can be searched through a datatbase of governemnt acts and regulations. Avoid extracting phrases that are not relevant to the semantic search through a database of government rules. for 3 : role : user : The query is: Based on the following search results, provide an answer to the query, after considering each result with respect to the query and checking if anything related to the query can be inferred from each result. Finally, comment on reason for your final interpreation, as well as any additional information that may not be contained in the text that may help answer the query. considering not only exact matches but also possible inferences about the expected action that can be made based on the results. You may also use the focus phrases : for better inference.:""" """def split_text(text, max_chunk_size=300, overlap=50): pattern = r'(\d+(\.\d+)*\s*\w*\.?)' chunks = re.split(pattern, text) combined_chunks = [] for i in range(0, len(chunks)-1, 2): chunk = '' if chunks[i]: chunk += chunks[i] if chunks[i+1]: chunk += chunks[i+1] combined_chunks.append(chunk) cleaned_chunks = [chunk.replace('\n', ' ').strip() for chunk in combined_chunks] smaller_chunks = [] for chunk in cleaned_chunks: rule_number_match = re.match(r'(\d+(\.\d+)*)', chunk) if rule_number_match: rule_number = rule_number_match.group(1) chunk = re.sub(r'\d+(\.\d+)*', '', chunk).strip() else: rule_number = '' tokens = chunk.split() for i in range(0, len(tokens), max_chunk_size - overlap): start = max(0, i - overlap) if i > 0 else i end = min(i + max_chunk_size, len(tokens)) small_chunk = ' '.join(tokens[start:end]) if rule_number: if start > 0: small_chunk = f"{rule_number} (cont'd) " + small_chunk else: small_chunk = f"{rule_number} " + small_chunk smaller_chunks.append(small_chunk) total_chunks = len(smaller_chunks) print(f"Total number of chunks created: {total_chunks}") return smaller_chunks Previous answer : with press or radio—Previous sanction of the Government shall not be required when the member of the service, in the bonafide discharge of his duties or otherwise, publishes a book or contributes to or participates in a public media. Provided that he shall observe the provisions of rules and at all times make it clear that the views expressed, are of his own and not those of the Government. 7. Criticism 0.84: should not be given to a moS to publish articles in the journals, souvenirs etc., of political parties: - A doubt has been raised whether members of the All India Services can be permitted to publish articles in the journals, souvenirs etc., of political parties. 2. The 0.81: 1995) 185 Provided that nothing in this rule shall apply to any statement made or views expressed by a member of the Service in his official capacity and in the due performance of the duties assigned to him. (GOI Instructions: D.P. & A.R. letter No. 11017/9/75—AlS(III), dated the 2nd March, 1976, reproduced under Miscellaneous Executive Instructions at the end of these Rules) 8. Evidence 0.81: Government may, however, at any time disallow the officer from pursuing his application for admission of financial assistance in cases where Govt. is of the view that 207 a member of the service has utilized his official influence to obtain the same or if the Government feels that such course of action is not in the interest of the Government. [Letter No. 11017/18/91-AIS(III)dated 1.7. 0.81: from literary, cultural or artistic efforts, which are not aided by the knowledge acquired by the member in the course of his service, is not ‘fee’ for the purpose of SR 12 and can be retained by the officer in full: - A question arose, whether a member of the service could accept royalty of the publication of a book of literary, artistic, or scientific character and also whether such royalties were to be treated as ‘Fee’ u Previous query: Can an officer publish an article in a journal? Previous final answer: Based on the search results, it can be inferred that a member of the service is allowed to publish a book or contribute to a public media in the bonafide discharge of his duties or otherwise, without requiring the previous sanction of the government. However, it is mandatory for the member to observe the provisions of rules and ensure that the views expressed are his own and not those of the government. It is not clear whether the query refers to a civil or police officer, but it does not seem to be prohibited unless it is a publication in a journal of a political party. It is important to note that any statements made or views expressed in the due performance of the duties assigned to him by a member of the service in his official capacity exempt him from the rule.""" """def translate_to_english_chatgpt(text): response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ {"role": "system", "content": "You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities,and for information on government services."}, {"role": "user", "content": f"Translate the following text to English: '{text}'.Then give answers on what is the context of the text and what is the action expected. Finally , convert the text into an appropriate query that can be used to search through a semantic database to search through embedded text to find the right text that can give an answer on whether the action expected can be peformed within the given context. Make sure that the query is appropriate for searching through a semantic database consisting of government, regulations,polcies, programmes and other government services, and avoid adding those phrases to the query that may unnecessarily confuse the search engine."} ] ) translated_text = response.choices[0].message['content'].strip() print(f"Translated text : '{translated_text}") return translated_text def extract_focus_phrases(translated_text): response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=[ {"role": "system", "content": "You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities,and for information on government services.."}, {"role": "user", "content": f"Extract key phrases from the following query: '{translated_text}'. Only extract the phrases related to actions that are expected (including giving permissions under existing regulations, asking for details within government schemes, looking for legal advice etc), that can be searched through a datatbase of governemnt acts, regulations, policies and welfare schemes. Avoid extracting phrases that are not relevant to the semantic search through such a database."} ] ) focus_phrases = response.choices[0].message['content'].strip() print(focus_phrases) print(f"Focus phrases : '{focus_phrases}") return focus_phrases"""
[ "The query is: 'PLACEHOLDER'. Based on the following search results, provide an answer to the query, after considering each result with respect to the query and checking if anything related to the query can be inferred from each result. Finally, comment on reason for your final interpreation, as well as any additional information that may not be contained in the text that may help answer the query. considering not only exact matches but also possible inferences about the expected action that can be made based on the results. :\n\nPLACEHOLDER", "You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities,and for information on government services.", "You are a helpful assistant working at a government help center facilities. People ask you questions related to permissable activities, and for information on government services." ]
2024-01-10
mukiblejlok/chatcli
chatcli.py
#!/usr/bin/env python3 import os import sys from datetime import datetime, timezone from pathlib import Path import openai from prompt_toolkit import PromptSession from prompt_toolkit.auto_suggest import AutoSuggestFromHistory from prompt_toolkit.history import FileHistory from prompt_toolkit.styles import Style from rich.console import Console from rich.live import Live from rich.markdown import Markdown from rich.status import Status from dotenv import load_dotenv prompt_style = Style.from_dict({ '': '#ff0066', 'pr': '#884444', }) def cli() -> int: console = Console(width=120) console.print(f'ChatCLI - OpenAI powered ChatCLI', style='blue bold', highlight=False) try: openai.api_key = os.environ['OPENAI_API_KEY'] except KeyError: console.print('You must set the OPENAI_API_KEY environment variable', style='red') return 1 now_utc = datetime.now(timezone.utc) setup = f""" Help the user by responding to their request, the output should be concise and always written in markdown. The current date and time is {datetime.now()} {now_utc.astimezone().tzinfo.tzname(now_utc)}. """ messages = [{'role': 'system', 'content': setup}] history = Path().home() / '.openai-prompt-history.txt' session = PromptSession(history=FileHistory(str(history))) while True: try: text = session.prompt([("class:pr", "chatcli ➤ ")], auto_suggest=AutoSuggestFromHistory(), style=prompt_style) except (KeyboardInterrupt, EOFError): return 0 if not text: continue status = Status('[dim]Thinking…[/dim]', console=console) status.start() messages.append({'role': 'user', 'content': text}) try: response = openai.ChatCompletion.create(model='gpt-4', messages=messages, stream=True) except KeyboardInterrupt: status.stop() return 0 status.stop() content = '' markdown = Markdown(content) try: with Live(markdown, refresh_per_second=15, console=console) as live: for chunk in response: if chunk['choices'][0]['finish_reason'] is not None: break chunk_text = chunk['choices'][0]['delta'].get('content', '') content += chunk_text live.update(Markdown(content)) except KeyboardInterrupt: console.print('[dim]Interrupted[/dim]') messages.append({'role': 'assistant', 'content': content}) if __name__ == '__main__': load_dotenv() sys.exit(cli())
[ "f\"\"\"\nHelp the user by responding to their request, the output should be concise and always written in markdown.\nThe current date and time is {datetime.now()} {now_utc.astimezone().tzinfo.tzname(now_utc)}.\n" ]
2024-01-10
the-virtual-brain/tvb-framework
tvb~config~init~datatypes_registry.py
# -*- coding: utf-8 -*- # # # TheVirtualBrain-Framework Package. This package holds all Data Management, and # Web-UI helpful to run brain-simulations. To use it, you also need do download # TheVirtualBrain-Scientific Package (for simulators). See content of the # documentation-folder for more details. See also http://www.thevirtualbrain.org # # (c) 2012-2017, Baycrest Centre for Geriatric Care ("Baycrest") and others # # This program is free software: you can redistribute it and/or modify it under the # terms of the GNU General Public License as published by the Free Software Foundation, # either version 3 of the License, or (at your option) any later version. # This program is distributed in the hope that it will be useful, but WITHOUT ANY # WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A # PARTICULAR PURPOSE. See the GNU General Public License for more details. # You should have received a copy of the GNU General Public License along with this # program. If not, see <http://www.gnu.org/licenses/>. # # # CITATION: # When using The Virtual Brain for scientific publications, please cite it as follows: # # Paula Sanz Leon, Stuart A. Knock, M. Marmaduke Woodman, Lia Domide, # Jochen Mersmann, Anthony R. McIntosh, Viktor Jirsa (2013) # The Virtual Brain: a simulator of primate brain network dynamics. # Frontiers in Neuroinformatics (7:10. doi: 10.3389/fninf.2013.00010) # # from tvb.datatypes.connectivity import Connectivity from tvb.datatypes.fcd import Fcd from tvb.datatypes.graph import ConnectivityMeasure, CorrelationCoefficients, Covariance from tvb.datatypes.local_connectivity import LocalConnectivity from tvb.datatypes.mode_decompositions import PrincipalComponents, IndependentComponents from tvb.datatypes.patterns import StimuliRegion, StimuliSurface from tvb.datatypes.projections import ProjectionMatrix from tvb.datatypes.region_mapping import RegionVolumeMapping, RegionMapping from tvb.datatypes.sensors import Sensors from tvb.adapters.datatypes.simulation_state import SimulationState from tvb.datatypes.spectral import CoherenceSpectrum, ComplexCoherenceSpectrum, FourierSpectrum, WaveletCoefficients from tvb.datatypes.structural import StructuralMRI from tvb.datatypes.surfaces import Surface from tvb.datatypes.temporal_correlations import CrossCorrelation from tvb.datatypes.time_series import TimeSeries, TimeSeriesRegion, TimeSeriesSurface, TimeSeriesVolume from tvb.datatypes.time_series import TimeSeriesEEG, TimeSeriesMEG, TimeSeriesSEEG from tvb.datatypes.tracts import Tracts from tvb.datatypes.volumes import Volume from tvb.datatypes.cortex import Cortex from tvb.core.entities.file.simulator.cortex_h5 import CortexH5 from tvb.adapters.datatypes.h5.annotation_h5 import ConnectivityAnnotationsH5, ConnectivityAnnotations from tvb.adapters.datatypes.h5.connectivity_h5 import ConnectivityH5 from tvb.adapters.datatypes.h5.fcd_h5 import FcdH5 from tvb.adapters.datatypes.h5.graph_h5 import ConnectivityMeasureH5, CorrelationCoefficientsH5, CovarianceH5 from tvb.adapters.datatypes.h5.local_connectivity_h5 import LocalConnectivityH5 from tvb.adapters.datatypes.h5.mapped_value_h5 import DatatypeMeasureH5, ValueWrapperH5 from tvb.adapters.datatypes.h5.mode_decompositions_h5 import PrincipalComponentsH5, IndependentComponentsH5 from tvb.adapters.datatypes.h5.patterns_h5 import StimuliRegionH5, StimuliSurfaceH5 from tvb.adapters.datatypes.h5.projections_h5 import ProjectionMatrixH5 from tvb.adapters.datatypes.h5.region_mapping_h5 import RegionMappingH5, RegionVolumeMappingH5 from tvb.adapters.datatypes.h5.sensors_h5 import SensorsH5 from tvb.adapters.datatypes.h5.simulation_state_h5 import SimulationStateH5 from tvb.adapters.datatypes.h5.spectral_h5 import CoherenceSpectrumH5, ComplexCoherenceSpectrumH5 from tvb.adapters.datatypes.h5.spectral_h5 import FourierSpectrumH5, WaveletCoefficientsH5 from tvb.adapters.datatypes.h5.structural_h5 import StructuralMRIH5 from tvb.adapters.datatypes.h5.surface_h5 import SurfaceH5 from tvb.adapters.datatypes.h5.temporal_correlations_h5 import CrossCorrelationH5 from tvb.adapters.datatypes.h5.time_series_h5 import TimeSeriesH5, TimeSeriesRegionH5, TimeSeriesSurfaceH5 from tvb.adapters.datatypes.h5.time_series_h5 import TimeSeriesVolumeH5, TimeSeriesEEGH5, TimeSeriesMEGH5 from tvb.adapters.datatypes.h5.time_series_h5 import TimeSeriesSEEGH5 from tvb.adapters.datatypes.h5.tracts_h5 import TractsH5 from tvb.adapters.datatypes.h5.volumes_h5 import VolumeH5 from tvb.adapters.datatypes.db.annotation import ConnectivityAnnotationsIndex from tvb.adapters.datatypes.db.connectivity import ConnectivityIndex from tvb.adapters.datatypes.db.fcd import FcdIndex from tvb.adapters.datatypes.db.graph import ConnectivityMeasureIndex, CorrelationCoefficientsIndex from tvb.adapters.datatypes.db.graph import CovarianceIndex from tvb.adapters.datatypes.db.local_connectivity import LocalConnectivityIndex from tvb.adapters.datatypes.db.mapped_value import DatatypeMeasureIndex, ValueWrapperIndex from tvb.adapters.datatypes.db.mode_decompositions import PrincipalComponentsIndex, IndependentComponentsIndex from tvb.adapters.datatypes.db.patterns import StimuliRegionIndex, StimuliSurfaceIndex from tvb.adapters.datatypes.db.projections import ProjectionMatrixIndex from tvb.adapters.datatypes.db.region_mapping import RegionVolumeMappingIndex, RegionMappingIndex from tvb.adapters.datatypes.db.sensors import SensorsIndex from tvb.adapters.datatypes.db.simulation_state import SimulationStateIndex from tvb.adapters.datatypes.db.spectral import CoherenceSpectrumIndex, ComplexCoherenceSpectrumIndex from tvb.adapters.datatypes.db.spectral import FourierSpectrumIndex, WaveletCoefficientsIndex from tvb.adapters.datatypes.db.structural import StructuralMRIIndex from tvb.adapters.datatypes.db.surface import SurfaceIndex from tvb.adapters.datatypes.db.temporal_correlations import CrossCorrelationIndex from tvb.adapters.datatypes.db.time_series import TimeSeriesIndex, TimeSeriesRegionIndex, TimeSeriesSurfaceIndex from tvb.adapters.datatypes.db.time_series import TimeSeriesVolumeIndex, TimeSeriesEEGIndex, TimeSeriesMEGIndex from tvb.adapters.datatypes.db.time_series import TimeSeriesSEEGIndex from tvb.adapters.datatypes.db.tracts import TractsIndex from tvb.adapters.datatypes.db.volume import VolumeIndex from tvb.core.neocom.h5 import REGISTRY # an alternative approach is to make each h5file declare if it has a corresponding datatype # then in a metaclass hook each class creation and populate a map def populate_datatypes_registry(): REGISTRY.register_datatype(Connectivity, ConnectivityH5, ConnectivityIndex) REGISTRY.register_datatype(LocalConnectivity, LocalConnectivityH5, LocalConnectivityIndex) REGISTRY.register_datatype(ProjectionMatrix, ProjectionMatrixH5, ProjectionMatrixIndex) REGISTRY.register_datatype(RegionVolumeMapping, RegionVolumeMappingH5, RegionVolumeMappingIndex) REGISTRY.register_datatype(RegionMapping, RegionMappingH5, RegionMappingIndex) REGISTRY.register_datatype(Sensors, SensorsH5, SensorsIndex) REGISTRY.register_datatype(SimulationState, SimulationStateH5, SimulationStateIndex) REGISTRY.register_datatype(CoherenceSpectrum, CoherenceSpectrumH5, CoherenceSpectrumIndex) REGISTRY.register_datatype(ComplexCoherenceSpectrum, ComplexCoherenceSpectrumH5, ComplexCoherenceSpectrumIndex) REGISTRY.register_datatype(FourierSpectrum, FourierSpectrumH5, FourierSpectrumIndex) REGISTRY.register_datatype(WaveletCoefficients, WaveletCoefficientsH5, WaveletCoefficientsIndex) REGISTRY.register_datatype(StructuralMRI, StructuralMRIH5, StructuralMRIIndex) REGISTRY.register_datatype(Surface, SurfaceH5, SurfaceIndex) REGISTRY.register_datatype(CrossCorrelation, CrossCorrelationH5, CrossCorrelationIndex) REGISTRY.register_datatype(TimeSeries, TimeSeriesH5, TimeSeriesIndex) REGISTRY.register_datatype(TimeSeriesRegion, TimeSeriesRegionH5, TimeSeriesRegionIndex) REGISTRY.register_datatype(TimeSeriesSurface, TimeSeriesSurfaceH5, TimeSeriesSurfaceIndex) REGISTRY.register_datatype(TimeSeriesVolume, TimeSeriesVolumeH5, TimeSeriesVolumeIndex) REGISTRY.register_datatype(TimeSeriesEEG, TimeSeriesEEGH5, TimeSeriesEEGIndex) REGISTRY.register_datatype(TimeSeriesMEG, TimeSeriesMEGH5, TimeSeriesMEGIndex) REGISTRY.register_datatype(TimeSeriesSEEG, TimeSeriesSEEGH5, TimeSeriesSEEGIndex) REGISTRY.register_datatype(Tracts, TractsH5, TractsIndex) REGISTRY.register_datatype(Volume, VolumeH5, VolumeIndex) REGISTRY.register_datatype(PrincipalComponents, PrincipalComponentsH5, PrincipalComponentsIndex) REGISTRY.register_datatype(IndependentComponents, IndependentComponentsH5, IndependentComponentsIndex) REGISTRY.register_datatype(ConnectivityMeasure, ConnectivityMeasureH5, ConnectivityMeasureIndex) REGISTRY.register_datatype(CorrelationCoefficients, CorrelationCoefficientsH5, CorrelationCoefficientsIndex) REGISTRY.register_datatype(Covariance, CovarianceH5, CovarianceIndex) REGISTRY.register_datatype(Fcd, FcdH5, FcdIndex) REGISTRY.register_datatype(StimuliRegion, StimuliRegionH5, StimuliRegionIndex) REGISTRY.register_datatype(StimuliSurface, StimuliSurfaceH5, StimuliSurfaceIndex) REGISTRY.register_datatype(None, DatatypeMeasureH5, DatatypeMeasureIndex) REGISTRY.register_datatype(ConnectivityAnnotations, ConnectivityAnnotationsH5, ConnectivityAnnotationsIndex) REGISTRY.register_datatype(None, ValueWrapperH5, ValueWrapperIndex) REGISTRY.register_datatype(Cortex, CortexH5, None)
[]
2024-01-10
the-virtual-brain/tvb-framework
tvb~adapters~visualizers~cross_coherence.py
# -*- coding: utf-8 -*- # # # TheVirtualBrain-Framework Package. This package holds all Data Management, and # Web-UI helpful to run brain-simulations. To use it, you also need do download # TheVirtualBrain-Scientific Package (for simulators). See content of the # documentation-folder for more details. See also http://www.thevirtualbrain.org # # (c) 2012-2017, Baycrest Centre for Geriatric Care ("Baycrest") and others # # This program is free software: you can redistribute it and/or modify it under the # terms of the GNU General Public License as published by the Free Software Foundation, # either version 3 of the License, or (at your option) any later version. # This program is distributed in the hope that it will be useful, but WITHOUT ANY # WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A # PARTICULAR PURPOSE. See the GNU General Public License for more details. # You should have received a copy of the GNU General Public License along with this # program. If not, see <http://www.gnu.org/licenses/>. # # # CITATION: # When using The Virtual Brain for scientific publications, please cite it as follows: # # Paula Sanz Leon, Stuart A. Knock, M. Marmaduke Woodman, Lia Domide, # Jochen Mersmann, Anthony R. McIntosh, Viktor Jirsa (2013) # The Virtual Brain: a simulator of primate brain network dynamics. # Frontiers in Neuroinformatics (7:10. doi: 10.3389/fninf.2013.00010) # # """ A displayer for the cross coherence of a time series. .. moduleauthor:: Marmaduke Woodman <[email protected]> """ import json from tvb.adapters.visualizers.matrix_viewer import MappedArraySVGVisualizerMixin from tvb.core.adapters.abcadapter import ABCAdapterForm from tvb.core.adapters.abcdisplayer import ABCDisplayer from tvb.adapters.datatypes.db.spectral import CoherenceSpectrumIndex from tvb.core.neotraits.forms import DataTypeSelectField class CrossCoherenceVisualizerForm(ABCAdapterForm): def __init__(self, prefix='', project_id=None): super(CrossCoherenceVisualizerForm, self).__init__(prefix, project_id) self.datatype = DataTypeSelectField(self.get_required_datatype(), self, name='datatype', required=True, label='Coherence spectrum:', conditions=self.get_filters()) @staticmethod def get_required_datatype(): return CoherenceSpectrumIndex @staticmethod def get_input_name(): return '_datatype' @staticmethod def get_filters(): return None class CrossCoherenceVisualizer(MappedArraySVGVisualizerMixin): _ui_name = "Cross Coherence Visualizer" _ui_subsection = "coherence" def get_form_class(self): return CrossCoherenceVisualizerForm def launch(self, datatype): """Construct data for visualization and launch it.""" datatype_h5_class, datatype_h5_path = self._load_h5_of_gid(datatype.gid) with datatype_h5_class(datatype_h5_path) as datatype_h5: # get data from coher datatype h5, convert to json frequency = ABCDisplayer.dump_with_precision(datatype_h5.frequency.load().flat) array_data = datatype_h5.array_data[:] params = self.compute_raw_matrix_params(array_data) params.update(frequency=frequency) params.update(matrix_strides=json.dumps([x / array_data.itemsize for x in array_data.strides])) return self.build_display_result("cross_coherence/view", params)
[]
2024-01-10
the-virtual-brain/tvb-framework
tvb~adapters~analyzers~node_coherence_adapter.py
# -*- coding: utf-8 -*- # # # TheVirtualBrain-Framework Package. This package holds all Data Management, and # Web-UI helpful to run brain-simulations. To use it, you also need do download # TheVirtualBrain-Scientific Package (for simulators). See content of the # documentation-folder for more details. See also http://www.thevirtualbrain.org # # (c) 2012-2017, Baycrest Centre for Geriatric Care ("Baycrest") and others # # This program is free software: you can redistribute it and/or modify it under the # terms of the GNU General Public License as published by the Free Software Foundation, # either version 3 of the License, or (at your option) any later version. # This program is distributed in the hope that it will be useful, but WITHOUT ANY # WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A # PARTICULAR PURPOSE. See the GNU General Public License for more details. # You should have received a copy of the GNU General Public License along with this # program. If not, see <http://www.gnu.org/licenses/>. # # # CITATION: # When using The Virtual Brain for scientific publications, please cite it as follows: # # Paula Sanz Leon, Stuart A. Knock, M. Marmaduke Woodman, Lia Domide, # Jochen Mersmann, Anthony R. McIntosh, Viktor Jirsa (2013) # The Virtual Brain: a simulator of primate brain network dynamics. # Frontiers in Neuroinformatics (7:10. doi: 10.3389/fninf.2013.00010) # # """ Adapter that uses the traits module to generate interfaces for FFT Analyzer. .. moduleauthor:: Stuart A. Knock <[email protected]> .. moduleauthor:: Lia Domide <[email protected]> """ import uuid import numpy from tvb.analyzers.node_coherence import NodeCoherence from tvb.core.adapters.abcadapter import ABCAsynchronous, ABCAdapterForm from tvb.datatypes.time_series import TimeSeries from tvb.core.entities.filters.chain import FilterChain from tvb.basic.logger.builder import get_logger from tvb.adapters.datatypes.h5.spectral_h5 import CoherenceSpectrumH5 from tvb.adapters.datatypes.db.spectral import CoherenceSpectrumIndex from tvb.adapters.datatypes.db.time_series import TimeSeriesIndex from tvb.core.neotraits.forms import ScalarField, DataTypeSelectField from tvb.core.neocom import h5 LOG = get_logger(__name__) class NodeCoherenceForm(ABCAdapterForm): def __init__(self, prefix='', project_id=None): super(NodeCoherenceForm, self).__init__(prefix, project_id) self.time_series = DataTypeSelectField(self.get_required_datatype(), self, name=self.get_input_name(), required=True, label=NodeCoherence.time_series.label, doc=NodeCoherence.time_series.doc, conditions=self.get_filters(), has_all_option=True) self.nfft = ScalarField(NodeCoherence.nfft, self) @staticmethod def get_required_datatype(): return TimeSeriesIndex @staticmethod def get_filters(): return FilterChain(fields=[FilterChain.datatype + '.data_ndim'], operations=["=="], values=[4]) @staticmethod def get_input_name(): return "time_series" def get_traited_datatype(self): return NodeCoherence() class NodeCoherenceAdapter(ABCAsynchronous): """ TVB adapter for calling the NodeCoherence algorithm. """ _ui_name = "Cross coherence of nodes" _ui_description = "Compute Node Coherence for a TimeSeries input DataType." _ui_subsection = "coherence" def get_form_class(self): return NodeCoherenceForm def get_output(self): return [CoherenceSpectrumIndex] def configure(self, time_series, nfft=None): """ Store the input shape to be later used to estimate memory usage. Also create the algorithm instance. """ self.input_time_series_index = time_series self.input_shape = (self.input_time_series_index.data_length_1d, self.input_time_series_index.data_length_2d, self.input_time_series_index.data_length_3d, self.input_time_series_index.data_length_4d) LOG.debug("Time series shape is %s" % str(self.input_shape)) # -------------------- Fill Algorithm for Analysis -------------------## self.algorithm = NodeCoherence() if nfft is not None: self.algorithm.nfft = nfft def get_required_memory_size(self, **kwargs): """ Return the required memory to run this algorithm. """ used_shape = (self.input_shape[0], 1, self.input_shape[2], self.input_shape[3]) input_size = numpy.prod(used_shape) * 8.0 output_size = self.algorithm.result_size(used_shape) return input_size + output_size def get_required_disk_size(self, **kwargs): """ Returns the required disk size to be able to run the adapter (in kB). """ used_shape = (self.input_shape[0], 1, self.input_shape[2], self.input_shape[3]) return self.array_size2kb(self.algorithm.result_size(used_shape)) def launch(self, time_series, nfft=None): """ Launch algorithm and build results. """ # --------- Prepare a CoherenceSpectrum object for result ------------## coherence_spectrum_index = CoherenceSpectrumIndex() time_series_h5 = h5.h5_file_for_index(time_series) dest_path = h5.path_for(self.storage_path, CoherenceSpectrumH5, coherence_spectrum_index.gid) coherence_h5 = CoherenceSpectrumH5(dest_path) coherence_h5.gid.store(uuid.UUID(coherence_spectrum_index.gid)) coherence_h5.source.store(time_series_h5.gid.load()) coherence_h5.nfft.store(self.algorithm.nfft) # ------------- NOTE: Assumes 4D, Simulator timeSeries. --------------## input_shape = time_series_h5.data.shape node_slice = [slice(input_shape[0]), None, slice(input_shape[2]), slice(input_shape[3])] # ---------- Iterate over slices and compose final result ------------## small_ts = TimeSeries() small_ts.sample_period = time_series_h5.sample_period.load() partial_coh = None for var in range(input_shape[1]): node_slice[1] = slice(var, var + 1) small_ts.data = time_series_h5.read_data_slice(tuple(node_slice)) self.algorithm.time_series = small_ts partial_coh = self.algorithm.evaluate() coherence_h5.write_data_slice(partial_coh) coherence_h5.frequency.store(partial_coh.frequency) coherence_h5.close() coherence_spectrum_index.ndim = len(coherence_h5.array_data.shape) time_series_h5.close() coherence_spectrum_index.source_gid = self.input_time_series_index.gid coherence_spectrum_index.nfft = partial_coh.nfft coherence_spectrum_index.frequencies = partial_coh.frequency return coherence_spectrum_index
[]
2024-01-10
the-virtual-brain/tvb-framework
tvb~core~services~introspector_registry.py
# -*- coding: utf-8 -*- # # # TheVirtualBrain-Framework Package. This package holds all Data Management, and # Web-UI helpful to run brain-simulations. To use it, you also need do download # TheVirtualBrain-Scientific Package (for simulators). See content of the # documentation-folder for more details. See also http://www.thevirtualbrain.org # # (c) 2012-2017, Baycrest Centre for Geriatric Care ("Baycrest") and others # # This program is free software: you can redistribute it and/or modify it under the # terms of the GNU General Public License as published by the Free Software Foundation, # either version 3 of the License, or (at your option) any later version. # This program is distributed in the hope that it will be useful, but WITHOUT ANY # WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A # PARTICULAR PURPOSE. See the GNU General Public License for more details. # You should have received a copy of the GNU General Public License along with this # program. If not, see <http://www.gnu.org/licenses/>. # # # CITATION: # When using The Virtual Brain for scientific publications, please cite it as follows: # # Paula Sanz Leon, Stuart A. Knock, M. Marmaduke Woodman, Lia Domide, # Jochen Mersmann, Anthony R. McIntosh, Viktor Jirsa (2013) # The Virtual Brain: a simulator of primate brain network dynamics. # Frontiers in Neuroinformatics (7:10. doi: 10.3389/fninf.2013.00010) # # from tvb.basic.profile import TvbProfile from tvb.datatypes.simulation_state import SimulationState from tvb.adapters.analyzers.analyze_algorithm_category_config import AnalyzeAlgorithmCategoryConfig from tvb.adapters.analyzers.cross_correlation_adapter import CrossCorrelateAdapter, PearsonCorrelationCoefficientAdapter from tvb.adapters.analyzers.fcd_adapter import FunctionalConnectivityDynamicsAdapter from tvb.adapters.analyzers.fmri_balloon_adapter import BalloonModelAdapter from tvb.adapters.analyzers.fourier_adapter import FourierAdapter from tvb.adapters.analyzers.ica_adapter import ICAAdapter from tvb.adapters.analyzers.metrics_group_timeseries import TimeseriesMetricsAdapter from tvb.adapters.analyzers.node_coherence_adapter import NodeCoherenceAdapter from tvb.adapters.analyzers.node_complex_coherence_adapter import NodeComplexCoherenceAdapter from tvb.adapters.analyzers.node_covariance_adapter import NodeCovarianceAdapter from tvb.adapters.analyzers.pca_adapter import PCAAdapter from tvb.adapters.analyzers.wavelet_adapter import ContinuousWaveletTransformAdapter from tvb.adapters.creators.allen_creator import AllenConnectomeBuilder from tvb.adapters.creators.connectivity_creator import ConnectivityCreator from tvb.adapters.creators.create_algorithm_category_config import CreateAlgorithmCategoryConfig from tvb.adapters.creators.local_connectivity_creator import LocalConnectivityCreator from tvb.adapters.creators.stimulus_creator import RegionStimulusCreator, SurfaceStimulusCreator from tvb.adapters.simulator.simulate_algorithm_category_config import SimulateAlgorithmCategoryConfig from tvb.adapters.simulator.simulator_adapter import SimulatorAdapter from tvb.adapters.uploaders.brco_importer import BRCOImporter from tvb.adapters.uploaders.cff_importer import CFF_Importer from tvb.adapters.uploaders.connectivity_measure_importer import ConnectivityMeasureImporter from tvb.adapters.uploaders.csv_connectivity_importer import CSVConnectivityImporter from tvb.adapters.uploaders.eeg_mat_timeseries_importer import EEGMatTimeSeriesImporter from tvb.adapters.uploaders.gifti_surface_importer import GIFTISurfaceImporter from tvb.adapters.uploaders.gifti_timeseries_importer import GIFTITimeSeriesImporter from tvb.adapters.uploaders.mat_timeseries_importer import MatTimeSeriesImporter from tvb.adapters.uploaders.networkx_importer import NetworkxConnectivityImporter from tvb.adapters.uploaders.nifti_importer import NIFTIImporter from tvb.adapters.uploaders.obj_importer import ObjSurfaceImporter from tvb.adapters.uploaders.projection_matrix_importer import ProjectionMatrixSurfaceEEGImporter from tvb.adapters.uploaders.region_mapping_importer import RegionMapping_Importer from tvb.adapters.uploaders.sensors_importer import Sensors_Importer from tvb.adapters.uploaders.tract_importer import TrackvizTractsImporter, ZipTxtTractsImporter from tvb.adapters.uploaders.tvb_importer import TVBImporter from tvb.adapters.uploaders.upload_algorithm_category_config import UploadAlgorithmCategoryConfig from tvb.adapters.uploaders.zip_connectivity_importer import ZIPConnectivityImporter from tvb.adapters.uploaders.zip_surface_importer import ZIPSurfaceImporter from tvb.adapters.visualizers.annotations_viewer import ConnectivityAnnotationsView from tvb.adapters.visualizers.brain import BrainViewer, DualBrainViewer from tvb.adapters.visualizers.complex_imaginary_coherence import ImaginaryCoherenceDisplay from tvb.adapters.visualizers.connectivity import ConnectivityViewer from tvb.adapters.visualizers.connectivity_edge_bundle import ConnectivityEdgeBundle from tvb.adapters.visualizers.covariance import CovarianceVisualizer from tvb.adapters.visualizers.cross_coherence import CrossCoherenceVisualizer from tvb.adapters.visualizers.cross_correlation import CrossCorrelationVisualizer from tvb.adapters.visualizers.eeg_monitor import EegMonitor from tvb.adapters.visualizers.fourier_spectrum import FourierSpectrumDisplay from tvb.adapters.visualizers.histogram import HistogramViewer from tvb.adapters.visualizers.ica import ICA from tvb.adapters.visualizers.local_connectivity_view import LocalConnectivityViewer from tvb.adapters.visualizers.matrix_viewer import MappedArrayVisualizer from tvb.adapters.visualizers.pca import PCA from tvb.adapters.visualizers.pearson_cross_correlation import PearsonCorrelationCoefficientVisualizer from tvb.adapters.visualizers.pearson_edge_bundle import PearsonEdgeBundle from tvb.adapters.visualizers.pse_discrete import DiscretePSEAdapter from tvb.adapters.visualizers.pse_isocline import IsoclinePSEAdapter from tvb.adapters.visualizers.region_volume_mapping import ConnectivityMeasureVolumeVisualizer, \ MappedArrayVolumeVisualizer, MriVolumeVisualizer, RegionVolumeMappingVisualiser from tvb.adapters.visualizers.sensors import SensorsViewer from tvb.adapters.visualizers.surface_view import SurfaceViewer, RegionMappingViewer, ConnectivityMeasureOnSurfaceViewer from tvb.adapters.visualizers.time_series import TimeSeries from tvb.adapters.visualizers.time_series_volume import TimeSeriesVolumeVisualiser from tvb.adapters.visualizers.topographic import TopographicViewer from tvb.adapters.visualizers.tract import TractViewer from tvb.adapters.visualizers.view_algorithm_category_config import ViewAlgorithmCategoryConfig from tvb.adapters.visualizers.wavelet_spectrogram import WaveletSpectrogramVisualizer import tvb.adapters.portlets as portlets_module from tvb.core.entities.model.datatypes.connectivity import ConnectivityIndex from tvb.core.entities.model.datatypes.fcd import FcdIndex from tvb.core.entities.model.datatypes.graph import ConnectivityMeasureIndex, CorrelationCoefficientsIndex, \ CovarianceIndex from tvb.core.entities.model.datatypes.local_connectivity import LocalConnectivityIndex from tvb.core.entities.model.datatypes.mode_decompositions import IndependentComponentsIndex, PrincipalComponentsIndex from tvb.core.entities.model.datatypes.patterns import StimuliRegionIndex, StimuliSurfaceIndex from tvb.core.entities.model.datatypes.projections import ProjectionMatrixIndex from tvb.core.entities.model.datatypes.region_mapping import RegionMappingIndex, RegionVolumeMappingIndex from tvb.core.entities.model.datatypes.sensors import SensorsIndex from tvb.core.entities.model.datatypes.spectral import CoherenceSpectrumIndex, ComplexCoherenceSpectrumIndex, \ FourierSpectrumIndex, WaveletCoefficientsIndex from tvb.core.entities.model.datatypes.structural import StructuralMRIIndex from tvb.core.entities.model.datatypes.surface import SurfaceIndex from tvb.core.entities.model.datatypes.temporal_correlations import CrossCorrelationIndex from tvb.core.entities.model.datatypes.time_series import TimeSeriesEEGIndex, TimeSeriesIndex, TimeSeriesMEGIndex, \ TimeSeriesRegionIndex, TimeSeriesSEEGIndex, TimeSeriesSurfaceIndex, TimeSeriesVolumeIndex from tvb.core.entities.model.datatypes.tracts import TractsIndex from tvb.core.entities.model.datatypes.volume import VolumeIndex if TvbProfile.current.MATLAB_EXECUTABLE and len(TvbProfile.current.MATLAB_EXECUTABLE) > 0: from tvb.adapters.analyzers.bct_adapters import DistanceDBIN, DistanceDWEI, DistanceNETW, DistanceRDA, DistanceRDM, \ ModularityOCSM, ModularityOpCSMU from tvb.adapters.analyzers.bct_centrality_adapters import CentralityEdgeBinary, CentralityEdgeWeighted, \ CentralityEigenVector, CentralityKCoreness, CentralityKCorenessBD, CentralityNodeBinary, CentralityNodeWeighted, \ CentralityShortcuts, FlowCoefficients, ParticipationCoefficient, ParticipationCoefficientSign, \ SubgraphCentrality from tvb.adapters.analyzers.bct_clustering_adapters import ClusteringCoefficient, ClusteringCoefficientBU, \ ClusteringCoefficientWD, ClusteringCoefficientWU, TransitivityBinaryDirected, TransitivityBinaryUnDirected, \ TransitivityWeightedDirected, TransitivityWeightedUnDirected from tvb.adapters.analyzers.bct_degree_adapters import Degree, DegreeIOD, DensityDirected, DensityUndirected, \ JointDegree, MatchingIndex, Strength, StrengthISOS, StrengthWeights class IntrospectionRegistry(object): """ This registry gathers classes that have a role in generating DB tables and rows. It is used at introspection time, for the following operations: - fill-in all rows in the ALGORITHM_CATEGORIES table - fill-in all rows in the ALGORITHMS table. Will add BCT algorithms only if Matlab/Octave path is set - generate DB tables for all datatype indexes - fill-in all rows in the PORTLETS table using data defined in XML files - keep an evidence of the datatype index removers All classes that subclass AlgorithmCategoryConfig, ABCAdapter, ABCRemover, HasTraitsIndex should be imported here and added to the proper dictionary/list. e.g. Each new class of type HasTraitsIndex should be imported here and added to the DATATYPES list. """ ADAPTERS = { AnalyzeAlgorithmCategoryConfig: [ CrossCorrelateAdapter, PearsonCorrelationCoefficientAdapter, FunctionalConnectivityDynamicsAdapter, BalloonModelAdapter, FourierAdapter, ICAAdapter, TimeseriesMetricsAdapter, NodeCoherenceAdapter, NodeComplexCoherenceAdapter, NodeCovarianceAdapter, PCAAdapter, ContinuousWaveletTransformAdapter ], SimulateAlgorithmCategoryConfig: [SimulatorAdapter], UploadAlgorithmCategoryConfig: [ BRCOImporter, ConnectivityMeasureImporter, GIFTISurfaceImporter, GIFTITimeSeriesImporter, CSVConnectivityImporter, MatTimeSeriesImporter, EEGMatTimeSeriesImporter, NetworkxConnectivityImporter, NIFTIImporter, ObjSurfaceImporter, ProjectionMatrixSurfaceEEGImporter, RegionMapping_Importer, Sensors_Importer, TVBImporter, TrackvizTractsImporter, ZipTxtTractsImporter, ZIPConnectivityImporter, ZIPSurfaceImporter ], ViewAlgorithmCategoryConfig: [ ConnectivityAnnotationsView, BrainViewer, DualBrainViewer, ImaginaryCoherenceDisplay, ConnectivityViewer, ConnectivityEdgeBundle, CovarianceVisualizer, CrossCoherenceVisualizer, CrossCorrelationVisualizer, EegMonitor, FourierSpectrumDisplay, HistogramViewer, ICA, LocalConnectivityViewer, MappedArrayVisualizer, PCA, PearsonCorrelationCoefficientVisualizer, PearsonEdgeBundle, DiscretePSEAdapter, IsoclinePSEAdapter, ConnectivityMeasureVolumeVisualizer, MappedArrayVolumeVisualizer, MriVolumeVisualizer, RegionVolumeMappingVisualiser, SensorsViewer, SurfaceViewer, RegionMappingViewer, ConnectivityMeasureOnSurfaceViewer, TimeSeries, TimeSeriesVolumeVisualiser, TractViewer, TopographicViewer, WaveletSpectrogramVisualizer ], CreateAlgorithmCategoryConfig: [ AllenConnectomeBuilder, ConnectivityCreator, LocalConnectivityCreator, RegionStimulusCreator, SurfaceStimulusCreator ], } if TvbProfile.current.MATLAB_EXECUTABLE and len(TvbProfile.current.MATLAB_EXECUTABLE) > 0: BCT_ADAPTERS = [ DistanceDBIN, DistanceDWEI, DistanceNETW, DistanceRDA, DistanceRDM, ModularityOCSM, ModularityOpCSMU, CentralityEdgeBinary, CentralityEdgeWeighted, CentralityEigenVector, CentralityKCoreness, CentralityKCorenessBD, CentralityNodeBinary, CentralityNodeWeighted, CentralityShortcuts, FlowCoefficients, ParticipationCoefficient, ParticipationCoefficientSign, SubgraphCentrality, ClusteringCoefficient, ClusteringCoefficientBU, ClusteringCoefficientWD, ClusteringCoefficientWU, TransitivityBinaryDirected, TransitivityBinaryUnDirected, TransitivityWeightedDirected, TransitivityWeightedUnDirected, Degree, DegreeIOD, DensityDirected, DensityUndirected, JointDegree, MatchingIndex, Strength, StrengthISOS, StrengthWeights, ] ADAPTERS[AnalyzeAlgorithmCategoryConfig].extend(BCT_ADAPTERS) DATATYPE_REMOVERS = { # TODO } DATATYPES = [ConnectivityIndex, FcdIndex, ConnectivityMeasureIndex, CorrelationCoefficientsIndex, CovarianceIndex, LocalConnectivityIndex, IndependentComponentsIndex, PrincipalComponentsIndex, StimuliRegionIndex, StimuliSurfaceIndex, ProjectionMatrixIndex, RegionMappingIndex, RegionVolumeMappingIndex, SensorsIndex, CoherenceSpectrumIndex, ComplexCoherenceSpectrumIndex, FourierSpectrumIndex, WaveletCoefficientsIndex, StructuralMRIIndex, SurfaceIndex, CrossCorrelationIndex, TimeSeriesEEGIndex, TimeSeriesIndex, TimeSeriesMEGIndex, TimeSeriesRegionIndex, TimeSeriesSEEGIndex, TimeSeriesSurfaceIndex, TimeSeriesVolumeIndex, TractsIndex, VolumeIndex] PORTLETS_MODULE = portlets_module SIMULATOR_MODULE = SimulatorAdapter.__module__ SIMULATOR_CLASS = SimulatorAdapter.__name__ SIMULATION_DATATYPE_MODULE = SimulationState.__module__ SIMULATION_DATATYPE_CLASS = SimulationState.__name__ CONNECTIVITY_MODULE = ConnectivityViewer.__module__ CONNECTIVITY_CLASS = ConnectivityViewer.__name__ ALLEN_CREATOR_MODULE = AllenConnectomeBuilder.__module__ ALLEN_CREATOR_CLASS = AllenConnectomeBuilder.__name__ MEASURE_METRICS_MODULE = TimeseriesMetricsAdapter.__module__ MEASURE_METRICS_CLASS = TimeseriesMetricsAdapter.__name__ DISCRETE_PSE_ADAPTER_MODULE = DiscretePSEAdapter.__module__ DISCRETE_PSE_ADAPTER_CLASS = DiscretePSEAdapter.__name__ ISOCLINE_PSE_ADAPTER_MODULE = IsoclinePSEAdapter.__module__ ISOCLINE_PSE_ADAPTER_CLASS = IsoclinePSEAdapter.__name__ DEFAULT_PORTLETS = {0: {0: 'TimeSeries'}} DEFAULT_PROJECT_GID = '2cc58a73-25c1-11e5-a7af-14109fe3bf71'
[]
2024-01-10
teo-ma/azure-open-ai-embeddings-qna
code~utilities~helper.py
import os import openai from dotenv import load_dotenv import logging import re import hashlib from langchain.embeddings.openai import OpenAIEmbeddings from langchain.llms import AzureOpenAI from langchain.vectorstores.base import VectorStore from langchain.chains import ChatVectorDBChain from langchain.chains import ConversationalRetrievalChain from langchain.chains.qa_with_sources import load_qa_with_sources_chain from langchain.chains.llm import LLMChain from langchain.chains.chat_vector_db.prompts import CONDENSE_QUESTION_PROMPT from langchain.prompts import PromptTemplate from langchain.document_loaders.base import BaseLoader from langchain.document_loaders import WebBaseLoader from langchain.text_splitter import TokenTextSplitter, TextSplitter from langchain.document_loaders.base import BaseLoader from langchain.document_loaders import TextLoader from langchain.chat_models import ChatOpenAI from langchain.schema import AIMessage, HumanMessage, SystemMessage from utilities.formrecognizer import AzureFormRecognizerClient from utilities.azureblobstorage import AzureBlobStorageClient from utilities.translator import AzureTranslatorClient from utilities.customprompt import PROMPT from utilities.redis import RedisExtended from utilities.azuresearch import AzureSearch import pandas as pd import urllib from fake_useragent import UserAgent class LLMHelper: def __init__(self, document_loaders : BaseLoader = None, text_splitter: TextSplitter = None, embeddings: OpenAIEmbeddings = None, llm: AzureOpenAI = None, temperature: float = None, max_tokens: int = None, custom_prompt: str = "", vector_store: VectorStore = None, k: int = None, pdf_parser: AzureFormRecognizerClient = None, blob_client: AzureBlobStorageClient = None, enable_translation: bool = False, translator: AzureTranslatorClient = None): load_dotenv() openai.api_type = "azure" openai.api_base = os.getenv('OPENAI_API_BASE') openai.api_version = "2023-03-15-preview" openai.api_key = os.getenv("OPENAI_API_KEY") # Azure OpenAI settings self.api_base = openai.api_base self.api_version = openai.api_version self.index_name: str = "embeddings" self.model: str = os.getenv('OPENAI_EMBEDDINGS_ENGINE_DOC', "text-embedding-ada-002") self.deployment_name: str = os.getenv("OPENAI_ENGINE", os.getenv("OPENAI_ENGINES", "text-davinci-003")) self.deployment_type: str = os.getenv("OPENAI_DEPLOYMENT_TYPE", "Text") self.temperature: float = float(os.getenv("OPENAI_TEMPERATURE", 0.7)) if temperature is None else temperature self.max_tokens: int = int(os.getenv("OPENAI_MAX_TOKENS", -1)) if max_tokens is None else max_tokens self.prompt = PROMPT if custom_prompt == '' else PromptTemplate(template=custom_prompt, input_variables=["summaries", "question"]) self.vector_store_type = os.getenv("VECTOR_STORE_TYPE") # Azure Search settings if self.vector_store_type == "AzureSearch": self.vector_store_address: str = os.getenv('AZURE_SEARCH_SERVICE_NAME') self.vector_store_password: str = os.getenv('AZURE_SEARCH_ADMIN_KEY') else: # Vector store settings self.vector_store_address: str = os.getenv('REDIS_ADDRESS', "localhost") self.vector_store_port: int= int(os.getenv('REDIS_PORT', 6379)) self.vector_store_protocol: str = os.getenv("REDIS_PROTOCOL", "redis://") self.vector_store_password: str = os.getenv("REDIS_PASSWORD", None) if self.vector_store_password: self.vector_store_full_address = f"{self.vector_store_protocol}:{self.vector_store_password}@{self.vector_store_address}:{self.vector_store_port}" else: self.vector_store_full_address = f"{self.vector_store_protocol}{self.vector_store_address}:{self.vector_store_port}" self.chunk_size = int(os.getenv('CHUNK_SIZE', 500)) self.chunk_overlap = int(os.getenv('CHUNK_OVERLAP', 100)) self.document_loaders: BaseLoader = WebBaseLoader if document_loaders is None else document_loaders self.text_splitter: TextSplitter = TokenTextSplitter(chunk_size=self.chunk_size, chunk_overlap=self.chunk_overlap) if text_splitter is None else text_splitter self.embeddings: OpenAIEmbeddings = OpenAIEmbeddings(model=self.model, chunk_size=1) if embeddings is None else embeddings if self.deployment_type == "Chat": self.llm: ChatOpenAI = ChatOpenAI(model_name=self.deployment_name, engine=self.deployment_name, temperature=self.temperature, max_tokens=self.max_tokens if self.max_tokens != -1 else None) if llm is None else llm else: self.llm: AzureOpenAI = AzureOpenAI(deployment_name=self.deployment_name, temperature=self.temperature, max_tokens=self.max_tokens) if llm is None else llm if self.vector_store_type == "AzureSearch": self.vector_store: VectorStore = AzureSearch(azure_cognitive_search_name=self.vector_store_address, azure_cognitive_search_key=self.vector_store_password, index_name=self.index_name, embedding_function=self.embeddings.embed_query) if vector_store is None else vector_store else: self.vector_store: RedisExtended = RedisExtended(redis_url=self.vector_store_full_address, index_name=self.index_name, embedding_function=self.embeddings.embed_query) if vector_store is None else vector_store self.k : int = 3 if k is None else k self.pdf_parser : AzureFormRecognizerClient = AzureFormRecognizerClient() if pdf_parser is None else pdf_parser self.blob_client: AzureBlobStorageClient = AzureBlobStorageClient() if blob_client is None else blob_client self.enable_translation : bool = False if enable_translation is None else enable_translation self.translator : AzureTranslatorClient = AzureTranslatorClient() if translator is None else translator self.user_agent: UserAgent() = UserAgent() self.user_agent.random def add_embeddings_lc(self, source_url): try: documents = self.document_loaders(source_url).load() # Convert to UTF-8 encoding for non-ascii text for(document) in documents: try: if document.page_content.encode("iso-8859-1") == document.page_content.encode("latin-1"): document.page_content = document.page_content.encode("iso-8859-1").decode("utf-8", errors="ignore") except: pass docs = self.text_splitter.split_documents(documents) # Remove half non-ascii character from start/end of doc content (langchain TokenTextSplitter may split a non-ascii character in half) pattern = re.compile(r'[\x00-\x1f\x7f\u0080-\u00a0\u2000-\u3000\ufff0-\uffff]') for(doc) in docs: doc.page_content = re.sub(pattern, '', doc.page_content) if doc.page_content == '': docs.remove(doc) keys = [] for i, doc in enumerate(docs): # Create a unique key for the document source_url = source_url.split('?')[0] filename = "/".join(source_url.split('/')[4:]) hash_key = hashlib.sha1(f"{source_url}_{i}".encode('utf-8')).hexdigest() hash_key = f"doc:{self.index_name}:{hash_key}" keys.append(hash_key) doc.metadata = {"source": f"[{source_url}]({source_url}_SAS_TOKEN_PLACEHOLDER_)" , "chunk": i, "key": hash_key, "filename": filename} if self.vector_store_type == 'AzureSearch': self.vector_store.add_documents(documents=docs, keys=keys) else: self.vector_store.add_documents(documents=docs, redis_url=self.vector_store_full_address, index_name=self.index_name, keys=keys) except Exception as e: logging.error(f"Error adding embeddings for {source_url}: {e}") raise e def convert_file_and_add_embeddings(self, source_url, filename, enable_translation=False): # Extract the text from the file text = self.pdf_parser.analyze_read(source_url) # Translate if requested text = list(map(lambda x: self.translator.translate(x), text)) if self.enable_translation else text # Upload the text to Azure Blob Storage converted_filename = f"converted/{filename}.txt" source_url = self.blob_client.upload_file("\n".join(text), f"converted/{filename}.txt", content_type='text/plain; charset=utf-8') print(f"Converted file uploaded to {source_url} with filename {filename}") # Update the metadata to indicate that the file has been converted self.blob_client.upsert_blob_metadata(filename, {"converted": "true"}) self.add_embeddings_lc(source_url=source_url) return converted_filename def get_all_documents(self, k: int = None): result = self.vector_store.similarity_search(query="*", k= k if k else self.k) return pd.DataFrame(list(map(lambda x: { 'key': x.metadata['key'], 'filename': x.metadata['filename'], 'source': urllib.parse.unquote(x.metadata['source']), 'content': x.page_content, 'metadata' : x.metadata, }, result))) def get_semantic_answer_lang_chain(self, question, chat_history): question_generator = LLMChain(llm=self.llm, prompt=CONDENSE_QUESTION_PROMPT, verbose=False) doc_chain = load_qa_with_sources_chain(self.llm, chain_type="stuff", verbose=True, prompt=self.prompt) chain = ConversationalRetrievalChain( retriever=self.vector_store.as_retriever(), question_generator=question_generator, combine_docs_chain=doc_chain, return_source_documents=True, # top_k_docs_for_context= self.k ) result = chain({"question": question, "chat_history": chat_history}) context = "\n".join(list(map(lambda x: x.page_content, result['source_documents']))) sources = "\n".join(set(map(lambda x: x.metadata["source"], result['source_documents']))) container_sas = self.blob_client.get_container_sas() result['answer'] = result['answer'].split('SOURCES:')[0].split('Sources:')[0].split('SOURCE:')[0].split('Source:')[0] sources = sources.replace('_SAS_TOKEN_PLACEHOLDER_', container_sas) return question, result['answer'], context, sources def get_embeddings_model(self): OPENAI_EMBEDDINGS_ENGINE_DOC = os.getenv('OPENAI_EMEBDDINGS_ENGINE', os.getenv('OPENAI_EMBEDDINGS_ENGINE_DOC', 'text-embedding-ada-002')) OPENAI_EMBEDDINGS_ENGINE_QUERY = os.getenv('OPENAI_EMEBDDINGS_ENGINE', os.getenv('OPENAI_EMBEDDINGS_ENGINE_QUERY', 'text-embedding-ada-002')) return { "doc": OPENAI_EMBEDDINGS_ENGINE_DOC, "query": OPENAI_EMBEDDINGS_ENGINE_QUERY } def get_completion(self, prompt, **kwargs): if self.deployment_type == 'Chat': return self.llm([HumanMessage(content=prompt)]).content else: return self.llm(prompt)
[]
2024-01-10
adeelkhan47/AI_Squad_poc
src~api~detector~endpoint.py
import json import os from os.path import join, dirname from pathlib import Path import openai import whisper from dotenv import load_dotenv from flask import request from flask_restx import Resource from . import api, schema from .speech_util import SpeechToText dotenv_path = join(dirname(__file__), '.env') load_dotenv(dotenv_path) OPENAPI_KEY = os.getenv('OPENAPI_KEY') ASSEMBLYAI_KEY = os.getenv('ASSEMBLYAI_KEY') ROOT_DIR = Path(__file__).parents[3] AUDIOS_PATH = os.path.join(ROOT_DIR, "audios") if OPENAPI_KEY: openai.api_key = OPENAPI_KEY def perform_task(query): response = openai.Completion.create( engine="text-davinci-003", prompt=query, max_tokens=200 ) return response.choices[0].text.strip() def get_speech_to_text(path): model = whisper.load_model("base") # load audio and pad/trim it to fit 30 seconds audio = whisper.load_audio(path) audio = whisper.pad_or_trim(audio) # make log-Mel spectrogram and move to the same device as the model mel = whisper.log_mel_spectrogram(audio).to(model.device) # detect the spoken language _, probs = model.detect_language(mel) print(f"Detected language: {max(probs, key=probs.get)}") # decode the audio options = whisper.DecodingOptions(fp16=False) result = whisper.decode(model, mel, options) # print the recognized text return result.text def translate_in_desired_language(text, language): prompt = f''' Your task is to identify the language (in ISO Language Code) the text enclosed in triple back ticks is written in. \ Then translate that piece of text into the langauge prescribed in <>. \ The output should be in JSON using 'translation' and 'detected_language' as keys. \ <{language}> ```{text}``` ''' params = { "model": "gpt-3.5-turbo", "messages": [{"role": "user", "content": prompt}], "temperature": 0, "max_tokens": 60, "top_p": 1, "frequency_penalty": 0, "presence_penalty": 0 } # Generate the translation using OpenAI's GPT-3 language model response = openai.ChatCompletion.create(**params) # Extract the translated text from the response output_dict = json.loads(response.choices[0].message.content) return output_dict @api.route("") class TranslatorOperation(Resource): @api.doc("Get Translated to Desired Language") @api.marshal_list_with(schema.GetTranslation) @api.param("text", required=True) @api.param("output_language", required=True) def get(self): try: args = request.args.copy() output_lan = args.get("output_language") text = args.get("text") output_dict = translate_in_desired_language(text, output_lan) return {"translation": output_dict['translation'], "detected_language": output_dict['detected_language'], "error": None}, 200 except Exception as e: return {"translation": None, "error": e.__str__()}, 200 @api.route("/speechtotext") class GetSpeechToText(Resource): @api.doc("Speech to Text endpoint") @api.marshal_list_with(schema.GetSpeechToText) @api.param("audio_path", required=True) def get(self): try: args = request.args.copy() audio_path = args.get("audio_path") upload_url = SpeechToText().upload_file(ASSEMBLYAI_KEY, audio_path) transcript = SpeechToText().create_transcript(ASSEMBLYAI_KEY, upload_url) return {"text": transcript['text'], 'language_code': transcript['language_code'], "error": None}, 200 except Exception as e: return {"translation": None, "error": e.__str__()}, 200 def post(self): try: file = request.files['file'] file_path = os.path.join(AUDIOS_PATH, file.filename) file.save(file_path) text = get_speech_to_text(file_path) response = perform_task(text) print(response) return {'text': response} except Exception as e: return {"translation": None, "error": e.__str__()}, 200
[ "\n Your task is to identify the language (in ISO Language Code) the text enclosed in triple back ticks is written in. Then translate that piece of text into the langauge prescribed in <>. The output should be in JSON using 'translation' and 'detected_language' as keys. \n <PLACEHOLDER>\n ```PLACEHOLDER```\n " ]
2024-01-10
apatankar22/LIGN167_Project
sql_session.py
from urllib.parse import urlparse import urllib.robotparser import requests import re from bs4 import BeautifulSoup import string import numpy as np import os import openai from io import StringIO import sys from pyfiglet import Figlet class SQLSession: w3_sql_topic_urls = {'Retrieve data from tables': ['https://www.w3resource.com/sql-exercises/sql-retrieve-exercise-{}.php', range(1, 34)], 'Boolean and Relational Operators': ['https://www.w3resource.com/sql-exercises/sql-boolean-operator-exercise-{}.php', range(1, 13)], 'Wildcard and Special operators': ['https://www.w3resource.com/sql-exercises/sql-wildcard-special-operator-exercise-{}.php', range(1, 23)], 'Aggregate Functions': ['https://www.w3resource.com/sql-exercises/sql-aggregate-function-exercise-{}.php', range(1, 26)], 'Formatting query output': ['https://www.w3resource.com/sql-exercises/sql-formatting-output-exercise-{}.php', range(1, 11)], 'SQL JOINS': ['https://www.w3resource.com/sql-exercises/sql-joins-exercise-{}.php', range(1, 30)] } w3_sql_topics = list(w3_sql_topic_urls.keys()) # PLEASE ENTER IN YOUR OWN OPEN AI API KEY openai.api_key = 'your-key-here' with open('sql_prompt.txt') as f: PROMPT = f.read() STDOUT = sys.stdout f = Figlet(font='slant') def __init__(self, username, new_user = True): new_session_message = "Welcome, {}, to your SQL Practice Hub! Here you'll be able to practice writing SQL queries in 6 different topics and get instant feedback." self.username = username self.question_tracker = {topic:set() for topic in SQLSession.w3_sql_topics} self.total_questions_completed = 0 self.current_topic = 0 self.current_q_num = 0 self.current_results = '' self.is_new_user = new_user print(SQLSession.f.renderText('The SQL Practice Hub')) print(new_session_message.format(self.username)) self.run_session(0) def run_session(self, flow): # exit session if flow == 4: print(SQLSession.f.renderText('Thank you!')) print('We hope The SQL Practice Hub helped you learn SQL or prepare for your next interview :)') return # repeat question elif flow == 1: # user solution user_input_query = self.get_user_input_query(self.current_results) # gpt3 comparison self.openai_api_call(self.current_topic, self.current_q_num, user_input_query, self.current_results) # ask for session input self.post_question(self.current_results) # new question elif flow == 0 or flow == 3: # get topic choice topic = self.get_user_topic_choice() # scrape w3 results q_num, exercise_url, topic = self.sql_exercises(topic) # new topic elif flow == 2: q_num, exercise_url, topic = self.sql_exercises(self.current_topic) results = self.get_sql_exercise_info(exercise_url) user_input_query = self.get_user_input_query(results) self.openai_api_call(self.current_topic, q_num, user_input_query, results) self.post_question() def get_user_topic_choice(self): print() print('Choose a Topic.') print() topic_list_message = '\n'.join(['{} for {}'.format(i+1, SQLSession.w3_sql_topics[i]) for i in range(len(SQLSession.w3_sql_topics))]) topic_prompt = 'Enter in the corresponding number for the topic you want to study. For example, if you would like to practice SQL JOINS, enter 6.\n' + topic_list_message + '\nYour choice: ' user_input = -1 while True: user_input = input(topic_prompt) if int(user_input) not in [1,2,3,4,5,6]: print('Invalid response. Try again.') continue else: break self.current_topic = int(user_input) return int(user_input) def sql_exercises(self, user_choice): topic = SQLSession.w3_sql_topics[user_choice - 1] url, q_nums = SQLSession.w3_sql_topic_urls[topic] completed = self.question_tracker[topic] q_nums_not_done = set(q_nums) - completed if len(q_nums_not_done) == 0: q_nums_not_done = q_nums random_choice = np.random.choice(list(q_nums_not_done)) self.current_q_num = random_choice self.question_tracker[topic].add(random_choice) return [random_choice, url.format(random_choice), user_choice] def get_sql_exercise_info(self, exercise_url): req_response = requests.get(exercise_url, timeout=5) html_content = BeautifulSoup(req_response.content, "html.parser") prompt = html_content.find_all('p')[0].text solution = html_content.find_all('code')[0].text table_strs = {} table_srcs = html_content.find_all("iframe", {"class": "span12"}) for lst in table_srcs: table = requests.get("https://www.w3resource.com/sql-exercises/{}".format(lst['src'])) table_name = lst['src'].strip('.php') table_str = BeautifulSoup(table.content, "html.parser").find('pre').text table_strs[table_name] = table_str results = {'prompt':prompt, 'solution':solution, 'tables':table_strs} self.current_results = results return results def get_user_input_query(self, results): print() print('************************************') print() print('Exercise:') print(results['prompt']) print() print('See sample table(s) below:') tables = results['tables'] for table in tables: print('Table name: ' + table) print(tables[table]) print('Please enter in your solution as one line (do not hit the Enter or Return key, and do not enter new lines)') user_input = input('Enter your solution here:') return user_input def openai_api_call(self, topic, q_num, user_input, results): sol = results['solution'] prompt = f'Solution SQL Query:\n{sol}\nInput SQL Query:{user_input}\nAnswer:\n```' response = openai.Completion.create( model="text-davinci-002", prompt=SQLSession.PROMPT+prompt, temperature=0, max_tokens=512, stop='```', ) self.total_questions_completed += 1 gpt3_feedback = response['choices'][0]['text'].strip() topic_str = SQLSession.w3_sql_topics[topic - 1] print('Topic: {}. Question #{}'.format(topic_str, q_num)) print('************************************') print('Example solution:') print(sol) print('************************************') print('Feedback:') print(gpt3_feedback) def post_question(self): post_question_prompt = 'If you would like to try this question again, enter 1. If you would like to try'\ ' a different question in this topic, enter 2. If you would like to try a new topic, enter 3. If you would like to'\ ' exit your session, enter 4: ' print('************************************') user_input = -1 while True: user_input = input(post_question_prompt) if int(user_input) not in [1,2,3,4]: print('Invalid response. Try again.') continue else: break user_input = int(user_input) self.run_session(user_input) class SQLInterview(SQLSession): def __init__(self, username, num_questions = 5): new_session_message = "Welcome, {}, to your SQL Interview Simulator. Here, you'll be asked 5 random SQL questions to assess your skills. At the end, you will be provided with feedback on each question. " self.username = username self.num_questions = num_questions self.question_tracker = {topic:set() for topic in SQLSession.w3_sql_topics} self.total_questions_completed = 0 self.current_topic = 0 print(new_session_message.format(self.username)) self.run_interview() def run_interview(self): exercises = self.get_random_questions() all_results = [self.get_sql_exercise_info(e[1]) for e in exercises] user_inputs = [] for result in all_results: user_inputs.append(self.get_user_input_query(result)) for i in range(len(all_results)): print('Question {} Results:'.format(i + 1)) self.openai_api_call(exercises[i][-1], exercises[i][0], user_inputs[i], all_results[i]) print(SQLSession.f.renderText('Thank you!')) print('We hope The SQL Interview Simulator helped you get one step closer to getting your dream job! Good luck :)') def get_random_questions(self): if self.num_questions <= 6: topics = np.random.choice(range(1, 7), self.num_questions, replace=False) exercises = [self.sql_exercises(topic) for topic in topics] else: exercises = [] for i in range(1, 7): for j in range(self.num_questions // 6): exercises.append(SQLSession.sql_exercises(i)) other_topics = np.random.choice(range(1, 7), self.num_questions % 6, replace=False) exercises += [self.sql_exercises(topic) for topic in other_topics] return exercises
[ "Solution SQL Query:\nPLACEHOLDER\nInput SQL Query:PLACEHOLDER\nAnswer:\n```", "Enter in the corresponding number for the topic you want to study. For example, if you would like to practice SQL JOINS, enter 6.\nPLACEHOLDER\nYour choice: ", "If you would like to try this question again, enter 1. If you would like to try a different question in this topic, enter 2. If you would like to try a new topic, enter 3. If you would like to exit your session, enter 4: " ]
2024-01-10
apatankar22/LIGN167_Project
ds_algo_gpt3.py
from urllib.parse import urlparse import urllib.robotparser import requests import re from bs4 import BeautifulSoup import string import numpy as np import os import openai from io import StringIO import sys from pyfiglet import Figlet class PythonCodingSession: all_string_exercises_url = 'https://www.w3resource.com/python-exercises/string/' all_ds_algo_exercises_url = 'https://www.w3resource.com/python-exercises/data-structures-and-algorithms/' all_ds_algo_response = requests.get(all_ds_algo_exercises_url, timeout=5) all_ds_algo_content = BeautifulSoup(all_ds_algo_response.content, "html.parser") all_ds_algo_text = [p.text for p in all_ds_algo_content.find_all('p')] all_ds_algo_exercises = [s for s in all_ds_algo_text if s[0].isdigit()] all_ds_algo_exercises_clean = list(map(lambda s: s.replace('Go to the editor', '').replace('Click me to see the sample solution', ''), all_ds_algo_exercises)) all_str_response = requests.get(all_string_exercises_url, timeout=5) all_str_content = BeautifulSoup(all_str_response.content, "html.parser") all_str_text = [p.text for p in all_str_content.find_all('p')] all_str_exercises = [s for s in all_str_text if s[0].isdigit()] all_str_exercises_clean = list(map(lambda s: s.replace('Go to the editor', '').replace('Click me to see the sample solution', ''), all_str_exercises)) all_questions = [all_str_exercises_clean, all_ds_algo_exercises_clean] # PLEASE ENTER IN YOUR OWN OPEN AI API KEY openai.api_key = 'your-key-here' with open('python_prompt.txt') as f: python_prompt = f.read() with open('python_comparison_prompt.txt') as f: python_comparison_prompt = f.read() STDOUT = sys.stdout f = Figlet(font='slant') def __init__(self, username): new_session_message = "Welcome, {}, to your Python Practice Hub! Here you'll be able to practice your Python coding skills in string manipulation, data structures, and algorithms and get instant feedback and sample solutions." self.username = username self.total_questions_completed = 0 self.current_topic = 0 self.current_question = 0 self.current_results = '' print(PythonCodingSession.f.renderText('The Python Practice Hub')) print(new_session_message.format(self.username)) self.run_session(0) def run_session(self, flow): # exit session if flow == 4: print(PythonCodingSession.f.renderText('Thank you!')) print('We hope The Python Practice Hub helped you learn SQL or prepare for your next interview :)') return # repeat question elif flow == 1: user_text = self.get_user_input_solution(self.current_question) self.get_gpt3_solution_and_comparison(self.current_question, user_text) self.post_question() # new question new topic elif flow == 0 or flow == 3: topic = self.get_user_topic_choice() random_choice = self.get_python_exercise_info(topic) # new question same topic elif flow == 2: random_choice = self.get_python_exercise_info(self.current_topic) user_text = self.get_user_input_solution(random_choice) self.get_gpt3_solution_and_comparison(random_choice, user_text) self.post_question() def get_user_topic_choice(self): print() print('Choose a Topic.') print() topic_list_message = '1 for String Manipulation.\n2 for Data Structures and Algorithms.' topic_prompt = 'Enter in the corresponding number for the topic you want to study. For example, if you would like to practice String Manipulation, enter 1.\n' + topic_list_message + '\nYour choice: ' user_input = -1 while True: user_input = input(topic_prompt) if int(user_input) not in [1,2]: print('Invalid response. Try again.') continue else: break self.current_topic = int(user_input) return int(user_input) def get_python_exercise_info(self, user_choice): topic = user_choice - 1 questions = PythonCodingSession.all_questions[topic] random_choice = np.random.choice(questions, replace = False) self.current_question = random_choice return random_choice def get_user_input_solution(self, random_choice): print() print('************************************') print() print('Exercise:') print(random_choice) print() print('Write your answer in a .py file. Then, enter in the filepath of your solution. Double check your filepath does not have any typos.') filepath = input('Enter in the .py filepath of your solution here (include file extension):') user_file = open(filepath, 'r') user_text = user_file.read() return user_text def get_gpt3_solution_and_comparison(self, random_choice, user_text): prompt = f'Prompt:\n{random_choice}\nAnswer:\n```' response = openai.Completion.create( model="text-davinci-002", prompt=PythonCodingSession.python_prompt+prompt, temperature=0, max_tokens=512, stop='```', ) self.total_questions_completed += 1 gpt3_solution = response['choices'][0]['text'].strip() print('Example solution:') print(gpt3_solution) print('************************************') comparison_prompt = f'Solution function:\n{gpt3_solution}\nInputted function:{user_text}\nFeedback:\n```' comparison_response = openai.Completion.create( model="text-davinci-002", prompt=PythonCodingSession.python_comparison_prompt+comparison_prompt, temperature=0, max_tokens=512, stop='```', ) gpt3_feedback = comparison_response['choices'][0]['text'].strip() print('Feedback:') print(gpt3_feedback) def post_question(self): post_question_prompt = 'If you would like to try this question again, enter 1. If you would like to try'\ ' a different question in this topic, enter 2. If you would like to try a new topic, enter 3. If you would like to'\ ' exit your session, enter 4: ' print('************************************') user_input = -1 while True: user_input = input(post_question_prompt) if int(user_input) not in [1,2,3,4]: print('Invalid response. Try again.') continue else: break user_input = int(user_input) self.run_session(user_input)
[ "Prompt:\nPLACEHOLDER\nAnswer:\n```", "Solution function:\nPLACEHOLDER\nInputted function:PLACEHOLDER\nFeedback:\n```", "If you would like to try this question again, enter 1. If you would like to try a different question in this topic, enter 2. If you would like to try a new topic, enter 3. If you would like to exit your session, enter 4: ", "Enter in the corresponding number for the topic you want to study. For example, if you would like to practice String Manipulation, enter 1.\n1 for String Manipulation.\n2 for Data Structures and Algorithms.\nYour choice: " ]
2024-01-10
LuxciumProject/monorepo-one
examples~open-ai~src~python~tts~speach.py
import os import openai from pydub import AudioSegment from pydub.playback import play import io # Set API key openai.api_key = os.getenv("OPENAI_API_KEY") # Create OpenAI client client = openai.OpenAI() # Generate speech audio response = client.audio.speech.create( model="tts-1", voice="alloy", input="Hello world! This is a streaming test." ) # Stream audio to speakers audio_stream = io.BytesIO(response.content) song = AudioSegment.from_file(audio_stream, format="mp3") play(song)
[]
2024-01-10
LuxciumProject/monorepo-one
examples~open-ai~src~python~davinci-002-prompter.py
import os import openai try: api_key = os.environ['KILO_TOKENS'] except KeyError: print("Error: The 'KILO_TOKENS' environment variable is not set.") exit(1) openai.api_key = api_key PROMPT = "Please generate a Python code snippet that converts a list of Fahrenheit temperatures to Celsius temperatures, with a temperature format of '{temperature}°F', and a maximum length of 1000 tokens." MODEL_ENGINE = "text-davinci-002" try: completions = openai.Completion.create( engine=MODEL_ENGINE, prompt=PROMPT, max_tokens=1000, n=1) output = completions.choices[0].text.strip() print(output) except Exception as e: print(f"Error: {e}")
[ "Please generate a Python code snippet that converts a list of Fahrenheit temperatures to Celsius temperatures, with a temperature format of '{temperature}°F', and a maximum length of 1000 tokens." ]
2024-01-10
lakshyagithub/Python-projects
HTML-IDE~HTML-IDE.py
from tkinter import * from PIL import Image, ImageTk from tkinter import filedialog from tkinter import messagebox import os import webbrowser import openai root = Tk() root.title("HTML editor - No file open") root.config(background="sky blue") root.minsize(600, 500) root.maxsize(600, 500) openai.api_key = "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" save_img = ImageTk.PhotoImage(Image.open("save1.png")) open_file_img = ImageTk.PhotoImage(Image.open("open1.png")) debug_img = ImageTk.PhotoImage(Image.open("run.png")) label_file_name = Label(root, text="File name: ") label_file_name.place(relx=0.6, rely=0.1, anchor=CENTER) input_file_name = Entry(root) input_file_name.place(relx=0.8, rely=0.1, anchor=CENTER) my_text = Text(root, height=20, width=60, background="grey", fg="white") my_text.place(relx=0.5, rely=0.55, anchor=CENTER) name = "" code = "" def open_file(): global name input_file_name.delete(0, END) my_text.delete(1.0, END) text_file = filedialog.askopenfilename(title="Select a html file (.html)", filetypes=(("HTML documents", "*.html"), )) print(text_file) name = os.path.basename(text_file) formated_name = name.split(".")[0] input_file_name.insert(END, formated_name) root.title("HTML editor - " + formated_name) text_file = open(name, "r") paragraph = text_file.read() my_text.insert(END, paragraph) text_file.close() def save_file(): input_name = input_file_name.get() file = open(input_name + ".html", "w") data = my_text.get("1.0", END) print(data) file.write(data) input_file_name.delete(0, END) my_text.delete(1.0, END) messagebox.showinfo("Done!", "Your file was saved!") def run_file(): global name webbrowser.open(name) def autocom(): global code code = openai.ChatCompletion.create(model="gpt-3.5-turbo", messages=[{"role": "user", "content": my_text.get("1.0", END)}]) my_text.insert(END, code.choices[0].message.content) open_button = Button(root, image=open_file_img, command=open_file) open_button.place(relx=0.05, rely=0.1, anchor=CENTER) save_button = Button(root, image=save_img, command=save_file) save_button.place(relx=0.11, rely=0.1, anchor=CENTER) exit_button = Button(root, image=debug_img, command=run_file) exit_button.place(relx=0.17, rely=0.1, anchor=CENTER) autocom_btn = Button(root, text="Autopilot", command=autocom) autocom_btn.place(relx=0.25, rely=0.1, anchor=CENTER) root.mainloop()
[]
2024-01-10
infiniteoo/audiobook_visualizer
visualizer.py
import os from dotenv import load_dotenv import time from openai import OpenAI import speech_recognition as sr from PIL import Image import matplotlib.pyplot as plt from pydub import AudioSegment import requests from io import BytesIO import threading import pygame load_dotenv() openai_api_key = os.getenv("OPENAI_API_KEY") client = OpenAI(api_key=openai_api_key) # Replace with your OpenAI API key chunk_duration = 10 # seconds def transcribe_audio(audio_file): recognizer = sr.Recognizer() recognizer.energy_threshold = 4000 with sr.AudioFile(audio_file) as source: audio = recognizer.record(source) try: text = recognizer.recognize_google(audio) return text except sr.UnknownValueError: print("Speech recognition could not understand audio") return "" except sr.RequestError as e: print(f"Could not request results from Google Speech Recognition service; {e}") return "" def get_current_audio_chunk(audio_file, start_time=0, end_time=10): audio = AudioSegment.from_file(audio_file) chunk = audio[start_time * 1000:end_time * 1000] chunk.export("current_chunk.wav", format="wav") return "current_chunk.wav" def generate_and_save_image(prompt, save_path): response = client.images.generate( model="dall-e-3", prompt=prompt, size="1024x1024", # Adjust the size based on what is supported by the model quality="standard", n=1, ) image_url = response.data[0].url image_data = requests.get(image_url).content with open(save_path, 'wb') as image_file: image_file.write(image_data) img = Image.open(BytesIO(image_data)) plt.imshow(img) plt.axis('off') plt.pause(0.001) # Add a small pause to allow the script to continue def display_images_concurrently(audio_file, start_time, end_time): audio_thread = threading.Thread(target=play_audio, args=(audio_file,)) audio_thread.start() while end_time <= len(AudioSegment.from_file(audio_file)) / 1000: current_audio_chunk = get_current_audio_chunk(audio_file, start_time, end_time) transcribed_text = transcribe_audio(current_audio_chunk) if transcribed_text: timestamp = time.strftime("%Y%m%d_%H%M%S") save_path = f"./images/generated_image_{timestamp}.png" generate_and_save_image(transcribed_text, save_path) start_time = end_time end_time += chunk_duration time.sleep(chunk_duration) def play_audio(audio_file): pygame.mixer.init() pygame.mixer.music.load(audio_file) pygame.mixer.music.play() while pygame.mixer.music.get_busy(): pygame.time.Clock().tick(10) if __name__ == "__main__": audio_file = "book.mp3" start_time = 0 end_time = 10 display_images_concurrently(audio_file, start_time, end_time)
[]
2024-01-10
brunotech/Binder
nsql~nsql_exec_python.py
# For sync the envs. import random import json import pandas as pd import pickle from nsql.qa_module.openai_qa import OpenAIQAModel import os import time from subprocess import PIPE, Popen import uuid # For Python execution. class Executor(object): def __init__(self, args, keys=None): self.new_col_name_id = 0 self.qa_model = OpenAIQAModel(args, keys) def nsql_exec(self, nsql: str, db: pd.DataFrame, verbose=True): # Add import part import_part = """import random import json import pandas as pd import pickle import numpy as np from collections.abc import Iterable from nsql.qa_module.openai_qa import OpenAIQAModel from nsql.database import NeuralDB import copy import os import time verbose = {}""".format(str(verbose)) # Add qa_map function qa_map_function_part = """def qa_map(db: pd.DataFrame, question, columns): new_db = NeuralDB([{"title": "", "table": {"header": db.columns.values.tolist(), "rows": db.values.tolist()}}]) sql_executed_sub_tables = [] for column in columns: column = f"`{column}`" sql_executed_sub_tables.append(new_db.execute_query(column)) sub_table = qa_model.qa(question, sql_executed_sub_tables, table_title=new_db.table_title, qa_type="map", new_col_name_s=[question], verbose=verbose) new_db.add_sub_table(sub_table, verbose=verbose) table = new_db.get_table() return pd.DataFrame(table["rows"], columns=table["header"])""" # Add qa_ans function qa_ans_function_part = """def qa_ans(db: pd.DataFrame, question, columns): new_db = NeuralDB([{"title": "", "table": {"header": db.columns.values.tolist(), "rows": db.values.tolist()}}]) sql_executed_sub_tables = [] for column in columns: column = f"`{column}`" sql_executed_sub_tables.append(new_db.execute_query(column)) answer = qa_model.qa(question,sql_executed_sub_tables,table_title=new_db.table_title,qa_type="ans",verbose=verbose) return answer""" # Convert np number type to python type convert_part = """def nested_to_python_number(x): if isinstance(x, np.int64): return int(x) if isinstance(x, np.float64): return float(x) if isinstance(x, Iterable) and not isinstance(x, (str, bytes)): return [nested_to_python_number(d) for d in x] return x""" # The prediction is a neural-python. # Add main function tmp_root_path = "tmp_python" os.makedirs(tmp_root_path, exist_ok=True) # Save the db db_file_path = '{}.db'.format(format(uuid.uuid4())) db_path = os.path.join(tmp_root_path, db_file_path) with open(db_path, "wb") as f: pickle.dump(db, f) # Save the qa_model model_file_path = '{}.model'.format(format(uuid.uuid4())) model_path = os.path.join(tmp_root_path, model_file_path) with open(model_path, "wb") as f: pickle.dump(self.qa_model, f) # Set the result path result_file_path = '{}.json'.format(format(uuid.uuid4())) result_path = os.path.join(tmp_root_path, result_file_path) # Read it and call solve function main_part = """if __name__ == '__main__': with open("{}", "rb") as f: db = pickle.load(f) with open("{}", "rb") as f: qa_model = pickle.load(f) result = solve(db) result = nested_to_python_number(result) with open("{}", "w") as f: json.dump(result, f)""".format(db_path, model_path, result_path) # Concat the code and execute the python all_code = "{}\n\n{}\n\n{}\n\n{}\n\n".format(import_part, qa_map_function_part, qa_ans_function_part, convert_part) + nsql + "\n\n" + main_part if verbose: print("----> Code <----") print(all_code) python_file_path = '{}.py'.format(format(uuid.uuid4())) python_path = os.path.join(tmp_root_path, python_file_path) with open(python_path, "w") as f: f.write(all_code) p = Popen("python " + python_path, shell=True, stdout=PIPE, stderr=PIPE) stdout, stderr = p.communicate() # Error in execution so that we didn't get result. if not os.path.exists(result_path): print("stderr: ", stderr) raise ValueError("Error execution!") # Read the result with open(result_path, "r") as f: result = json.load(f) return result
[]
2024-01-10
brunotech/Binder
nsql~nsql_exec.py
from typing import List, Dict from nsql.qa_module.openai_qa import OpenAIQAModel from nsql.qa_module.vqa import vqa_call from nsql.database import NeuralDB from nsql.parser import get_cfg_tree, get_steps, remove_duplicate, TreeNode, parse_question_paras, nsql_role_recognize, \ extract_answers class Executor(object): def __init__(self, args, keys=None): self.new_col_name_id = 0 self.qa_model = OpenAIQAModel(args, keys) def generate_new_col_names(self, number): col_names = ["col_{}".format(i) for i in range(self.new_col_name_id, self.new_col_name_id + number)] self.new_col_name_id += number return col_names def sql_exec(self, sql: str, db: NeuralDB, verbose=True): if verbose: print("Exec SQL '{}' with additional row_id on {}".format(sql, db)) result = db.execute_query(sql) return result def nsql_exec(self, nsql: str, db: NeuralDB, verbose=True): steps = [] root_node = get_cfg_tree(nsql) # Parse execution tree from nsql. get_steps(root_node, steps) # Flatten the execution tree and get the steps. steps = remove_duplicate(steps) # Remove the duplicate steps. if verbose: print("Steps:", [s.rename for s in steps]) col_idx = 0 for step in steps: # All steps should be formatted as 'QA()' except for last step which could also be normal SQL. assert isinstance(step, TreeNode), "step must be treenode" nsql = step.rename if nsql.startswith('QA('): question, sql_s = parse_question_paras(nsql, self.qa_model) sql_executed_sub_tables = [] # Execute all SQLs and get the results as parameters for sql_item in sql_s: role, sql_item = nsql_role_recognize(sql_item, db.get_header(), db.get_passages_titles(), db.get_images_titles()) if role in ['col', 'complete_sql']: sql_executed_sub_table = self.sql_exec(sql_item, db, verbose=verbose) sql_executed_sub_tables.append(sql_executed_sub_table) elif role == 'val': val = eval(sql_item) sql_executed_sub_tables.append({ "header": ["row_id", "val"], "rows": [["0", val]] }) elif role == 'passage_title_and_image_title': sql_executed_sub_tables.append({ "header": ["row_id", "{}".format(sql_item)], "rows": [["0", db.get_passage_by_title(sql_item) + db.get_image_caption_by_title(sql_item) # "{} (The answer of '{}' is {})".format( # sql_item, # # Add image qa result as backup info # question[len("***@"):], # vqa_call(question=question[len("***@"):], # image_path=db.get_image_by_title(sql_item))) ]] }) elif role == 'passage_title': sql_executed_sub_tables.append({ "header": ["row_id", "{}".format(sql_item)], "rows": [["0", db.get_passage_by_title(sql_item)]] }) elif role == 'image_title': sql_executed_sub_tables.append({ "header": ["row_id", "{}".format(sql_item)], "rows": [["0", db.get_image_caption_by_title(sql_item)]], # "rows": [["0", "{} (The answer of '{}' is {})".format( # sql_item, # # Add image qa result as backup info # question[len("***@"):], # vqa_call(question=question[len("***@"):], # image_path=db.get_image_by_title(sql_item)))]], }) # If the sub_tables to execute with link, append it to the cell. passage_linker = db.get_passage_linker() image_linker = db.get_image_linker() for _sql_executed_sub_table in sql_executed_sub_tables: for i in range(len(_sql_executed_sub_table['rows'])): for j in range(len(_sql_executed_sub_table['rows'][i])): _cell = _sql_executed_sub_table['rows'][i][j] if _cell in passage_linker.keys(): _sql_executed_sub_table['rows'][i][j] += " ({})".format( # Add passage text as backup info db.get_passage_by_title(passage_linker[_cell])) if _cell in image_linker.keys(): _sql_executed_sub_table['rows'][i][j] += " ({})".format( # Add image caption as backup info db.get_image_caption_by_title(image_linker[_cell])) # _sql_executed_sub_table['rows'][i][j] += " (The answer of '{}' is {})".format( # # Add image qa result as backup info # question[len("***@"):], # vqa_call(question=question[len("***@"):], # image_path=db.get_image_by_title(image_linker[_cell]))) pass if question.lower().startswith("map@"): # When the question is a type of mapping, we return the mapped column. question = question[len("map@"):] if step.father: step.rename_father_col(col_idx=col_idx) sub_table: Dict = self.qa_model.qa(question, sql_executed_sub_tables, table_title=db.table_title, qa_type="map", new_col_name_s=step.produced_col_name_s, verbose=verbose) db.add_sub_table(sub_table, verbose=verbose) col_idx += 1 else: # This step is the final step sub_table: Dict = self.qa_model.qa(question, sql_executed_sub_tables, table_title=db.table_title, qa_type="map", new_col_name_s=["col_{}".format(col_idx)], verbose=verbose) return extract_answers(sub_table) elif question.lower().startswith("ans@"): # When the question is a type of answering, we return an answer list. question = question[len("ans@"):] answer: List = self.qa_model.qa(question, sql_executed_sub_tables, table_title=db.table_title, qa_type="ans", verbose=verbose) if step.father: step.rename_father_val(answer) else: # This step is the final step return answer else: raise ValueError( "Except for operators or NL question must start with 'map@' or 'ans@'!, check '{}'".format( question)) else: sub_table = self.sql_exec(nsql, db, verbose=verbose) return extract_answers(sub_table)
[]
2024-01-10
PacktPublishing/fastText-Quick-Start-Guide
chapter5~Training_visualizations.py
# coding: utf-8 # # Setup Visdom # # Install it with: # # `pip install visdom` # # Start the server: # # `python -m visdom.server` # # Visdom now can be accessed at http://localhost:8097 in the browser. # # # # LDA Training Visualization # # Knowing about the progress and performance of a model, as we train them, could be very helpful in understanding it’s learning process and makes it easier to debug and optimize them. In this notebook, we will learn how to visualize training statistics for LDA topic model in gensim. To monitor the training, a list of Metrics is passed to the LDA function call for plotting their values live as the training progresses. # # # <img src="visdom_graph.png"> # # # Let's plot the training stats for an LDA model being trained on kaggle's [fake news dataset](https://www.kaggle.com/mrisdal/fake-news). We will use the four evaluation metrics available for topic models in gensim: Coherence, Perplexity, Topic diff and Convergence. (using separate hold_out and test corpus for evaluating the perplexity) # In[*] from gensim.models.fasttext import FastText from gensim.corpora import Dictionary import pandas as pd import re from gensim.parsing.preprocessing import remove_stopwords, strip_punctuation import numpy as np df_fake = pd.read_csv('fake.csv') df_fake[['title', 'text', 'language']].head() df_fake = df_fake.loc[(pd.notnull(df_fake.text)) & (df_fake.language == 'english')] # remove stopwords and punctuations def preprocess(row): return strip_punctuation(remove_stopwords(row.lower())) df_fake['text'] = df_fake['text'].apply(preprocess) # Convert data to required input format by LDA texts = [] for line in df_fake.text: lowered = line.lower() words = re.findall(r'\w+', lowered) texts.append(words) # In[*] from gensim.test.utils import common_texts as sentences from gensim.models.callbacks import CallbackAny2Vec from gensim.models import Word2Vec from gensim.test.utils import get_tmpfile class EpochSaver(CallbackAny2Vec): "Callback to save model after every epoch" def __init__(self, path_prefix): self.path_prefix = path_prefix self.epoch = 0 def on_epoch_end(self, model): output_path = '{}_epoch{}.model'.format(self.path_prefix, self.epoch) print("Save model to {}".format(output_path)) model.save(output_path) self.epoch += 1 # to save the similarity scores similarity = [] class EpochLogger(CallbackAny2Vec): "Callback to log information about training" def __init__(self): self.epoch = 0 def on_epoch_begin(self, model): print("Epoch #{} start".format(self.epoch)) def on_epoch_end(self, model): print("Epoch #{} end".format(self.epoch)) self.epoch += 1 def on_batch_begin(self, model): similarity.append(model.wv.similarity('woman', 'man')) # In[*] epoch_saver = EpochSaver(get_tmpfile("temporary_model")) epoch_logger = EpochLogger() # In[*] # from gensim.models.callbacks import CoherenceMetric, DiffMetric, PerplexityMetric, ConvergenceMetric # # define perplexity callback for hold_out and test corpus # # pl_holdout = PerplexityMetric(texts=texts, logger="visdom", title="Perplexity (hold_out)") # # pl_test = PerplexityMetric(texts=texts, logger="visdom", title="Perplexity (test)") # # define other remaining metrics available # ch_umass = CoherenceMetric(texts=texts, coherence="u_mass", logger="visdom", title="Coherence (u_mass)") # # ch_cv = CoherenceMetric(texts=texts, texts=training_texts, coherence="c_v", logger="visdom", title="Coherence (c_v)") # ch_cv = CoherenceMetric(texts=texts, coherence="c_v", logger="visdom", title="Coherence (c_v)") # diff_kl = DiffMetric(distance="kullback_leibler", logger="visdom", title="Diff (kullback_leibler)") # convergence_kl = ConvergenceMetric(distance="jaccard", logger="visdom", title="Convergence (jaccard)") # callbacks = [ch_umass] # # training LDA model # # model = ldamodel.LdaModel(corpus=training_corpus, id2word=dictionary, num_topics=35, passes=50, chunksize=1500, iterations=200, alpha='auto', callbacks=callbacks) # In[*] import gensim import os from gensim.models.word2vec import LineSentence from gensim.models.fasttext import FastText as FT_gensim # Set file names for train and test data lee_train_file = './gensim/gensim/test/test_data/lee_background.cor' lee_data = LineSentence(lee_train_file) model_gensim = FT_gensim(size=100) # build the vocabulary model_gensim.build_vocab(lee_data) # train the model model_gensim.train(lee_data, total_examples=model_gensim.corpus_count, epochs=model_gensim.epochs, callbacks=[epoch_saver, epoch_logger]) print(model_gensim) # When the model is set for training, you can open http://localhost:8097 to see the training progress. # # Training Logs # # We can also log the metric values after every epoch to the shell apart from visualizing them in Visdom. The only change is to define `logger="shell"` instead of `"visdom"` in the input callbacks. # In[*] import visdom vis = visdom.Visdom() trace = dict(x=[1, 2, 3], y=[4, 5, 6], mode="markers+lines", type='custom', marker={'color': 'red', 'symbol': 104, 'size': "10"}, text=["one", "two", "three"], name='1st Trace') layout = dict(title="First Plot", xaxis={'title': 'x1'}, yaxis={'title': 'x2'}) vis._send({'data': [trace], 'layout': layout, 'win': 'mywin'}) # In[*] len(similarity) # In[*] import visdom vis = visdom.Visdom() trace = dict(x=list(range(len(similarity))), y=similarity, mode="markers+lines", type='custom', marker={'color': 'red', 'symbol': 104, 'size': "10"}, text=["one", "two", "three"], name='1st Trace') layout = dict(title="First Plot", xaxis={'title': 'x1'}, yaxis={'title': 'x2'}) vis._send({'data': [trace], 'layout': layout, 'win': 'mywin'}) # In[*] similarity
[]
2024-01-10
Marshal-AM/AI_Phase1
AI_Phase3.py
import openai openai.api_key = '' def load_qa_pairs(file_path): qa_pairs = [] with open(file_path, 'r') as file: lines = file.readlines() for line in lines: question, answer = line.strip().split(',') qa_pairs.append({'question': question, 'answer': answer}) return qa_pairs qa_pairs = load_qa_pairs('fitness_qa.txt') def find_answer_in_dataset(user_input, qa_pairs): for qa in qa_pairs: if user_input.lower() in qa['question'].lower(): return qa['answer'] return None def get_gpt3_response(user_input): response = openai.Completion.create( engine="text-davinci-002", prompt=user_input, max_tokens=50 ) return response.choices[0].text def chatbot_response(user_input): dataset_answer = find_answer_in_dataset(user_input, qa_pairs) if dataset_answer: return dataset_answer else: return get_gpt3_response(user_input) user_input = input("You: ") response = chatbot_response(user_input) print("FitnessBot:", response)
[]
2024-01-10
AndreasKaratzas/td3
common~structures.py
import sys sys.path.append('../') import operator class SegmentTree: """ Create SegmentTree. Taken from OpenAI baselines github repository: https://github.com/openai/baselines/blob/master/baselines/common/segment_tree.py Attributes: capacity (int) tree (list) operation (function) """ def __init__(self, capacity, operation, init_value): """Initialization. Args: capacity (int) operation (function) init_value (float) """ assert ( capacity > 0 and capacity & (capacity - 1) == 0 ), "capacity must be positive and a power of 2." self.capacity = capacity self.tree = [init_value for _ in range(2 * capacity)] self.operation = operation def _operate_helper(self, start, end, node, node_start, node_end): """Returns result of operation in segment.""" if start == node_start and end == node_end: return self.tree[node] mid = (node_start + node_end) // 2 if end <= mid: return self._operate_helper(start, end, 2 * node, node_start, mid) else: if mid + 1 <= start: return self._operate_helper(start, end, 2 * node + 1, mid + 1, node_end) else: return self.operation( self._operate_helper( start, mid, 2 * node, node_start, mid), self._operate_helper( mid + 1, end, 2 * node + 1, mid + 1, node_end), ) def operate(self, start=0, end=0): """Returns result of applying `self.operation`.""" if end <= 0: end += self.capacity end -= 1 return self._operate_helper(start, end, 1, 0, self.capacity - 1) def __setitem__(self, idx, val): """Set value in tree.""" idx += self.capacity self.tree[idx] = val idx //= 2 while idx >= 1: self.tree[idx] = self.operation( self.tree[2 * idx], self.tree[2 * idx + 1]) idx //= 2 def __getitem__(self, idx): """Get real value in leaf node of tree.""" assert 0 <= idx < self.capacity return self.tree[self.capacity + idx] class SumSegmentTree(SegmentTree): """ Create SumSegmentTree. Taken from OpenAI baselines github repository: https://github.com/openai/baselines/blob/master/baselines/common/segment_tree.py """ def __init__(self, capacity): """Initialization. Args: capacity (int) """ super(SumSegmentTree, self).__init__( capacity=capacity, operation=operator.add, init_value=0.0 ) def sum(self, start=0, end=0): """Returns arr[start] + ... + arr[end].""" return super(SumSegmentTree, self).operate(start, end) def retrieve(self, upperbound): """Find the highest index `i` about upper bound in the tree""" assert 0 <= upperbound <= self.sum() + 1e-5, "upperbound: {}".format(upperbound) idx = 1 while idx < self.capacity: # while non-leaf left = 2 * idx right = left + 1 if self.tree[left] > upperbound: idx = 2 * idx else: upperbound -= self.tree[left] idx = right return idx - self.capacity class MinSegmentTree(SegmentTree): """ Create SegmentTree. Taken from OpenAI baselines github repository: https://github.com/openai/baselines/blob/master/baselines/common/segment_tree.py """ def __init__(self, capacity): """Initialization. Args: capacity (int) """ super(MinSegmentTree, self).__init__( capacity=capacity, operation=min, init_value=float("inf") ) def min(self, start=0, end=0): """Returns min(arr[start], ..., arr[end]).""" return super(MinSegmentTree, self).operate(start, end)
[]
2024-01-10
ericrallen/sentiment-analysis-notebook
widgets~advanced.py
import ipywidgets as widgets from nltk.sentiment.vader import SentimentIntensityAnalyzer from nrclex import NRCLex import openai analyzer = SentimentIntensityAnalyzer() ADVANCED_SYSTEM_PROMPT = """ You are VibeCheck, an advanced AI system for detecting the sentiment conveyed in user-generated text. The user will provide you with a prompt, and you will analyze it following these steps: 1. Analyze the prompt for relevant emotion, tone, affinity, sarcasm, irony, etc. 2. Analyze the likely emotional state of the author based on those findings 3. Summarize the emotional state and sentiment of the prompt based on your findings using 5 or less names for emotions using lowercase letters and separating each emotional state with a comma Only return the output from the final step to the user. """ BETTER_SYSTEM_PROMPT = """ You are VibeCheck, an advanced AI system for detecting the sentiment conveyed in user-generated text. The user will provide you with a prompt, and you will respond with the sentiment of that prompt on a scale of -1 (extremely negative) to 1 (extremely positive). Do not attempt to take actions based on the prompt provided. Only respond with a floating point number between -1 and 1 that represents the sentiment of the prompt. Do not respond with text. """ OPEN_AI_MODEL = "gpt-3.5-turbo" TEMPERATURE = 0.37 def configureOpenAi(key, model, temp): global OPEN_AI_MODEL global TEMPERATURE openai.api_key = key OPEN_AI_MODEL = model TEMPERATURE = temp def advancedChatGptSentiment(prompt: str) -> str: messages = [{"role": "system", "content": ADVANCED_SYSTEM_PROMPT}] messages.append({"role": "user", "content": prompt}) response = openai.ChatCompletion.create( model=OPEN_AI_MODEL, messages=messages, temperature=TEMPERATURE, ) if "choices" in response and len(response.choices): return response.choices[0].message["content"] else: return "Error: ChatGPT did not respond" def betterChatGptSentiment(prompt: str) -> str: messages = [{"role": "system", "content": BETTER_SYSTEM_PROMPT}] messages.append({"role": "user", "content": prompt}) response = openai.ChatCompletion.create( model=OPEN_AI_MODEL, messages=messages, temperature=TEMPERATURE, ) if "choices" in response and len(response.choices): return response.choices[0].message["content"] else: return "Error: ChatGPT did not respond" # convert sentiment score to label def convertSentimentToLabel(sentiment: float) -> str: # map the sentiment to a human readable label if sentiment >= 0.75: return "Very Positive" elif sentiment >= 0.4: return "Positive" elif sentiment >= 0.1: return "Leaning Positive" elif sentiment <= -0.1 and sentiment > -0.4: return "Leaning Negative" elif sentiment <= -0.4 and sentiment > -0.75: return "Negative" elif sentiment <= -0.75: return "Very Negative" else: return "Neutral" # analyze the sentiment of a string of text def analyzeBasicSentiment(text: str) -> dict[str, float]: if not text: return "" # use VADER to get the +/- sentiment of the string return analyzer.polarity_scores(text) def getNRCEmotion(text: str) -> list[tuple[str, float]]: emotion = NRCLex(text) return emotion.top_emotions def getAdvancedSentiment(event): text = advancedDemoString.value.strip() # Get the sentiment sentiment = analyzeBasicSentiment(text)["compound"] emotionAnalysis = getNRCEmotion(text) openAiSentimentScore = betterChatGptSentiment(text) openAiSentimentEmotion = advancedChatGptSentiment(text) emotions = [] for emotion, value in emotionAnalysis: if value > 0.00: emotions.append(emotion) if sentiment: with basicAnalysis: basicAnalysis.clear_output(wait=True) print(f"VADER: {sentiment}: {convertSentimentToLabel(sentiment)}") else: basicAnalysis.clear_output() if emotions: with nrcLexAnalysis: nrcLexAnalysis.clear_output(wait=True) print(f"NRC: {emotions}") else: nrcLexAnalysis.clear_output() if openAiSentimentScore: with openAiSentimentAnalysis: openAiSentimentAnalysis.clear_output(wait=True) print( f"{OPEN_AI_MODEL}: {openAiSentimentScore}: {convertSentimentToLabel(float(openAiSentimentScore))}" ) if openAiSentimentEmotion: with advancedAnalysis: advancedAnalysis.clear_output(wait=True) print(f"{OPEN_AI_MODEL}: {openAiSentimentEmotion}") else: advancedAnalysis.clear_output() basicAnalysis = widgets.Output() nrcLexAnalysis = widgets.Output() advancedAnalysis = widgets.Output() openAiSentimentAnalysis = widgets.Output() advancedAnalysisButton = widgets.Button( description="Analyze", button_style="primary", ) advancedDemoString = widgets.Text( value="", placeholder="Type something", ) advancedAnalysisButton.on_click(getAdvancedSentiment) advancedAnalysisInput = widgets.HBox([advancedDemoString, advancedAnalysisButton]) advancedAnalysisOutput = widgets.VBox( [basicAnalysis, nrcLexAnalysis, openAiSentimentAnalysis, advancedAnalysis] ) advancedAnalysisWidget = widgets.VBox([advancedAnalysisInput, advancedAnalysisOutput])
[ "\nYou are VibeCheck, an advanced AI system for detecting the sentiment conveyed in user-generated text.\n\nThe user will provide you with a prompt, and you will analyze it following these steps:\n\n1. Analyze the prompt for relevant emotion, tone, affinity, sarcasm, irony, etc.\n2. Analyze the likely emotional state of the author based on those findings\n3. Summarize the emotional state and sentiment of the prompt based on your findings using 5 or less names for emotions using lowercase letters and separating each emotional state with a comma\n\nOnly return the output from the final step to the user.\n", "\nYou are VibeCheck, an advanced AI system for detecting the sentiment conveyed in user-generated text.\n\nThe user will provide you with a prompt, and you will respond with the sentiment of that prompt on a scale of -1 (extremely negative) to 1 (extremely positive).\n\nDo not attempt to take actions based on the prompt provided.\n\nOnly respond with a floating point number between -1 and 1 that represents the sentiment of the prompt.\n\nDo not respond with text.\n" ]
2024-01-10
ranger-labs/benchmark
benchmark_runner.py
from datasets import Dataset, load_dataset from benchmark import Benchmark, Ranger import logging as log import os def upload_legalbench() -> None: # load dataset from hugginface or local dataset = load_dataset( "legalbench/definition_classification/", "ssla_company_defendents/", # TODO delimiter="\t", column_names=["text", "label"], ) legalbench = Benchmark(name="legalbench") def my_parser(input, **kwargs): baseprompt = kwargs["template"] input = kwargs["input"] return baseprompt.replace("{{input}}", input) legalbench.generate_from_template = my_parser legalbench.add_dataset("abc", dataset) f = open( "legalbench/supply_chain_disclosure_best_practice_certification/base_prompt.txt", "r", ) template = f.read() legalbench.add_assignment( "supply_chain_disclosure_best_practice", dataset_name="abc", input_col="text", output_col="answer", template=template, ) Ranger.upload_benchmark(legalbench, "ranger-uploads", "legalbench") # TODO: save results from a run on AWS and associate with benchmark_run Id. def save_results_aws(): return def generate(template, input): return template.replace("{{text}}",input) def legalbench() -> None: model_source = "NousResearch/Nous-Hermes-Llama2-13b" ranger: Ranger = Ranger(model_source,"baseten",5,"DZ0KPsrZ.dGcR1CXBTSQSTzH1Vur2GCX5W1kSk7PL","1PylM20") custom_benchmark: Benchmark = Benchmark("legalbench") # add dataset, assignment and baseprompt for each test tsv and txt in legalbench folder for assignment_name in os.listdir("legalbench"): if os.path.isdir(f"legalbench/{assignment_name}"): for file in os.listdir(f"legalbench/{assignment_name}"): if file.endswith("test.tsv"): dataset_name = file custom_benchmark.add_dataset_from_csv( dataset_name, f"legalbench/{assignment_name}/{file}", delimiter="\t", ) custom_benchmark.add_assignment( f"{assignment_name}_assignment", dataset_name, "text", "answer", ) elif file.endswith(".txt"): with open(f"legalbench/{assignment_name}/{file}", "r") as f: template = f.read() for assignment in custom_benchmark.assignments: if assignment.name == f"{assignment_name}_assignment": assignment.template = template assignment.generate_from_template = generate ranger.add_benchmark(custom_benchmark) ranger.run_benchmark("legalbench") for result in ranger.get_results(): log.info(result) def openai_evals(): # TODO import OPENAIEVALS from openai_utils import OpenAIEvals import json evals = OpenAIEvals() benchmark: Benchmark = Benchmark("openai_evals") input_col = "input" output_col = "ideal" model_source = "baseten" ranger: Ranger = Ranger( model_source, "baseten", 500, "DZ0KPsrZ.dGcR1CXBTSQSTzH1Vur2GCX5W1kSk7PL", "1PylM20", ) # Iterating over all of the assignents and their data for i, directory in enumerate(evals.DATA_PATH.iterdir()): # for i, directory in enumerate((evals.DATA_PATH / 'sarcasm').iterdir()): # XXX: For debugging, delete later # {{{ Toggle fold if i > 4: break # }}} if not directory.is_dir(): log.debug(f"Weird file: {directory} in {evals.DATA_PATH}") continue name = directory.absolute().name # basename of the file dataset_name = f"{name}_dataset" # Iterate over each file; create a dataset for each file; create an assignmnet dataset_files = list(directory.glob("*.jsonl")) for file in dataset_files: with open(file) as f: log.debug(f"Adding file: {file}") frames = list() # FIXME: This is likely EXTREMELY slow for line in f: json_data = json.loads(line) actual_input = "" # {{{ Convert Input column to strings for input_data in json_data[input_col]: content = input_data["content"] actual_input = f"{actual_input}\n\n{content}" # }}} json_data["input"] = actual_input frames.append(json_data) dataset = Dataset.from_list(frames) benchmark.add_dataset(dataset_name, dataset) # type: ignore[code] # If @OrestesK adds adding multiple datasets indent the add_assignment line, create an array of the datasets # and then add them in the array as a second argument benchmark.add_assignment(name, dataset_name, input_col, output_col) ranger.add_benchmark(benchmark) ranger.run_benchmark("openai_evals") for result in ranger.get_results(): log.info(result) def openai_testing(): from openai_utils import OpenAIEvals import json evals = OpenAIEvals() benchmark: Benchmark = Benchmark("openai_evals") input_col = "input" output_col = "ideal" model_source = "baseten" ranger: Ranger = Ranger( model_source, "baseten", "DZ0KPsrZ.dGcR1CXBTSQSTzH1Vur2GCX5W1kSk7PL", "1PylM20" ) name = "sarcasm" file = ( evals.DATA_PATH / f"{name}/samples.jsonl" ) # change this if you want to test it out on a different file # Iterating over all the assignents and their data with open(file) as f: log.debug(f"Adding file: {file}") frames = list() # FIXME: This is PROBABLY slow for line in f: json_data = json.loads(line) actual_input = "" # {{{ Convert Input column to strings for input_data in json_data[input_col]: content = input_data["content"] actual_input = f"{actual_input}\n\n{content}" # }}} json_data["input"] = actual_input frames.append(json_data) dataset = Dataset.from_list(frames) benchmark.add_dataset(f"{name}_dataset", dataset) # type: ignore[code] # If @OrestesK adds multiple datasets indent the add_assignment line, create an array of the datasets # and then add them in the array as a second argument benchmark.add_assignment(name, f"{name}_dataset", input_col, output_col) # TODO: @OrestesK, can you add logging for which assignment is being run? ranger.add_benchmark(benchmark) ranger.run_benchmark("openai_evals") for result in ranger.get_results(): log.info(result) if __name__ == "__main__": log.basicConfig(level=log.DEBUG, filename="benchmark_runner.log") # upload_legalbench() legalbench() # openai_evals() #openai_testing()
[]
2024-01-10
mwang87/exmol
exmol~exmol.py
from functools import reduce import inspect from typing import * import io import math import requests # type: ignore import numpy as np import matplotlib.pyplot as plt # type: ignore from matplotlib.patches import Rectangle, FancyBboxPatch # type: ignore from matplotlib.offsetbox import AnnotationBbox # type: ignore import matplotlib as mpl # type: ignore import selfies as sf # type: ignore import tqdm # type: ignore import textwrap # type: ignore import skunk # type: ignore from ratelimit import limits, sleep_and_retry # type: ignore from sklearn.cluster import DBSCAN # type: ignore from sklearn.decomposition import PCA # type: ignore import scipy.stats as ss # type: ignore from rdkit.Chem import MolFromSmiles as smi2mol # type: ignore from rdkit.Chem import MolFromSmarts # type: ignore from rdkit.Chem import MolToSmiles as mol2smi # type: ignore from rdkit.Chem import rdchem, MACCSkeys, AllChem # type: ignore from rdkit.Chem.Draw import MolToImage as mol2img, DrawMorganBit # type: ignore from rdkit.Chem import rdchem # type: ignore from rdkit.DataStructs.cDataStructs import BulkTanimotoSimilarity, TanimotoSimilarity # type: ignore from . import stoned from .plot_utils import _mol_images, _image_scatter, _bit2atoms from .data import * def _fp_dist_matrix(smiles, fp_type, _pbar): mols = [(smi2mol(s), _pbar.update(0.5))[0] for s in smiles] # Sorry about the one-line. Just sneaky insertion of progressbar update fp = [(stoned.get_fingerprint(m, fp_type), _pbar.update(0.5))[0] for m in mols] M = np.array([BulkTanimotoSimilarity(f, fp) for f in fp]) # 1 - similarity because we want distance return 1 - M def _check_multiple_bases(examples): return sum([e.is_origin for e in examples]) > 1 def _ecfp_names(examples, joint_bits): # add names for given descriptor indices multiple_bases = _check_multiple_bases(examples) # need to get base molecule(s) for naming bitInfo = {} # Type Dict[Any, Any] base_mol = [smi2mol(e.smiles) for e in examples if e.is_origin == True] if multiple_bases: multiBitInfo = {} # type: Dict[int, Tuple[Any, int, int]] for b in base_mol: bitInfo = {} AllChem.GetMorganFingerprint(b, 3, bitInfo=bitInfo) for bit in bitInfo: if bit not in multiBitInfo: multiBitInfo[bit] = (b, bit, {bit: bitInfo[bit]}) else: base_mol = smi2mol(examples[0].smiles) bitInfo = {} # type: Dict[Any, Any] AllChem.GetMorganFingerprint(base_mol, 3, bitInfo=bitInfo) result = [] # type: List[str] for i in range(len(joint_bits)): k = joint_bits[i] if multiple_bases: m = multiBitInfo[k][0] b = multiBitInfo[k][2] name = name_morgan_bit(m, b, k) else: name = name_morgan_bit(base_mol, bitInfo, k) result.append(name) return tuple(result) def _calculate_rdkit_descriptors(mol): from rdkit.ML.Descriptors import MoleculeDescriptors # type: ignore dlist = [ "NumHDonors", "NumHAcceptors", "MolLogP", "NumHeteroatoms", "RingCount", "NumRotatableBonds", ] # , 'NumHeteroatoms'] c = MoleculeDescriptors.MolecularDescriptorCalculator(dlist) d = c.CalcDescriptors(mol) def calc_aromatic_bonds(mol): return sum(1 for b in mol.GetBonds() if b.GetIsAromatic()) def _create_smarts(SMARTS): s = ",".join("$(" + s + ")" for s in SMARTS) _mol = MolFromSmarts("[" + s + "]") return _mol def calc_acid_groups(mol): acid_smarts = ( "[O;H1]-[C,S,P]=O", "[*;-;!$(*~[*;+])]", "[NH](S(=O)=O)C(F)(F)F", "n1nnnc1", ) pat = _create_smarts(acid_smarts) return len(mol.GetSubstructMatches(pat)) def calc_basic_groups(mol): basic_smarts = ( "[NH2]-[CX4]", "[NH](-[CX4])-[CX4]", "N(-[CX4])(-[CX4])-[CX4]", "[*;+;!$(*~[*;-])]", "N=C-N", "N-C=N", ) pat = _create_smarts(basic_smarts) return len(mol.GetSubstructMatches(pat)) def calc_apol(mol, includeImplicitHs=True): # atomic polarizabilities available here: # https://github.com/mordred-descriptor/mordred/blob/develop/mordred/data/polarizalibity78.txt from importlib_resources import files # type: ignore import exmol.lime_data # type: ignore ap = files(exmol.lime_data).joinpath("atom_pols.txt") with open(ap, "r") as f: atom_pols = [float(x) for x in next(f).split(",")] res = 0.0 for atom in mol.GetAtoms(): anum = atom.GetAtomicNum() if anum <= len(atom_pols): apol = atom_pols[anum] if includeImplicitHs: apol += atom_pols[1] * atom.GetTotalNumHs(includeNeighbors=False) res += apol else: raise ValueError(f"atomic number {anum} not found") return res d = d + ( calc_aromatic_bonds(mol), calc_acid_groups(mol), calc_basic_groups(mol), calc_apol(mol), ) return d def _get_joint_ecfp_descriptors(examples): """Create a union of ECFP bits from all base molecules""" # get reference bases = [smi2mol(e.smiles) for e in examples if e.is_origin] ecfp_joint = set() for m in bases: # Get bitinfo and create a union b = {} # type: Dict[Any, Any] temp_fp = AllChem.GetMorganFingerprint(m, 3, bitInfo=b) # add if radius greater than 0 ecfp_joint |= set([(k, v[0][1]) for k, v in b.items() if v[0][1] > 0]) # want to go in order of radius so when # we drop non-unique names, we keep smaller fragments ecfp_joint = list(ecfp_joint) ecfp_joint.sort(key=lambda x: x[1]) ecfp_joint = [x[0] for x in ecfp_joint] names = _ecfp_names(examples, ecfp_joint) # downselect to only unique names unique_names = set(names) output_ecfp = [] output_names = [] for b, n in zip(ecfp_joint, names): if n in unique_names and n is not None: unique_names.remove(n) output_ecfp.append(b) output_names.append(n) return output_ecfp, output_names _SMARTS = None def _load_smarts(path, rank_cutoff=500): # we have a rank cut for SMARTS that match too often smarts = {} with open(path) as f: for line in f.readlines(): if line[0] == "#": continue i1 = line.find(":") i2 = line.find(":", i1 + 1) m = MolFromSmarts(line[i2 + 1 :].strip()) rank = int(line[i1 + 1 : i2]) if rank > rank_cutoff: continue name = line[:i1] if m is None: print(f"Could not parse SMARTS: {line}") print(line[i2:].strip()) smarts[name] = (m, rank) return smarts def name_morgan_bit(m: Any, bitInfo: Dict[Any, Any], key: int) -> str: """Get the name of a Morgan bit using a SMARTS dictionary :param m: RDKit molecule :param bitInfo: bitInfo dictionary from rdkit.Chem.AllChem.GetMorganFingerprint :param key: bit key corresponding to the fingerprint you want to have named """ global _SMARTS if _SMARTS is None: from importlib_resources import files # type: ignore import exmol.lime_data # type: ignore sp = files(exmol.lime_data).joinpath("smarts.txt") _SMARTS = _load_smarts(sp) morgan_atoms = _bit2atoms(m, bitInfo, key) heteroatoms = set() for a in morgan_atoms: if m.GetAtomWithIdx(a).GetAtomicNum() > 6: heteroatoms.add(a) names = [] for name, (sm, r) in _SMARTS.items(): matches = m.GetSubstructMatches(sm) for match in matches: # check if match is in morgan bit match = set(match) if match.issubset(morgan_atoms): names.append((r, name, match)) names.sort(key=lambda x: x[0]) if len(names) == 0: return None umatch = names[0][2] name = names[0][1][0].lower() + names[0][1][1:].replace("_", " ") unique_names = set([names[0][1]]) for _, n, m in names: if len(m.intersection(umatch)) == 0: if n not in unique_names: name += "/" + n[0].lower() + n[1:].replace("_", " ") umatch |= m unique_names.add(n) if "/" in name and "fragment" not in name.split("/")[-1]: name = name + " group" # if we failed to match all heteroatoms, fail if len(heteroatoms.difference(umatch)) > 0: return None return name def clear_descriptors( examples: List[Example], ) -> List[Example]: """Clears all descriptors from examples :param examples: list of examples :param descriptor_type: type of descriptor to clear, if None, all descriptors are cleared """ for e in examples: e.descriptors = None # type: ignore return examples def add_descriptors( examples: List[Example], descriptor_type: str = "MACCS", mols: List[Any] = None, ) -> List[Example]: """Add descriptors to passed examples :param examples: List of example :param descriptor_type: Kind of descriptors to return, choose between 'Classic', 'ECFP', or 'MACCS'. Default is 'MACCS'. :param mols: Can be used if you already have rdkit Mols computed. :return: List of examples with added descriptors """ from importlib_resources import files import exmol.lime_data if mols is None: mols = [smi2mol(m.smiles) for m in examples] if descriptor_type.lower() == "classic": names = tuple( [ "number of hydrogen bond donor", "number of hydrogen bond acceptor", "Wildman-Crippen LogP", "number of heteroatoms", "ring count", "number of rotatable bonds", "aromatic bonds count", "acidic group count", "basic group count", "atomic polarizability", ] ) for e, m in zip(examples, mols): descriptors = _calculate_rdkit_descriptors(m) descriptor_names = names e.descriptors = Descriptors( descriptor_type=descriptor_type, descriptors=descriptors, descriptor_names=descriptor_names, plotting_names=descriptor_names, ) return examples elif descriptor_type.lower() == "maccs": mk = files(exmol.lime_data).joinpath("MACCSkeys.txt") with open(str(mk), "r") as f: names = tuple([x.strip().split("\t")[-1] for x in f.readlines()[1:]]) for e, m in zip(examples, mols): # rdkit sets fps[0] to 0 and starts keys at 1! fps = list(MACCSkeys.GenMACCSKeys(m).ToBitString()) descriptors = tuple(int(i) for i in fps) descriptor_names = names e.descriptors = Descriptors( descriptor_type=descriptor_type, descriptors=descriptors, descriptor_names=descriptor_names, plotting_names=descriptor_names, ) return examples elif descriptor_type.lower() == "ecfp": descriptor_bits, plotting_names = _get_joint_ecfp_descriptors(examples) for e, m in zip(examples, mols): bitInfo = {} # type: Dict[Any, Any] AllChem.GetMorganFingerprint(m, 3, bitInfo=bitInfo) descriptors = tuple( [1 if x in bitInfo.keys() else 0 for x in descriptor_bits] ) e.descriptors = Descriptors( descriptor_type=descriptor_type, descriptors=descriptors, descriptor_names=descriptor_bits, plotting_names=plotting_names, ) return examples else: raise ValueError( "Invalid descriptor string. Valid descriptor strings are 'Classic', 'ECFP', or 'MACCS'." ) def get_basic_alphabet() -> Set[str]: """Returns set of interpretable SELFIES tokens Generated by removing P and most ionization states from :func:`selfies.get_semantic_robust_alphabet` :return: Set of interpretable SELFIES tokens """ a = sf.get_semantic_robust_alphabet() # remove cations/anions except oxygen anion to_remove = [] for ai in a: if "+1" in ai: to_remove.append(ai) elif "-1" in ai: to_remove.append(ai) # remove [P],[#P],[=P] to_remove.extend(["[P]", "[#P]", "[=P]"]) a -= set(to_remove) a.add("[O-1]") return a def run_stoned( start_smiles: str, fp_type: str = "ECFP4", num_samples: int = 2000, max_mutations: int = 2, min_mutations: int = 1, alphabet: Union[List[str], Set[str]] = None, return_selfies: bool = False, _pbar: Any = None, ) -> Union[Tuple[List[str], List[float]], Tuple[List[str], List[str], List[float]]]: """Run ths STONED SELFIES algorithm. Typically not used, call :func:`sample_space` instead. :param start_smiles: SMILES string to start from :param fp_type: Fingerprint type :param num_samples: Number of total molecules to generate :param max_mutations: Maximum number of mutations :param min_mutations: Minimum number of mutations :param alphabet: Alphabet to use for mutations, typically from :func:`get_basic_alphabet()` :param return_selfies: If SELFIES should be returned as well :return: SELFIES, SMILES, and SCORES generated or SMILES and SCORES generated """ if alphabet is None: alphabet = get_basic_alphabet() if type(alphabet) == set: alphabet = list(alphabet) num_mutation_ls = list(range(min_mutations, max_mutations + 1)) start_mol = smi2mol(start_smiles) if start_mol == None: raise Exception("Invalid starting structure encountered") # want it so after sampling have num_samples randomized_smile_orderings = [ stoned.randomize_smiles(smi2mol(start_smiles)) for _ in range(num_samples // len(num_mutation_ls)) ] # Convert all the molecules to SELFIES selfies_ls = [sf.encoder(x) for x in randomized_smile_orderings] all_smiles_collect: List[str] = [] all_selfies_collect: List[str] = [] for num_mutations in num_mutation_ls: # Mutate the SELFIES: if _pbar: _pbar.set_description(f"🥌STONED🥌 Mutations: {num_mutations}") selfies_mut = stoned.get_mutated_SELFIES( selfies_ls.copy(), num_mutations=num_mutations, alphabet=alphabet ) # Convert back to SMILES: smiles_back = [sf.decoder(x) for x in selfies_mut] all_smiles_collect = all_smiles_collect + smiles_back all_selfies_collect = all_selfies_collect + selfies_mut if _pbar: _pbar.update(len(smiles_back)) if _pbar: _pbar.set_description(f"🥌STONED🥌 Filtering") # filter out duplicates all_mols = [smi2mol(s) for s in all_smiles_collect] all_canon = [mol2smi(m, canonical=True) if m else None for m in all_mols] seen = set() to_keep = [False for _ in all_canon] for i in range(len(all_canon)): if all_canon[i] and all_canon[i] not in seen: to_keep[i] = True seen.add(all_canon[i]) # now do filter filter_mols = [m for i, m in enumerate(all_mols) if to_keep[i]] filter_selfies = [s for i, s in enumerate(all_selfies_collect) if to_keep[i]] filter_smiles = [s for i, s in enumerate(all_smiles_collect) if to_keep[i]] # compute similarity scores base_fp = stoned.get_fingerprint(start_mol, fp_type=fp_type) fps = [stoned.get_fingerprint(m, fp_type) for m in filter_mols] scores = BulkTanimotoSimilarity(base_fp, fps) # type: List[float] if _pbar: _pbar.set_description(f"🥌STONED🥌 Done") if return_selfies: return filter_selfies, filter_smiles, scores else: return filter_smiles, scores @sleep_and_retry @limits(calls=2, period=30) def run_chemed( origin_smiles: str, num_samples: int, similarity: float = 0.1, fp_type: str = "ECFP4", _pbar: Any = None, ) -> Tuple[List[str], List[float]]: """ This method is similar to STONED but works by quering PubChem :param origin_smiles: Base SMILES :param num_samples: Minimum number of returned molecules. May return less due to network timeout or exhausting tree :param similarity: Tanimoto similarity to use in query (float between 0 to 1) :param fp_type: Fingerprint type :return: SMILES and SCORES """ if _pbar: _pbar.set_description("⚡CHEMED⚡") else: print("⚡CHEMED⚡") url = f"https://pubchem.ncbi.nlm.nih.gov/rest/pug/compound/fastsimilarity_2d/smiles/{requests.utils.quote(origin_smiles)}/property/CanonicalSMILES/JSON" try: reply = requests.get( url, params={"Threshold": int(similarity * 100), "MaxRecords": num_samples}, headers={"accept": "text/json"}, timeout=10, ) except requests.exceptions.Timeout: print("Pubchem seems to be down right now ️☠️☠️") return [], [] try: data = reply.json() except: return [], [] smiles = [d["CanonicalSMILES"] for d in data["PropertyTable"]["Properties"]] smiles = list(set(smiles)) if _pbar: _pbar.set_description(f"Received {len(smiles)} similar molecules") mol0 = smi2mol(origin_smiles) mols = [smi2mol(s) for s in smiles] fp0 = stoned.get_fingerprint(mol0, fp_type) scores = [] # drop Nones smiles = [s for s, m in zip(smiles, mols) if m is not None] for m in mols: if m is None: continue fp = stoned.get_fingerprint(m, fp_type) scores.append(TanimotoSimilarity(fp0, fp)) if _pbar: _pbar.update() return smiles, scores def run_custom( origin_smiles: str, data: List[Union[str, rdchem.Mol]], fp_type: str = "ECFP4", _pbar: Any = None, **kwargs, ) -> Tuple[List[str], List[float]]: """ This method is similar to STONED but uses a custom dataset provided by the user :param origin_smiles: Base SMILES :param data: List of SMILES or RDKit molecules :param fp_type: Fingerprint type :return: SMILES and SCORES """ if _pbar: _pbar.set_description("⚡CUSTOM⚡") else: print("⚡CUSTOM⚡") mol0 = smi2mol(origin_smiles) fp0 = stoned.get_fingerprint(mol0, fp_type) scores = [] smiles = [] # drop invalid molecules for d in data: if isinstance(d, str): m = smi2mol(d) else: m = d if m is None: continue smiles.append(mol2smi(m)) fp = stoned.get_fingerprint(m, fp_type) scores.append(TanimotoSimilarity(fp0, fp)) if _pbar: _pbar.update() return smiles, scores def sample_space( origin_smiles: str, f: Union[ Callable[[str, str], List[float]], Callable[[str], List[float]], Callable[[List[str], List[str]], List[float]], Callable[[List[str]], List[float]], ], batched: bool = True, preset: str = "medium", data: List[Union[str, rdchem.Mol]] = None, method_kwargs: Dict = None, num_samples: int = None, stoned_kwargs: Dict = None, quiet: bool = False, use_selfies: bool = False, sanitize_smiles: bool = True, ) -> List[Example]: """Sample chemical space around given SMILES This will evaluate the given function and run the :func:`run_stoned` function over chemical space around molecule. ``num_samples`` will be set to 3,000 by default if using STONED and 150 if using ``chemed``. :param origin_smiles: starting SMILES :param f: A function which takes in SMILES or SELFIES and returns predicted value. Assumed to work with lists of SMILES/SELFIES unless `batched = False` :param batched: If `f` is batched :param preset: Can be wide, medium, or narrow. Determines how far across chemical space is sampled. Try `"chemed"` preset to only sample commerically available compounds. :param data: If not None and preset is `"custom"` will use this data instead of generating new ones. :param method_kwargs: More control over STONED, CHEMED and CUSTOM can be set here. See :func:`run_stoned`, :func:`run_chemed` and :func:`run_custom` :param num_samples: Number of desired samples. Can be set in `method_kwargs` (overrides) or here. `None` means default for preset :param stoned_kwargs: Backwards compatible alias for `methods_kwargs` :param quiet: If True, will not print progress bar :param use_selfies: If True, will use SELFIES instead of SMILES for `f` :param sanitize_smiles: If True, will sanitize all SMILES :return: List of generated :obj:`Example` """ wrapped_f = f # if f only takes in 1 arg, wrap it in a function that takes in 2 # count args with no default value. Looks fancy because of possible objects/partials argcount = len( [ i for i in inspect.signature(f).parameters.values() if i.default == inspect.Parameter.empty ] ) if argcount == 1: if use_selfies: def wrapped_f(sm, sf): return f(sf) else: def wrapped_f(sm, sf): return f(sm) batched_f: Any = wrapped_f if not batched: def batched_f(sm, se): return np.array([wrapped_f(smi, sei) for smi, sei in zip(sm, se)]) if sanitize_smiles: origin_smiles = stoned.sanitize_smiles(origin_smiles)[1] if origin_smiles is None: raise ValueError("Given SMILES does not appear to be valid") smi_yhat = np.asarray(batched_f([origin_smiles], [sf.encoder(origin_smiles)])) try: iter(smi_yhat) except TypeError: raise ValueError("Your model function does not appear to be batched") smi_yhat = np.squeeze(smi_yhat[0]) if stoned_kwargs is not None: method_kwargs = stoned_kwargs if method_kwargs is None: method_kwargs = {} if preset == "medium": method_kwargs["num_samples"] = 3000 if num_samples is None else num_samples method_kwargs["max_mutations"] = 2 method_kwargs["alphabet"] = get_basic_alphabet() elif preset == "narrow": method_kwargs["num_samples"] = 3000 if num_samples is None else num_samples method_kwargs["max_mutations"] = 1 method_kwargs["alphabet"] = get_basic_alphabet() elif preset == "wide": method_kwargs["num_samples"] = 3000 if num_samples is None else num_samples method_kwargs["max_mutations"] = 5 method_kwargs["alphabet"] = sf.get_semantic_robust_alphabet() elif preset == "chemed": method_kwargs["num_samples"] = 150 if num_samples is None else num_samples elif preset == "custom" and data is not None: method_kwargs["num_samples"] = len(data) else: raise ValueError(f'Unknown preset "{preset}"') try: num_samples = method_kwargs["num_samples"] except KeyError as e: if num_samples is None: num_samples = 150 method_kwargs["num_samples"] = num_samples pbar = tqdm.tqdm(total=num_samples, disable=quiet) # STONED if preset.startswith("chem"): smiles, scores = run_chemed(origin_smiles, _pbar=pbar, **method_kwargs) selfies = [sf.encoder(s) for s in smiles] elif preset == "custom": smiles, scores = run_custom( origin_smiles, data=cast(Any, data), _pbar=pbar, **method_kwargs ) selfies = [sf.encoder(s) for s in smiles] else: result = run_stoned( origin_smiles, _pbar=pbar, return_selfies=True, **method_kwargs ) selfies, smiles, scores = cast(Tuple[List[str], List[str], List[float]], result) pbar.set_description("😀Calling your model function😀") if sanitize_smiles: smiles = [stoned.sanitize_smiles(s)[1] for s in smiles] fxn_values = batched_f(smiles, selfies) # pack them into data structure with filtering out identical # and nan exps = [ Example( origin_smiles, sf.encoder(origin_smiles), 1.0, cast(Any, smi_yhat), index=0, is_origin=True, ) ] + [ Example(sm, se, s, cast(Any, np.squeeze(y)), index=0) for i, (sm, se, s, y) in enumerate(zip(smiles, selfies, scores, fxn_values)) if s < 1.0 and np.isfinite(np.squeeze(y)) ] for i, e in enumerate(exps): # type: ignore e.index = i # type: ignore pbar.reset(len(exps)) pbar.set_description("🔭Projecting...🔭") # compute distance matrix full_dmat = _fp_dist_matrix( [e.smiles for e in exps], method_kwargs["fp_type"] if ("fp_type" in method_kwargs) else "ECFP4", _pbar=pbar, ) pbar.set_description("🥰Finishing up🥰") # compute PCA pca = PCA(n_components=2) proj_dmat = pca.fit_transform(full_dmat) for e in exps: # type: ignore e.position = proj_dmat[e.index, :] # type: ignore # do clustering everywhere (maybe do counter/same separately?) # clustering = AgglomerativeClustering( # n_clusters=max_k, affinity='precomputed', linkage='complete').fit(full_dmat) # Just do it on projected so it looks prettier. clustering = DBSCAN(eps=0.15, min_samples=5).fit(proj_dmat) for i, e in enumerate(exps): # type: ignore e.cluster = clustering.labels_[i] # type: ignore pbar.set_description("🤘Done🤘") pbar.close() return exps def _select_examples(cond, examples, nmols): result = [] # similarity filtered by if cluster/counter def cluster_score(e, i): return (e.cluster == i) * cond(e) * e.similarity clusters = set([e.cluster for e in examples]) for i in clusters: close_counter = max(examples, key=lambda e, i=i: cluster_score(e, i)) # check if actually is (since call could have been zero) if cluster_score(close_counter, i): result.append(close_counter) # trim, in case we had too many cluster result = sorted(result, key=lambda v: v.similarity * cond(v), reverse=True)[:nmols] # fill in remaining ncount = sum([cond(e) for e in result]) fill = max(0, nmols - ncount) result.extend( sorted(examples, key=lambda v: v.similarity * cond(v), reverse=True)[:fill] ) return list(filter(cond, result)) def lime_explain( examples: List[Example], descriptor_type: str = "MACCS", return_beta: bool = True, ): """From given :obj:`Examples<Example>`, find descriptor t-statistics (see :doc: `index`) :param examples: Output from :func: `sample_space` :param descriptor_type: Desired descriptors, choose from 'Classic', 'ECFP' 'MACCS' :return_beta: Whether or not the function should return regression coefficient values """ # add descriptors examples = add_descriptors(examples, descriptor_type) # weighted tanimoto similarities w = np.array([1 / (1 + (1 / (e.similarity + 0.000001) - 1) ** 5) for e in examples]) # Only keep nonzero weights non_zero = w > 10 ** (-6) nonzero_w = w[non_zero] # create a diagonal matrix of w N = nonzero_w.shape[0] diag_w = np.zeros((N, N)) np.fill_diagonal(diag_w, nonzero_w) # get feature matrix x_mat = np.array([list(e.descriptors.descriptors) for e in examples])[ non_zero ].reshape(N, -1) # remove zero variance columns y = ( np.array([e.yhat for e in examples]) .reshape(len(examples))[non_zero] .astype(float) ) # remove bias y -= np.mean(y) # compute least squares fit xtinv = np.linalg.pinv( (x_mat.T @ diag_w @ x_mat) + 0.001 * np.identity(len(examples[0].descriptors.descriptors)) ) beta = xtinv @ x_mat.T @ (y * nonzero_w) # compute standard error in beta yhat = x_mat @ beta resids = yhat - y SSR = np.sum(resids**2) se2_epsilon = SSR / (len(examples) - len(beta)) se2_beta = se2_epsilon * xtinv # now compute t-statistic for existence of coefficients tstat = beta * np.sqrt(1 / np.diag(se2_beta)) # Set tstats for bases, to be used later # TODO: Used to put them on examples[0] only, # but now copy them to all examples for e in examples: e.descriptors.tstats = tstat # Return beta (feature weights) which are the fits if asked for if return_beta: return beta else: return None def cf_explain(examples: List[Example], nmols: int = 3) -> List[Example]: """From given :obj:`Examples<Example>`, find closest counterfactuals (see :doc:`index`) :param examples: Output from :func:`sample_space` :param nmols: Desired number of molecules """ def is_counter(e): return e.yhat != examples[0].yhat result = _select_examples(is_counter, examples[1:], nmols) for i, r in enumerate(result): r.label = f"Counterfactual {i+1}" return examples[:1] + result def rcf_explain( examples: List[Example], delta: Union[Any, Tuple[float, float]] = (-1, 1), nmols: int = 4, ) -> List[Example]: """From given :obj:`Examples<Example>`, find closest counterfactuals (see :doc:`index`) This version works with regression, so that a counterfactual is if the given example is higher or lower than base. :param examples: Output from :func:`sample_space` :param delta: float or tuple of hi/lo indicating margin for what is counterfactual :param nmols: Desired number of molecules """ if type(delta) is float: delta = (-delta, delta) def is_high(e): return e.yhat + delta[0] >= examples[0].yhat def is_low(e): return e.yhat + delta[1] <= examples[0].yhat hresult = ( [] if delta[0] is None else _select_examples(is_high, examples[1:], nmols // 2) ) for i, h in enumerate(hresult): h.label = f"Increase ({i+1})" lresult = ( [] if delta[1] is None else _select_examples(is_low, examples[1:], nmols // 2) ) for i, l in enumerate(lresult): l.label = f"Decrease ({i+1})" return examples[:1] + lresult + hresult def plot_space( examples: List[Example], exps: List[Example], figure_kwargs: Dict = None, mol_size: Tuple[int, int] = (200, 200), highlight_clusters: bool = False, mol_fontsize: int = 8, offset: int = 0, ax: Any = None, cartoon: bool = False, rasterized: bool = False, ): """Plot chemical space around example and annotate given examples. :param examples: Large list of :obj:Example which make-up points :param exps: Small list of :obj:Example which will be annotated :param figure_kwargs: kwargs to pass to :func:`plt.figure<matplotlib.pyplot.figure>` :param mol_size: size of rdkit molecule rendering, in pixles :param highlight_clusters: if `True`, cluster indices are rendered instead of :obj:Example.yhat :param mol_fontsize: minimum font size passed to rdkit :param offset: offset annotations to allow colorbar or other elements to fit into plot. :param ax: axis onto which to plot :param cartoon: do cartoon outline on points? :param rasterized: raster the scatter? """ imgs = _mol_images(exps, mol_size, mol_fontsize) # , True) if figure_kwargs is None: figure_kwargs = {"figsize": (12, 8)} base_color = "gray" if ax is None: ax = plt.figure(**figure_kwargs).gca() if highlight_clusters: colors = [e.cluster for e in examples] def normalizer(x): return x cmap = "Accent" else: colors = cast(Any, [e.yhat for e in examples]) normalizer = plt.Normalize(min(colors), max(colors)) cmap = "viridis" space_x = [e.position[0] for e in examples] space_y = [e.position[1] for e in examples] if cartoon: # plot shading, lines, front ax.scatter(space_x, space_y, 50, "0.0", lw=2, rasterized=rasterized) ax.scatter(space_x, space_y, 50, "1.0", lw=0, rasterized=rasterized) ax.scatter( space_x, space_y, 40, c=normalizer(colors), cmap=cmap, lw=2, alpha=0.1, rasterized=rasterized, ) else: ax.scatter( space_x, space_y, c=normalizer(colors), cmap=cmap, alpha=0.5, edgecolors="none", rasterized=rasterized, ) # now plot cfs/annotated points ax.scatter( [e.position[0] for e in exps], [e.position[1] for e in exps], c=normalizer([e.cluster if highlight_clusters else e.yhat for e in exps]), cmap=cmap, edgecolors="black", ) x = [e.position[0] for e in exps] y = [e.position[1] for e in exps] titles = [] colors = [] for e in exps: if not e.is_origin: titles.append(f"Similarity = {e.similarity:.2f}\n{e.label}") colors.append(cast(Any, base_color)) else: titles.append("Base") colors.append(cast(Any, base_color)) _image_scatter(x, y, imgs, titles, colors, ax, offset=offset) ax.axis("off") ax.set_aspect("auto") def plot_cf( exps: List[Example], fig: Any = None, figure_kwargs: Dict = None, mol_size: Tuple[int, int] = (200, 200), mol_fontsize: int = 10, nrows: int = None, ncols: int = None, ): """Draw the given set of Examples in a grid :param exps: Small list of :obj:`Example` which will be drawn :param fig: Figure to plot onto :param figure_kwargs: kwargs to pass to :func:`plt.figure<matplotlib.pyplot.figure>` :param mol_size: size of rdkit molecule rendering, in pixles :param mol_fontsize: minimum font size passed to rdkit :param nrows: number of rows to draw in grid :param ncols: number of columns to draw in grid """ imgs = _mol_images(exps, mol_size, mol_fontsize) if nrows is not None: R = nrows else: R = math.ceil(math.sqrt(len(imgs))) if ncols is not None: C = ncols else: C = math.ceil(len(imgs) / R) if fig is None: if figure_kwargs is None: figure_kwargs = {"figsize": (12, 8)} fig, axs = plt.subplots(R, C, **figure_kwargs) else: axs = fig.subplots(R, C) if type(axs) != np.ndarray: # Happens if nrows=ncols=1 axs = np.array([[axs]]) axs = axs.flatten() for i, (img, e) in enumerate(zip(imgs, exps)): title = "Base" if e.is_origin else f"Similarity = {e.similarity:.2f}\n{e.label}" title += f"\nf(x) = {e.yhat:.3f}" axs[i].set_title(title) axs[i].imshow(np.asarray(img), gid=f"rdkit-img-{i}") axs[i].axis("off") for j in range(i, C * R): axs[j].axis("off") axs[j].set_facecolor("white") plt.tight_layout() def plot_descriptors( examples: List[Example], output_file: str = None, fig: Any = None, figure_kwargs: Dict = None, title: str = None, return_svg: bool = False, ): """Plot descriptor attributions from given set of Examples. :param examples: Output from :func:`sample_space` :param output_file: Output file name to save the plot - optional except for ECFP :param fig: Figure to plot on to :param figure_kwargs: kwargs to pass to :func:`plt.figure<matplotlib.pyplot.figure>` :param title: Title for the plot :param return_svg: Whether to return svg for plot """ from importlib_resources import files import exmol.lime_data import pickle # type: ignore # infer descriptor_type from examples descriptor_type = examples[0].descriptors.descriptor_type.lower() multiple_bases = _check_multiple_bases(examples) if output_file is None and descriptor_type == "ecfp" and not return_svg: raise ValueError("No filename provided to save the plot") space_tstats = list(examples[0].descriptors.tstats) if fig is None: if figure_kwargs is None: figure_kwargs = ( {"figsize": (5, 5)} if descriptor_type.lower() == "classic" else {"figsize": (8, 5)} ) fig, ax = plt.subplots(nrows=1, ncols=1, dpi=180, **figure_kwargs) # find important descriptors d_importance = { a: [b, i, n] for i, (a, b, n) in enumerate( zip( examples[0].descriptors.descriptor_names, space_tstats, examples[0].descriptors.plotting_names, ) ) if not np.isnan(b) } d_importance = dict( sorted(d_importance.items(), key=lambda item: abs(item[1][0]), reverse=True) ) t = [a[0] for a in list(d_importance.values())][:5] key_ids = [a[1] for a in list(d_importance.values())][:5] keys = [a for a in list(d_importance.keys())] names = [a[2] for a in list(d_importance.values())][:5] # set colors colors = [] for ti in t: if ti < 0: colors.append("#F06060") if ti > 0: colors.append("#1BBC9B") # plot the bars bar1 = ax.barh(range(len(t)), t, color=colors, height=0.75) new_patches = [] for patch in reversed(ax.patches): bb = patch.get_bbox() color = patch.get_facecolor() p_bbox = FancyBboxPatch( (bb.xmin, bb.ymin), abs(bb.width), abs(bb.height), boxstyle="round,pad=-0.040,rounding_size=0.015", ec="none", fc=color, mutation_aspect=4, ) patch.remove() new_patches.append(p_bbox) for patch in new_patches: ax.add_patch(patch) count = 0 sk_dict, key_imgs = {}, {} if descriptor_type == "maccs": # Load svg/png images mk = files(exmol.lime_data).joinpath("keys.pb") with open(str(mk), "rb") as f: key_imgs = pickle.load(f) if descriptor_type == "ecfp": # get reference for ECFP if multiple_bases: bases = [smi2mol(e.smiles) for e in examples if e.is_origin == True] bi = {} # type: Dict[Any, Any] for b in bases: bit_info = {} # type: Dict[Any, Any] fp = AllChem.GetMorganFingerprint(b, 3, bitInfo=bit_info) for bit in bit_info: if bit not in bi: bi[bit] = (b, bit, bit_info) else: bi = {} m = smi2mol(examples[0].smiles) fp = AllChem.GetMorganFingerprint(m, 3, bitInfo=bi) for rect, ti, k, ki, n in zip(bar1, t, keys, key_ids, names): # account for Nones if n is None: n = "" # annotate patches with text desciption y = rect.get_y() + rect.get_height() / 2.0 n = textwrap.fill(str(n), 20) if ti < 0: x = 0.25 skx = ( np.max(np.absolute(t)) + 2 if descriptor_type == "maccs" else np.max(np.absolute(t)) ) box_x = 0.98 ax.text( x, y, n, ha="left", va="center", wrap=True, fontsize=12, ) else: x = -0.25 skx = ( -np.max(np.absolute(t)) - 2 if descriptor_type == "maccs" else np.max(np.absolute(t)) ) box_x = 0.02 ax.text( x, y, n, ha="right", va="center", wrap=True, fontsize=12, ) # add SMARTS annotation where applicable if descriptor_type == "maccs" or descriptor_type == "ecfp": if descriptor_type == "maccs": key_img = plt.imread(io.BytesIO(key_imgs[ki]["png"])) box = skunk.ImageBox(f"sk{count}", key_img, zoom=1) else: box = skunk.Box(130, 50, f"sk{count}") ab = AnnotationBbox( box, xy=(skx, count), xybox=(box_x, (5 - count) * 0.2 - 0.1), # Invert axis xycoords="data", boxcoords="axes fraction", bboxprops=dict(lw=0.5), ) ax.add_artist(ab) if descriptor_type == "maccs": sk_dict[f"sk{count}"] = key_imgs[ki]["svg"] if descriptor_type == "ecfp": if multiple_bases: m = bi[int(k)][0] b = bi[int(k)][2] else: b = bi svg = DrawMorganBit( m, int(k), b, molSize=(300, 200), centerColor=None, aromaticColor=None, ringColor=None, extraColor=(0.8, 0.8, 0.8), useSVG=True, ) # TODO: Why? try: svgdata = svg.data except AttributeError: svgdata = svg sk_dict[f"sk{count}"] = svgdata count += 1 ax.axvline(x=0, color="grey", linewidth=0.5) # calculate significant T w = np.array([1 / (1 + (1 / (e.similarity + 0.000001) - 1) ** 5) for e in examples]) effective_n = np.sum(w) ** 2 / np.sum(w**2) T = ss.t.ppf(0.975, df=effective_n) # plot T ax.axvline(x=T, color="#f5ad4c", linewidth=0.75, linestyle="--", zorder=0) ax.axvline(x=-T, color="#f5ad4c", linewidth=0.75, linestyle="--", zorder=0) # set axis ax.set_yticks([]) ax.invert_yaxis() ax.set_xlabel("Descriptor t-statistics", fontsize=12) if title is None: ax.set_title(f"{descriptor_type} descriptors", fontsize=12) else: ax.set_title(f"{title}", fontsize=12) # inset SMARTS svg images for MACCS descriptors if descriptor_type == "maccs" or descriptor_type == "ecfp": if descriptor_type == "maccs": print( "SMARTS annotations for MACCS descriptors were created using SMARTSviewer (smartsview.zbh.uni-hamburg.de, Copyright: ZBH, Center for Bioinformatics Hamburg) developed by K. Schomburg et. al. (J. Chem. Inf. Model. 2010, 50, 9, 1529–1535)" ) xlim = np.max(np.absolute(t)) + 6 ax.set_xlim(-xlim, xlim) svg = skunk.insert(sk_dict) if output_file is not None: plt.tight_layout() with open(output_file, "w") as f: # type: ignore f.write(svg) if return_svg: plt.close() return svg elif descriptor_type == "classic": xlim = max(np.max(np.absolute(t)), T + 1) ax.set_xlim(-xlim, xlim) if output_file is not None: plt.tight_layout() plt.savefig(output_file, dpi=180, bbox_inches="tight") def check_multiple_aromatic_rings(mol): ri = mol.GetRingInfo() count = 0 for bondRing in ri.BondRings(): flag = True for id in bondRing: if not mol.GetBondWithIdx(id).GetIsAromatic(): flag = False continue if flag: count += 1 return True if count > 1 else False def merge_text_explains( *args: List[Tuple[str, float]], filter: Optional[float] = None ) -> List[Tuple[str, float]]: """Merge multiple text explanations into one and sort.""" # sort them by T value, putting negative examples at the end joint = reduce(lambda x, y: x + y, args) if len(joint) == 0: return [] # get the highest (hopefully) positive m = max([x[1] for x in joint if x[1] > 0]) pos = [x for x in joint if x[1] == m] joint = [x for x in joint if x[1] != m] joint = sorted(joint, key=lambda x: np.absolute(x[1]), reverse=True) return pos + joint _text_prompt = """ The following are a series of questions about molecules that connect their structure to a property, along with how important each question is for the molecular property. An answer of "Yes" means that the question was true and that attribute of structure contributed to the molecular property. An answer of "Counterfactual" means the lack of that attribute contributed to the molecular property. A summary paragraph is given below, which only summarizes on the most important structure-property relationships. Property: [PROPERTY] [TEXT] Summary: The molecular property "[PROPERTY]" can be explained""" def text_prompt( text_explanations: List[Tuple[str, float]], property_name: str, open_ai_key: Optional[str] = None, ) -> str: """Insert text explanations into template, and optionally send to OpenAI.""" result = _text_prompt.replace("[PROPERTY]", property_name) # want to have negative examples at the end text_explanations.sort(key=lambda x: x[1], reverse=True) result = result.replace("[TEXT]", "".join([f"{t[0]}" for t in text_explanations])) if open_ai_key is not None: import openai openai.api_key = open_ai_key response = openai.Completion.create( model="text-davinci-003", prompt=result, temperature=0.7, max_tokens=256, top_p=1, frequency_penalty=0, presence_penalty=0, ) completion = response["choices"][0]["text"] return ( 'The molecular property "' + property_name + '" can be explained' + completion ) return result def text_explain( examples: List[Example], descriptor_type: str = "maccs", count: int = 5, presence_thresh: float = 0.2, include_weak: Optional[bool] = None, ) -> List[Tuple[str, float]]: """Take an example and convert t-statistics into text explanations :param examples: Output from :func:`sample_space` :param descriptor_type: Type of descriptor, either "maccs", or "ecfp". :param count: Number of text explanations to return :param presence_thresh: Threshold for presence of descriptor in examples :param include_weak: Include weak descriptors. If not set, the function will be first have this set to False, and if no descriptors are found, will be set to True and function will be re-run """ descriptor_type = descriptor_type.lower() # populate lime explanation if examples[-1].descriptors is None: lime_explain(examples, descriptor_type=descriptor_type) nbases = sum([1 for e in examples if e.is_origin]) # Take t-statistics, rank them d_importance = [ (n, t, i) # name, t-stat, index for i, (n, t) in enumerate( zip( examples[0].descriptors.plotting_names, examples[0].descriptors.tstats, ) ) # don't want NANs and want match (if not multiple bases) if not np.isnan(t) ] d_importance = sorted(d_importance, key=lambda x: abs(x[1]), reverse=True) # get significance value - if >significance, then important else weakly important? w = np.array([1 / (1 + (1 / (e.similarity + 0.000001) - 1) ** 5) for e in examples]) effective_n = np.sum(w) ** 2 / np.sum(w**2) if np.isnan(effective_n): effective_n = len(examples) T = ss.t.ppf(0.975, df=effective_n) pos_count = 0 neg_count = 0 result = [] existing_names = set() for k, v, i in d_importance: if pos_count + neg_count == count: break name = k if name is None or name in existing_names: continue existing_names.add(name) if abs(v) > 4: imp = "This is very important for the property\n" elif abs(v) >= T: imp = "This is important for the property\n" elif include_weak: imp = "This could be relevent for the property\n" else: continue # check if it's present in majority of base molecules present = sum( [1 for e in examples if e.descriptors.descriptors[i] != 0 and e.is_origin] ) if present / nbases < (1 - presence_thresh) and v < 0: if neg_count == count - 2: # don't want to have only negative examples continue kind = "No (Counterfactual)." neg_count += 1 elif present / nbases > presence_thresh and v > 0: kind = "Yes." pos_count += 1 else: continue # adjust name to be question if name[-1] != "?": name = "Is there " + name + "?" s = f"{name} {kind} {imp}" result.append((s, v)) if len(result) == 0 or pos_count == 0 and include_weak is None: return text_explain( examples, descriptor_type=descriptor_type, count=count, presence_thresh=presence_thresh, include_weak=True, ) return result
[ "\nThe following are a series of questions about molecules that connect their structure to a property, along with how important each question is for the molecular property. An answer of \"Yes\" means that the question was true and that attribute of structure contributed to the molecular property. An answer of \"Counterfactual\" means the lack of that attribute contributed to the molecular property. A summary paragraph is given below, which only summarizes on the most important structure-property relationships.\n\nProperty: [PROPERTY]\n[TEXT]\nSummary: The molecular property \"[PROPERTY]\" can be explained" ]
2024-01-10
daily-demos/llm-translator
server~services~open_ai_service.py
from services.ai_service import AIService import requests import io import openai import os class OpenAIService(AIService): def __init__(self): # we handle all the api config directly in the calls below pass def run_llm(self, messages, stream=True): model = os.getenv("OPEN_AI_MODEL") if not model: model = "gpt-4" response = openai.ChatCompletion.create( api_type='openai', api_version='2020-11-07', api_base="https://api.openai.com/v1", api_key=os.getenv("OPEN_AI_KEY"), model=model, stream=stream, messages=messages ) return response
[]
2024-01-10
duckietown/gym-gazebo
gym_gazebo~envs~gazebo_env.py
import gym import rospy #import roslaunch import os import signal import subprocess from os import path from std_srvs.srv import Empty class GazeboEnv(gym.Env): """Superclass for all Gazebo environments. """ metadata = {'render.modes': ['human']} def __init__(self, launchfile): #start roscore subprocess.Popen("roscore") print ("Roscore launched!") # Launch the simulation with the given launchfile name rospy.init_node('gym', anonymous=True) if launchfile.startswith("/"): fullpath = launchfile else: fullpath = os.path.join(os.path.dirname(__file__), "assets","launch", launchfile) if not path.exists(fullpath): raise IOError("File "+fullpath+" does not exist") subprocess.Popen(["roslaunch",fullpath]) print ("Gazebo launched!") self.gzclient_pid = 0 def _step(self, action): # Implement this method in every subclass # Perform a step in gazebo. E.g. move the robot raise NotImplementedError def _reset(self): # Implemented in subclass raise NotImplementedError def _render(self, mode="human", close=False): if close: tmp = os.popen("ps -Af").read() proccount = tmp.count('gzclient') if proccount > 0: if self.gzclient_pid != 0: os.kill(self.gzclient_pid, signal.SIGTERM) os.wait() return tmp = os.popen("ps -Af").read() proccount = tmp.count('gzclient') if proccount < 1: subprocess.Popen("gzclient") self.gzclient_pid = int(subprocess.check_output(["pidof","-s","gzclient"])) else: self.gzclient_pid = 0 def _close(self): # Kill gzclient, gzserver and roscore tmp = os.popen("ps -Af").read() gzclient_count = tmp.count('gzclient') gzserver_count = tmp.count('gzserver') roscore_count = tmp.count('roscore') rosmaster_count = tmp.count('rosmaster') if gzclient_count > 0: os.system("killall -9 gzclient") if gzserver_count > 0: os.system("killall -9 gzserver") if rosmaster_count > 0: os.system("killall -9 rosmaster") if roscore_count > 0: os.system("killall -9 roscore") if (gzclient_count or gzserver_count or roscore_count or rosmaster_count >0): os.wait() def _configure(self): # TODO # From OpenAI API: Provides runtime configuration to the enviroment # Maybe set the Real Time Factor? pass def _seed(self): # TODO # From OpenAI API: Sets the seed for this env's random number generator(s) pass
[]
2024-01-10
Sway2003/test
openagent~finetune~LLMFinetune.py
from abc import ABC, abstractmethod from logging import Logger import openai class LLMFinetune(ABC): def __init__(self, logger: Logger, openai_key: str): self.logger = logger openai.api_key = openai_key @abstractmethod def transform_data(self, train_csv_file: str, val_csv_file: str , train_output_file: str, val_output_file: str) -> str: pass @abstractmethod def finetune(self, **kwargs): pass
[]
2024-01-10
aws-banjo/intro_bedrock_workshop
chat_module~chat_bedrock_st.py
import streamlit as st from langchain.chains import ConversationChain from langchain.memory import ConversationBufferMemory from langchain.llms.bedrock import Bedrock import time st.title("ChatBedrock")
[]
2024-01-10
aws-banjo/intro_bedrock_workshop
agent_module~bedrock_langchain_agent.py
import boto3 from langchain.agents import AgentType, initialize_agent, tool from langchain.llms import Bedrock from langchain.tools import StructuredTool from langchain.embeddings import BedrockEmbeddings from langchain.vectorstores import FAISS from typing import List, Dict, Any import boto3 # Setup bedrock bedrock_runtime = boto3.client( service_name="bedrock-runtime", region_name="us-east-1", ) llm = Bedrock(client=bedrock_runtime, model_id="anthropic.claude-v2") # Demo tools def get_word_length(word: str) -> int: """Returns the length of a word.""" return len(word) def adder(a: float, b: float) -> float: """Multiply the provided floats.""" return a + b def multiplier(a: float, b: float) -> float: """Multiply the provided floats.""" # TODO: Implement this function return None def get_current_time() -> str: """Returns the current time.""" # TODO: Implement this function return None # AWS Well-Architected Framework tool def well_arch_tool(query: str) -> Dict[str, Any]: """Returns text from AWS Well-Architected Framework releated to the query""" embeddings = BedrockEmbeddings( client=bedrock_runtime, model_id="amazon.titan-embed-text-v1", ) vectorstore = FAISS.load_local("local_index", embeddings) docs = vectorstore.similarity_search(query) resp_json = {"docs": docs} return resp_json adder_tool = StructuredTool.from_function(adder) time_tool = StructuredTool.from_function(get_current_time) multiplier_tool = StructuredTool.from_function(multiplier) word_tool = StructuredTool.from_function(get_word_length) aws_well_arch_tool = StructuredTool.from_function(well_arch_tool) # Start Agent agent_executor = initialize_agent( [multiplier_tool, word_tool, aws_well_arch_tool, adder_tool, time_tool], llm, agent=AgentType.STRUCTURED_CHAT_ZERO_SHOT_REACT_DESCRIPTION, verbose=True, ) # Run some queries resp = agent_executor.run("How many letters is in 'educa'?") print(resp) resp = agent_executor.run("What is 3 plus 4?") print(resp) resp = agent_executor.run( "What does the AWS Well-Architected Framework say about how to create secure VPCs?" ) print(resp) resp = agent_executor.run("What is the current time?") print(resp) resp = agent_executor.run("What is 3 times 4?") print(resp)
[]
2024-01-10
aws-banjo/intro_bedrock_workshop
full_code~agent_full_code.py
import datetime from typing import Any, Dict, List import boto3 from langchain.agents import AgentType, initialize_agent, tool from langchain.embeddings import BedrockEmbeddings from langchain.llms import Bedrock from langchain.tools import StructuredTool from langchain.vectorstores import FAISS # Setup bedrock bedrock_runtime = boto3.client( service_name="bedrock-runtime", region_name="us-east-1", ) llm = Bedrock(client=bedrock_runtime, model_id="anthropic.claude-v2") # Demo tools def get_word_length(word: str) -> int: """Returns the length of a word.""" return len(word) def adder(a: float, b: float) -> float: """Multiply the provided floats.""" return a + b def multiplier(a: float, b: float) -> float: """Multiply the provided floats.""" return a * b def get_current_time() -> str: """Returns the current time""" return datetime.datetime.now().strftime("%H:%M:%S") # AWS Well-Architected Framework tool def well_arch_tool(query: str) -> Dict[str, Any]: """Returns text from AWS Well-Architected Framework releated to the query""" embeddings = BedrockEmbeddings( client=bedrock_runtime, model_id="amazon.titan-embed-text-v1", ) vectorstore = FAISS.load_local("local_index", embeddings) docs = vectorstore.similarity_search(query) resp_json = {"docs": docs} return resp_json adder_tool = StructuredTool.from_function(adder) time_tool = StructuredTool.from_function(get_current_time) multiplier_tool = StructuredTool.from_function(multiplier) word_tool = StructuredTool.from_function(get_word_length) aws_well_arch_tool = StructuredTool.from_function(well_arch_tool) # Start Agent agent_executor = initialize_agent( [multiplier_tool, word_tool, aws_well_arch_tool, adder_tool, time_tool], llm, agent=AgentType.STRUCTURED_CHAT_ZERO_SHOT_REACT_DESCRIPTION, verbose=True, ) # Run some queries resp = agent_executor.run("How many letters is in 'educa'?") print(resp) resp = agent_executor.run("What is 3 plus 4?") print(resp) resp = agent_executor.run( "What does the AWS Well-Architected Framework say about how to create secure VPCs?" ) print(resp) resp = agent_executor.run("What is the current time?") print(resp) resp = agent_executor.run("What is 3 times 4?") print(resp)
[]
2024-01-10
aws-banjo/intro_bedrock_workshop
full_code~chat_full_bedrock_st.py
import time import boto3 import streamlit as st from langchain.chains import ConversationChain from langchain.llms.bedrock import Bedrock from langchain.memory import ConversationBufferMemory st.title("ChatBedrock") # Setup bedrock bedrock_runtime = boto3.client( service_name="bedrock-runtime", region_name="us-east-1", ) @st.cache_resource def load_llm(): llm = Bedrock(client=bedrock_runtime, model_id="anthropic.claude-v2") llm.model_kwargs = {"temperature": 0.7, "max_tokens_to_sample": 2048} model = ConversationChain(llm=llm, verbose=True, memory=ConversationBufferMemory()) return model model = load_llm() if "messages" not in st.session_state: st.session_state.messages = [] for message in st.session_state.messages: with st.chat_message(message["role"]): st.markdown(message["content"]) if prompt := st.chat_input("What is up?"): st.session_state.messages.append({"role": "user", "content": prompt}) with st.chat_message("user"): st.markdown(prompt) with st.chat_message("assistant"): message_placeholder = st.empty() full_response = "" # prompt = prompt_fixer(prompt) result = model.predict(input=prompt) # Simulate stream of response with milliseconds delay for chunk in result.split(): full_response += chunk + " " time.sleep(0.05) # Add a blinking cursor to simulate typing message_placeholder.markdown(full_response + "▌") message_placeholder.markdown(full_response) st.session_state.messages.append({"role": "assistant", "content": full_response})
[]
2024-01-10
eltociear/TaskWeaver
project~plugins~sql_pull_data.py
from operator import itemgetter import pandas as pd from langchain.chat_models import AzureChatOpenAI, ChatOpenAI from langchain.prompts import ChatPromptTemplate from langchain.schema.output_parser import StrOutputParser from langchain.schema.runnable import RunnableLambda, RunnableMap from langchain.utilities import SQLDatabase from taskweaver.plugin import Plugin, register_plugin @register_plugin class SqlPullData(Plugin): def __call__(self, query: str): api_type = self.config.get("api_type", "azure") if api_type == "azure": model = AzureChatOpenAI( azure_endpoint=self.config.get("api_base"), openai_api_key=self.config.get("api_key"), openai_api_version=self.config.get("api_version"), azure_deployment=self.config.get("deployment_name"), temperature=0, verbose=True, ) elif api_type == "openai": model = ChatOpenAI( openai_api_key=self.config.get("api_key"), model_name=self.config.get("deployment_name"), temperature=0, verbose=True, ) else: raise ValueError("Invalid API type. Please check your config file.") template = """Based on the table schema below, write a SQL query that would answer the user's question: {schema} Question: {question} SQL Query:""" prompt = ChatPromptTemplate.from_template(template) db = SQLDatabase.from_uri(self.config.get("sqlite_db_path")) def get_schema(_): return db.get_table_info() inputs = { "schema": RunnableLambda(get_schema), "question": itemgetter("question"), } sql_response = RunnableMap(inputs) | prompt | model.bind(stop=["\nSQLResult:"]) | StrOutputParser() sql = sql_response.invoke({"question": query}) result = db._execute(sql, fetch="all") df = pd.DataFrame(result) if len(df) == 0: return df, ( f"I have generated a SQL query based on `{query}`.\nThe SQL query is {sql}.\n" f"The result is empty." ) else: return df, ( f"I have generated a SQL query based on `{query}`.\nThe SQL query is {sql}.\n" f"There are {len(df)} rows in the result.\n" f"The first {min(5, len(df))} rows are:\n{df.head(min(5, len(df))).to_markdown()}" )
[ "Based on the table schema below, write a SQL query that would answer the user's question:\n {schema}\n\n Question: {question}\n SQL Query:" ]