code
stringlengths
141
79.4k
apis
sequencelengths
1
23
extract_api
stringlengths
126
73.2k
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session from langchain.utils import get_from_env try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento RETURN_VAL_TYPE = Sequence[Generation] def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) class BaseCache(ABC): """Base interface for cache.""" @abstractmethod def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" @abstractmethod def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" @abstractmethod def clear(self, **kwargs: Any) -> None: """Clear cache that can take additional keyword arguments.""" class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" # TODO - implement a TTL policy in Redis def __init__(self, redis_: Any): """Initialize by passing in Redis instance.""" try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) self.redis.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, ) except ValueError: redis = RedisVectorstore( embedding_function=self.embedding.embed_query, index_name=index_name, redis_url=self.redis_url, ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations = [] # Read from a Hash results = llm_cache.similarity_search_limit_score( query=prompt, k=1, score_threshold=self.score_threshold, ) if results: for document in results: for text in document.metadata["return_val"]: generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) # Write to vectorstore metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": [generation.text for generation in return_val], } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((1483, 1510), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1500, 1510), False, 'import logging\n'), ((3955, 3973), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3971, 3973), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((4130, 4162), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4136, 4162), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4173, 4205), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4179, 4205), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4216, 4249), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (4222, 4249), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4265, 4279), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (4271, 4279), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2453, 2481), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2463, 2481), False, 'import json\n'), ((6683, 6726), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (6696, 6726), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((15254, 15261), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (15259, 15261), False, 'from gptcache import Cache\n'), ((16587, 16619), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (16590, 16619), False, 'from gptcache.adapter.api import get\n'), ((17506, 17552), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (17509, 17552), False, 'from gptcache.adapter.api import put\n'), ((21401, 21443), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (21431, 21443), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21467, 21523), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (21478, 21523), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((2498, 2527), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2508, 2527), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5029, 5049), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (5036, 5049), False, 'from sqlalchemy.orm import Session\n'), ((6136, 6156), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6143, 6156), False, 'from sqlalchemy.orm import Session\n'), ((6338, 6358), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6345, 6358), False, 'from sqlalchemy.orm import Session\n'), ((10691, 10807), 'langchain.vectorstores.redis.Redis.from_existing_index', 'RedisVectorstore.from_existing_index', ([], {'embedding': 'self.embedding', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding=self.embedding, index_name=\n index_name, redis_url=self.redis_url)\n', (10727, 10807), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((15328, 15370), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (15345, 15370), False, 'import inspect\n'), ((17775, 17805), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (17779, 17805), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast\n'), ((18705, 18725), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (18714, 18725), False, 'from datetime import timedelta\n'), ((21268, 21294), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (21292, 21294), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21330, 21378), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (21342, 21378), False, 'from langchain.utils import get_from_env\n'), ((8403, 8522), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (8416, 8522), False, 'import warnings\n'), ((10913, 11030), 'langchain.vectorstores.redis.Redis', 'RedisVectorstore', ([], {'embedding_function': 'self.embedding.embed_query', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding_function=self.embedding.embed_query, index_name=\n index_name, redis_url=self.redis_url)\n', (10929, 11030), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((12815, 12943), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (12828, 12943), False, 'import warnings\n'), ((16673, 16702), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (16683, 16702), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6046, 6056), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (6051, 6056), False, 'from langchain.load.dump import dumps\n'), ((7862, 7883), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (7872, 7883), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((15714, 15752), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (15730, 15752), False, 'from gptcache.manager.factory import get_data_manager\n'), ((16726, 16741), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (16736, 16741), False, 'import json\n'), ((5184, 5197), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (5189, 5197), False, 'from langchain.load.load import loads\n'), ((12266, 12287), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (12276, 12287), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5754, 5777), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (5764, 5777), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((4799, 4833), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (4805, 4833), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
import os import json import base64 import requests import langchain from flask import jsonify from langchain.chat_models import ChatOpenAI from typing import Sequence from threading import Thread from queue import Queue, Empty from langchain.callbacks.base import BaseCallbackHandler from typing import Any, Callable from langchain.vectorstores.pinecone import Pinecone from dotenv import load_dotenv from langchain.chains import LLMChain from langchain.prompts import PromptTemplate, ChatPromptTemplate import pinecone from langchain.embeddings.openai import OpenAIEmbeddings from langchain.chains import RetrievalQA from openai import OpenAI from langchain.chains.openai_functions import ( create_structured_output_runnable, ) from langchain.pydantic_v1 import BaseModel, Field load_dotenv() PINECONE_API_KEY = os.getenv('PINECONE_API_KEY') PINECONE_ENV = os.getenv('PINECONE_ENV') OPENAI_API_KEY = os.getenv('OPENAI_API_KEY') PINECONE_INDEX_NAME = os.getenv('PINECONE_INDEX_NAME') class Person(BaseModel): """Identifying information about a person.""" name: str = Field(..., description="The person's name") class People(BaseModel): """Identifying information about all people in a text.""" people: Sequence[Person] = Field(..., description="The people in the text") class QueueCallback(BaseCallbackHandler): """Callback handler for streaming LLM responses to a queue.""" def __init__(self, q): self.q = q def on_llm_new_token(self, token: str, **kwargs: Any) -> None: self.q.put(token) def on_llm_end(self, *args, **kwargs: Any) -> None: return self.q.empty() def get_name_from_prompt(query): llm = ChatOpenAI(model="gpt-4-1106-preview", temperature=0) prompt = ChatPromptTemplate.from_messages( [ ( "system", "You are a world class algorithm for extracting information in structured formats.", ), ( "human", "Use the given format to extract information from the following input: {input}", ), ("human", "Tip: Make sure to answer in the correct format"), ] ) runnable = create_structured_output_runnable(People, llm, prompt) result = runnable.invoke({"input": query}) peoples = result.people names = [person.name for person in peoples] return names def generate_message(query, behavior, temp, model, chat, template, openai_api_key = None): if openai_api_key is None: openai_api_key = os.getenv('OPENAI_API_KEY') q = Queue() job_done = object() prompt = PromptTemplate( input_variables=["context", "question"], template=template) chain_type_kwargs = {"prompt": prompt} if model == "1": llm = ChatOpenAI(model_name="gpt-3.5-turbo-1106", temperature=temp, streaming=True, callbacks=[QueueCallback(q)], openai_api_key=openai_api_key) elif model == "2": llm = ChatOpenAI(model_name="gpt-3.5-turbo-1106", temperature=temp, streaming=True, callbacks=[QueueCallback(q)], openai_api_key=openai_api_key) elif model == "3": llm = ChatOpenAI(model_name="gpt-4-1106-preview", temperature=temp, streaming=True, callbacks=[QueueCallback(q)], openai_api_key=openai_api_key) embeddings = OpenAIEmbeddings(openai_api_key=os.getenv("OPENAI_API_KEY")) docsearch = Pinecone.from_existing_index( index_name=PINECONE_INDEX_NAME, embedding=embeddings) docs = docsearch.similarity_search_with_score(" ", filter={"chat": str(chat)}) qa = RetrievalQA.from_chain_type(llm=llm, chain_type="stuff", retriever=docsearch.as_retriever(search_kwargs={'filter': {"chat": str(chat)}}), chain_type_kwargs=chain_type_kwargs) def task(): try: qa.run(query) except Exception as e: print(f"generate_message task error: {e}") finally: q.put(job_done) t = Thread(target=task) t.start() content = "" # Get each new token from the queue and yield for our generator while True: try: next_token = q.get(True, timeout=0) if next_token is job_done: break content += next_token yield next_token, content except Empty: continue def generate_AI_message_langchain(query, history, behavior, temp, model, openai_api_key): if openai_api_key is None: openai_api_key = os.getenv('OPENAI_API_KEY') q = Queue() job_done = object() template = """ {behavior} ========== {history} ========== Human: {human_input} Assistant:""" prompt = PromptTemplate( input_variables=["history", "human_input", "behavior"], template=template) if model == "1": llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=temp, streaming=True, callbacks=[QueueCallback(q)], openai_api_key=openai_api_key) elif model == "2": llm = ChatOpenAI(model_name="gpt-3.5-turbo-1106", temperature=temp, streaming=True, callbacks=[QueueCallback(q)], openai_api_key=openai_api_key) elif model == "3": llm = ChatOpenAI(model_name="gpt-4-1106-preview", temperature=temp, streaming=True, callbacks=[QueueCallback(q)], openai_api_key=openai_api_key) conversation = LLMChain( llm=llm, verbose=True, prompt=prompt ) def task(): try: response = conversation.run( human_input=query, history=history, behavior=behavior, ) except Exception as e: print(f"generate_AI_message_langchain task error: {e}") finally: q.put(job_done) t = Thread(target=task) t.start() content = "" # Get each new token from the queue and yield for our generator while True: try: next_token = q.get(True, timeout=0) if next_token is job_done: break content += next_token yield next_token, content except Empty: continue def generate_AI_message(query, history, behavior, temp, model, openai_api_key): client = OpenAI(api_key=os.getenv('OPENAI_API_KEY')) messages = [] messages.append({'role': 'system', 'content': behavior}) for item in history: if item['role'] == 'human': messages.append({'role': 'user', 'content': item['content']}) else: messages.append({'role': 'assistant', 'content': item['content']}) messages.append({'role': 'user', 'content': query}) if model == "1": model_name="gpt-3.5-turbo" elif model == "2": model_name="gpt-3.5-turbo-1106" elif model == "3": model_name="gpt-4-1106-preview" response = client.chat.completions.create( model=model_name, temperature=temp, stream=True, messages=messages ) content = '' for chunk in response: next_token = chunk.choices[0].delta.content if next_token is not None: content += next_token yield next_token, content def generate_Bubble_message(query): template = "Generate a title for a fantasy animal, character or fairy in {query}. A title must two words, first is adjective and second is noun. Do not provide any explanations. Do not respond with anything except the output of the title." prompt = PromptTemplate( input_variables=["query"], template=template) llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=1, streaming=True, openai_api_key=os.getenv('OPENAI_API_KEY')) conversation = LLMChain( llm=llm, verbose=True, prompt=prompt ) response = conversation.run( query=query ) response = response.replace('"', '') return response def generate_system_prompt_role(role): try: template = '''You are a OpenAI GPT system role expert. Your job is to analyze the needs of users and generate system roles for users' 'Interactive Tutors' that they embed to change the role of the 'tutor' powered by OpenAI API to deliver on what they need. The user will give you details on what they need the system behavior prompt to deliver. Once you have this information, do not ask any further questions and please provide JSON object: A short name, description, a conversation starter and the system role written out in full. {{ "name": "", "system_role": "", "starter" : "", "description": "" }} The system role should be well detailed, clearly detail what steps the AI should take and to use British English if communicating in English. Most importantly, the system role's maximum character length must be less than 65500. Consider this: The output must be a JSON object. ========================= user: {role} ''' prompt = PromptTemplate( input_variables=["role"], template=template) llm = ChatOpenAI(model_name="gpt-3.5-turbo", temperature=0.2, openai_api_key=os.getenv('OPENAI_API_KEY')) conversation = LLMChain( llm=llm, prompt=prompt ) response = conversation.run( role=role ) if type(response) == str: response = json.loads(response) return response except Exception as e: print(f"generate_system_prompt_role error: {e}") return jsonify({ "name": "", "system_role": "", "starter" : "", "description": "" , 'message': e.message, 'success': False}) def generate_part_file(prompt, data): template = '''Answer using the sentences below Context. If you cannot find an appropriate answer to the question in the Context, return "". Context: {{ {text} }} ========================= user: {prompt} ''' prompt = PromptTemplate( input_variables=["text", "prompt"], template=template) llm = ChatOpenAI(model_name="gpt-3.5-turbo-1106", temperature=0.2, openai_api_key=os.getenv('OPENAI_API_KEY')) conversation = LLMChain( llm=llm, prompt=prompt ) response = conversation.run( text=data, prompt=prompt ) return response def image_understanding(image_files, prompt): base64_images = [base64.b64encode(image_file.read()).decode('utf-8') for image_file in image_files] client = OpenAI(api_key=os.getenv('OPENAI_API_KEY')) image_prompt = [{ "type": "text", "text": prompt }] for base64_image in base64_images: image_prompt.append({ "type": "image_url", "image_url": { "url": f"data:image/jpeg;base64,{base64_image}" } }) content = "" response = client.chat.completions.create( model="gpt-4-vision-preview", temperature=0.3, stream=True, max_tokens=4000, messages=[ { "role": "user", "content": image_prompt } ]) for chunk in response: next_token = chunk.choices[0].delta.content if next_token is not None: content += next_token yield next_token, content
[ "langchain.pydantic_v1.Field", "langchain.chat_models.ChatOpenAI", "langchain.prompts.ChatPromptTemplate.from_messages", "langchain.chains.openai_functions.create_structured_output_runnable", "langchain.chains.LLMChain", "langchain.vectorstores.pinecone.Pinecone.from_existing_index", "langchain.prompts.PromptTemplate" ]
[((786, 799), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (797, 799), False, 'from dotenv import load_dotenv\n'), ((820, 849), 'os.getenv', 'os.getenv', (['"""PINECONE_API_KEY"""'], {}), "('PINECONE_API_KEY')\n", (829, 849), False, 'import os\n'), ((865, 890), 'os.getenv', 'os.getenv', (['"""PINECONE_ENV"""'], {}), "('PINECONE_ENV')\n", (874, 890), False, 'import os\n'), ((908, 935), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (917, 935), False, 'import os\n'), ((958, 990), 'os.getenv', 'os.getenv', (['"""PINECONE_INDEX_NAME"""'], {}), "('PINECONE_INDEX_NAME')\n", (967, 990), False, 'import os\n'), ((1085, 1128), 'langchain.pydantic_v1.Field', 'Field', (['...'], {'description': '"""The person\'s name"""'}), '(..., description="The person\'s name")\n', (1090, 1128), False, 'from langchain.pydantic_v1 import BaseModel, Field\n'), ((1249, 1297), 'langchain.pydantic_v1.Field', 'Field', (['...'], {'description': '"""The people in the text"""'}), "(..., description='The people in the text')\n", (1254, 1297), False, 'from langchain.pydantic_v1 import BaseModel, Field\n'), ((1680, 1733), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-4-1106-preview"""', 'temperature': '(0)'}), "(model='gpt-4-1106-preview', temperature=0)\n", (1690, 1733), False, 'from langchain.chat_models import ChatOpenAI\n'), ((1747, 2049), 'langchain.prompts.ChatPromptTemplate.from_messages', 'ChatPromptTemplate.from_messages', (["[('system',\n 'You are a world class algorithm for extracting information in structured formats.'\n ), ('human',\n 'Use the given format to extract information from the following input: {input}'\n ), ('human', 'Tip: Make sure to answer in the correct format')]"], {}), "([('system',\n 'You are a world class algorithm for extracting information in structured formats.'\n ), ('human',\n 'Use the given format to extract information from the following input: {input}'\n ), ('human', 'Tip: Make sure to answer in the correct format')])\n", (1779, 2049), False, 'from langchain.prompts import PromptTemplate, ChatPromptTemplate\n'), ((2202, 2256), 'langchain.chains.openai_functions.create_structured_output_runnable', 'create_structured_output_runnable', (['People', 'llm', 'prompt'], {}), '(People, llm, prompt)\n', (2235, 2256), False, 'from langchain.chains.openai_functions import create_structured_output_runnable\n'), ((2588, 2595), 'queue.Queue', 'Queue', ([], {}), '()\n', (2593, 2595), False, 'from queue import Queue, Empty\n'), ((2633, 2707), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['context', 'question']", 'template': 'template'}), "(input_variables=['context', 'question'], template=template)\n", (2647, 2707), False, 'from langchain.prompts import PromptTemplate, ChatPromptTemplate\n'), ((3683, 3770), 'langchain.vectorstores.pinecone.Pinecone.from_existing_index', 'Pinecone.from_existing_index', ([], {'index_name': 'PINECONE_INDEX_NAME', 'embedding': 'embeddings'}), '(index_name=PINECONE_INDEX_NAME, embedding=\n embeddings)\n', (3711, 3770), False, 'from langchain.vectorstores.pinecone import Pinecone\n'), ((4243, 4262), 'threading.Thread', 'Thread', ([], {'target': 'task'}), '(target=task)\n', (4249, 4262), False, 'from threading import Thread\n'), ((4803, 4810), 'queue.Queue', 'Queue', ([], {}), '()\n', (4808, 4810), False, 'from queue import Queue, Empty\n'), ((4973, 5066), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['history', 'human_input', 'behavior']", 'template': 'template'}), "(input_variables=['history', 'human_input', 'behavior'],\n template=template)\n", (4987, 5066), False, 'from langchain.prompts import PromptTemplate, ChatPromptTemplate\n'), ((5915, 5961), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'verbose': '(True)', 'prompt': 'prompt'}), '(llm=llm, verbose=True, prompt=prompt)\n', (5923, 5961), False, 'from langchain.chains import LLMChain\n'), ((6333, 6352), 'threading.Thread', 'Thread', ([], {'target': 'task'}), '(target=task)\n', (6339, 6352), False, 'from threading import Thread\n'), ((8043, 8103), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['query']", 'template': 'template'}), "(input_variables=['query'], template=template)\n", (8057, 8103), False, 'from langchain.prompts import PromptTemplate, ChatPromptTemplate\n'), ((8320, 8366), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'verbose': '(True)', 'prompt': 'prompt'}), '(llm=llm, verbose=True, prompt=prompt)\n', (8328, 8366), False, 'from langchain.chains import LLMChain\n'), ((10739, 10808), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['text', 'prompt']", 'template': 'template'}), "(input_variables=['text', 'prompt'], template=template)\n", (10753, 10808), False, 'from langchain.prompts import PromptTemplate, ChatPromptTemplate\n'), ((10995, 11027), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt)\n', (11003, 11027), False, 'from langchain.chains import LLMChain\n'), ((2551, 2578), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (2560, 2578), False, 'import os\n'), ((4765, 4792), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (4774, 4792), False, 'import os\n'), ((9713, 9772), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['role']", 'template': 'template'}), "(input_variables=['role'], template=template)\n", (9727, 9772), False, 'from langchain.prompts import PromptTemplate, ChatPromptTemplate\n'), ((9972, 10004), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt)\n', (9980, 10004), False, 'from langchain.chains import LLMChain\n'), ((3638, 3665), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (3647, 3665), False, 'import os\n'), ((6816, 6843), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (6825, 6843), False, 'import os\n'), ((8272, 8299), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (8281, 8299), False, 'import os\n'), ((10165, 10185), 'json.loads', 'json.loads', (['response'], {}), '(response)\n', (10175, 10185), False, 'import json\n'), ((10309, 10427), 'flask.jsonify', 'jsonify', (["{'name': '', 'system_role': '', 'starter': '', 'description': '', 'message':\n e.message, 'success': False}"], {}), "({'name': '', 'system_role': '', 'starter': '', 'description': '',\n 'message': e.message, 'success': False})\n", (10316, 10427), False, 'from flask import jsonify\n'), ((10947, 10974), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (10956, 10974), False, 'import os\n'), ((11329, 11356), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (11338, 11356), False, 'import os\n'), ((9920, 9947), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (9929, 9947), False, 'import os\n')]
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import asyncio import functools import inspect import itertools import logging import uuid import warnings from enum import Enum from typing import ( TYPE_CHECKING, Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, cast, ) from urllib.parse import urlparse, urlunparse from langsmith import Client, RunEvaluator from langsmith.schemas import Dataset, DataType, Example from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.manager import Callbacks from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.chains.base import Chain from langchain.evaluation.loading import load_evaluator from langchain.evaluation.schema import EvaluatorType, StringEvaluator from langchain.schema import ChatResult, LLMResult from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import BaseMessage, messages_from_dict from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda from langchain.smith.evaluation.config import EvalConfig, RunEvalConfig from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain if TYPE_CHECKING: import pandas as pd logger = logging.getLogger(__name__) MODEL_OR_CHAIN_FACTORY = Union[ Callable[[], Union[Chain, Runnable]], BaseLanguageModel, Callable[[dict], Any], Runnable, Chain, ] MCF = Union[Callable[[], Union[Chain, Runnable]], BaseLanguageModel] class InputFormatError(Exception): """Raised when the input format is invalid.""" ## Shared Utilities class TestResult(dict): """A dictionary of the results of a single test run.""" def to_dataframe(self) -> pd.DataFrame: """Convert the results to a dataframe.""" try: import pandas as pd except ImportError as e: raise ImportError( "Pandas is required to convert the results to a dataframe." " to install pandas, run `pip install pandas`." ) from e indices = [] records = [] for example_id, result in self["results"].items(): feedback = result["feedback"] records.append( {**{f.key: f.score for f in feedback}, "output": result["output"]} ) indices.append(example_id) return pd.DataFrame(records, index=indices) def _get_eval_project_url(api_url: str, project_id: str) -> str: """Get the project url from the api url.""" parsed = urlparse(api_url) hostname = parsed.hostname or "" if "api." in hostname: hostname = hostname.replace("api.", "", 1) if "localhost" in hostname: # Remove the port hostname = "localhost" url = urlunparse(parsed._replace(netloc=hostname)) return f"{url}/projects/p/{project_id}?eval=true" def _wrap_in_chain_factory( llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, dataset_name: str = "<my_dataset>", ) -> MCF: """Forgive the user if they pass in a chain without memory instead of a chain factory. It's a common mistake. Raise a more helpful error message as well.""" if isinstance(llm_or_chain_factory, Chain): chain = llm_or_chain_factory chain_class = chain.__class__.__name__ if llm_or_chain_factory.memory is not None: memory_class = chain.memory.__class__.__name__ raise ValueError( "Cannot directly evaluate a chain with stateful memory." " To evaluate this chain, pass in a chain constructor" " that initializes fresh memory each time it is called." " This will safegaurd against information" " leakage between dataset examples." "\nFor example:\n\n" "def chain_constructor():\n" f" new_memory = {memory_class}(...)\n" f" return {chain_class}" "(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) logger.warning( "Directly passing in a chain is not recommended as chains may have state." " This can lead to unexpected behavior as the " "same chain instance could be used across multiple datasets. Instead," " please pass a chain constructor that creates a new " "chain with fresh memory each time it is called. This will safeguard" " against information leakage between dataset examples. " "\nFor example:\n\n" "def chain_constructor():\n" f" return {chain_class}(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) return lambda: chain elif isinstance(llm_or_chain_factory, BaseLanguageModel): return llm_or_chain_factory elif isinstance(llm_or_chain_factory, Runnable): # Memory may exist here, but it's not elegant to check all those cases. lcf = llm_or_chain_factory return lambda: lcf elif callable(llm_or_chain_factory): try: _model = llm_or_chain_factory() # type: ignore[call-arg] except TypeError: # It's an arbitrary function, wrap it in a RunnableLambda user_func = cast(Callable, llm_or_chain_factory) sig = inspect.signature(user_func) logger.info(f"Wrapping function {sig} as RunnableLambda.") wrapped = RunnableLambda(user_func) return lambda: wrapped constructor = cast(Callable, llm_or_chain_factory) if isinstance(_model, BaseLanguageModel): # It's not uncommon to do an LLM constructor instead of raw LLM, # so we'll unpack it for the user. return _model elif not isinstance(_model, Runnable): # This is unlikely to happen - a constructor for a model function return lambda: RunnableLambda(constructor) else: # Typical correct case return constructor # noqa return llm_or_chain_factory def _first_example(examples: Iterator[Example]) -> Tuple[Example, Iterator[Example]]: """Get the first example while chaining it back and preserving the iterator.""" try: example: Example = next(examples) except StopIteration: raise ValueError("No examples provided.") return example, itertools.chain([example], examples) def _get_prompt(inputs: Dict[str, Any]) -> str: """Get prompt from inputs. Args: inputs: The input dictionary. Returns: A string prompt. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") prompts = [] if "prompt" in inputs: if not isinstance(inputs["prompt"], str): raise InputFormatError( "Expected string for 'prompt', got" f" {type(inputs['prompt']).__name__}" ) prompts = [inputs["prompt"]] elif "prompts" in inputs: if not isinstance(inputs["prompts"], list) or not all( isinstance(i, str) for i in inputs["prompts"] ): raise InputFormatError( "Expected list of strings for 'prompts'," f" got {type(inputs['prompts']).__name__}" ) prompts = inputs["prompts"] elif len(inputs) == 1: prompt_ = next(iter(inputs.values())) if isinstance(prompt_, str): prompts = [prompt_] elif isinstance(prompt_, list) and all(isinstance(i, str) for i in prompt_): prompts = prompt_ else: raise InputFormatError(f"LLM Run expects string prompt input. Got {inputs}") else: raise InputFormatError( f"LLM Run expects 'prompt' or 'prompts' in inputs. Got {inputs}" ) if len(prompts) == 1: return prompts[0] else: raise InputFormatError( f"LLM Run expects single prompt input. Got {len(prompts)} prompts." ) def _get_messages(inputs: Dict[str, Any]) -> List[BaseMessage]: """Get Chat Messages from inputs. Args: inputs: The input dictionary. Returns: A list of chat messages. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") if "messages" in inputs: single_input = inputs["messages"] elif len(inputs) == 1: single_input = next(iter(inputs.values())) else: raise InputFormatError( f"Chat Run expects 'messages' in inputs when example has multiple" f" input keys. Got {inputs}" ) if isinstance(single_input, list) and all( isinstance(i, dict) for i in single_input ): raw_messages = [single_input] elif isinstance(single_input, list) and all( isinstance(i, list) for i in single_input ): raw_messages = single_input else: raise InputFormatError( f"Chat Run expects List[dict] or List[List[dict]] values for" f" 'messages' key input. Got {inputs}" ) if len(raw_messages) == 1: return messages_from_dict(raw_messages[0]) else: raise InputFormatError( f"Chat Run expects single List[dict] or List[List[dict]] 'messages'" f" input. Got {len(raw_messages)} messages from inputs {inputs}" ) def _get_project_name( project_name: Optional[str], llm_or_chain_factory: MCF, ) -> str: """ Get the project name. Args: project_name: The project name if manually specified. llm_or_chain_factory: The Chain or language model constructor. Returns: The project name. """ if project_name is not None: return project_name if isinstance(llm_or_chain_factory, BaseLanguageModel): model_name = llm_or_chain_factory.__class__.__name__ else: model_name = llm_or_chain_factory().__class__.__name__ hex = uuid.uuid4().hex return f"{hex}-{model_name}" ## Shared Validation Utilities def _validate_example_inputs_for_language_model( first_example: Example, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: if input_mapper: prompt_input = input_mapper(first_example.inputs) if not isinstance(prompt_input, str) and not ( isinstance(prompt_input, list) and all(isinstance(msg, BaseMessage) for msg in prompt_input) ): raise InputFormatError( "When using an input_mapper to prepare dataset example inputs" " for an LLM or chat model, the output must a single string or" " a list of chat messages." f"\nGot: {prompt_input} of type {type(prompt_input)}." ) else: try: _get_prompt(first_example.inputs) except InputFormatError: try: _get_messages(first_example.inputs) except InputFormatError: raise InputFormatError( "Example inputs do not match language model input format. " "Expected a dictionary with messages or a single prompt." f" Got: {first_example.inputs}" " Please update your dataset OR provide an input_mapper" " to convert the example.inputs to a compatible format" " for the llm or chat model you wish to evaluate." ) def _validate_example_inputs_for_chain( first_example: Example, chain: Chain, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: """Validate that the example inputs match the chain input keys.""" if input_mapper: first_inputs = input_mapper(first_example.inputs) missing_keys = set(chain.input_keys).difference(first_inputs) if not isinstance(first_inputs, dict): raise InputFormatError( "When using an input_mapper to prepare dataset example" " inputs for a chain, the mapped value must be a dictionary." f"\nGot: {first_inputs} of type {type(first_inputs)}." ) if missing_keys: raise InputFormatError( "Missing keys after loading example using input_mapper." f"\nExpected: {chain.input_keys}. Got: {first_inputs.keys()}" ) else: first_inputs = first_example.inputs missing_keys = set(chain.input_keys).difference(first_inputs) if len(first_inputs) == 1 and len(chain.input_keys) == 1: # We can pass this through the run method. # Refrain from calling to validate. pass elif missing_keys: raise InputFormatError( "Example inputs missing expected chain input keys." " Please provide an input_mapper to convert the example.inputs" " to a compatible format for the chain you wish to evaluate." f"Expected: {chain.input_keys}. " f"Got: {first_inputs.keys()}" ) def _validate_example_inputs( examples: Iterator[Example], llm_or_chain_factory: MCF, input_mapper: Optional[Callable[[Dict], Any]], ) -> Iterator[Example]: """Validate that the example inputs are valid for the model.""" first_example, examples = _first_example(examples) if isinstance(llm_or_chain_factory, BaseLanguageModel): _validate_example_inputs_for_language_model(first_example, input_mapper) else: chain = llm_or_chain_factory() if isinstance(chain, Chain): # Otherwise it's a runnable _validate_example_inputs_for_chain(first_example, chain, input_mapper) elif isinstance(chain, Runnable): logger.debug(f"Skipping input validation for {chain}") return examples ## Shared Evaluator Setup Utilities def _setup_evaluation( llm_or_chain_factory: MCF, examples: Iterator[Example], evaluation: Optional[RunEvalConfig], data_type: DataType, ) -> Tuple[Optional[List[RunEvaluator]], Iterator[Example]]: """Configure the evaluators to run on the results of the chain.""" if evaluation: first_example, examples = _first_example(examples) if isinstance(llm_or_chain_factory, BaseLanguageModel): run_inputs, run_outputs = None, None run_type = "llm" else: run_type = "chain" if data_type in (DataType.chat, DataType.llm): val = data_type.value if isinstance(data_type, Enum) else data_type raise ValueError( "Cannot evaluate a chain on dataset with " f"data_type={val}. " "Please specify a dataset with the default 'kv' data type." ) chain = llm_or_chain_factory() run_inputs = chain.input_keys if isinstance(chain, Chain) else None run_outputs = chain.output_keys if isinstance(chain, Chain) else None run_evaluators = _load_run_evaluators( evaluation, run_type, data_type, list(first_example.outputs) if first_example.outputs else None, run_inputs, run_outputs, ) else: # TODO: Create a default helpfulness evaluator run_evaluators = None return run_evaluators, examples def _determine_input_key( config: RunEvalConfig, run_inputs: Optional[List[str]], ) -> Optional[str]: input_key = None if config.input_key: input_key = config.input_key if run_inputs and input_key not in run_inputs: raise ValueError(f"Input key {input_key} not in run inputs {run_inputs}") elif run_inputs and len(run_inputs) == 1: input_key = run_inputs[0] elif run_inputs is not None and len(run_inputs) > 1: raise ValueError( f"Must specify input key for model with multiple inputs: {run_inputs}" ) return input_key def _determine_prediction_key( config: RunEvalConfig, run_outputs: Optional[List[str]], ) -> Optional[str]: prediction_key = None if config.prediction_key: prediction_key = config.prediction_key if run_outputs and prediction_key not in run_outputs: raise ValueError( f"Prediction key {prediction_key} not in run outputs {run_outputs}" ) elif run_outputs and len(run_outputs) == 1: prediction_key = run_outputs[0] elif run_outputs is not None and len(run_outputs) > 1: raise ValueError( f"Must specify prediction key for model" f" with multiple outputs: {run_outputs}" ) return prediction_key def _determine_reference_key( config: RunEvalConfig, example_outputs: Optional[List[str]], ) -> Optional[str]: if config.reference_key: reference_key = config.reference_key if example_outputs and reference_key not in example_outputs: raise ValueError( f"Reference key {reference_key} not in Dataset" f" example outputs: {example_outputs}" ) elif example_outputs and len(example_outputs) == 1: reference_key = list(example_outputs)[0] else: reference_key = None return reference_key def _construct_run_evaluator( eval_config: Union[EvaluatorType, str, EvalConfig], eval_llm: Optional[BaseLanguageModel], run_type: str, data_type: DataType, example_outputs: Optional[List[str]], reference_key: Optional[str], input_key: Optional[str], prediction_key: Optional[str], ) -> RunEvaluator: if isinstance(eval_config, (EvaluatorType, str)): if not isinstance(eval_config, EvaluatorType): eval_config = EvaluatorType(eval_config) evaluator_ = load_evaluator(eval_config, llm=eval_llm) eval_type_tag = eval_config.value else: kwargs = {"llm": eval_llm, **eval_config.get_kwargs()} evaluator_ = load_evaluator(eval_config.evaluator_type, **kwargs) eval_type_tag = eval_config.evaluator_type.value if isinstance(evaluator_, StringEvaluator): if evaluator_.requires_reference and reference_key is None: raise ValueError( f"Must specify reference_key in RunEvalConfig to use" f" evaluator of type {eval_type_tag} with" f" dataset with multiple output keys: {example_outputs}." ) run_evaluator = StringRunEvaluatorChain.from_run_and_data_type( evaluator_, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, tags=[eval_type_tag], ) else: raise NotImplementedError( f"Run evaluator for {eval_type_tag} is not implemented" ) return run_evaluator def _get_keys( config: RunEvalConfig, run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], example_outputs: Optional[List[str]], ) -> Tuple[Optional[str], Optional[str], Optional[str]]: input_key = _determine_input_key(config, run_inputs) prediction_key = _determine_prediction_key(config, run_outputs) reference_key = _determine_reference_key(config, example_outputs) return input_key, prediction_key, reference_key def _load_run_evaluators( config: RunEvalConfig, run_type: str, data_type: DataType, example_outputs: Optional[List[str]], run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], ) -> List[RunEvaluator]: """ Load run evaluators from a configuration. Args: config: Configuration for the run evaluators. Returns: A list of run evaluators. """ run_evaluators = [] input_key, prediction_key, reference_key = None, None, None if ( config.evaluators or any([isinstance(e, EvaluatorType) for e in config.evaluators]) or ( config.custom_evaluators and any([isinstance(e, StringEvaluator) for e in config.custom_evaluators]) ) ): input_key, prediction_key, reference_key = _get_keys( config, run_inputs, run_outputs, example_outputs ) for eval_config in config.evaluators: run_evaluator = _construct_run_evaluator( eval_config, config.eval_llm, run_type, data_type, example_outputs, reference_key, input_key, prediction_key, ) run_evaluators.append(run_evaluator) custom_evaluators = config.custom_evaluators or [] for custom_evaluator in custom_evaluators: if isinstance(custom_evaluator, RunEvaluator): run_evaluators.append(custom_evaluator) elif isinstance(custom_evaluator, StringEvaluator): run_evaluators.append( StringRunEvaluatorChain.from_run_and_data_type( custom_evaluator, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, ) ) else: raise ValueError( f"Unsupported custom evaluator: {custom_evaluator}." f" Expected RunEvaluator or StringEvaluator." ) return run_evaluators ### Async Helpers async def _arun_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], *, tags: Optional[List[str]] = None, callbacks: Callbacks = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """Asynchronously run the language model. Args: llm: The language model to run. inputs: The input dictionary. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map inputs to the expected format. Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): return await llm.apredict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): return await llm.apredict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format" f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: prompt = _get_prompt(inputs) llm_output: Union[str, BaseMessage] = await llm.apredict( prompt, callbacks=callbacks, tags=tags ) except InputFormatError: messages = _get_messages(inputs) llm_output = await llm.apredict_messages( messages, callbacks=callbacks, tags=tags ) return llm_output async def _arun_chain( chain: Union[Chain, Runnable], inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str]: """Run a chain asynchronously on inputs.""" inputs_ = inputs if input_mapper is None else input_mapper(inputs) if isinstance(chain, Chain): if isinstance(inputs_, dict) and len(inputs_) == 1: val = next(iter(inputs_.values())) output = await chain.acall(val, callbacks=callbacks, tags=tags) else: output = await chain.acall(inputs_, callbacks=callbacks, tags=tags) else: runnable_config = RunnableConfig(tags=tags or [], callbacks=callbacks) output = await chain.ainvoke(inputs_, config=runnable_config) return output async def _arun_llm_or_chain( example: Example, llm_or_chain_factory: MCF, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str, LLMResult, ChatResult]: """Asynchronously run the Chain or language model. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map the input to the expected format. Returns: A list of outputs. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) result = None try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = await _arun_llm( llm_or_chain_factory, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = await _arun_chain( chain, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) result = output except Exception as e: logger.warning(f"{chain_or_llm} failed for example {example.id}. Error: {e}") result = {"Error": str(e)} if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return result async def _gather_with_concurrency( n: int, initializer: Callable[[], Coroutine[Any, Any, Any]], *async_funcs: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ], ) -> List[Any]: """Run coroutines with a concurrency limit. Args: n: The maximum number of concurrent tasks. initializer: A coroutine that initializes shared resources for the tasks. async_funcs: The async_funcs to be run concurrently. Returns: A list of results from the coroutines. """ semaphore = asyncio.Semaphore(n) job_state = {"num_processed": 0} callback_queue: asyncio.Queue[Sequence[BaseCallbackHandler]] = asyncio.Queue() for _ in range(n): callback_queue.put_nowait(await initializer()) async def run_coroutine_with_semaphore( async_func: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ] ) -> Any: async with semaphore: callbacks = await callback_queue.get() try: result = await async_func(callbacks, job_state) finally: callback_queue.put_nowait(callbacks) return result results = await asyncio.gather( *(run_coroutine_with_semaphore(function) for function in async_funcs) ) while callback_queue: try: callbacks = callback_queue.get_nowait() except asyncio.QueueEmpty: break for callback in callbacks: if isinstance(callback, (LangChainTracer, EvaluatorCallbackHandler)): callback.wait_for_futures() return results async def _callbacks_initializer( project_name: Optional[str], client: Client, run_evaluators: Sequence[RunEvaluator], evaluation_handler_collector: List[EvaluatorCallbackHandler], ) -> List[BaseTracer]: """ Initialize a tracer to share across tasks. Args: project_name: The project name for the tracer. client: The client to use for the tracer. run_evaluators: The evaluators to run. evaluation_handler_collector: A list to collect the evaluators. Used to wait for the evaluators to finish. Returns: The callbacks for this thread. """ callbacks: List[BaseTracer] = [] if project_name: callbacks.append( LangChainTracer( project_name=project_name, client=client, use_threading=False ) ) if run_evaluators: callback = EvaluatorCallbackHandler( client=client, evaluators=run_evaluators, # We already have concurrency, don't want to overload the machine max_workers=1, ) callbacks.append(callback) evaluation_handler_collector.append(callback) return callbacks async def _arun_on_examples( client: Client, examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, data_type: DataType = DataType.kv, ) -> Dict[str, Any]: """ Asynchronously run the chain on examples and store traces to the specified project name. Args: client: LangSmith client to use to log feedback and runs. examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating concurrency_level: The number of async tasks to run concurrently. project_name: Project name to use when tracing runs. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. data_type: The dataset's data type. This is used to determine determine how to deserialize the reference data and model compatibility. Returns: A dictionary mapping example ids to the model outputs. """ wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory) project_name = _get_project_name(project_name, wrapped_model) run_evaluators, examples = _setup_evaluation( wrapped_model, examples, evaluation, data_type ) examples = _validate_example_inputs(examples, wrapped_model, input_mapper) results: Dict[str, dict] = {} async def process_example( example: Example, callbacks: List[BaseCallbackHandler], job_state: dict ) -> None: """Process a single example.""" result = await _arun_llm_or_chain( example, wrapped_model, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) results[str(example.id)] = {"output": result} job_state["num_processed"] += 1 if verbose: print( f"Processed examples: {job_state['num_processed']}", end="\r", flush=True, ) evaluation_handlers: List[EvaluatorCallbackHandler] = [] await _gather_with_concurrency( concurrency_level, functools.partial( _callbacks_initializer, project_name=project_name, client=client, evaluation_handler_collector=evaluation_handlers, run_evaluators=run_evaluators or [], ), *(functools.partial(process_example, e) for e in examples), ) all_feedback = {} for handler in evaluation_handlers: handler.wait_for_futures() all_feedback.update(handler.logged_feedback) # join the results and feedback on the example id for example_id, output_dict in results.items(): feedback = all_feedback.get(example_id, []) output_dict["feedback"] = feedback return results ## Sync Utilities def _run_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """ Run the language model on the example. Args: llm: The language model to run. inputs: The input dictionary. callbacks: The callbacks to use during the run. tags: Optional tags to add to the run. input_mapper: function to map to the inputs dictionary from an Example Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): llm_output: Union[str, BaseMessage] = llm.predict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): llm_output = llm.predict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format: " f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: llm_prompts = _get_prompt(inputs) llm_output = llm.predict(llm_prompts, callbacks=callbacks, tags=tags) except InputFormatError: llm_messages = _get_messages(inputs) llm_output = llm.predict_messages(llm_messages, callbacks=callbacks) return llm_output def _run_chain( chain: Union[Chain, Runnable], inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[Dict, str]: """Run a chain on inputs.""" inputs_ = inputs if input_mapper is None else input_mapper(inputs) if isinstance(chain, Chain): if isinstance(inputs_, dict) and len(inputs_) == 1: val = next(iter(inputs_.values())) output = chain(val, callbacks=callbacks, tags=tags) else: output = chain(inputs_, callbacks=callbacks, tags=tags) else: runnable_config = RunnableConfig(tags=tags or [], callbacks=callbacks) output = chain.invoke(inputs_, config=runnable_config) return output def _run_llm_or_chain( example: Example, llm_or_chain_factory: MCF, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str, LLMResult, ChatResult]: """ Run the Chain or language model synchronously. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. Returns: Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: The outputs of the model or chain. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) result = None try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = _run_llm( llm_or_chain_factory, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = _run_chain( chain, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) result = output except Exception as e: logger.warning( f"{chain_or_llm} failed for example {example.id} with inputs:" f" {example.inputs}.\nError: {e}", ) result = {"Error": str(e)} if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return result def _run_on_examples( client: Client, examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, data_type: DataType = DataType.kv, ) -> Dict[str, Any]: """ Run the Chain or language model on examples and store traces to the specified project name. Args: client: LangSmith client to use to log feedback and runs. examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. data_type: The dataset's data type. This is used to determine determine how to deserialize the reference data and model compatibility. Returns: A dictionary mapping example ids to the model outputs. """ results: Dict[str, dict] = {} wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory) project_name = _get_project_name(project_name, wrapped_model) tracer = LangChainTracer( project_name=project_name, client=client, use_threading=False ) run_evaluators, examples = _setup_evaluation( wrapped_model, examples, evaluation, data_type ) examples = _validate_example_inputs(examples, wrapped_model, input_mapper) evaluation_handler = EvaluatorCallbackHandler( evaluators=run_evaluators or [], client=client, ) callbacks: List[BaseCallbackHandler] = [tracer, evaluation_handler] for i, example in enumerate(examples): result = _run_llm_or_chain( example, wrapped_model, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) if verbose: print(f"{i+1} processed", flush=True, end="\r") results[str(example.id)] = {"output": result} tracer.wait_for_futures() evaluation_handler.wait_for_futures() all_feedback = evaluation_handler.logged_feedback # join the results and feedback on the example id for example_id, output_dict in results.items(): feedback = all_feedback.get(example_id, []) output_dict["feedback"] = feedback return results ## Public API def _prepare_eval_run( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, project_name: Optional[str], ) -> Tuple[MCF, str, Dataset, Iterator[Example]]: wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory, dataset_name) project_name = _get_project_name(project_name, wrapped_model) try: project = client.create_project(project_name) except ValueError as e: if "already exists " not in str(e): raise e raise ValueError( f"Project {project_name} already exists. Please use a different name." ) project_url = _get_eval_project_url(client.api_url, project.id) print( f"View the evaluation results for project '{project_name}' at:\n{project_url}" ) dataset = client.read_dataset(dataset_name=dataset_name) examples = client.list_examples(dataset_id=str(dataset.id)) return wrapped_model, project_name, dataset, examples async def arun_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, **kwargs: Any, ) -> Dict[str, Any]: """ Asynchronously run the Chain or language model on a dataset and store traces to the specified project name. Args: client: LangSmith client to use to read the dataset, and to log feedback and run traces. dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating concurrency_level: The number of async tasks to run concurrently. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. For the synchronous version, see :func:`run_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import RunEvalConfig, arun_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", RunEvalConfig.Criteria("helpfulness"), RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() await arun_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) await arun_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 if kwargs: warnings.warn( "The following arguments are deprecated and will " "be removed in a future release: " f"{kwargs.keys()}.", DeprecationWarning, ) wrapped_model, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name ) results = await _arun_on_examples( client, examples, wrapped_model, concurrency_level=concurrency_level, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) return TestResult( project_name=project_name, results=results, ) def _handle_coroutine(coro: Coroutine) -> Any: """ Handles a coroutine from a sync context. Args: coro (asyncio.coroutine): The coroutine to be handled. Returns: any: The result of the executed coroutine. """ # Check if there's a running event loop try: loop = asyncio.get_event_loop() except RuntimeError: # No event loop return asyncio.run(coro) if loop.is_running(): return loop.run_until_complete(coro) else: return asyncio.run(coro) def run_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, **kwargs: Any, ) -> Dict[str, Any]: """ Run the Chain or language model on a dataset and store traces to the specified project name. Args: client: LangSmith client to use to access the dataset and to log feedback and run traces. dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Configuration for evaluators to run on the results of the chain concurrency_level: The number of async tasks to run concurrently. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. For the (usually faster) async version of this function, see :func:`arun_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import RunEvalConfig, run_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", RunEvalConfig.Criteria("helpfulness"), RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 if kwargs: warnings.warn( "The following arguments are deprecated and " "will be removed in a future release: " f"{kwargs.keys()}.", DeprecationWarning, ) wrapped_model, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name ) if concurrency_level in (0, 1): results = _run_on_examples( client, examples, wrapped_model, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) else: # TODO: Use runnables and the batch method coro = _arun_on_examples( client, examples, wrapped_model, concurrency_level=concurrency_level, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) results = _handle_coroutine(coro) return TestResult( project_name=project_name, results=results, )
[ "langchain.schema.messages.messages_from_dict", "langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type", "langchain.evaluation.schema.EvaluatorType", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler", "langchain.schema.runnable.RunnableLambda", "langchain.schema.runnable.RunnableConfig", "langchain.evaluation.loading.load_evaluator" ]
[((1500, 1527), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1517, 1527), False, 'import logging\n'), ((2799, 2816), 'urllib.parse.urlparse', 'urlparse', (['api_url'], {}), '(api_url)\n', (2807, 2816), False, 'from urllib.parse import urlparse, urlunparse\n'), ((27519, 27539), 'asyncio.Semaphore', 'asyncio.Semaphore', (['n'], {}), '(n)\n', (27536, 27539), False, 'import asyncio\n'), ((27645, 27660), 'asyncio.Queue', 'asyncio.Queue', ([], {}), '()\n', (27658, 27660), False, 'import asyncio\n'), ((40275, 40353), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)'}), '(project_name=project_name, client=client, use_threading=False)\n', (40290, 40353), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((40583, 40655), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'evaluators': '(run_evaluators or [])', 'client': 'client'}), '(evaluators=run_evaluators or [], client=client)\n', (40607, 40655), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler\n'), ((2634, 2670), 'pandas.DataFrame', 'pd.DataFrame', (['records'], {'index': 'indices'}), '(records, index=indices)\n', (2646, 2670), True, 'import pandas as pd\n'), ((6729, 6765), 'itertools.chain', 'itertools.chain', (['[example]', 'examples'], {}), '([example], examples)\n', (6744, 6765), False, 'import itertools\n'), ((9608, 9643), 'langchain.schema.messages.messages_from_dict', 'messages_from_dict', (['raw_messages[0]'], {}), '(raw_messages[0])\n', (9626, 9643), False, 'from langchain.schema.messages import BaseMessage, messages_from_dict\n'), ((10444, 10456), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (10454, 10456), False, 'import uuid\n'), ((18358, 18399), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config'], {'llm': 'eval_llm'}), '(eval_config, llm=eval_llm)\n', (18372, 18399), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((18536, 18588), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config.evaluator_type'], {}), '(eval_config.evaluator_type, **kwargs)\n', (18550, 18588), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((19034, 19224), 'langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type', 'StringRunEvaluatorChain.from_run_and_data_type', (['evaluator_', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key', 'tags': '[eval_type_tag]'}), '(evaluator_, run_type,\n data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key, tags=[eval_type_tag])\n', (19080, 19224), False, 'from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain\n'), ((24688, 24740), 'langchain.schema.runnable.RunnableConfig', 'RunnableConfig', ([], {'tags': '(tags or [])', 'callbacks': 'callbacks'}), '(tags=tags or [], callbacks=callbacks)\n', (24702, 24740), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((29506, 29591), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'client': 'client', 'evaluators': 'run_evaluators', 'max_workers': '(1)'}), '(client=client, evaluators=run_evaluators,\n max_workers=1)\n', (29530, 29591), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler\n'), ((36030, 36082), 'langchain.schema.runnable.RunnableConfig', 'RunnableConfig', ([], {'tags': '(tags or [])', 'callbacks': 'callbacks'}), '(tags=tags or [], callbacks=callbacks)\n', (36044, 36082), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((47863, 47887), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (47885, 47887), False, 'import asyncio\n'), ((48059, 48076), 'asyncio.run', 'asyncio.run', (['coro'], {}), '(coro)\n', (48070, 48076), False, 'import asyncio\n'), ((18310, 18336), 'langchain.evaluation.schema.EvaluatorType', 'EvaluatorType', (['eval_config'], {}), '(eval_config)\n', (18323, 18336), False, 'from langchain.evaluation.schema import EvaluatorType, StringEvaluator\n'), ((29345, 29423), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)'}), '(project_name=project_name, client=client, use_threading=False)\n', (29360, 29423), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((32818, 32997), 'functools.partial', 'functools.partial', (['_callbacks_initializer'], {'project_name': 'project_name', 'client': 'client', 'evaluation_handler_collector': 'evaluation_handlers', 'run_evaluators': '(run_evaluators or [])'}), '(_callbacks_initializer, project_name=project_name, client\n =client, evaluation_handler_collector=evaluation_handlers,\n run_evaluators=run_evaluators or [])\n', (32835, 32997), False, 'import functools\n'), ((47945, 47962), 'asyncio.run', 'asyncio.run', (['coro'], {}), '(coro)\n', (47956, 47962), False, 'import asyncio\n'), ((5873, 5909), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (5877, 5909), False, 'from typing import TYPE_CHECKING, Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, cast\n'), ((21499, 21673), 'langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type', 'StringRunEvaluatorChain.from_run_and_data_type', (['custom_evaluator', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key'}), '(custom_evaluator, run_type,\n data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key)\n', (21545, 21673), False, 'from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain\n'), ((33071, 33108), 'functools.partial', 'functools.partial', (['process_example', 'e'], {}), '(process_example, e)\n', (33088, 33108), False, 'import functools\n'), ((5613, 5649), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (5617, 5649), False, 'from typing import TYPE_CHECKING, Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, cast\n'), ((5668, 5696), 'inspect.signature', 'inspect.signature', (['user_func'], {}), '(user_func)\n', (5685, 5696), False, 'import inspect\n'), ((5790, 5815), 'langchain.schema.runnable.RunnableLambda', 'RunnableLambda', (['user_func'], {}), '(user_func)\n', (5804, 5815), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((6262, 6289), 'langchain.schema.runnable.RunnableLambda', 'RunnableLambda', (['constructor'], {}), '(constructor)\n', (6276, 6289), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n')]
from langchain.llms import HuggingFacePipeline, CTransformers import langchain from ingest import load_db from langchain.cache import InMemoryCache from langchain.schema import prompt from langchain.chains import RetrievalQA from langchain.callbacks import StdOutCallbackHandler from langchain import PromptTemplate import chainlit as cl DB_FAISS_PATH = 'vectorstoredb/db_faiss' langchain.llm_cache = InMemoryCache() PROMPT_TEMPLATE = ''' Use the following pieces of information to answer the user's question. If you don't know the answer, just say that you don't know, don't try to make up an answer. Context: {context} Question: {question} Do provide only helpful answers Helpful answer: ''' handler = StdOutCallbackHandler() def set_custom_prompt(): input_variables = ['context', 'question'] prompt = PromptTemplate(template=PROMPT_TEMPLATE, input_variables=input_variables) return prompt def load_retriever(): return load_db() def load_llm(): llm = CTransformers(model="models/mistral-7b-instruct-v0.1.Q2_K.gguf", config={'max_new_tokens': 128, 'temperature': 0.01}) return llm def retrieval_qa_chain(llm, prompt, db): qa_chain = RetrievalQA.from_chain_type( llm=llm, chain_type="stuff", retriever = db, verbose=True, callbacks=[handler], chain_type_kwargs={"prompt": prompt}, return_source_documents=True ) return qa_chain def qa_bot(): db = load_retriever() llm = load_llm() qa_prompt = set_custom_prompt() qa = retrieval_qa_chain(llm, qa_prompt, db) return qa def final_result(query): qa_result = qa_bot() response = qa_result({"query": query}) return response @cl.on_chat_start async def start(): chain = qa_bot() msg = cl.Message(content="Starting the bot...") await msg.send() msg.content = "Hi, Welcome to Medical Bot. What is your query?" await msg.update() cl.user_session.set("chain", chain) @cl.on_message async def main(message: cl.Message): chain = cl.user_session.get("chain") cb = cl.AsyncLangchainCallbackHandler( stream_final_answer=True, answer_prefix_tokens=["FINAL", "ANSWER"] ) cb.answer_reached = True res = await chain.acall(message.content, callbacks=[cb]) answer = res["result"] sources = res["source_documents"] if sources: answer += f"\nSources:" + str(sources) else: answer += "\nNo sources found" await cl.Message(content=answer).send()
[ "langchain.chains.RetrievalQA.from_chain_type", "langchain.llms.CTransformers", "langchain.callbacks.StdOutCallbackHandler", "langchain.cache.InMemoryCache", "langchain.PromptTemplate" ]
[((403, 418), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (416, 418), False, 'from langchain.cache import InMemoryCache\n'), ((709, 732), 'langchain.callbacks.StdOutCallbackHandler', 'StdOutCallbackHandler', ([], {}), '()\n', (730, 732), False, 'from langchain.callbacks import StdOutCallbackHandler\n'), ((817, 890), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': 'PROMPT_TEMPLATE', 'input_variables': 'input_variables'}), '(template=PROMPT_TEMPLATE, input_variables=input_variables)\n', (831, 890), False, 'from langchain import PromptTemplate\n'), ((943, 952), 'ingest.load_db', 'load_db', ([], {}), '()\n', (950, 952), False, 'from ingest import load_db\n'), ((980, 1102), 'langchain.llms.CTransformers', 'CTransformers', ([], {'model': '"""models/mistral-7b-instruct-v0.1.Q2_K.gguf"""', 'config': "{'max_new_tokens': 128, 'temperature': 0.01}"}), "(model='models/mistral-7b-instruct-v0.1.Q2_K.gguf', config={\n 'max_new_tokens': 128, 'temperature': 0.01})\n", (993, 1102), False, 'from langchain.llms import HuggingFacePipeline, CTransformers\n'), ((1170, 1351), 'langchain.chains.RetrievalQA.from_chain_type', 'RetrievalQA.from_chain_type', ([], {'llm': 'llm', 'chain_type': '"""stuff"""', 'retriever': 'db', 'verbose': '(True)', 'callbacks': '[handler]', 'chain_type_kwargs': "{'prompt': prompt}", 'return_source_documents': '(True)'}), "(llm=llm, chain_type='stuff', retriever=db,\n verbose=True, callbacks=[handler], chain_type_kwargs={'prompt': prompt},\n return_source_documents=True)\n", (1197, 1351), False, 'from langchain.chains import RetrievalQA\n'), ((1745, 1786), 'chainlit.Message', 'cl.Message', ([], {'content': '"""Starting the bot..."""'}), "(content='Starting the bot...')\n", (1755, 1786), True, 'import chainlit as cl\n'), ((1904, 1939), 'chainlit.user_session.set', 'cl.user_session.set', (['"""chain"""', 'chain'], {}), "('chain', chain)\n", (1923, 1939), True, 'import chainlit as cl\n'), ((2005, 2033), 'chainlit.user_session.get', 'cl.user_session.get', (['"""chain"""'], {}), "('chain')\n", (2024, 2033), True, 'import chainlit as cl\n'), ((2044, 2148), 'chainlit.AsyncLangchainCallbackHandler', 'cl.AsyncLangchainCallbackHandler', ([], {'stream_final_answer': '(True)', 'answer_prefix_tokens': "['FINAL', 'ANSWER']"}), "(stream_final_answer=True,\n answer_prefix_tokens=['FINAL', 'ANSWER'])\n", (2076, 2148), True, 'import chainlit as cl\n'), ((2438, 2464), 'chainlit.Message', 'cl.Message', ([], {'content': 'answer'}), '(content=answer)\n', (2448, 2464), True, 'import chainlit as cl\n')]
#!/usr/bin/env python # coding: utf-8 # # Building hotel room search with self-querying retrieval # # In this example we'll walk through how to build and iterate on a hotel room search service that leverages an LLM to generate structured filter queries that can then be passed to a vector store. # # For an introduction to self-querying retrieval [check out the docs](https://python.langchain.com/docs/modules/data_connection/retrievers/self_query). # ## Imports and data prep # # In this example we use `ChatOpenAI` for the model and `ElasticsearchStore` for the vector store, but these can be swapped out with an LLM/ChatModel and [any VectorStore that support self-querying](https://python.langchain.com/docs/integrations/retrievers/self_query/). # # Download data from: https://www.kaggle.com/datasets/keshavramaiah/hotel-recommendation import langchain # from langchain_openai import ChatOpenAI import pandas as pd import json import os data_root = r'./data/' def load(data_root): details = (pd.read_csv(os.path.join(data_root, 'Hotel_details.csv')). drop_duplicates(subset="hotelid").set_index("hotelid")) attributes = pd.read_csv(os.path.join(data_root, 'Hotel_Room_attributes.csv'), index_col='id') price = pd.read_csv(os.path.join(data_root, 'hotels_RoomPrice.csv'), index_col='id') return details, attributes, price details, attributes, price = load(data_root=data_root) latest_price = price.drop_duplicates(subset="refid", keep="last")[ [ "hotelcode", "roomtype", "onsiterate", "roomamenities", "maxoccupancy", "mealinclusiontype", ] ] latest_price["ratedescription"] = attributes.loc[latest_price.index]["ratedescription"] latest_price = latest_price.join( details[["hotelname", "city", "country", "starrating"]], on="hotelcode" ) latest_price = latest_price.rename({"ratedescription": "roomdescription"}, axis=1) latest_price["mealsincluded"] = ~latest_price["mealinclusiontype"].isnull() latest_price.pop("hotelcode") latest_price.pop("mealinclusiontype") latest_price = latest_price.reset_index(drop=True) latest_price.head() # ## Describe data attributes # # We'll use a self-query retriever, which requires us to describe the metadata we can filter on. # # Or if we're feeling lazy we can have a model write a draft of the descriptions for us :) from langchain_openai import ChatOpenAI model = ChatOpenAI(model="gpt-4") res = model.predict( "Below is a table with information about hotel rooms. " "Return a JSON list with an entry for each column. Each entry should have " '{"name": "column name", "description": "column description", "type": "column data type"}' f"\n\n{latest_price.head()}\n\nJSON:\n" ) attribute_info = json.loads(res) # For low cardinality features, let's include the valid values in the description latest_price.nunique()[latest_price.nunique() < 40] attribute_info[-2]["description" ] += f". Valid values are {sorted(latest_price['starrating'].value_counts().index.tolist())}" attribute_info[3][ "description" ] += f". Valid values are {sorted(latest_price['maxoccupancy'].value_counts().index.tolist())}" attribute_info[-3][ "description" ] += f". Valid values are {sorted(latest_price['country'].value_counts().index.tolist())}" # ## Creating a query constructor chain # # Let's take a look at the chain that will convert natural language requests into structured queries. # # To start we can just load the prompt and see what it looks like # In[ ]: from langchain.chains.query_constructor.base import ( get_query_constructor_prompt, load_query_constructor_runnable, ) doc_contents = "Detailed description of a hotel room" prompt = get_query_constructor_prompt(doc_contents, attribute_info) print(prompt.format(query="{query}")) # In[ ]: chain = load_query_constructor_runnable( ChatOpenAI(model="gpt-3.5-turbo", temperature=0), doc_contents, attribute_info ) # In[ ]: chain.invoke({"query": "I want a hotel in Southern Europe and my budget is 200 bucks."}) # In[ ]: chain.invoke( { "query": "Find a 2-person room in Vienna or London, preferably with meals included and AC" } ) # ## Refining attribute descriptions # # We can see that at least two issues above. First is that when we ask for a Southern European destination we're only getting a filter for Italy, and second when we ask for AC we get a literal string lookup for AC (which isn't so bad but will miss things like 'Air conditioning'). # # As a first step, let's try to update our description of the 'country' attribute to emphasize that equality should only be used when a specific country is mentioned. # In[ ]: attribute_info[-3][ "description" ] += ". NOTE: Only use the 'eq' operator if a specific country is mentioned. If a region is mentioned, include all relevant countries in filter." chain = load_query_constructor_runnable( ChatOpenAI(model="gpt-3.5-turbo", temperature=0), doc_contents, attribute_info, ) # In[ ]: chain.invoke({"query": "I want a hotel in Southern Europe and my budget is 200 bucks."}) # ## Refining which attributes to filter on # # This seems to have helped! Now let's try to narrow the attributes we're filtering on. More freeform attributes we can leave to the main query, which is better for capturing semantic meaning than searching for specific substrings. # In[ ]: content_attr = ["roomtype", "roomamenities", "roomdescription", "hotelname"] doc_contents = "A detailed description of a hotel room, including information about the room type and room amenities." filter_attribute_info = tuple( ai for ai in attribute_info if ai["name"] not in content_attr ) chain = load_query_constructor_runnable( ChatOpenAI(model="gpt-3.5-turbo", temperature=0), doc_contents, filter_attribute_info, ) # In[ ]: chain.invoke( { "query": "Find a 2-person room in Vienna or London, preferably with meals included and AC" } ) # ## Adding examples specific to our use case # # We've removed the strict filter for 'AC' but it's still not being included in the query string. Our chain prompt is a few-shot prompt with some default examples. Let's see if adding use case-specific examples will help: # In[ ]: examples = [ ( "I want a hotel in the Balkans with a king sized bed and a hot tub. Budget is $300 a night", { "query": "king-sized bed, hot tub", "filter": 'and(in("country", ["Bulgaria", "Greece", "Croatia", "Serbia"]), lte("onsiterate", 300))', }, ), ( "A room with breakfast included for 3 people, at a Hilton", { "query": "Hilton", "filter": 'and(eq("mealsincluded", true), gte("maxoccupancy", 3))', }, ), ] prompt = get_query_constructor_prompt( doc_contents, filter_attribute_info, examples=examples ) print(prompt.format(query="{query}")) # In[ ]: chain = load_query_constructor_runnable( ChatOpenAI(model="gpt-3.5-turbo", temperature=0), doc_contents, filter_attribute_info, examples=examples, ) # In[ ]: chain.invoke( { "query": "Find a 2-person room in Vienna or London, preferably with meals included and AC" } ) # This seems to have helped! Let's try another complex query: # In[ ]: chain.invoke( { "query": "I want to stay somewhere highly rated along the coast. I want a room with a patio and a fireplace." } ) # ## Automatically ignoring invalid queries # # It seems our model get's tripped up on this more complex query and tries to search over an attribute ('description') that doesn't exist. By setting `fix_invalid=True` in our query constructor chain, we can automatically remove any parts of the filter that is invalid (meaning it's using disallowed operations, comparisons or attributes). # In[ ]: chain = load_query_constructor_runnable( ChatOpenAI(model="gpt-3.5-turbo", temperature=0), doc_contents, filter_attribute_info, examples=examples, fix_invalid=True, ) # In[ ]: chain.invoke( { "query": "I want to stay somewhere highly rated along the coast. I want a room with a patio and a fireplace." } ) # ## Using with a self-querying retriever # # Now that our query construction chain is in a decent place, let's try using it with an actual retriever. For this example we'll use the [ElasticsearchStore](https://python.langchain.com/docs/integrations/vectorstores/elasticsearch). # In[ ]: from langchain_community.vectorstores import ElasticsearchStore from langchain_openai import OpenAIEmbeddings embeddings = OpenAIEmbeddings() # ## Populating vectorstore # # The first time you run this, uncomment the below cell to first index the data. # In[ ]: # docs = [] # for _, room in latest_price.fillna("").iterrows(): # doc = Document( # page_content=json.dumps(room.to_dict(), indent=2), # metadata=room.to_dict() # ) # docs.append(doc) # vecstore = ElasticsearchStore.from_documents( # docs, # embeddings, # es_url="http://localhost:9200", # index_name="hotel_rooms", # # strategy=ElasticsearchStore.ApproxRetrievalStrategy( # # hybrid=True, # # ) # ) # In[ ]: vecstore = ElasticsearchStore( "hotel_rooms", embedding=embeddings, es_url="http://localhost:9200", # strategy=ElasticsearchStore.ApproxRetrievalStrategy(hybrid=True) # seems to not be available in community version ) # In[ ]: from langchain.retrievers import SelfQueryRetriever retriever = SelfQueryRetriever( query_constructor=chain, vectorstore=vecstore, verbose=True ) # In[ ]: results = retriever.get_relevant_documents( "I want to stay somewhere highly rated along the coast. I want a room with a patio and a fireplace." ) for res in results: print(res.page_content) print("\n" + "-" * 20 + "\n") # In[ ]:
[ "langchain_openai.ChatOpenAI", "langchain.chains.query_constructor.base.get_query_constructor_prompt", "langchain_openai.OpenAIEmbeddings", "langchain.retrievers.SelfQueryRetriever", "langchain_community.vectorstores.ElasticsearchStore" ]
[((2434, 2459), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-4"""'}), "(model='gpt-4')\n", (2444, 2459), False, 'from langchain_openai import ChatOpenAI\n'), ((2780, 2795), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (2790, 2795), False, 'import json\n'), ((3742, 3800), 'langchain.chains.query_constructor.base.get_query_constructor_prompt', 'get_query_constructor_prompt', (['doc_contents', 'attribute_info'], {}), '(doc_contents, attribute_info)\n', (3770, 3800), False, 'from langchain.chains.query_constructor.base import get_query_constructor_prompt, load_query_constructor_runnable\n'), ((6834, 6923), 'langchain.chains.query_constructor.base.get_query_constructor_prompt', 'get_query_constructor_prompt', (['doc_contents', 'filter_attribute_info'], {'examples': 'examples'}), '(doc_contents, filter_attribute_info, examples=\n examples)\n', (6862, 6923), False, 'from langchain.chains.query_constructor.base import get_query_constructor_prompt, load_query_constructor_runnable\n'), ((8668, 8686), 'langchain_openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (8684, 8686), False, 'from langchain_openai import OpenAIEmbeddings\n'), ((9298, 9390), 'langchain_community.vectorstores.ElasticsearchStore', 'ElasticsearchStore', (['"""hotel_rooms"""'], {'embedding': 'embeddings', 'es_url': '"""http://localhost:9200"""'}), "('hotel_rooms', embedding=embeddings, es_url=\n 'http://localhost:9200')\n", (9316, 9390), False, 'from langchain_community.vectorstores import ElasticsearchStore\n'), ((9599, 9678), 'langchain.retrievers.SelfQueryRetriever', 'SelfQueryRetriever', ([], {'query_constructor': 'chain', 'vectorstore': 'vecstore', 'verbose': '(True)'}), '(query_constructor=chain, vectorstore=vecstore, verbose=True)\n', (9617, 9678), False, 'from langchain.retrievers import SelfQueryRetriever\n'), ((3897, 3945), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (3907, 3945), False, 'from langchain_openai import ChatOpenAI\n'), ((4954, 5002), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (4964, 5002), False, 'from langchain_openai import ChatOpenAI\n'), ((5779, 5827), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (5789, 5827), False, 'from langchain_openai import ChatOpenAI\n'), ((7021, 7069), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (7031, 7069), False, 'from langchain_openai import ChatOpenAI\n'), ((7949, 7997), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model='gpt-3.5-turbo', temperature=0)\n", (7959, 7997), False, 'from langchain_openai import ChatOpenAI\n'), ((1182, 1234), 'os.path.join', 'os.path.join', (['data_root', '"""Hotel_Room_attributes.csv"""'], {}), "(data_root, 'Hotel_Room_attributes.csv')\n", (1194, 1234), False, 'import os\n'), ((1277, 1324), 'os.path.join', 'os.path.join', (['data_root', '"""hotels_RoomPrice.csv"""'], {}), "(data_root, 'hotels_RoomPrice.csv')\n", (1289, 1324), False, 'import os\n'), ((1022, 1066), 'os.path.join', 'os.path.join', (['data_root', '"""Hotel_details.csv"""'], {}), "(data_root, 'Hotel_details.csv')\n", (1034, 1066), False, 'import os\n')]
from llama_index.prompts import PromptTemplate from llama_index import VectorStoreIndex, SimpleDirectoryReader from llama_index.vector_stores import WeaviateVectorStore import weaviate from llama_index.node_parser import ( SentenceWindowNodeParser, ) from llama_index import ( GPTVectorStoreIndex, ServiceContext, StorageContext ) from llama_index.embeddings import LangchainEmbedding from langchain.embeddings.huggingface import ( HuggingFaceBgeEmbeddings, ) from llama_index.postprocessor.sbert_rerank import SentenceTransformerRerank from llama_index.indices.postprocessor import MetadataReplacementPostProcessor from langchain_community.chat_models.huggingface import ChatHuggingFace from langchain_community.llms import HuggingFaceTextGenInference from llama_index.llms import LangChainLLM from dataclasses import dataclass # from llama_index.indices.document_summary import DocumentSummaryIndex import asyncio import json from loguru import logger f = open('ntc.json') ntc_data = json.load(f) f.close() @dataclass class NtcData: title: str tag: list link: str text: list visited_sites = [] def read_data(data, ntc_datas): for key, value in data.items(): key = key value = value ntc_data = NtcData( title=value["title"], tag=value["tags"], link=key, text=value["text"]) # ntc_data = NtcData( # title="", tag=0, link=key, text=[]) if ntc_data.link in visited_sites: return ntc_datas visited_sites.append(ntc_data.link) ntc_datas.append(ntc_data) children = value["children"] if children is None: return ntc_datas for child in children: read_data(data=child, ntc_datas=ntc_datas) return ntc_datas ntc_datas = read_data(ntc_data, ntc_datas=[]) print(ntc_datas[0]) template = """ <s><INST> Defn: An entity is a person(person), institute (institute), organisation (organisation), services(services), packages(packages), information(information), organization(organization) ,,,event(event), award(award), or theory(theory). Abstract communication concepts can be entities if they have a name associated with them. If an entity does not fit the types above it is (misc). Dates, times, adjectives and verbs are not entities. Example: ____________________________________________________ Q: Given the paragraph below, identify a list of possible entities and for each entry explain why it either is or not an entity: Paragraph: He attended the U.S Air Force Institute of Technology for a year, earning a bachelor's degree in aeromechanics, and received his test pilot training at Edwards Air Force Base in California before his assignment as a test pilot at Wright-Patterson Air Force Base in Ohio. Answer: 1. U.S. Air Force Institute of Technology | True | as he attended this institue is likely a institute (institute) 2. bachelor's degree | False | as it is not a university, award or any other entity type 3. aeromechanics | True | as it is a discipline (discipline) 4. Edwards Air Force Base | True | as an Air Force Base is an organised unit (organisation) 5. California | True | as in this cas California refers to location (location) 6. Wright-Patterson Air Force Base | True | as an Air Force Base is an organisation (organisation) 7. Ohio | True | as it is a state (location) ____________________________________________________ Q: Given the paragraph below, identify a list of possible entities and for each entry explain why it either is or not an entity: Paragraph: {text} </INST> Answer:\n """ embed_model = LangchainEmbedding( HuggingFaceBgeEmbeddings(model_name="BAAI/bge-large-en-v1.5")) client = weaviate.Client("http://localhost:8029") # print(client) vector_store = WeaviateVectorStore( weaviate_client=client, index_name="NTC_BOT_NO_ENTITY_TABLE_Metadata_tokens", ) # required_exts = [".mmd"] # docs = SimpleDirectoryReader( # input_dir="./wiseyak_file", required_exts=required_exts).load_data() from llama_index.node_parser.text.utils import truncate_text from llama_index.node_parser import TokenTextSplitter sentence_splitter = TokenTextSplitter.from_defaults(chunk_size = 100, chunk_overlap = 20) def split_by_token(): from typing import List def split(text: str) -> List[str]: sentences = sentence_splitter.split_text(text) return sentences return split node_parser = SentenceWindowNodeParser.from_defaults( sentence_splitter = split_by_token(), window_size=8, window_metadata_key="window", original_text_metadata_key="original_text", ) # nodes = node_parser.get_nodes_from_documents(docs) class ExtractEntity: def __init__(self, datas, node_parser): self.prompt_template = PromptTemplate(template) self.llm = LangChainLLM(llm=HuggingFaceTextGenInference(inference_server_url="http://192.168.88.10:8024", max_new_tokens=2048, top_k=10, top_p=0.95, typical_p=0.95, temperature=0.01, repetition_penalty=1.03, streaming=True,), ) self.docs = self.format_data_in_docs(ntc_datas) self.node_parser = node_parser self.nodes = self.get_nodes_from_documents(self.docs, self.node_parser) # self.nodes_with_entity = self.extract_entity_from_nodes() self.nodes_with_entity = self.nodes self.bge_rerank_postprocessor = SentenceTransformerRerank( model="BAAI/bge-reranker-large", top_n=3 ) self.storage_context = StorageContext.from_defaults( vector_store=vector_store) self.service_context = ServiceContext.from_defaults( embed_model=embed_model, llm=self.llm) def build_index(self): self.index = VectorStoreIndex( nodes=self.nodes_with_entity, storage_context=self.storage_context, service_context=self.service_context, show_progress=True) # self.index = DocumentSummaryIndex( # nodes=self.nodes_with_entity, storage_context=self.storage_context, service_context=self.service_context, show_progress=True) def format_data_in_docs(self, data): from llama_index import Document, VectorStoreIndex docs = [] for ntc_data in ntc_datas: # metadata = {"title": ntc_data.title, "str_tag": str(ntc_data.tag), # "link": ntc_data.link, "text": str(ntc_data.text)} metadata = {"title": ntc_data.title, "str_tag": str(ntc_data.tag), "link": ntc_data.link,} docs.append(Document(text=" ".join( ntc_data.text), metadata=metadata, excluded_embed_metadata_keys=list(metadata.keys()) ,excluded_llm_metadata_keys=list(metadata.keys()))) return docs def get_nodes_from_documents(self, docs, node_parser): nodes = node_parser.get_nodes_from_documents(docs) return nodes def parse_entity(self, response): list_of_entities = response.text.split("\n") entities = [] for entity in list_of_entities: entity_value = entity.split("|") try: if entity_value[1].replace(" ", "") == "True": value = entity_value[0].split(" ", 1)[1] if len(value) < 60: entities.append(value) except: continue return entities async def extract_entity_from_nodes(self): self.nodes_with_entity = [] word_count = [] max_word_length = 1050 for i, node in enumerate(self.nodes): entities = [] words = node.metadata["window"].split(" ") if len(words) > max_word_length: count = 0 while len(words) > max_word_length: if count > 3: break texts = " ".join(words[:max_word_length]) response = await self.llm.acomplete( self.prompt_template.format(text=texts)) logger.debug(f"Working on Node: {i+1}") entities.extend(self.parse_entity(response)) words = words[max_word_length:] count += 1 texts = " ".join(words[:max_word_length]) response = await self.llm.acomplete( self.prompt_template.format(text=texts)) logger.debug(f"Working on Node: {i+1}") entities.extend(self.parse_entity(response)) else: response = await self.llm.acomplete( self.prompt_template.format(text=node.metadata["window"])) logger.debug(f"Working on Node: {i+1}") entities = self.parse_entity(response) node.metadata["entity"] = entities self.nodes_with_entity.append(node) return self.nodes_with_entity def query_anything(self, query): query_engine = self.index.as_query_engine(similarity_top_k=8, node_postprocessors=[self.bge_rerank_postprocessor, MetadataReplacementPostProcessor( target_metadata_key="window")]) # node_postprocessors=[self.bge_rerank_postprocessor] # Make the query asynchronously streaming_response = query_engine.query(query) return streaming_response if __name__ == "__main__": ee = ExtractEntity(datas=ntc_datas, node_parser=node_parser) # _ = asyncio.run(ee.extract_entity_from_nodes()) ee.build_index() response = ee.query_anything("What are fixed lines?") print(response) print(response.source_nodes[0].metadata)
[ "langchain_community.llms.HuggingFaceTextGenInference", "langchain.embeddings.huggingface.HuggingFaceBgeEmbeddings" ]
[((1010, 1022), 'json.load', 'json.load', (['f'], {}), '(f)\n', (1019, 1022), False, 'import json\n'), ((3664, 3704), 'weaviate.Client', 'weaviate.Client', (['"""http://localhost:8029"""'], {}), "('http://localhost:8029')\n", (3679, 3704), False, 'import weaviate\n'), ((3736, 3838), 'llama_index.vector_stores.WeaviateVectorStore', 'WeaviateVectorStore', ([], {'weaviate_client': 'client', 'index_name': '"""NTC_BOT_NO_ENTITY_TABLE_Metadata_tokens"""'}), "(weaviate_client=client, index_name=\n 'NTC_BOT_NO_ENTITY_TABLE_Metadata_tokens')\n", (3755, 3838), False, 'from llama_index.vector_stores import WeaviateVectorStore\n'), ((4112, 4177), 'llama_index.node_parser.TokenTextSplitter.from_defaults', 'TokenTextSplitter.from_defaults', ([], {'chunk_size': '(100)', 'chunk_overlap': '(20)'}), '(chunk_size=100, chunk_overlap=20)\n', (4143, 4177), False, 'from llama_index.node_parser import TokenTextSplitter\n'), ((3592, 3653), 'langchain.embeddings.huggingface.HuggingFaceBgeEmbeddings', 'HuggingFaceBgeEmbeddings', ([], {'model_name': '"""BAAI/bge-large-en-v1.5"""'}), "(model_name='BAAI/bge-large-en-v1.5')\n", (3616, 3653), False, 'from langchain.embeddings.huggingface import HuggingFaceBgeEmbeddings\n'), ((4721, 4745), 'llama_index.prompts.PromptTemplate', 'PromptTemplate', (['template'], {}), '(template)\n', (4735, 4745), False, 'from llama_index.prompts import PromptTemplate\n'), ((5789, 5856), 'llama_index.postprocessor.sbert_rerank.SentenceTransformerRerank', 'SentenceTransformerRerank', ([], {'model': '"""BAAI/bge-reranker-large"""', 'top_n': '(3)'}), "(model='BAAI/bge-reranker-large', top_n=3)\n", (5814, 5856), False, 'from llama_index.postprocessor.sbert_rerank import SentenceTransformerRerank\n'), ((5922, 5977), 'llama_index.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {'vector_store': 'vector_store'}), '(vector_store=vector_store)\n', (5950, 5977), False, 'from llama_index import GPTVectorStoreIndex, ServiceContext, StorageContext\n'), ((6022, 6089), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'embed_model': 'embed_model', 'llm': 'self.llm'}), '(embed_model=embed_model, llm=self.llm)\n', (6050, 6089), False, 'from llama_index import GPTVectorStoreIndex, ServiceContext, StorageContext\n'), ((6152, 6299), 'llama_index.VectorStoreIndex', 'VectorStoreIndex', ([], {'nodes': 'self.nodes_with_entity', 'storage_context': 'self.storage_context', 'service_context': 'self.service_context', 'show_progress': '(True)'}), '(nodes=self.nodes_with_entity, storage_context=self.\n storage_context, service_context=self.service_context, show_progress=True)\n', (6168, 6299), False, 'from llama_index import Document, VectorStoreIndex\n'), ((4782, 4986), 'langchain_community.llms.HuggingFaceTextGenInference', 'HuggingFaceTextGenInference', ([], {'inference_server_url': '"""http://192.168.88.10:8024"""', 'max_new_tokens': '(2048)', 'top_k': '(10)', 'top_p': '(0.95)', 'typical_p': '(0.95)', 'temperature': '(0.01)', 'repetition_penalty': '(1.03)', 'streaming': '(True)'}), "(inference_server_url=\n 'http://192.168.88.10:8024', max_new_tokens=2048, top_k=10, top_p=0.95,\n typical_p=0.95, temperature=0.01, repetition_penalty=1.03, streaming=True)\n", (4809, 4986), False, 'from langchain_community.llms import HuggingFaceTextGenInference\n'), ((8841, 8882), 'loguru.logger.debug', 'logger.debug', (['f"""Working on Node: {i + 1}"""'], {}), "(f'Working on Node: {i + 1}')\n", (8853, 8882), False, 'from loguru import logger\n'), ((9108, 9149), 'loguru.logger.debug', 'logger.debug', (['f"""Working on Node: {i + 1}"""'], {}), "(f'Working on Node: {i + 1}')\n", (9120, 9149), False, 'from loguru import logger\n'), ((8465, 8506), 'loguru.logger.debug', 'logger.debug', (['f"""Working on Node: {i + 1}"""'], {}), "(f'Working on Node: {i + 1}')\n", (8477, 8506), False, 'from loguru import logger\n'), ((9496, 9558), 'llama_index.indices.postprocessor.MetadataReplacementPostProcessor', 'MetadataReplacementPostProcessor', ([], {'target_metadata_key': '"""window"""'}), "(target_metadata_key='window')\n", (9528, 9558), False, 'from llama_index.indices.postprocessor import MetadataReplacementPostProcessor\n')]
# Import langchain modules from langchain.memory import Memory from langchain.tools import VectorStore # Import other modules import os import requests # Define code memory class class CodeMemory(Memory): def __init__(self): # Initialize the memory with an empty dictionary super().__init__() # Initialize the vector store with a default model and index name self.vector_store = VectorStore(model="codebert-base", index_name="code") def store(self, key, value): # Override this method to store a key-value pair in the memory # The key should be a string representing the name or path of the code file # The value should be a string representing the content of the code file # Check if the key is a valid string if not isinstance(key, str): raise TypeError("Key must be a string") # Check if the value is a valid string if not isinstance(value, str): raise TypeError("Value must be a string") # Store the key-value pair in the memory dictionary self.memory[key] = value # Generate an embedding for the value using the vector store embedding = self.vector_store.encode(value) # Store the embedding in the vector store index using the key as an id self.vector_store.index(embedding, key) def retrieve(self, key): # Override this method to retrieve a value from the memory given a key # The key should be a string representing the name or path of the code file or a query for semantic search # The value should be a string representing the content of the code file or None if not found # Check if the key is a valid string if not isinstance(key, str): raise TypeError("Key must be a string") # Try to get the value from the memory dictionary using the key value = self.memory.get(key) # If the value is not None, return it if value is not None: return value # Otherwise, try to perform a semantic search using the vector store and the key as a query results = self.vector_store.search(key) # If there are any results, get the first one and its id if results: result = results[0] result_id = result["id"] # Try to get the value from the memory dictionary using the result id as a key value = self.memory.get(result_id) # If the value is not None, return it if value is not None: return value # Otherwise, try to get the value from GitHub using requests and result id as a url else: try: response = requests.get(result_id) response.raise_for_status() value = response.text # If successful, store and return the value self.store(result_id, value) return value # If unsuccessful, return None except requests.exceptions.RequestException: return None # If there are no results, return None else: return None def load(self, path): # Define a method to load code files from a local or remote path and store them in memory # Check if path is a valid string if not isinstance(path, str): raise TypeError("Path must be a string") # Check if path is a local file path if os.path.isfile(path): # If path is a local file path, read the file and store its contents in memory using the file name as a key with open(path, "r") as f: content = f.read() key = os.path.basename(path) self.store(key, content) # Check if path is a remote file path elif path.startswith("http"): # If path is a remote file path, use requests to get the file contents and store them in memory using the file name as a key try: response = requests.get(path) response.raise_for_status() content = response.text key = os.path.basename(path) self.store(key, content) # If unsuccessful, raise an exception except requests.exceptions.RequestException: raise ValueError("Invalid remote file path") # If path is not a valid file path, raise an exception else: raise ValueError("Invalid file path")
[ "langchain.tools.VectorStore" ]
[((435, 488), 'langchain.tools.VectorStore', 'VectorStore', ([], {'model': '"""codebert-base"""', 'index_name': '"""code"""'}), "(model='codebert-base', index_name='code')\n", (446, 488), False, 'from langchain.tools import VectorStore\n'), ((3648, 3668), 'os.path.isfile', 'os.path.isfile', (['path'], {}), '(path)\n', (3662, 3668), False, 'import os\n'), ((3874, 3896), 'os.path.basename', 'os.path.basename', (['path'], {}), '(path)\n', (3890, 3896), False, 'import os\n'), ((2834, 2857), 'requests.get', 'requests.get', (['result_id'], {}), '(result_id)\n', (2846, 2857), False, 'import requests\n'), ((4188, 4206), 'requests.get', 'requests.get', (['path'], {}), '(path)\n', (4200, 4206), False, 'import requests\n'), ((4304, 4326), 'os.path.basename', 'os.path.basename', (['path'], {}), '(path)\n', (4320, 4326), False, 'import os\n')]
import streamlit as st from langchain.chat_models import ChatOpenAI from langchain.chains import ConversationalRetrievalChain from langchain.prompts.prompt import PromptTemplate from langchain.callbacks import get_openai_callback from langchain.chains import FlareChain from langchain.prompts import MessagesPlaceholder #fix Error: module 'langchain' has no attribute 'verbose' import langchain langchain.verbose = False class Chatbot: def __init__(self, model_name, temperature, vectors, qa_template): self.model_name = model_name self.temperature = temperature self.vectors = vectors self.retriever = self.vectors.as_retriever() self.qa_template = qa_template self.QA_PROMPT = PromptTemplate( template=self.qa_template, input_variables=["context", "question"], ) def conversational_chat(self, query): """ Start a conversational chat with a model via Langchain """ llm = ChatOpenAI(model_name=self.model_name, temperature=self.temperature) chain = ConversationalRetrievalChain.from_llm(llm=llm, retriever=self.retriever, verbose=True, return_source_documents=True, max_tokens_limit=4097, combine_docs_chain_kwargs={'prompt': self.QA_PROMPT}) chain_input = {"question": query, "chat_history": st.session_state["history"]} result = chain(chain_input) st.session_state["history"].append((query, result["answer"])) return result["answer"] def conversational_chatagent(self, query): """ Start a conversational chat with a model via Langchain """ llm = ChatOpenAI(model_name=self.model_name, temperature=self.temperature) chain = FlareChain.from_llm( llm=llm, retriever=self.retriever, max_generation_len=1000, # Adjusted from 4000 to 1000 min_prob=0.3, ) chain_input = {"question": query, "chat_history": st.session_state["history"]} result = chain.run(chain_input['question']) st.session_state["history"].append((query, result["answer"])) return result["answer"] def count_tokens_chain(chain, query): with get_openai_callback() as cb: result = chain.run(query) st.write(f'###### Tokens used in this conversation : {cb.total_tokens} tokens') return result
[ "langchain.prompts.prompt.PromptTemplate", "langchain.chains.ConversationalRetrievalChain.from_llm", "langchain.chat_models.ChatOpenAI", "langchain.callbacks.get_openai_callback", "langchain.chains.FlareChain.from_llm" ]
[((736, 822), 'langchain.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'template': 'self.qa_template', 'input_variables': "['context', 'question']"}), "(template=self.qa_template, input_variables=['context',\n 'question'])\n", (750, 822), False, 'from langchain.prompts.prompt import PromptTemplate\n'), ((1003, 1071), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': 'self.model_name', 'temperature': 'self.temperature'}), '(model_name=self.model_name, temperature=self.temperature)\n', (1013, 1071), False, 'from langchain.chat_models import ChatOpenAI\n'), ((1089, 1290), 'langchain.chains.ConversationalRetrievalChain.from_llm', 'ConversationalRetrievalChain.from_llm', ([], {'llm': 'llm', 'retriever': 'self.retriever', 'verbose': '(True)', 'return_source_documents': '(True)', 'max_tokens_limit': '(4097)', 'combine_docs_chain_kwargs': "{'prompt': self.QA_PROMPT}"}), "(llm=llm, retriever=self.retriever,\n verbose=True, return_source_documents=True, max_tokens_limit=4097,\n combine_docs_chain_kwargs={'prompt': self.QA_PROMPT})\n", (1126, 1290), False, 'from langchain.chains import ConversationalRetrievalChain\n'), ((1675, 1743), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': 'self.model_name', 'temperature': 'self.temperature'}), '(model_name=self.model_name, temperature=self.temperature)\n', (1685, 1743), False, 'from langchain.chat_models import ChatOpenAI\n'), ((1761, 1859), 'langchain.chains.FlareChain.from_llm', 'FlareChain.from_llm', ([], {'llm': 'llm', 'retriever': 'self.retriever', 'max_generation_len': '(1000)', 'min_prob': '(0.3)'}), '(llm=llm, retriever=self.retriever, max_generation_len=\n 1000, min_prob=0.3)\n', (1780, 1859), False, 'from langchain.chains import FlareChain\n'), ((2234, 2255), 'langchain.callbacks.get_openai_callback', 'get_openai_callback', ([], {}), '()\n', (2253, 2255), False, 'from langchain.callbacks import get_openai_callback\n'), ((2305, 2384), 'streamlit.write', 'st.write', (['f"""###### Tokens used in this conversation : {cb.total_tokens} tokens"""'], {}), "(f'###### Tokens used in this conversation : {cb.total_tokens} tokens')\n", (2313, 2384), True, 'import streamlit as st\n')]
# Import Langchain dependencies from langchain.document_loaders import PyPDFLoader from langchain.indexes import VectorstoreIndexCreator from langchain.chains import RetrievalQA from langchain.embeddings import HuggingFaceEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter # Bring in streamlit for UI dev import streamlit as st # Bring in watsonx interface from watsonxlangchain import LangChainInterface # Setup credentials dictionary creds = { "apiKey": "", "url": "", } # Create LLM using langchain llm = LangChainInterface( credentials = creds, model = "meta-llama/llama-2-70b-chat", params = { "decoding_method": "sample", "max_new_tokens": 200, "temperature": 0.5 }, project_id="" ) # This function loads a PDF of your choosing @st.cache_resource def load_pdf(): # Update PDF name here to whatever you like pdf_name= "what is generative ai.pdf" loaders = [PyPDFLoader(pdf_name)] # Create index - aka vector database - aka chromadb index = VectorstoreIndexCreator( embedding = HuggingFaceEmbeddings(model_name="all-MiniLM-L12-v2"), text_splitter=RecursiveCharacterTextSplitter(chunk_size=100, chunk_overlap=0) ).from_loaders(loaders) # Return the vector database return index # Load er on up index = load_pdf() # Create a Q&A chain chain = RetrievalQA.from_chain_type( llm=llm, chain_type="stuff", retriever=index.vectorstore.as_retriever() ) # Setup the app title st.title("Ask watsonx") # Setup a session state message variable to hild all the old messages if "messages" not in st.session_state: st.session_state.messages = [] # Display all the historical messages for message in st.session_state.messages: st.chat_message(message["role"]) # Build a prompt input template to display the prompts prompt = st.chat_input("Pass Your Prompt Here") # If the user hits enter then if prompt: # Display the prompt st.chat_message("user").markdown(prompt) # Store the user prompt in state st.session_state.messages.append({"role":"user", "content":prompt}) # Send the prompt to the llm st.chat_message("assistant").markdown(response) # Store the LLM response in state st.session_state.messages.append( {"role": "assistant", "content": response} )
[ "langchain.document_loaders.PyPDFLoader", "langchain.embeddings.HuggingFaceEmbeddings", "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((539, 717), 'watsonxlangchain.LangChainInterface', 'LangChainInterface', ([], {'credentials': 'creds', 'model': '"""meta-llama/llama-2-70b-chat"""', 'params': "{'decoding_method': 'sample', 'max_new_tokens': 200, 'temperature': 0.5}", 'project_id': '""""""'}), "(credentials=creds, model='meta-llama/llama-2-70b-chat',\n params={'decoding_method': 'sample', 'max_new_tokens': 200,\n 'temperature': 0.5}, project_id='')\n", (557, 717), False, 'from watsonxlangchain import LangChainInterface\n'), ((1432, 1455), 'streamlit.title', 'st.title', (['"""Ask watsonx"""'], {}), "('Ask watsonx')\n", (1440, 1455), True, 'import streamlit as st\n'), ((1778, 1816), 'streamlit.chat_input', 'st.chat_input', (['"""Pass Your Prompt Here"""'], {}), "('Pass Your Prompt Here')\n", (1791, 1816), True, 'import streamlit as st\n'), ((1680, 1712), 'streamlit.chat_message', 'st.chat_message', (["message['role']"], {}), "(message['role'])\n", (1695, 1712), True, 'import streamlit as st\n'), ((1958, 2027), 'streamlit.session_state.messages.append', 'st.session_state.messages.append', (["{'role': 'user', 'content': prompt}"], {}), "({'role': 'user', 'content': prompt})\n", (1990, 2027), True, 'import streamlit as st\n'), ((2141, 2217), 'streamlit.session_state.messages.append', 'st.session_state.messages.append', (["{'role': 'assistant', 'content': response}"], {}), "({'role': 'assistant', 'content': response})\n", (2173, 2217), True, 'import streamlit as st\n'), ((908, 929), 'langchain.document_loaders.PyPDFLoader', 'PyPDFLoader', (['pdf_name'], {}), '(pdf_name)\n', (919, 929), False, 'from langchain.document_loaders import PyPDFLoader\n'), ((1882, 1905), 'streamlit.chat_message', 'st.chat_message', (['"""user"""'], {}), "('user')\n", (1897, 1905), True, 'import streamlit as st\n'), ((2057, 2085), 'streamlit.chat_message', 'st.chat_message', (['"""assistant"""'], {}), "('assistant')\n", (2072, 2085), True, 'import streamlit as st\n'), ((1032, 1085), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': '"""all-MiniLM-L12-v2"""'}), "(model_name='all-MiniLM-L12-v2')\n", (1053, 1085), False, 'from langchain.embeddings import HuggingFaceEmbeddings\n'), ((1103, 1166), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(100)', 'chunk_overlap': '(0)'}), '(chunk_size=100, chunk_overlap=0)\n', (1133, 1166), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n')]
import streamlit as st import langchain from langchain_community.document_loaders import RecursiveUrlLoader, TextLoader, JSONLoader from langchain_community.document_transformers import Html2TextTransformer from langchain.docstore.document import Document from langchain_community.embeddings.openai import OpenAIEmbeddings from langchain.text_splitter import RecursiveCharacterTextSplitter, CharacterTextSplitter from langchain.vectorstores.faiss import FAISS from langchain_community.vectorstores import Chroma from langchain_community.callbacks import StreamlitCallbackHandler from langchain.tools import tool from langchain.tools.json.tool import JsonSpec from langchain.agents import OpenAIFunctionsAgent, AgentExecutor, load_tools #from langchain_experimental.tools.python.tool import PythonREPLTool from langchain_community.agent_toolkits import JsonToolkit from langchain.tools.retriever import create_retriever_tool from langchain.agents.openai_functions_agent.agent_token_buffer_memory import (AgentTokenBufferMemory,) from langchain_community.chat_models.openai import ChatOpenAI from langchain.schema import SystemMessage, AIMessage, HumanMessage from langchain.prompts import MessagesPlaceholder from langsmith import Client from langchain_community.vectorstores.redis import Redis from langchain.vectorstores.redis import RedisText, RedisNum, RedisTag from langchain.memory import RedisChatMessageHistory from langchain.globals import set_llm_cache import os, openai, requests, json, zeep, datetime, pandas as pd from requests.auth import HTTPBasicAuth #from dotenv import load_dotenv, find_dotenv from zeep.wsse.username import UsernameToken #_ = load_dotenv(find_dotenv()) # read local .env file openai.api_key = os.getenv('OPENAI_API_KEY') model = "gpt-4-1106-preview" #model = "gpt-3.5-turbo-16k" #model = "gpt-3.5-turbo" embeddings = OpenAIEmbeddings() TENANT = 'wdmarketdesk_dpt1' WD_USER_ID = os.getenv('WD_USER_ID') WD_PWD = os.getenv('WD_PWD') WD_Worker_URL = "https://impl-services1.wd12.myworkday.com/ccx/service/customreport2/wdmarketdesk_dpt1/xjin-impl/Worker_Data?format=json" WD_Absence_URL = "https://impl-services1.wd12.myworkday.com/ccx/service/customreport2/wdmarketdesk_dpt1/xjin-impl/Worker_Absence_Data_2?format=json&employee_id=" WD_COMP_URL = "https://impl-services1.wd12.myworkday.com/ccx/service/customreport2/wdmarketdesk_dpt1/xjin-impl/Worker_Comp_Data_2?format=json&employee_id=" WD_STAFFING_WSDL_URL = "https://impl-services1.wd12.myworkday.com/ccx/service/wdmarketdesk_dpt1/Staffing/v41.1?wsdl" WD_HR_WSDL_URL = "https://impl-services1.wd12.myworkday.com/ccx/service/wdmarketdesk_dpt1/Human_Resources/v42.0?wsdl" basicAuth = HTTPBasicAuth(WD_USER_ID, WD_PWD) wd_hr_client = zeep.Client(WD_HR_WSDL_URL, wsse=UsernameToken(WD_USER_ID + '@' + TENANT, WD_PWD)) wd_staffing_client = zeep.Client(WD_STAFFING_WSDL_URL, wsse=UsernameToken(WD_USER_ID + '@' + TENANT, WD_PWD)) #client = Client() rds = Redis.from_existing_index( embeddings, index_name="worker_hr", redis_url="redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042", password="1iI48215k0GAEC3gzmpfPrXD2UDXYOYN", schema="worker_hr.yaml" ) retriever = None st.set_page_config( page_title="Ask HR", page_icon="🦜", layout="wide", initial_sidebar_state="collapsed", ) "# Ask HR🦜🔗" with st.form(key='login_form'): col1, col2, col3 = st.columns([1,2,1]) with col1: input_number = st.number_input(label='Enter Employee ID', min_value=0, max_value=1000000) login_submit = st.form_submit_button(label='Submit') def get_cache(): # construct cache implementation based on env var CACHE_TYPE = os.getenv("CACHE_TYPE", "semantic") REDIS_URL = os.getenv("REDIS_URL", "redis://localhost:6379") if CACHE_TYPE == "semantic": from langchain.cache import RedisSemanticCache print("Using semantic cache") return RedisSemanticCache( redis_url=REDIS_URL, embedding=embeddings, score_threshold=0.001, ) else: from langchain.cache import RedisCache from langchain_community.utilities import redis print("Using regular cache") return RedisCache(redis_= redis.get_client(REDIS_URL)) def get_raas_data(): response = requests.get(WD_Worker_URL, auth = basicAuth) responseJson = json.dumps(json.loads(response.content)) return responseJson def get_worker_data(): employee_id = '21082' worker_request_dict = { 'Worker_Reference': { 'ID': { 'type': 'Employee_ID', '_value_1': employee_id }, 'Descriptor': None }, 'Skip_Non_Existing_Instances': None, 'Ignore_Invalid_References': None } response = zeep_client.service.Get_Workers(worker_request_dict) return response.Response_Data def get_worker_name(employee_id=0): print("Trying to get Worker Name") is_emp_id = RedisNum("Employee_ID") == employee_id results = rds.similarity_search_with_score("", k=3, filter=is_emp_id if is_emp_id!=0 else None) return results[0][0].metadata['Employee_Legal_Full_Name'] #@st.cache_resource(ttl="4h") def initialize_retriever_redis(employee_id=0): """Initializes with all of worker data. If any information is not found, \ please use this tool as the default tool to look for the data needed. \ Do not try to get the same data more than 2 times. """ print("Initializing with Worker data in Redis", employee_id) rds2 = Redis.from_existing_index( embeddings, index_name="worker_hr", redis_url="redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042", password="1iI48215k0GAEC3gzmpfPrXD2UDXYOYN", schema="worker_hr.yaml" ) print("returning from Redis") return rds2.as_retriever(search_type="similarity", search_kwargs={"k": 10, "filter": employee_id if employee_id!=0 else None}) #@st.cache_resource(ttl="4h") def initialize_policies(): """Initializes with all policies data for GMS. If any information is not found, \ please say you don't know. Do not make up answers. \ For each answer, provide source in brackets. \ Do not repeat the same source information in the same line. In the final response, always replace word KPPRA with GMS """ print("Initializing with Policies") rds = Redis.from_existing_index( embeddings, index_name="hr_policies", redis_url="redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042", password="1iI48215k0GAEC3gzmpfPrXD2UDXYOYN", schema="hr_policies.yaml", ) print("returning from Redis") return rds.as_retriever(search_type="similarity", search_kwargs={"k": 5}) #@st.cache_resource(ttl="4h") def initialize_retriever(): """Initializes with all of worker data. If any information is not found, \ please use this tool as the default tool to look for the data needed. \ Do not try to get the same data more than 2 times. """ print("Initializing with all of worker data") txt = get_raas_data() # Split text text_splitter = CharacterTextSplitter() texts = text_splitter.split_text(txt) # Create multiple documents docs = [Document(page_content=t) for t in texts] embeddings = OpenAIEmbeddings() vectorstore = FAISS.from_documents(docs, embeddings) return vectorstore.as_retriever(search_kwargs={"k": 4}) @tool def init_json_loader(): """ Initializes with all of worker data. If any information is not found, \ please use this tool as the default tool to look for the data needed. \ """ json_data = get_worker_data() loader = JSONLoader(json_data, jq_schema='.messages[].content') extracted_data = loader.load() embeddings = OpenAIEmbeddings() vector_store = Chroma.from_documents(extracted_data, embeddings) return vector_store.as_retriever(search_kwargs={"k": 4}) #@st.cache_resource(ttl="4h") def get_vectorstore(txt): # Split text text_splitter = CharacterTextSplitter() texts = text_splitter.split_text(txt) # Create multiple documents docs = [Document(page_content=t) for t in texts] embeddings = OpenAIEmbeddings() vectorstore = FAISS.from_documents(docs, embeddings) return vectorstore.as_retriever(search_kwargs={"k": 4}) #@st.cache_resource(ttl="4h") def get_Absence_Data(Employee_ID: str) -> str: """Returns PTO or absence or time off data. \ Use this for any questions related to knowing about PTO or absence or time off data. \ Use Employee ID as the input, \ and this function will return data as string or JSON structure \ """ print(f"Getting absence data from Workday {WD_Absence_URL + str(input_number)}") response = requests.get(WD_Absence_URL + str(input_number), auth = basicAuth) responseJson = json.dumps(json.loads(response.content)) print(f"Absence Data: {responseJson}") return get_vectorstore(responseJson) @tool #@st.cache_resource(ttl="4h") def get_Comp_Data(text: str) -> str: """Returns Compensation or comp or salary data. \ Use this for any questions related to knowing about compensation or salary information. \ The input of text can not be empty string. \ and this function will return data as string or JSON structure \ """ print(f"Getting comp data from Workday {WD_COMP_URL + str(input_number)}") response = requests.get(WD_COMP_URL + str(input_number), auth = basicAuth) responseJson = json.dumps(json.loads(response.content)) print(f"Comp Data: {responseJson}") return get_vectorstore(responseJson) @tool def update_business_title(business_title: str) -> str: """Updates business title of the employee. \ Use this for any requests to update or change business title. \ The input can be an empty string or employee ID, \ and this function will return data as JSON structure \ """ employee_id = '21072' effective_date = '2023-11-12' Business_Process_Parameters = { 'Auto_Complete': True, 'Run_Now': True, 'Discard_On_Exit_Validation_Error': True } Change_Business_Title_Business_Process_Data = { 'Worker_Reference': { 'ID': { 'type': 'Employee_ID', '_value_1': employee_id }, 'Descriptor': None }, 'Change_Business_Title_Data': { 'Event_Effective_Date': effective_date, 'Proposed_Business_Title': business_title } } try: responseJson = wd_hr_client.service.Change_Business_Title(Business_Process_Parameters, Change_Business_Title_Business_Process_Data) except zeep.exceptions.Fault as error: responseJson = error except BaseException as error: responseJson = error return (responseJson) @tool def add_additional_job(Job_Requisition_ID: str, effective_date: datetime.date) -> str: """Adds additional job to the employee. \ Use this for any requests to add more jobs to existing employees. \ The input can be an empty string or employee ID, \ and this function will return data as JSON structure \ """ employee_id = '21072' Organization_Reference_ID = 'SUPERVISORY_Finance_and_Administration' #effective_date = '2023-11-17' #Job_Requisition_ID = 'R-00306' Event_Classification_Subcategory_ID = 'Add_Additional_Employee_Job_Secondment_Leave_of_Absence_Backfill' true = True Business_Process_Parameters = { 'Auto_Complete': True, 'Run_Now': True, 'Discard_On_Exit_Validation_Error': True } Add_Additional_Job_Data = { 'Employee_Reference': { 'ID': { 'type': 'Employee_ID', '_value_1': employee_id }, 'Descriptor': None }, 'Organization_Reference': { 'ID': { 'type': 'Organization_Reference_ID', '_value_1': Organization_Reference_ID } }, 'Job_Requisition_Reference': { 'ID': { 'type': 'Job_Requisition_ID', '_value_1': Job_Requisition_ID } }, 'Add_Additional_Job_Event_Data': { 'Additional_Job_Reason_Reference': { 'ID': { 'type': 'Event_Classification_Subcategory_ID', '_value_1': Event_Classification_Subcategory_ID } }, 'Position_Details': { } }, 'Event_Effective_Date': effective_date } try: responseJson = wd_staffing_client.service.Add_Additional_Job(Business_Process_Parameters, Add_Additional_Job_Data) except zeep.exceptions.Fault as error: responseJson = error return (responseJson) tool = create_retriever_tool( initialize_retriever_redis(input_number), "Ask_HR", """Initializes with all of worker data. If any information is not found, \ please use this tool as the default tool to look for the data needed. \ """ ) policies = create_retriever_tool( initialize_policies(), "Policies", """Initializes with all policies data for GMS. If any information is not found, \ please say you don't know. Do not make up answers. \ For each answer, provide source in brackets. \ Do not repeat the same source information in the same line. \ Do not call this tool more than 2 times. \ In the final response, always replace word KPPRA with GMS """ ) absence = create_retriever_tool( get_Absence_Data(""), "get_Absence_Data", """Returns PTO or absence or time off data. \ Use this for any questions related to knowing about PTO or absence or time off data. \ The input can be an empty string or employee ID, \ and this function will return data as string or JSON structure \ """) comp = create_retriever_tool( get_Comp_Data(""), "get_Comp_Data", """Returns Compensation or comp or salary data. \ Use this for any questions related to knowing about compensation or salary information. \ If what is needed is not found, please use Ask_HR tool as the default tool instead. \ """) tools = [tool, policies, absence, comp, update_business_title, add_additional_job] try: set_llm_cache(get_cache()) print("*** Retrieved cache ***") except Exception as e: print("*** Could not retrieve cache ***") print(e) pass chat_llm = ChatOpenAI(temperature=0, streaming=True, model=model) message = SystemMessage( content=( """ You are HR Bot, an automated service to handle HR requests for an organization. \ Please "DO NOT" hallucinate. You first greet the employee, then ask for what task he is looking to perform. \ If the request is related to PTO or Absences, Check to see if he has enough balance of hours by substracting requested hours from available balance. \ If he has enough balance, ask if he wants to submit for approval. \ If the balance is not enough let him know to adjust the request. \ You wait to collect all the details, then summarize it and check for a final \ time if the employee wants to change anything. \ Make sure to clarify all options, days and hours to uniquely identify the PTO from the options.\ You respond in a short, very conversational friendly style. \ When the user is ready to submit the request, Do not respond with plain human language but \ respond with type of PTO, hours and comments in JSON format. \ For type: Identify what type of PTO is being requested for submission. Use 'Reference_ID' value of the PTO plan for \ final JSON format. Do not show 'Reference_ID' for any intermediate human friendly updates or messages \ For hours: If he has enough balance, extract just numeric value of how many hours are being requested or submitted. \ If the balance is not enough consider 0. \ For comments: Extract the reason why employee is requesting or submitting the PTO \ The PTO options with available balances in hours includes as below in JSON format \ If the questions is on Policies and Rules provide very brief summary \ For each answer, if available, provide source including page number in brackets. \ In the final response, always replace word KPPRA with GMS """ ) ) prompt = OpenAIFunctionsAgent.create_prompt( system_message=message, extra_prompt_messages=[MessagesPlaceholder(variable_name="history")], ) agent = OpenAIFunctionsAgent(llm=chat_llm, tools=tools, prompt=prompt) agent_executor = AgentExecutor( agent=agent, tools=tools, verbose=True, return_intermediate_steps=True, ) memory = AgentTokenBufferMemory(llm=chat_llm) #redis_memory = RedisChatMessageHistory() starter_message = f"Hello and Welcome. I am here to help you with your HR needs!!" if "messages" not in st.session_state or st.sidebar.button("Clear message history") or login_submit: starter_message = f"Hello and Welcome. I am here to help you with your HR needs!!" st.session_state["messages"] = [AIMessage(content=starter_message)] #def send_feedback(run_id, score): # client.create_feedback(run_id, "user_score", score=score) for msg in st.session_state.messages: if isinstance(msg, AIMessage): st.chat_message("assistant").write(msg.content) elif isinstance(msg, HumanMessage): st.chat_message("user").write(msg.content) memory.chat_memory.add_message(msg) if prompt := st.chat_input(placeholder=starter_message): st.chat_message("user").write(prompt) with st.chat_message("assistant"): st_callback = StreamlitCallbackHandler(st.container()) response = agent_executor( {"input": prompt, "history": st.session_state.messages, "employee_id": input_number}, callbacks=[st_callback], include_run_info=True, ) print("returned from agent executor") st.session_state.messages.append(AIMessage(content=response["output"])) st.write(response["output"]) memory.save_context({"input": prompt}, response) st.session_state["messages"] = memory.buffer run_id = response["__run"].run_id #col_blank, col_text, col1, col2 = st.columns([10, 2, 1, 1]) #with col_text: # st.text('Feedback:') #with col1: # st.button('👍', on_click=send_feedback, args=(run_id, 1)) #with col2: # st.button('👎', on_click=send_feedback, args=(run_id, 0))
[ "langchain.agents.openai_functions_agent.agent_token_buffer_memory.AgentTokenBufferMemory", "langchain.text_splitter.CharacterTextSplitter", "langchain.agents.AgentExecutor", "langchain_community.embeddings.openai.OpenAIEmbeddings", "langchain_community.vectorstores.Chroma.from_documents", "langchain.docstore.document.Document", "langchain.vectorstores.redis.RedisNum", "langchain.prompts.MessagesPlaceholder", "langchain_community.utilities.redis.get_client", "langchain_community.vectorstores.redis.Redis.from_existing_index", "langchain.cache.RedisSemanticCache", "langchain.schema.AIMessage", "langchain.schema.SystemMessage", "langchain.agents.OpenAIFunctionsAgent", "langchain.vectorstores.faiss.FAISS.from_documents", "langchain_community.document_loaders.JSONLoader", "langchain_community.chat_models.openai.ChatOpenAI" ]
[((1732, 1759), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (1741, 1759), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((1857, 1875), 'langchain_community.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (1873, 1875), False, 'from langchain_community.embeddings.openai import OpenAIEmbeddings\n'), ((1919, 1942), 'os.getenv', 'os.getenv', (['"""WD_USER_ID"""'], {}), "('WD_USER_ID')\n", (1928, 1942), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((1952, 1971), 'os.getenv', 'os.getenv', (['"""WD_PWD"""'], {}), "('WD_PWD')\n", (1961, 1971), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((2676, 2709), 'requests.auth.HTTPBasicAuth', 'HTTPBasicAuth', (['WD_USER_ID', 'WD_PWD'], {}), '(WD_USER_ID, WD_PWD)\n', (2689, 2709), False, 'from requests.auth import HTTPBasicAuth\n'), ((2945, 3167), 'langchain_community.vectorstores.redis.Redis.from_existing_index', 'Redis.from_existing_index', (['embeddings'], {'index_name': '"""worker_hr"""', 'redis_url': '"""redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042"""', 'password': '"""1iI48215k0GAEC3gzmpfPrXD2UDXYOYN"""', 'schema': '"""worker_hr.yaml"""'}), "(embeddings, index_name='worker_hr', redis_url=\n 'redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042',\n password='1iI48215k0GAEC3gzmpfPrXD2UDXYOYN', schema='worker_hr.yaml')\n", (2970, 3167), False, 'from langchain_community.vectorstores.redis import Redis\n'), ((3224, 3332), 'streamlit.set_page_config', 'st.set_page_config', ([], {'page_title': '"""Ask HR"""', 'page_icon': '"""🦜"""', 'layout': '"""wide"""', 'initial_sidebar_state': '"""collapsed"""'}), "(page_title='Ask HR', page_icon='🦜', layout='wide',\n initial_sidebar_state='collapsed')\n", (3242, 3332), True, 'import streamlit as st\n'), ((14656, 14710), 'langchain_community.chat_models.openai.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'streaming': '(True)', 'model': 'model'}), '(temperature=0, streaming=True, model=model)\n', (14666, 14710), False, 'from langchain_community.chat_models.openai import ChatOpenAI\n'), ((14722, 16688), 'langchain.schema.SystemMessage', 'SystemMessage', ([], {'content': '"""\n You are HR Bot, an automated service to handle HR requests for an organization. Please "DO NOT" hallucinate.\n You first greet the employee, then ask for what task he is looking to perform. If the request is related to PTO or Absences, Check to see if he has enough balance of hours by substracting requested hours from available balance. If he has enough balance, ask if he wants to submit for approval. If the balance is not enough let him know to adjust the request. You wait to collect all the details, then summarize it and check for a final time if the employee wants to change anything. Make sure to clarify all options, days and hours to uniquely identify the PTO from the options. You respond in a short, very conversational friendly style. \n When the user is ready to submit the request, Do not respond with plain human language but respond with type of PTO, hours and comments in JSON format. \n For type: Identify what type of PTO is being requested for submission. Use \'Reference_ID\' value of the PTO plan for final JSON format. Do not show \'Reference_ID\' for any intermediate human friendly updates or messages For hours: If he has enough balance, extract just numeric value of how many hours are being requested or submitted. If the balance is not enough consider 0. For comments: Extract the reason why employee is requesting or submitting the PTO \n The PTO options with available balances in hours includes as below in JSON format \n If the questions is on Policies and Rules provide very brief summary For each answer, if available, provide source including page number in brackets. In the final response, always replace word KPPRA with GMS\n """'}), '(content=\n """\n You are HR Bot, an automated service to handle HR requests for an organization. Please "DO NOT" hallucinate.\n You first greet the employee, then ask for what task he is looking to perform. If the request is related to PTO or Absences, Check to see if he has enough balance of hours by substracting requested hours from available balance. If he has enough balance, ask if he wants to submit for approval. If the balance is not enough let him know to adjust the request. You wait to collect all the details, then summarize it and check for a final time if the employee wants to change anything. Make sure to clarify all options, days and hours to uniquely identify the PTO from the options. You respond in a short, very conversational friendly style. \n When the user is ready to submit the request, Do not respond with plain human language but respond with type of PTO, hours and comments in JSON format. \n For type: Identify what type of PTO is being requested for submission. Use \'Reference_ID\' value of the PTO plan for final JSON format. Do not show \'Reference_ID\' for any intermediate human friendly updates or messages For hours: If he has enough balance, extract just numeric value of how many hours are being requested or submitted. If the balance is not enough consider 0. For comments: Extract the reason why employee is requesting or submitting the PTO \n The PTO options with available balances in hours includes as below in JSON format \n If the questions is on Policies and Rules provide very brief summary For each answer, if available, provide source including page number in brackets. In the final response, always replace word KPPRA with GMS\n """\n )\n', (14735, 16688), False, 'from langchain.schema import SystemMessage, AIMessage, HumanMessage\n'), ((16898, 16960), 'langchain.agents.OpenAIFunctionsAgent', 'OpenAIFunctionsAgent', ([], {'llm': 'chat_llm', 'tools': 'tools', 'prompt': 'prompt'}), '(llm=chat_llm, tools=tools, prompt=prompt)\n', (16918, 16960), False, 'from langchain.agents import OpenAIFunctionsAgent, AgentExecutor, load_tools\n'), ((16979, 17068), 'langchain.agents.AgentExecutor', 'AgentExecutor', ([], {'agent': 'agent', 'tools': 'tools', 'verbose': '(True)', 'return_intermediate_steps': '(True)'}), '(agent=agent, tools=tools, verbose=True,\n return_intermediate_steps=True)\n', (16992, 17068), False, 'from langchain.agents import OpenAIFunctionsAgent, AgentExecutor, load_tools\n'), ((17094, 17130), 'langchain.agents.openai_functions_agent.agent_token_buffer_memory.AgentTokenBufferMemory', 'AgentTokenBufferMemory', ([], {'llm': 'chat_llm'}), '(llm=chat_llm)\n', (17116, 17130), False, 'from langchain.agents.openai_functions_agent.agent_token_buffer_memory import AgentTokenBufferMemory\n'), ((3368, 3393), 'streamlit.form', 'st.form', ([], {'key': '"""login_form"""'}), "(key='login_form')\n", (3375, 3393), True, 'import streamlit as st\n'), ((3418, 3439), 'streamlit.columns', 'st.columns', (['[1, 2, 1]'], {}), '([1, 2, 1])\n', (3428, 3439), True, 'import streamlit as st\n'), ((3570, 3607), 'streamlit.form_submit_button', 'st.form_submit_button', ([], {'label': '"""Submit"""'}), "(label='Submit')\n", (3591, 3607), True, 'import streamlit as st\n'), ((3698, 3733), 'os.getenv', 'os.getenv', (['"""CACHE_TYPE"""', '"""semantic"""'], {}), "('CACHE_TYPE', 'semantic')\n", (3707, 3733), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((3750, 3798), 'os.getenv', 'os.getenv', (['"""REDIS_URL"""', '"""redis://localhost:6379"""'], {}), "('REDIS_URL', 'redis://localhost:6379')\n", (3759, 3798), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((4326, 4369), 'requests.get', 'requests.get', (['WD_Worker_URL'], {'auth': 'basicAuth'}), '(WD_Worker_URL, auth=basicAuth)\n', (4338, 4369), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((5599, 5821), 'langchain_community.vectorstores.redis.Redis.from_existing_index', 'Redis.from_existing_index', (['embeddings'], {'index_name': '"""worker_hr"""', 'redis_url': '"""redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042"""', 'password': '"""1iI48215k0GAEC3gzmpfPrXD2UDXYOYN"""', 'schema': '"""worker_hr.yaml"""'}), "(embeddings, index_name='worker_hr', redis_url=\n 'redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042',\n password='1iI48215k0GAEC3gzmpfPrXD2UDXYOYN', schema='worker_hr.yaml')\n", (5624, 5821), False, 'from langchain_community.vectorstores.redis import Redis\n'), ((6490, 6716), 'langchain_community.vectorstores.redis.Redis.from_existing_index', 'Redis.from_existing_index', (['embeddings'], {'index_name': '"""hr_policies"""', 'redis_url': '"""redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042"""', 'password': '"""1iI48215k0GAEC3gzmpfPrXD2UDXYOYN"""', 'schema': '"""hr_policies.yaml"""'}), "(embeddings, index_name='hr_policies', redis_url=\n 'redis://redis-10042.c280.us-central1-2.gce.cloud.redislabs.com:10042',\n password='1iI48215k0GAEC3gzmpfPrXD2UDXYOYN', schema='hr_policies.yaml')\n", (6515, 6716), False, 'from langchain_community.vectorstores.redis import Redis\n'), ((7258, 7281), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {}), '()\n', (7279, 7281), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter, CharacterTextSplitter\n'), ((7431, 7449), 'langchain_community.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (7447, 7449), False, 'from langchain_community.embeddings.openai import OpenAIEmbeddings\n'), ((7468, 7506), 'langchain.vectorstores.faiss.FAISS.from_documents', 'FAISS.from_documents', (['docs', 'embeddings'], {}), '(docs, embeddings)\n', (7488, 7506), False, 'from langchain.vectorstores.faiss import FAISS\n'), ((7813, 7867), 'langchain_community.document_loaders.JSONLoader', 'JSONLoader', (['json_data'], {'jq_schema': '""".messages[].content"""'}), "(json_data, jq_schema='.messages[].content')\n", (7823, 7867), False, 'from langchain_community.document_loaders import RecursiveUrlLoader, TextLoader, JSONLoader\n'), ((7920, 7938), 'langchain_community.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (7936, 7938), False, 'from langchain_community.embeddings.openai import OpenAIEmbeddings\n'), ((7958, 8007), 'langchain_community.vectorstores.Chroma.from_documents', 'Chroma.from_documents', (['extracted_data', 'embeddings'], {}), '(extracted_data, embeddings)\n', (7979, 8007), False, 'from langchain_community.vectorstores import Chroma\n'), ((8168, 8191), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {}), '()\n', (8189, 8191), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter, CharacterTextSplitter\n'), ((8341, 8359), 'langchain_community.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (8357, 8359), False, 'from langchain_community.embeddings.openai import OpenAIEmbeddings\n'), ((8378, 8416), 'langchain.vectorstores.faiss.FAISS.from_documents', 'FAISS.from_documents', (['docs', 'embeddings'], {}), '(docs, embeddings)\n', (8398, 8416), False, 'from langchain.vectorstores.faiss import FAISS\n'), ((17299, 17341), 'streamlit.sidebar.button', 'st.sidebar.button', (['"""Clear message history"""'], {}), "('Clear message history')\n", (17316, 17341), True, 'import streamlit as st\n'), ((17892, 17934), 'streamlit.chat_input', 'st.chat_input', ([], {'placeholder': 'starter_message'}), '(placeholder=starter_message)\n', (17905, 17934), True, 'import streamlit as st\n'), ((2758, 2806), 'zeep.wsse.username.UsernameToken', 'UsernameToken', (["(WD_USER_ID + '@' + TENANT)", 'WD_PWD'], {}), "(WD_USER_ID + '@' + TENANT, WD_PWD)\n", (2771, 2806), False, 'from zeep.wsse.username import UsernameToken\n'), ((2869, 2917), 'zeep.wsse.username.UsernameToken', 'UsernameToken', (["(WD_USER_ID + '@' + TENANT)", 'WD_PWD'], {}), "(WD_USER_ID + '@' + TENANT, WD_PWD)\n", (2882, 2917), False, 'from zeep.wsse.username import UsernameToken\n'), ((3476, 3550), 'streamlit.number_input', 'st.number_input', ([], {'label': '"""Enter Employee ID"""', 'min_value': '(0)', 'max_value': '(1000000)'}), "(label='Enter Employee ID', min_value=0, max_value=1000000)\n", (3491, 3550), True, 'import streamlit as st\n'), ((3940, 4028), 'langchain.cache.RedisSemanticCache', 'RedisSemanticCache', ([], {'redis_url': 'REDIS_URL', 'embedding': 'embeddings', 'score_threshold': '(0.001)'}), '(redis_url=REDIS_URL, embedding=embeddings,\n score_threshold=0.001)\n', (3958, 4028), False, 'from langchain.cache import RedisSemanticCache\n'), ((4402, 4430), 'json.loads', 'json.loads', (['response.content'], {}), '(response.content)\n', (4412, 4430), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((5026, 5049), 'langchain.vectorstores.redis.RedisNum', 'RedisNum', (['"""Employee_ID"""'], {}), "('Employee_ID')\n", (5034, 5049), False, 'from langchain.vectorstores.redis import RedisText, RedisNum, RedisTag\n'), ((7368, 7392), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': 't'}), '(page_content=t)\n', (7376, 7392), False, 'from langchain.docstore.document import Document\n'), ((8278, 8302), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': 't'}), '(page_content=t)\n', (8286, 8302), False, 'from langchain.docstore.document import Document\n'), ((9006, 9034), 'json.loads', 'json.loads', (['response.content'], {}), '(response.content)\n', (9016, 9034), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((9660, 9688), 'json.loads', 'json.loads', (['response.content'], {}), '(response.content)\n', (9670, 9688), False, 'import os, openai, requests, json, zeep, datetime, pandas as pd\n'), ((17482, 17516), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'starter_message'}), '(content=starter_message)\n', (17491, 17516), False, 'from langchain.schema import SystemMessage, AIMessage, HumanMessage\n'), ((17987, 18015), 'streamlit.chat_message', 'st.chat_message', (['"""assistant"""'], {}), "('assistant')\n", (18002, 18015), True, 'import streamlit as st\n'), ((18429, 18457), 'streamlit.write', 'st.write', (["response['output']"], {}), "(response['output'])\n", (18437, 18457), True, 'import streamlit as st\n'), ((16840, 16884), 'langchain.prompts.MessagesPlaceholder', 'MessagesPlaceholder', ([], {'variable_name': '"""history"""'}), "(variable_name='history')\n", (16859, 16884), False, 'from langchain.prompts import MessagesPlaceholder\n'), ((17940, 17963), 'streamlit.chat_message', 'st.chat_message', (['"""user"""'], {}), "('user')\n", (17955, 17963), True, 'import streamlit as st\n'), ((18064, 18078), 'streamlit.container', 'st.container', ([], {}), '()\n', (18076, 18078), True, 'import streamlit as st\n'), ((18382, 18419), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': "response['output']"}), "(content=response['output'])\n", (18391, 18419), False, 'from langchain.schema import SystemMessage, AIMessage, HumanMessage\n'), ((4256, 4283), 'langchain_community.utilities.redis.get_client', 'redis.get_client', (['REDIS_URL'], {}), '(REDIS_URL)\n', (4272, 4283), False, 'from langchain_community.utilities import redis\n'), ((17699, 17727), 'streamlit.chat_message', 'st.chat_message', (['"""assistant"""'], {}), "('assistant')\n", (17714, 17727), True, 'import streamlit as st\n'), ((17795, 17818), 'streamlit.chat_message', 'st.chat_message', (['"""user"""'], {}), "('user')\n", (17810, 17818), True, 'import streamlit as st\n')]
# Using flask to make an api # import necessary libraries and functions from flask import Flask, jsonify, request, render_template from pydantic import BaseModel from ast import literal_eval import os import openai openai.api_key = os.getenv("OPENAI_API_KEY") import langchain from langchain.vectorstores import FAISS from langchain.embeddings.openai import OpenAIEmbeddings # creating a Flask app app = Flask(__name__) # on the terminal type: curl http://127.0.0.1:5000/ # returns hello world when we use GET. # returns the data that we send when we use POST. @app.route('/related', methods = ['GET', 'POST']) def home(): if(request.method == 'GET'): data = "its up" return jsonify({'data': data}) if(request.method == 'POST'): temp= request.get_json() print(temp) vector = temp["ve"] ticker = temp["ticker"] embeddings = OpenAIEmbeddings(openai.api_key) print(str(ticker)) new_db = FAISS.load_local(str(ticker), embeddings) ar = literal_eval(vector) docs_and_scores = new_db.similarity_search_by_vector(ar) size = len(docs_and_scores) data = "" for i in range(0,size): data+=(docs_and_scores[i].page_content + "\n\n") print(data) return jsonify({'data': data}) def relatedwords(ticker, vembeddings): embeddings = OpenAIEmbeddings() print(str(ticker)) new_db = FAISS.load_local(ticker, embeddings) # ar = literal_eval(vector) docs_and_scores = new_db.similarity_search_by_vector(vembeddings) size = len(docs_and_scores) data = "" for i in range(0,size): data+=(docs_and_scores[i].page_content + "\n\n") print(data) return data def completion(context): completion = openai.ChatCompletion.create( model="gpt-3.5-turbo", temperature = 0, messages=[ { "role": "system", "content": "You are a financial assistant that VERY ACCURATELY and CONCISELY answers queries using publicly filled annual reports of publicly traded companies. Use the CONTEXT provided to form your answer. Keep your answer under 5 sentences. Be accurate, helpful, concise, and clear, while performing accurate mathematical calculations." }, { "role": "user", "content": context } ] ) return (completion.choices[0].message["content"]) def embeddingsgenerate(embeddingsof): embeddings = openai.Embedding.create( model="text-embedding-ada-002", input=embeddingsof ) return embeddings["data"][0]["embedding"] @app.route('/generate_embeddings', methods= ['GET', 'POST']) def get_embeddings(): if(request.method == 'GET'): data = "its up" return jsonify({'data': data}) if(request.method == 'POST'): data = request.get_json() texts = data['texts'] embeddings = openai.Encodings.create(texts) return jsonify({'embeddings': embeddings}) @app.route('/<path_name>',methods=['GET', 'POST']) def answer_question(path_name): urlforprice = ("https://s.tradingview.com/embed-widget/symbol-info/?locale=en&colorTheme=dark&symbol=BSE:"+path_name) if request.method == 'POST': # User submitted a question, so process it and get an answer from OpenAI question = request.form['question'] # openai_api_key = request.form['openai_api_key'] prompt = f"{question}" embeddings_recieved = embeddingsgenerate(question) context = relatedwords(path_name, embeddings_recieved) questionforgpt = "CONTEXT\n\n" + context + "\nQUESTION\n" + prompt answerfromgpt = completion(questionforgpt) print(answerfromgpt) return render_template('chat.html', path_name=path_name, questionasked=prompt, urlforprice=urlforprice, embeddings_recieved = context, answer_received = answerfromgpt) else: answer = "not yet used" # Render the answer on an HTML template return render_template('chat.html', path_name=path_name, answer=answer,urlforprice=urlforprice) @app.route('/',methods=['GET']) def homepage(): return render_template('home.html') # driver function if __name__ == '__main__': app.run(debug = True)
[ "langchain.vectorstores.FAISS.load_local", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((235, 262), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (244, 262), False, 'import os\n'), ((412, 427), 'flask.Flask', 'Flask', (['__name__'], {}), '(__name__)\n', (417, 427), False, 'from flask import Flask, jsonify, request, render_template\n'), ((1396, 1414), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (1412, 1414), False, 'from langchain.embeddings.openai import OpenAIEmbeddings\n'), ((1451, 1487), 'langchain.vectorstores.FAISS.load_local', 'FAISS.load_local', (['ticker', 'embeddings'], {}), '(ticker, embeddings)\n', (1467, 1487), False, 'from langchain.vectorstores import FAISS\n'), ((1807, 2295), 'openai.ChatCompletion.create', 'openai.ChatCompletion.create', ([], {'model': '"""gpt-3.5-turbo"""', 'temperature': '(0)', 'messages': "[{'role': 'system', 'content':\n 'You are a financial assistant that VERY ACCURATELY and CONCISELY answers queries using publicly filled annual reports of publicly traded companies. Use the CONTEXT provided to form your answer. Keep your answer under 5 sentences. Be accurate, helpful, concise, and clear, while performing accurate mathematical calculations.'\n }, {'role': 'user', 'content': context}]"}), "(model='gpt-3.5-turbo', temperature=0, messages\n =[{'role': 'system', 'content':\n 'You are a financial assistant that VERY ACCURATELY and CONCISELY answers queries using publicly filled annual reports of publicly traded companies. Use the CONTEXT provided to form your answer. Keep your answer under 5 sentences. Be accurate, helpful, concise, and clear, while performing accurate mathematical calculations.'\n }, {'role': 'user', 'content': context}])\n", (1835, 2295), False, 'import openai\n'), ((2502, 2577), 'openai.Embedding.create', 'openai.Embedding.create', ([], {'model': '"""text-embedding-ada-002"""', 'input': 'embeddingsof'}), "(model='text-embedding-ada-002', input=embeddingsof)\n", (2525, 2577), False, 'import openai\n'), ((4211, 4239), 'flask.render_template', 'render_template', (['"""home.html"""'], {}), "('home.html')\n", (4226, 4239), False, 'from flask import Flask, jsonify, request, render_template\n'), ((709, 732), 'flask.jsonify', 'jsonify', (["{'data': data}"], {}), "({'data': data})\n", (716, 732), False, 'from flask import Flask, jsonify, request, render_template\n'), ((785, 803), 'flask.request.get_json', 'request.get_json', ([], {}), '()\n', (801, 803), False, 'from flask import Flask, jsonify, request, render_template\n'), ((906, 938), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', (['openai.api_key'], {}), '(openai.api_key)\n', (922, 938), False, 'from langchain.embeddings.openai import OpenAIEmbeddings\n'), ((1038, 1058), 'ast.literal_eval', 'literal_eval', (['vector'], {}), '(vector)\n', (1050, 1058), False, 'from ast import literal_eval\n'), ((1310, 1333), 'flask.jsonify', 'jsonify', (["{'data': data}"], {}), "({'data': data})\n", (1317, 1333), False, 'from flask import Flask, jsonify, request, render_template\n'), ((2806, 2829), 'flask.jsonify', 'jsonify', (["{'data': data}"], {}), "({'data': data})\n", (2813, 2829), False, 'from flask import Flask, jsonify, request, render_template\n'), ((2879, 2897), 'flask.request.get_json', 'request.get_json', ([], {}), '()\n', (2895, 2897), False, 'from flask import Flask, jsonify, request, render_template\n'), ((2949, 2979), 'openai.Encodings.create', 'openai.Encodings.create', (['texts'], {}), '(texts)\n', (2972, 2979), False, 'import openai\n'), ((2995, 3030), 'flask.jsonify', 'jsonify', (["{'embeddings': embeddings}"], {}), "({'embeddings': embeddings})\n", (3002, 3030), False, 'from flask import Flask, jsonify, request, render_template\n'), ((3792, 3957), 'flask.render_template', 'render_template', (['"""chat.html"""'], {'path_name': 'path_name', 'questionasked': 'prompt', 'urlforprice': 'urlforprice', 'embeddings_recieved': 'context', 'answer_received': 'answerfromgpt'}), "('chat.html', path_name=path_name, questionasked=prompt,\n urlforprice=urlforprice, embeddings_recieved=context, answer_received=\n answerfromgpt)\n", (3807, 3957), False, 'from flask import Flask, jsonify, request, render_template\n'), ((4059, 4152), 'flask.render_template', 'render_template', (['"""chat.html"""'], {'path_name': 'path_name', 'answer': 'answer', 'urlforprice': 'urlforprice'}), "('chat.html', path_name=path_name, answer=answer,\n urlforprice=urlforprice)\n", (4074, 4152), False, 'from flask import Flask, jsonify, request, render_template\n')]
# -*- coding: utf-8 -*- import random import streamlit as st from langchain.llms import OpenAI from langchain.text_splitter import RecursiveCharacterTextSplitter #from langchain.text_splitter import CharacterTextSplitter from langchain.embeddings import HuggingFaceEmbeddings from langchain.vectorstores import FAISS import openai # from playsound import playsound from gtts import gTTS from PyPDF2 import PdfReader from utils import text_to_docs from langchain import PromptTemplate, LLMChain #import os #from io import StringIO #from langchain.embeddings.openai import OpenAIEmbeddings #from langchain.chains.summarize import load_summarize_chain #import os #import pyaudio #import wave #import langchain #from langchain.document_loaders import UnstructuredPDFLoader #from io import BytesIO # import streamlit.components.v1 as components #from st_custom_components import st_audiorec, text_to_docs #import sounddevice as sd #from scipy.io.wavfile import write from usellm import Message, Options, UseLLM llm = OpenAI(model_name='text-davinci-003', temperature=0.2, max_tokens=512, openai_api_key=st.secrets["api_key"]) def record_audio(filename): duration=5 sample_rate = 44100 # Record audio audio_data = sd.rec(int(duration * sample_rate), samplerate=sample_rate, channels=6) st.write('Recording started. Speak into the microphone...') # Wait for recording to complete sd.wait() # Save the recorded audio to a WAV file write(filename, sample_rate, audio_data) st.write(f'Recording saved as {filename}') def transcribe_audio(filename): with open(filename, "rb") as audio_file: transcript = openai.Audio.transcribe("whisper-1", filename) return transcript["text"] st.title("Suspicious Activity Reporting") st.subheader('Evidence Processor') model_name = "sentence-transformers/all-MiniLM-L6-v2" @st.cache_data def usellm(prompt): service = UseLLM(service_url="https://usellm.org/api/llm") messages = [ Message(role="system", content="You are a fraud analyst"), Message(role="user", content=f"{prompt}"), ] options = Options(messages=messages) response = service.chat(options) return response.content @st.cache_resource def embed(model_name): hf_embeddings = HuggingFaceEmbeddings(model_name=model_name) return hf_embeddings # @st.cache_resource # def chain(_llm): # chain = load_summarize_chain(llm=_llm, chain_type="map_reduce") # chain = load_summarize_chain(llm=llm, chain_type="map_reduce") hf_embeddings = embed(model_name) # File Upload file = st.file_uploader("Upload a file") st.write(file) text_splitter = RecursiveCharacterTextSplitter( # Set a really small chunk size, just to show. chunk_size = 1500, chunk_overlap = 100, length_function = len, separators=["\n\n", "\n", " ", ""] ) #text_splitter = CharacterTextSplitter.from_tiktoken_encoder(chunk_size=100, chunk_overlap=0) #texts = '' @st.cache_data def embedding_store(file): # save file pdf_reader = PdfReader(file) text = "" for page in pdf_reader.pages: text += page.extract_text() #st.write(text) texts = text_splitter.split_text(text) docs = text_to_docs(texts) #st.write(texts) docsearch = FAISS.from_documents(docs, hf_embeddings) return docs, docsearch # Submit Button if st.button("Submit"): if file is not None: # File handling logic st.write("File Uploaded...") _, docsearch = embedding_store(file) queries ="Any fraud encountered in the passed document?\ if any." contexts = docsearch.similarity_search(queries, k=1) prompts = f"Give concise answer to the below questions as truthfully as possible as per given context only,\n\n\ Context: {contexts}\n\ Response (in readable bullet points): " response = usellm(prompts) # Print the assistant's response # st.subheader("Few informative points in the uploaded document are as follows:") # st.write(response) #st.write("Uploaded File Contents:") if file is not None: docs, docsearch = embedding_store(file) #docs = text_to_docs(texts) #st.write(docs) #summary = chain.run(docs) #st.write(summary) #st.header("Speech and Text Input") # Speech Input # st.subheader("Speech Input") # wav_audio_data = st_audiorec() # if wav_audio_data is not None: # st.audio(wav_audio_data, format='audio/wav') # audio_text = transcribe_audio(wav_audio_data) # st.write(audio_text) # if st.button("Voice Input"): # rand = random.randint(1, 10000)*random.randint(10001,20000) # audio_filename = f"recorded_audio{rand}.wav" # record_audio(filename=audio_filename) # aud = transcribe_audio(audio_filename) # st.write(aud) # Text Input st.subheader("Ask Questions") query = st.text_input('your queries will go here...') def LLM_Response(): llm_chain = LLMChain(prompt=prompt, llm=llm) response = llm_chain.run({"query":query, "context":context}) return response if query: # Text input handling logic #st.write("Text Input:") #st.write(text_input) context = docsearch.similarity_search(query, k=5) prompt = f"Act as a financial analyst and give concise answer to below Question as truthfully as possible, with given Context\n\n\ Question: {query}\n\ Context: {context}\n\ Response: " #prompt = PromptTemplate(template=prompt, input_variables=["query", "context"]) response = usellm(prompt) #LLM_Response() st.write(response) # language = 'en' # Create a gTTS object # tts = gTTS(text=response, lang=language) # Save the audio file # rand = random.randint(1, 10000)*random.randint(10001,20000) # audio_file = f'output{rand}.mp3' # tts.save(audio_file) # playsound(audio_file)
[ "langchain.LLMChain", "langchain.embeddings.HuggingFaceEmbeddings", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.llms.OpenAI", "langchain.vectorstores.FAISS.from_documents" ]
[((1047, 1159), 'langchain.llms.OpenAI', 'OpenAI', ([], {'model_name': '"""text-davinci-003"""', 'temperature': '(0.2)', 'max_tokens': '(512)', 'openai_api_key': "st.secrets['api_key']"}), "(model_name='text-davinci-003', temperature=0.2, max_tokens=512,\n openai_api_key=st.secrets['api_key'])\n", (1053, 1159), False, 'from langchain.llms import OpenAI\n'), ((1791, 1832), 'streamlit.title', 'st.title', (['"""Suspicious Activity Reporting"""'], {}), "('Suspicious Activity Reporting')\n", (1799, 1832), True, 'import streamlit as st\n'), ((1834, 1868), 'streamlit.subheader', 'st.subheader', (['"""Evidence Processor"""'], {}), "('Evidence Processor')\n", (1846, 1868), True, 'import streamlit as st\n'), ((2673, 2706), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload a file"""'], {}), "('Upload a file')\n", (2689, 2706), True, 'import streamlit as st\n'), ((2710, 2724), 'streamlit.write', 'st.write', (['file'], {}), '(file)\n', (2718, 2724), True, 'import streamlit as st\n'), ((2742, 2869), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1500)', 'chunk_overlap': '(100)', 'length_function': 'len', 'separators': "['\\n\\n', '\\n', ' ', '']"}), "(chunk_size=1500, chunk_overlap=100,\n length_function=len, separators=['\\n\\n', '\\n', ' ', ''])\n", (2772, 2869), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((3468, 3487), 'streamlit.button', 'st.button', (['"""Submit"""'], {}), "('Submit')\n", (3477, 3487), True, 'import streamlit as st\n'), ((4989, 5018), 'streamlit.subheader', 'st.subheader', (['"""Ask Questions"""'], {}), "('Ask Questions')\n", (5001, 5018), True, 'import streamlit as st\n'), ((5028, 5073), 'streamlit.text_input', 'st.text_input', (['"""your queries will go here..."""'], {}), "('your queries will go here...')\n", (5041, 5073), True, 'import streamlit as st\n'), ((1343, 1402), 'streamlit.write', 'st.write', (['"""Recording started. Speak into the microphone..."""'], {}), "('Recording started. Speak into the microphone...')\n", (1351, 1402), True, 'import streamlit as st\n'), ((1554, 1596), 'streamlit.write', 'st.write', (['f"""Recording saved as {filename}"""'], {}), "(f'Recording saved as {filename}')\n", (1562, 1596), True, 'import streamlit as st\n'), ((1982, 2030), 'usellm.UseLLM', 'UseLLM', ([], {'service_url': '"""https://usellm.org/api/llm"""'}), "(service_url='https://usellm.org/api/llm')\n", (1988, 2030), False, 'from usellm import Message, Options, UseLLM\n'), ((2189, 2215), 'usellm.Options', 'Options', ([], {'messages': 'messages'}), '(messages=messages)\n', (2196, 2215), False, 'from usellm import Message, Options, UseLLM\n'), ((2352, 2396), 'langchain.embeddings.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': 'model_name'}), '(model_name=model_name)\n', (2373, 2396), False, 'from langchain.embeddings import HuggingFaceEmbeddings\n'), ((3135, 3150), 'PyPDF2.PdfReader', 'PdfReader', (['file'], {}), '(file)\n', (3144, 3150), False, 'from PyPDF2 import PdfReader\n'), ((3316, 3335), 'utils.text_to_docs', 'text_to_docs', (['texts'], {}), '(texts)\n', (3328, 3335), False, 'from utils import text_to_docs\n'), ((3375, 3416), 'langchain.vectorstores.FAISS.from_documents', 'FAISS.from_documents', (['docs', 'hf_embeddings'], {}), '(docs, hf_embeddings)\n', (3395, 3416), False, 'from langchain.vectorstores import FAISS\n'), ((5114, 5146), 'langchain.LLMChain', 'LLMChain', ([], {'prompt': 'prompt', 'llm': 'llm'}), '(prompt=prompt, llm=llm)\n', (5122, 5146), False, 'from langchain import PromptTemplate, LLMChain\n'), ((5779, 5797), 'streamlit.write', 'st.write', (['response'], {}), '(response)\n', (5787, 5797), True, 'import streamlit as st\n'), ((1704, 1750), 'openai.Audio.transcribe', 'openai.Audio.transcribe', (['"""whisper-1"""', 'filename'], {}), "('whisper-1', filename)\n", (1727, 1750), False, 'import openai\n'), ((2056, 2113), 'usellm.Message', 'Message', ([], {'role': '"""system"""', 'content': '"""You are a fraud analyst"""'}), "(role='system', content='You are a fraud analyst')\n", (2063, 2113), False, 'from usellm import Message, Options, UseLLM\n'), ((2122, 2163), 'usellm.Message', 'Message', ([], {'role': '"""user"""', 'content': 'f"""{prompt}"""'}), "(role='user', content=f'{prompt}')\n", (2129, 2163), False, 'from usellm import Message, Options, UseLLM\n'), ((3555, 3583), 'streamlit.write', 'st.write', (['"""File Uploaded..."""'], {}), "('File Uploaded...')\n", (3563, 3583), True, 'import streamlit as st\n')]
import langchain from langchain.agents import load_tools, initialize_agent, AgentType from langchain.chat_models import ChatOpenAI langchain.verbose = True langchain.debug = True def get_chat(): return ChatOpenAI(model_name="gpt-3.5-turbo", temperature=0) if __name__ == "__main__": chat = get_chat() tools = load_tools(["terminal"]) agent_chain = initialize_agent( tools, chat, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION ) result = agent_chain.run("sample_data ディレクトリにあるファイルの一覧を教えて") print(result) # { # "prompts": [ # "Human: Answer the following questions as best you can. You have access to the following tools: # terminal: Run shell commands on this Linux machine. # Use the following format: # # Question: the input question you must answer # Thought: you should always think about what to do # Action: the action to take, should be one of [terminal] # Action Input: the input to the action # Observation: the result of the action # ... (this Thought/Action/Action Input/Observation can repeat N times) # Thought: I now know the final answer # Final Answer: the final answer to the original input question # # Begin! # # Question: sample_data ディレクトリにあるファイルの一覧を教えて # Thought:I need to list the files in the sample_data directory. # Action: terminal # Action Input: ls sample_data # Observation: ls: cannot access 'sample_data': No such file or directory # # Thought:The sample_data directory does not exist. I need to find the correct directory. # Action: terminal # Action Input: ls # Observation: Dockerfile # app # docker-compose.yml # langchain # requirements.txt # venv # # Thought:The sample_data directory is not in the current directory. I need to search for it. # Action: terminal # Action Input: find / -name sample_data # Observation: /usr/src/app/sample_data # /opt/project/app/sample_data # # Thought:" # ] # }
[ "langchain.agents.initialize_agent", "langchain.agents.load_tools", "langchain.chat_models.ChatOpenAI" ]
[((209, 262), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model_name='gpt-3.5-turbo', temperature=0)\n", (219, 262), False, 'from langchain.chat_models import ChatOpenAI\n'), ((326, 350), 'langchain.agents.load_tools', 'load_tools', (["['terminal']"], {}), "(['terminal'])\n", (336, 350), False, 'from langchain.agents import load_tools, initialize_agent, AgentType\n'), ((369, 443), 'langchain.agents.initialize_agent', 'initialize_agent', (['tools', 'chat'], {'agent': 'AgentType.ZERO_SHOT_REACT_DESCRIPTION'}), '(tools, chat, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION)\n', (385, 443), False, 'from langchain.agents import load_tools, initialize_agent, AgentType\n')]
from llama_index import ( GPTVectorStoreIndex, ServiceContext, ) from llama_index.postprocessor import SentenceTransformerRerank from llama_index.embeddings import LangchainEmbedding from langchain.embeddings.huggingface import ( HuggingFaceBgeEmbeddings, ) from llama_index.vector_stores import WeaviateVectorStore from llama_index.postprocessor import MetadataReplacementPostProcessor import weaviate from llama_index.prompts.base import PromptTemplate from prompt import MISTRAL_V2_CHAT from langchain_community.llms import HuggingFaceTextGenInference from llama_index.llms import LangChainLLM from llama_index.memory import ChatMemoryBuffer import re from llama_index import ( ServiceContext, set_global_handler, ) from typing import List from llama_index.bridge.langchain import ( HumanMessage, ) from llama_index.bridge.langchain import BaseMessage from langchain_community.chat_models.huggingface import ChatHuggingFace from huggingface_hub.commands.user import login token = "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" if login(token=token): print("Login success") client = weaviate.Client("http://192.168.88.10:8080") vector_store = WeaviateVectorStore( weaviate_client=client, index_name="LlamaIndex" ) class CustomChatHuggingFace(ChatHuggingFace): def _to_chat_prompt( self, messages: List[BaseMessage], ) -> str: """Convert a list of messages into a prompt format expected by wrapped LLM.""" if not messages: raise ValueError("at least one HumanMessage must be provided") if not isinstance(messages[-1], HumanMessage): raise ValueError("last message must be a HumanMessage") messages_dicts = [self._to_chatml_format(m) for m in messages] system_str = "<s>[INST] "+ messages_dicts.pop(0)['content'] + " </INST>" chat_str = self.tokenizer.apply_chat_template( messages_dicts, tokenize=False, add_generation_prompt=True ) return system_str + chat_str class ConversationalAI: def __init__(self): self.embed_model = LangchainEmbedding(HuggingFaceBgeEmbeddings(model_name="BAAI/bge-large-en-v1.5",cache_folder='./all_model/bge-large')) self.llm = LangChainLLM(llm = CustomChatHuggingFace( llm = HuggingFaceTextGenInference(inference_server_url="http://192.168.88.10:8024", max_new_tokens=2048, top_k=10, top_p=0.95, typical_p=0.95, temperature=0.01, repetition_penalty=1.03, streaming=True,), model_id ="TheBloke/Mistral-7B-Instruct-v0.2-AWQ" )) # Initialize ServiceContext self.service_context = ServiceContext.from_defaults( embed_model=self.embed_model, llm=self.llm ) self.bge_rerank_postprocessor = SentenceTransformerRerank( model="./all_model/rerank/bge-reranker-large", top_n=3 ) # Initialize Memory self.memory = ChatMemoryBuffer.from_defaults(token_limit=4000) # Initialize vectorstore self.index = GPTVectorStoreIndex.from_vector_store(vector_store=vector_store, service_context=self.service_context) self.chat_prompt = PromptTemplate(MISTRAL_V2_CHAT) self.pattern = re.compile(r"<\|im_end\|>") def get_memory_token_count(self): chat_history = self.memory.get_all() msg_str = " ".join(str(m.content) for m in chat_history) return len(self.memory.tokenizer_fn(msg_str)) def prune_memory(self, chat_limit): chat_history = self.memory.get_all() if len(chat_history) > chat_limit: self.memory.chat_store.set_messages(self.memory.chat_store_key, chat_history[2:]) def _response(self, query): chat_engine = self.index.as_chat_engine( chat_mode="context", memory=self.memory, system_prompt= MISTRAL_V2_CHAT, # context_template= self.context_prompt, similarity_top_k=8, node_postprocessors=[self.bge_rerank_postprocessor, MetadataReplacementPostProcessor(target_metadata_key="window")], verbose=True) streaming_response = chat_engine.chat(query) for i in range(len(streaming_response.source_nodes)): print("NODE No", i) print("-------------------------------------------") print("NODE :", streaming_response.source_nodes[i]) self.prune_memory(chat_limit = 4) print("Response :", streaming_response.response) return streaming_response.response
[ "langchain_community.llms.HuggingFaceTextGenInference", "langchain.embeddings.huggingface.HuggingFaceBgeEmbeddings" ]
[((1058, 1076), 'huggingface_hub.commands.user.login', 'login', ([], {'token': 'token'}), '(token=token)\n', (1063, 1076), False, 'from huggingface_hub.commands.user import login\n'), ((1116, 1160), 'weaviate.Client', 'weaviate.Client', (['"""http://192.168.88.10:8080"""'], {}), "('http://192.168.88.10:8080')\n", (1131, 1160), False, 'import weaviate\n'), ((1176, 1244), 'llama_index.vector_stores.WeaviateVectorStore', 'WeaviateVectorStore', ([], {'weaviate_client': 'client', 'index_name': '"""LlamaIndex"""'}), "(weaviate_client=client, index_name='LlamaIndex')\n", (1195, 1244), False, 'from llama_index.vector_stores import WeaviateVectorStore\n'), ((2769, 2841), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'embed_model': 'self.embed_model', 'llm': 'self.llm'}), '(embed_model=self.embed_model, llm=self.llm)\n', (2797, 2841), False, 'from llama_index import ServiceContext, set_global_handler\n'), ((2918, 3003), 'llama_index.postprocessor.SentenceTransformerRerank', 'SentenceTransformerRerank', ([], {'model': '"""./all_model/rerank/bge-reranker-large"""', 'top_n': '(3)'}), "(model='./all_model/rerank/bge-reranker-large',\n top_n=3)\n", (2943, 3003), False, 'from llama_index.postprocessor import SentenceTransformerRerank\n'), ((3078, 3126), 'llama_index.memory.ChatMemoryBuffer.from_defaults', 'ChatMemoryBuffer.from_defaults', ([], {'token_limit': '(4000)'}), '(token_limit=4000)\n', (3108, 3126), False, 'from llama_index.memory import ChatMemoryBuffer\n'), ((3183, 3289), 'llama_index.GPTVectorStoreIndex.from_vector_store', 'GPTVectorStoreIndex.from_vector_store', ([], {'vector_store': 'vector_store', 'service_context': 'self.service_context'}), '(vector_store=vector_store,\n service_context=self.service_context)\n', (3220, 3289), False, 'from llama_index import GPTVectorStoreIndex, ServiceContext\n'), ((3315, 3346), 'llama_index.prompts.base.PromptTemplate', 'PromptTemplate', (['MISTRAL_V2_CHAT'], {}), '(MISTRAL_V2_CHAT)\n', (3329, 3346), False, 'from llama_index.prompts.base import PromptTemplate\n'), ((3370, 3398), 're.compile', 're.compile', (['"""<\\\\|im_end\\\\|>"""'], {}), "('<\\\\|im_end\\\\|>')\n", (3380, 3398), False, 'import re\n'), ((2126, 2230), 'langchain.embeddings.huggingface.HuggingFaceBgeEmbeddings', 'HuggingFaceBgeEmbeddings', ([], {'model_name': '"""BAAI/bge-large-en-v1.5"""', 'cache_folder': '"""./all_model/bge-large"""'}), "(model_name='BAAI/bge-large-en-v1.5', cache_folder=\n './all_model/bge-large')\n", (2150, 2230), False, 'from langchain.embeddings.huggingface import HuggingFaceBgeEmbeddings\n'), ((4389, 4451), 'llama_index.postprocessor.MetadataReplacementPostProcessor', 'MetadataReplacementPostProcessor', ([], {'target_metadata_key': '"""window"""'}), "(target_metadata_key='window')\n", (4421, 4451), False, 'from llama_index.postprocessor import MetadataReplacementPostProcessor\n'), ((2310, 2514), 'langchain_community.llms.HuggingFaceTextGenInference', 'HuggingFaceTextGenInference', ([], {'inference_server_url': '"""http://192.168.88.10:8024"""', 'max_new_tokens': '(2048)', 'top_k': '(10)', 'top_p': '(0.95)', 'typical_p': '(0.95)', 'temperature': '(0.01)', 'repetition_penalty': '(1.03)', 'streaming': '(True)'}), "(inference_server_url=\n 'http://192.168.88.10:8024', max_new_tokens=2048, top_k=10, top_p=0.95,\n typical_p=0.95, temperature=0.01, repetition_penalty=1.03, streaming=True)\n", (2337, 2514), False, 'from langchain_community.llms import HuggingFaceTextGenInference\n')]
import os import time import pickle as pkl # import re # import yaml import toml import logging from datetime import date # import aiohttp import pandas as pd from pytrends.request import TrendReq import serpapi from serpapi import GoogleSearch import asyncio import streamlit as st import streamlit.components.v1 as components import streamlit_authenticator as stauth import langchain # from langchain.utilities import GoogleSerperAPIWrapper from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.chat_models import ChatOpenAI from langchain.agents import initialize_agent, Tool, AgentType from langchain.docstore.document import Document from langchain.chains.qa_with_sources import load_qa_with_sources_chain from langchain.llms import OpenAI from langchain.prompts.prompt import PromptTemplate import openai from dotenv import load_dotenv load_dotenv() ########### ===== config ===== ############# config_path = '.streamlit/secrets.toml' # print(f" os.listdir('.') -> { os.listdir('.')}") if os.path.exists(config_path): logging.info(f"{config_path} exists") config = toml.load(open( config_path, 'r')) else: logging.info( f"secrets -> {st.secrets}" ) config = dict(st.secrets.items()) print( f"config -> {config}") for k in ['name', 'authentication_status', 'username' ]: st.session_state[k] = None logging.info(f"session state -> {st.session_state}") openai.api_key = os.environ.get("OPENAI_API_KEY") or config["settings"]["OPENAI_API_KEY"] os.environ["OPENAI_API_KEY"] = openai.api_key secret_key = os.environ.get('SERP_APIKEY') or config["settings"]["SERP_APIKEY"] logging.info(f"serp api key -> {secret_key}") # serp_client = serpapi.Client(api_key=secret_key) pytrends = TrendReq(hl='zh-TW', tz=480, timeout=(10,25), retries=2, backoff_factor=0.1, requests_args={'verify':False}) ########### ==================== ############# ### ===== Creating a login widget ===== ### authenticator = stauth.Authenticate( config['credentials'], config['cookie']['name'], config['cookie']['key'], config['cookie']['expiry_days'], config['preauthorized'] ) name, authentication_status, username = authenticator.login('Login', 'main') is_production = os.environ.get("PRODUCTION") or config["settings"]["PRODUCTION"] logging.info(f"is_production -> {is_production}") ########### ==================== ############# @st.cache_data def get_trending_searches(today) -> pd.DataFrame: """ Get trending searches from Google Trends """ results = pytrends.trending_searches(pn='taiwan') return results @st.cache_data def get_search_results( secret_key, query: str, today: str) -> dict: """ Get search results from serpapi.com """ print(f"secret_key -> {secret_key[-4:]}, query -> {query}, today -> {today}") serp_client=serpapi.Client(api_key=secret_key) results = serp_client.search({ 'engine': 'google', 'q': query, }) return results.as_dict() audio_input_dir = "data/audio" audio_output_dir = "data/audio" transcript_input_dir = "data/transcripts" transcript_output_dir = "data/transcripts" summary_input_dir = "data/summary" summary_output_dir = "data/summary" ### === load model & utilities === ### from model import get_answer, aget_answer from postprocess import parse_url, parse_element from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary ### ===== Authenticating users ===== ### if authentication_status: col1, col2, col3 , col4, col5, col6, col7 = st.columns(7) with col4: # st.write(f'Welcome *{name}*') st.write(f'Welcome!') # st.write(f'*{name}*') # st.markdown(f"<p style='text-align: center;'> {name} </p>", unsafe_allow_html=True) authenticator.logout('Logout', 'main') # st.write(f'\n\n\n\n') st.title("生產力小幫手") # st.divider() tab1, tab2, tab3, tab4, tab5, tab6, tab7 = st.tabs(["疼痛小幫手", "切割錄音檔", "錄音轉文字", "文字改寫", "總結摘要", "錄音檔摘要", "火熱話題"]) with tab1: question = st.text_input("病患描述問題", placeholder="肩膀不太舒服,前幾天有去打球,會不會是韌帶拉傷?") if question: if is_production: res = get_answer(question) else: res = asyncio.run(aget_answer(question)) urls = res['urls'] st.write("【建議回覆】") st.write(res["output_text"]) print( f"found urls -> {urls}" ) html = "" if len(urls)>0: for url in urls: html += f"""<br><a href=\"{url}\" style="color:gray">開啟網頁:{url}</a>""" html = f"<div>{html}</div>" components.html( html, height=600, scrolling=True) with tab2: ## --- upload file --- ## with st.form("upload-audio-form", clear_on_submit=True): uploaded_file = st.file_uploader("Upload an audio clip", key='split_audio', type=["wav", "mp3", "m4a"]) submitted = st.form_submit_button("UPLOAD") if submitted and uploaded_file: ## -- read file -- ## sound_bytes = uploaded_file.getvalue() # bytes file_obj = compose_file_path( uploaded_file.name, output_prefix = None, output_dir = audio_input_dir, output_ext = None) audio_file_path = file_obj['output_file_path'] print( f"audio_file_path: { audio_file_path}") with open( audio_file_path, "wb") as f: f.write(sound_bytes) ## --- split --- ## with st.spinner(f"reading from { audio_file_path}"): multi_file_paths = split( in_file_path = audio_file_path, length = 10) with tab3: ## --- upload file --- ## transcript_file_path = "" transcripted = 0 with st.form("upload-audio-clip-form", clear_on_submit=True): uploaded_file = st.file_uploader("Upload an audio clip", key = 'transcribe_audio', type=["wav", "mp3", "m4a"]) submitted = st.form_submit_button("UPLOAD") if submitted and uploaded_file: ## -- read file -- ## sound_bytes = uploaded_file.getvalue() # bytes file_obj = compose_file_path( input_file_path=uploaded_file.name, output_prefix = None, output_dir = audio_input_dir, output_ext = None) audio_file_path = file_obj['output_file_path'] print( f"audio_file_path: { audio_file_path}") with open( audio_file_path, "wb") as f: f.write(sound_bytes) ## --- transcribe --- ## start_time = time.time() st.divider() file_obj = compose_file_path( audio_file_path, output_prefix = None, output_dir = transcript_input_dir, output_ext = "txt") transcript_file_path = file_obj['output_file_path'] if os.path.exists(transcript_file_path): print(f"{transcript_file_path} already exists. will skip this file.") with st.spinner(f"written transcript to { transcript_file_path}"): transcript = sound2txt( audio_file_path, transcript_file_path) st.write(transcript['text']) print( f"Elapsed time: {time.time() - start_time} seconds") transcripted = 1 if transcripted and os.path.exists(transcript_file_path): with open(transcript_file_path) as f: st.download_button('下載逐字稿', f, file_name = os.path.basename(transcript_file_path), ) # Defaults to 'text/plain' with tab4: rewrite_chain = get_rewrite_chain() rewrite_file_path = "" rewritten = 0 question = st.text_input( "你希望整理會議逐字稿的方式(用簡短的英文描述)", "correct typos, remove redundant sentences/words and add punctuations", placeholder="correct typos, remove redundant sentences/words and add punctuations", # disabled=not uploaded_file, ) with st.form("upload-transcript-clip-form", clear_on_submit=True): rewrite_uploaded_file = st.file_uploader("Upload a transcript or a document (txt file)", type="txt") submitted = st.form_submit_button("UPLOAD") if submitted and rewrite_uploaded_file and question and not openai.api_key: st.info("Please add your OPENAI API KEY to environment to continue.") if submitted and rewrite_uploaded_file and question and openai.api_key: ## -- read file -- ## article = rewrite_uploaded_file.read().decode() file_obj = compose_file_path( input_file_path = rewrite_uploaded_file.name, output_prefix = "rewrite", output_dir = transcript_output_dir, output_ext = "txt" ) rewrite_file_path = file_obj['output_file_path'] # assert not os.path.exists( transcript_file_path), f"File { transcript_file_path} already exists." ## -- split into chunks -- ## text_splitter = RecursiveCharacterTextSplitter( chunk_size=1024, chunk_overlap=20, ) paragraphs = text_splitter.split_text(article) ## --- anthropic --- ## # prompt = f""" # {anthropic.HUMAN_PROMPT} Here's an article:\n\n<article> # {article}\n\n</article>\n\n{question}{anthropic.AI_PROMPT} # """ # client = anthropic.Client(anthropic_api_key) # #client = anthropic.Client(st.secrets.anthropic_api_key) # response = client.completion( # prompt=prompt, # stop_sequences=[anthropic.HUMAN_PROMPT], # model="claude-v1", # max_tokens_to_sample=100, # ) # st.write("### Answer") # st.write(response["completion"]) ## --- openai --- ## st.write("### Rewritten transcript") with st.spinner(f"written to { rewrite_file_path}"): for paragraph in paragraphs: try: res = rewrite_chain({"functions": question, "paragraph": paragraph}, return_only_outputs = True) except Exception as e: print(e) continue st.write(res["text"]) st.write("---") with open( rewrite_file_path, "a") as f: f.write(res["text"]) # rewrite_file_paths.append( rewrite_file_path) rewritten = 1 if rewritten and os.path.exists(rewrite_file_path): with open(rewrite_file_path) as f: st.download_button('下載加入標點符號後的逐字稿', f, file_name = os.path.basename(rewrite_file_path), key = f"download-button-{rewrite_file_path}") # Defaults to 'text/plain' with tab5: instruction = st.text_input( "你希望總結會議逐字稿的方式(用簡短的英文描述)", "give helpful and concise summary with action items in traditional chinese.", placeholder="give helpful and concise summary with action items in traditional chinese.", # disabled=not summary_uploaded_file, ) summarized = 0 summary_file_path = "" with st.form("upload-summarize-clip-form", clear_on_submit=True): summary_uploaded_file = st.file_uploader("Upload a transcript or a document (txt file)", type="txt") summarize_submitted = st.form_submit_button("UPLOAD") if summarize_submitted and summary_uploaded_file and instruction and not openai.api_key: st.info("Please add your OPENAI API KEY to environment to continue.") if summarize_submitted and summary_uploaded_file and instruction and openai.api_key: ## -- read file -- ## article = summary_uploaded_file.read().decode() file_obj = compose_file_path( input_file_path = summary_uploaded_file.name, output_prefix = "summary", output_dir = summary_output_dir, output_ext = "txt" ) summary_file_path = file_obj['output_file_path'] ## -- split into chunks -- ## text_splitter = RecursiveCharacterTextSplitter( chunk_size=1024, chunk_overlap=100, ) paragraphs = text_splitter.split_text(article) ## -- summarize -- ## st.write("### Summarize transcript") with st.spinner(f"written to { summary_file_path}"): summarize_chain = get_summarize_chain() with open( summary_file_path, "w") as f: for paragraph in paragraphs: try: res = summarize_chain( {"input_documents": [ Document( page_content=paragraph) ], "functions": instruction} ) summary = res['output_text'] except Exception as e: print(e) continue st.write(summary) st.write("---") f.write(summary) summarized = 1 if summarized and os.path.exists(summary_file_path): with open(summary_file_path) as f: st.download_button('下載摘要', f, file_name = os.path.basename(summary_file_path), key = f"download-button-{summary_file_path}") # Defaults to 'text/plain' with tab6: # summarize_chain = load_summarize_chain() # data/audio/GMT20230613-051643_Recording.m4a_0.m4a # with open("data/transcripts/GMT20230613-051643_Recording.m4a_0.txt") as f: # transcript = f.read() # st.markdown( f'<p style="background-color: rgba(100,100,200,.2);">{transcript}</p>', unsafe_allow_html=True) ## --- upload file --- ## with st.form("end-to-end-upload-audio-form", clear_on_submit=True): uploaded_file = st.file_uploader("Upload an audio clip", key='end_to_end', type=["wav", "mp3", "m4a"]) submitted = st.form_submit_button("UPLOAD") if submitted and uploaded_file: ## -- read file -- ## with st.spinner(f"uploaded { uploaded_file.name}"): sound_bytes = uploaded_file.getvalue() # bytes audio_file_path = compose_file_path( input_file_path = uploaded_file.name, output_prefix = None, output_dir = audio_input_dir, output_ext = None)['output_file_path'] print( f"audio_file_path: {audio_file_path}") with open( audio_file_path, "wb") as f: f.write(sound_bytes) ## --- split --- ## audio_file_paths = split( in_file_path = audio_file_path, length = 10) ## --- transcribe --- ## st.write("### Transcribing audio") with st.spinner(f"transcribe { ', '.join(audio_file_paths) }"): transcripts = [] transcript_file_paths = [] for i, audio_file_path in enumerate(audio_file_paths): start_time = time.time() transcript_file_path = compose_file_path( input_file_path = audio_file_path, output_prefix = None, output_dir = transcript_input_dir, output_ext = "txt")['output_file_path'] transcript_file_paths.append(transcript_file_path) st.write(f"Trascribing {i}-th clip out of total {len(audio_file_paths)} clips.") transcript = sound2txt( audio_file_path, transcript_file_path) st.write(transcript['text'][:50]) st.divider() transcripts.append( transcript['text']) print( f"[transcription] Elapsed time: {time.time() - start_time} seconds") # print( f"Transcript: {transcript}") ## --- Rewrite --- ## st.write("### Rewritten transcript") with st.spinner(f"rewritten { ','.join(transcript_file_paths)}"): text_splitter = RecursiveCharacterTextSplitter( chunk_size=1024, chunk_overlap=20, ) rewrite_instruction = "correct typos, remove redundant sentences/words and add punctuations" rewrite_file_paths = [] rewrites = [] for i, blob in enumerate(zip(transcript_file_paths, transcripts)): transcript_file_path, transcript = blob paragraphs = text_splitter.split_text( "\n".join(transcript) ) rewrite_file_path = compose_file_path( input_file_path = transcript_file_path, output_prefix = "rewrite", output_dir = transcript_output_dir, output_ext = "txt")['output_file_path'] rewrite_file_paths.append(rewrite_file_path) st.write(f"Rewriting {i}-th transcript out of total {len(transcript_file_paths)} transcripts.") for paragraph in paragraphs: try: res = rewrite_chain({"functions": rewrite_instruction, "paragraph": paragraph}, return_only_outputs = True) except Exception as e: print(e) continue with open( rewrite_file_path, "a") as f: f.write(res["text"]) st.write(res["text"][:50]) st.divider() ## --- summarize --- ## with st.spinner(f"summarize { ', '.join(rewrite_file_paths) }"): summary_text = "" for rewrite, rewrite_file_path in zip( rewrites, rewrite_file_paths): start_time = time.time() summary_file_path = compose_file_path( input_file_path = rewrite_file_path, output_prefix = "summary", output_dir = summary_output_dir, output_ext = "txt")['output_file_path'] summary = text2summary( rewrite, summary_file_path, instruction = "give helpful and concise summary with action items in traditional chinese.") st.write(summary) st.divider() summary_text += f"{summary}\n" print( f"[summary] Elapsed time: {time.time() - start_time} seconds") with tab7: if False: # today = date.today() # topn_results = 3 # topn_articles = 1 # trend_results = get_trending_searches(today) # highlights = [] # for query in trend_results[0].values[:topn_results]: # search_results = get_search_results( secret_key, query, str(today) ) # highlights += search_results['organic_results'][:topn_articles] # highlights = pd.DataFrame(highlights) params = { "api_key": secret_key, "engine": "google_trends_trending_now", "frequency": "daily", "geo": "TW", "hl": "zh-tw" } search = GoogleSearch(params) results = search.get_dict() content = [] for ds in results['daily_searches']: for s in ds['searches']: q = s['query'] if 'articles' in s and len(s['articles'])>0: link = s['articles'][0]['link'] title = s['articles'][0]['title'] snippet = s['articles'][0]['snippet'] else: continue content.append({'query': q, 'title': title, 'snippet': snippet, 'link': link}) st.dataframe(pd.DataFrame(content)) else: st.write("Coming soon...") elif authentication_status == False: st.error('Username/password is incorrect') elif authentication_status == None: st.warning('Please enter your username and password')
[ "langchain.docstore.document.Document", "langchain.text_splitter.RecursiveCharacterTextSplitter" ]
[((872, 885), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (883, 885), False, 'from dotenv import load_dotenv\n'), ((1028, 1055), 'os.path.exists', 'os.path.exists', (['config_path'], {}), '(config_path)\n', (1042, 1055), False, 'import os\n'), ((1341, 1393), 'logging.info', 'logging.info', (['f"""session state -> {st.session_state}"""'], {}), "(f'session state -> {st.session_state}')\n", (1353, 1393), False, 'import logging\n'), ((1611, 1656), 'logging.info', 'logging.info', (['f"""serp api key -> {secret_key}"""'], {}), "(f'serp api key -> {secret_key}')\n", (1623, 1656), False, 'import logging\n'), ((1719, 1834), 'pytrends.request.TrendReq', 'TrendReq', ([], {'hl': '"""zh-TW"""', 'tz': '(480)', 'timeout': '(10, 25)', 'retries': '(2)', 'backoff_factor': '(0.1)', 'requests_args': "{'verify': False}"}), "(hl='zh-TW', tz=480, timeout=(10, 25), retries=2, backoff_factor=\n 0.1, requests_args={'verify': False})\n", (1727, 1834), False, 'from pytrends.request import TrendReq\n'), ((1938, 2099), 'streamlit_authenticator.Authenticate', 'stauth.Authenticate', (["config['credentials']", "config['cookie']['name']", "config['cookie']['key']", "config['cookie']['expiry_days']", "config['preauthorized']"], {}), "(config['credentials'], config['cookie']['name'], config\n ['cookie']['key'], config['cookie']['expiry_days'], config['preauthorized']\n )\n", (1957, 2099), True, 'import streamlit_authenticator as stauth\n'), ((2260, 2309), 'logging.info', 'logging.info', (['f"""is_production -> {is_production}"""'], {}), "(f'is_production -> {is_production}')\n", (2272, 2309), False, 'import logging\n'), ((1058, 1095), 'logging.info', 'logging.info', (['f"""{config_path} exists"""'], {}), "(f'{config_path} exists')\n", (1070, 1095), False, 'import logging\n'), ((1148, 1188), 'logging.info', 'logging.info', (['f"""secrets -> {st.secrets}"""'], {}), "(f'secrets -> {st.secrets}')\n", (1160, 1188), False, 'import logging\n'), ((1411, 1443), 'os.environ.get', 'os.environ.get', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (1425, 1443), False, 'import os\n'), ((1544, 1573), 'os.environ.get', 'os.environ.get', (['"""SERP_APIKEY"""'], {}), "('SERP_APIKEY')\n", (1558, 1573), False, 'import os\n'), ((2195, 2223), 'os.environ.get', 'os.environ.get', (['"""PRODUCTION"""'], {}), "('PRODUCTION')\n", (2209, 2223), False, 'import os\n'), ((2797, 2831), 'serpapi.Client', 'serpapi.Client', ([], {'api_key': 'secret_key'}), '(api_key=secret_key)\n', (2811, 2831), False, 'import serpapi\n'), ((3525, 3538), 'streamlit.columns', 'st.columns', (['(7)'], {}), '(7)\n', (3535, 3538), True, 'import streamlit as st\n'), ((3791, 3809), 'streamlit.title', 'st.title', (['"""生產力小幫手"""'], {}), "('生產力小幫手')\n", (3799, 3809), True, 'import streamlit as st\n'), ((3873, 3942), 'streamlit.tabs', 'st.tabs', (["['疼痛小幫手', '切割錄音檔', '錄音轉文字', '文字改寫', '總結摘要', '錄音檔摘要', '火熱話題']"], {}), "(['疼痛小幫手', '切割錄音檔', '錄音轉文字', '文字改寫', '總結摘要', '錄音檔摘要', '火熱話題'])\n", (3880, 3942), True, 'import streamlit as st\n'), ((1206, 1224), 'streamlit.secrets.items', 'st.secrets.items', ([], {}), '()\n', (1222, 1224), True, 'import streamlit as st\n'), ((3587, 3608), 'streamlit.write', 'st.write', (['f"""Welcome!"""'], {}), "(f'Welcome!')\n", (3595, 3608), True, 'import streamlit as st\n'), ((3969, 4032), 'streamlit.text_input', 'st.text_input', (['"""病患描述問題"""'], {'placeholder': '"""肩膀不太舒服,前幾天有去打球,會不會是韌帶拉傷?"""'}), "('病患描述問題', placeholder='肩膀不太舒服,前幾天有去打球,會不會是韌帶拉傷?')\n", (3982, 4032), True, 'import streamlit as st\n'), ((6949, 6968), 'utils.get_rewrite_chain', 'get_rewrite_chain', ([], {}), '()\n', (6966, 6968), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((7023, 7232), 'streamlit.text_input', 'st.text_input', (['"""你希望整理會議逐字稿的方式(用簡短的英文描述)"""', '"""correct typos, remove redundant sentences/words and add punctuations"""'], {'placeholder': '"""correct typos, remove redundant sentences/words and add punctuations"""'}), "('你希望整理會議逐字稿的方式(用簡短的英文描述)',\n 'correct typos, remove redundant sentences/words and add punctuations',\n placeholder=\n 'correct typos, remove redundant sentences/words and add punctuations')\n", (7036, 7232), True, 'import streamlit as st\n'), ((9715, 9942), 'streamlit.text_input', 'st.text_input', (['"""你希望總結會議逐字稿的方式(用簡短的英文描述)"""', '"""give helpful and concise summary with action items in traditional chinese."""'], {'placeholder': '"""give helpful and concise summary with action items in traditional chinese."""'}), "('你希望總結會議逐字稿的方式(用簡短的英文描述)',\n 'give helpful and concise summary with action items in traditional chinese.'\n , placeholder=\n 'give helpful and concise summary with action items in traditional chinese.'\n )\n", (9728, 9942), True, 'import streamlit as st\n'), ((17034, 17076), 'streamlit.error', 'st.error', (['"""Username/password is incorrect"""'], {}), "('Username/password is incorrect')\n", (17042, 17076), True, 'import streamlit as st\n'), ((4182, 4200), 'streamlit.write', 'st.write', (['"""【建議回覆】"""'], {}), "('【建議回覆】')\n", (4190, 4200), True, 'import streamlit as st\n'), ((4204, 4232), 'streamlit.write', 'st.write', (["res['output_text']"], {}), "(res['output_text'])\n", (4212, 4232), True, 'import streamlit as st\n'), ((4437, 4486), 'streamlit.components.v1.html', 'components.html', (['html'], {'height': '(600)', 'scrolling': '(True)'}), '(html, height=600, scrolling=True)\n', (4452, 4486), True, 'import streamlit.components.v1 as components\n'), ((4537, 4587), 'streamlit.form', 'st.form', (['"""upload-audio-form"""'], {'clear_on_submit': '(True)'}), "('upload-audio-form', clear_on_submit=True)\n", (4544, 4587), True, 'import streamlit as st\n'), ((4608, 4699), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload an audio clip"""'], {'key': '"""split_audio"""', 'type': "['wav', 'mp3', 'm4a']"}), "('Upload an audio clip', key='split_audio', type=['wav',\n 'mp3', 'm4a'])\n", (4624, 4699), True, 'import streamlit as st\n'), ((4711, 4742), 'streamlit.form_submit_button', 'st.form_submit_button', (['"""UPLOAD"""'], {}), "('UPLOAD')\n", (4732, 4742), True, 'import streamlit as st\n'), ((5411, 5466), 'streamlit.form', 'st.form', (['"""upload-audio-clip-form"""'], {'clear_on_submit': '(True)'}), "('upload-audio-clip-form', clear_on_submit=True)\n", (5418, 5466), True, 'import streamlit as st\n'), ((5487, 5584), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload an audio clip"""'], {'key': '"""transcribe_audio"""', 'type': "['wav', 'mp3', 'm4a']"}), "('Upload an audio clip', key='transcribe_audio', type=[\n 'wav', 'mp3', 'm4a'])\n", (5503, 5584), True, 'import streamlit as st\n'), ((5597, 5628), 'streamlit.form_submit_button', 'st.form_submit_button', (['"""UPLOAD"""'], {}), "('UPLOAD')\n", (5618, 5628), True, 'import streamlit as st\n'), ((6719, 6755), 'os.path.exists', 'os.path.exists', (['transcript_file_path'], {}), '(transcript_file_path)\n', (6733, 6755), False, 'import os\n'), ((7282, 7342), 'streamlit.form', 'st.form', (['"""upload-transcript-clip-form"""'], {'clear_on_submit': '(True)'}), "('upload-transcript-clip-form', clear_on_submit=True)\n", (7289, 7342), True, 'import streamlit as st\n'), ((7371, 7447), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload a transcript or a document (txt file)"""'], {'type': '"""txt"""'}), "('Upload a transcript or a document (txt file)', type='txt')\n", (7387, 7447), True, 'import streamlit as st\n'), ((7463, 7494), 'streamlit.form_submit_button', 'st.form_submit_button', (['"""UPLOAD"""'], {}), "('UPLOAD')\n", (7484, 7494), True, 'import streamlit as st\n'), ((9445, 9478), 'os.path.exists', 'os.path.exists', (['rewrite_file_path'], {}), '(rewrite_file_path)\n', (9459, 9478), False, 'import os\n'), ((10034, 10093), 'streamlit.form', 'st.form', (['"""upload-summarize-clip-form"""'], {'clear_on_submit': '(True)'}), "('upload-summarize-clip-form', clear_on_submit=True)\n", (10041, 10093), True, 'import streamlit as st\n'), ((10122, 10198), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload a transcript or a document (txt file)"""'], {'type': '"""txt"""'}), "('Upload a transcript or a document (txt file)', type='txt')\n", (10138, 10198), True, 'import streamlit as st\n'), ((10224, 10255), 'streamlit.form_submit_button', 'st.form_submit_button', (['"""UPLOAD"""'], {}), "('UPLOAD')\n", (10245, 10255), True, 'import streamlit as st\n'), ((11610, 11643), 'os.path.exists', 'os.path.exists', (['summary_file_path'], {}), '(summary_file_path)\n', (11624, 11643), False, 'import os\n'), ((12208, 12269), 'streamlit.form', 'st.form', (['"""end-to-end-upload-audio-form"""'], {'clear_on_submit': '(True)'}), "('end-to-end-upload-audio-form', clear_on_submit=True)\n", (12215, 12269), True, 'import streamlit as st\n'), ((12290, 12380), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload an audio clip"""'], {'key': '"""end_to_end"""', 'type': "['wav', 'mp3', 'm4a']"}), "('Upload an audio clip', key='end_to_end', type=['wav',\n 'mp3', 'm4a'])\n", (12306, 12380), True, 'import streamlit as st\n'), ((12392, 12423), 'streamlit.form_submit_button', 'st.form_submit_button', (['"""UPLOAD"""'], {}), "('UPLOAD')\n", (12413, 12423), True, 'import streamlit as st\n'), ((16478, 16498), 'serpapi.GoogleSearch', 'GoogleSearch', (['params'], {}), '(params)\n', (16490, 16498), False, 'from serpapi import GoogleSearch\n'), ((16967, 16993), 'streamlit.write', 'st.write', (['"""Coming soon..."""'], {}), "('Coming soon...')\n", (16975, 16993), True, 'import streamlit as st\n'), ((17116, 17169), 'streamlit.warning', 'st.warning', (['"""Please enter your username and password"""'], {}), "('Please enter your username and password')\n", (17126, 17169), True, 'import streamlit as st\n'), ((4080, 4100), 'model.get_answer', 'get_answer', (['question'], {}), '(question)\n', (4090, 4100), False, 'from model import get_answer, aget_answer\n'), ((4875, 4982), 'utils.compose_file_path', 'compose_file_path', (['uploaded_file.name'], {'output_prefix': 'None', 'output_dir': 'audio_input_dir', 'output_ext': 'None'}), '(uploaded_file.name, output_prefix=None, output_dir=\n audio_input_dir, output_ext=None)\n', (4892, 4982), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((5756, 5878), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'uploaded_file.name', 'output_prefix': 'None', 'output_dir': 'audio_input_dir', 'output_ext': 'None'}), '(input_file_path=uploaded_file.name, output_prefix=None,\n output_dir=audio_input_dir, output_ext=None)\n', (5773, 5878), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((6103, 6114), 'time.time', 'time.time', ([], {}), '()\n', (6112, 6114), False, 'import time\n'), ((6119, 6131), 'streamlit.divider', 'st.divider', ([], {}), '()\n', (6129, 6131), True, 'import streamlit as st\n'), ((6148, 6258), 'utils.compose_file_path', 'compose_file_path', (['audio_file_path'], {'output_prefix': 'None', 'output_dir': 'transcript_input_dir', 'output_ext': '"""txt"""'}), "(audio_file_path, output_prefix=None, output_dir=\n transcript_input_dir, output_ext='txt')\n", (6165, 6258), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((6324, 6360), 'os.path.exists', 'os.path.exists', (['transcript_file_path'], {}), '(transcript_file_path)\n', (6338, 6360), False, 'import os\n'), ((7582, 7651), 'streamlit.info', 'st.info', (['"""Please add your OPENAI API KEY to environment to continue."""'], {}), "('Please add your OPENAI API KEY to environment to continue.')\n", (7589, 7651), True, 'import streamlit as st\n'), ((7830, 7973), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'rewrite_uploaded_file.name', 'output_prefix': '"""rewrite"""', 'output_dir': 'transcript_output_dir', 'output_ext': '"""txt"""'}), "(input_file_path=rewrite_uploaded_file.name, output_prefix\n ='rewrite', output_dir=transcript_output_dir, output_ext='txt')\n", (7847, 7973), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((8201, 8266), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1024)', 'chunk_overlap': '(20)'}), '(chunk_size=1024, chunk_overlap=20)\n', (8231, 8266), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((8924, 8960), 'streamlit.write', 'st.write', (['"""### Rewritten transcript"""'], {}), "('### Rewritten transcript')\n", (8932, 8960), True, 'import streamlit as st\n'), ((10356, 10425), 'streamlit.info', 'st.info', (['"""Please add your OPENAI API KEY to environment to continue."""'], {}), "('Please add your OPENAI API KEY to environment to continue.')\n", (10363, 10425), True, 'import streamlit as st\n'), ((10617, 10757), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'summary_uploaded_file.name', 'output_prefix': '"""summary"""', 'output_dir': 'summary_output_dir', 'output_ext': '"""txt"""'}), "(input_file_path=summary_uploaded_file.name, output_prefix\n ='summary', output_dir=summary_output_dir, output_ext='txt')\n", (10634, 10757), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((10881, 10947), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1024)', 'chunk_overlap': '(100)'}), '(chunk_size=1024, chunk_overlap=100)\n', (10911, 10947), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((11047, 11083), 'streamlit.write', 'st.write', (['"""### Summarize transcript"""'], {}), "('### Summarize transcript')\n", (11055, 11083), True, 'import streamlit as st\n'), ((13026, 13060), 'streamlit.write', 'st.write', (['"""### Transcribing audio"""'], {}), "('### Transcribing audio')\n", (13034, 13060), True, 'import streamlit as st\n'), ((13941, 13977), 'streamlit.write', 'st.write', (['"""### Rewritten transcript"""'], {}), "('### Rewritten transcript')\n", (13949, 13977), True, 'import streamlit as st\n'), ((16933, 16954), 'pandas.DataFrame', 'pd.DataFrame', (['content'], {}), '(content)\n', (16945, 16954), True, 'import pandas as pd\n'), ((4132, 4153), 'model.aget_answer', 'aget_answer', (['question'], {}), '(question)\n', (4143, 4153), False, 'from model import get_answer, aget_answer\n'), ((5191, 5236), 'streamlit.spinner', 'st.spinner', (['f"""reading from {audio_file_path}"""'], {}), "(f'reading from {audio_file_path}')\n", (5201, 5236), True, 'import streamlit as st\n'), ((5263, 5309), 'utils.split', 'split', ([], {'in_file_path': 'audio_file_path', 'length': '(10)'}), '(in_file_path=audio_file_path, length=10)\n', (5268, 5309), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((6449, 6508), 'streamlit.spinner', 'st.spinner', (['f"""written transcript to {transcript_file_path}"""'], {}), "(f'written transcript to {transcript_file_path}')\n", (6459, 6508), True, 'import streamlit as st\n'), ((6529, 6577), 'utils.sound2txt', 'sound2txt', (['audio_file_path', 'transcript_file_path'], {}), '(audio_file_path, transcript_file_path)\n', (6538, 6577), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((6584, 6612), 'streamlit.write', 'st.write', (["transcript['text']"], {}), "(transcript['text'])\n", (6592, 6612), True, 'import streamlit as st\n'), ((8970, 9015), 'streamlit.spinner', 'st.spinner', (['f"""written to {rewrite_file_path}"""'], {}), "(f'written to {rewrite_file_path}')\n", (8980, 9015), True, 'import streamlit as st\n'), ((11093, 11138), 'streamlit.spinner', 'st.spinner', (['f"""written to {summary_file_path}"""'], {}), "(f'written to {summary_file_path}')\n", (11103, 11138), True, 'import streamlit as st\n'), ((11164, 11185), 'utils.get_summarize_chain', 'get_summarize_chain', ([], {}), '()\n', (11183, 11185), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((12494, 12538), 'streamlit.spinner', 'st.spinner', (['f"""uploaded {uploaded_file.name}"""'], {}), "(f'uploaded {uploaded_file.name}')\n", (12504, 12538), True, 'import streamlit as st\n'), ((12938, 12984), 'utils.split', 'split', ([], {'in_file_path': 'audio_file_path', 'length': '(10)'}), '(in_file_path=audio_file_path, length=10)\n', (12943, 12984), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((14069, 14134), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1024)', 'chunk_overlap': '(20)'}), '(chunk_size=1024, chunk_overlap=20)\n', (14099, 14134), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((6855, 6893), 'os.path.basename', 'os.path.basename', (['transcript_file_path'], {}), '(transcript_file_path)\n', (6871, 6893), False, 'import os\n'), ((9235, 9256), 'streamlit.write', 'st.write', (["res['text']"], {}), "(res['text'])\n", (9243, 9256), True, 'import streamlit as st\n'), ((9263, 9278), 'streamlit.write', 'st.write', (['"""---"""'], {}), "('---')\n", (9271, 9278), True, 'import streamlit as st\n'), ((9599, 9634), 'os.path.basename', 'os.path.basename', (['rewrite_file_path'], {}), '(rewrite_file_path)\n', (9615, 9634), False, 'import os\n'), ((11737, 11772), 'os.path.basename', 'os.path.basename', (['summary_file_path'], {}), '(summary_file_path)\n', (11753, 11772), False, 'import os\n'), ((12616, 12738), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'uploaded_file.name', 'output_prefix': 'None', 'output_dir': 'audio_input_dir', 'output_ext': 'None'}), '(input_file_path=uploaded_file.name, output_prefix=None,\n output_dir=audio_input_dir, output_ext=None)\n', (12633, 12738), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((13262, 13273), 'time.time', 'time.time', ([], {}), '()\n', (13271, 13273), False, 'import time\n'), ((13625, 13673), 'utils.sound2txt', 'sound2txt', (['audio_file_path', 'transcript_file_path'], {}), '(audio_file_path, transcript_file_path)\n', (13634, 13673), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((13682, 13715), 'streamlit.write', 'st.write', (["transcript['text'][:50]"], {}), "(transcript['text'][:50])\n", (13690, 13715), True, 'import streamlit as st\n'), ((13722, 13734), 'streamlit.divider', 'st.divider', ([], {}), '()\n', (13732, 13734), True, 'import streamlit as st\n'), ((15403, 15414), 'time.time', 'time.time', ([], {}), '()\n', (15412, 15414), False, 'import time\n'), ((15613, 15753), 'utils.text2summary', 'text2summary', (['rewrite', 'summary_file_path'], {'instruction': '"""give helpful and concise summary with action items in traditional chinese."""'}), "(rewrite, summary_file_path, instruction=\n 'give helpful and concise summary with action items in traditional chinese.'\n )\n", (15625, 15753), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((15753, 15770), 'streamlit.write', 'st.write', (['summary'], {}), '(summary)\n', (15761, 15770), True, 'import streamlit as st\n'), ((15777, 15789), 'streamlit.divider', 'st.divider', ([], {}), '()\n', (15787, 15789), True, 'import streamlit as st\n'), ((11506, 11523), 'streamlit.write', 'st.write', (['summary'], {}), '(summary)\n', (11514, 11523), True, 'import streamlit as st\n'), ((11531, 11546), 'streamlit.write', 'st.write', (['"""---"""'], {}), "('---')\n", (11539, 11546), True, 'import streamlit as st\n'), ((13310, 13435), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'audio_file_path', 'output_prefix': 'None', 'output_dir': 'transcript_input_dir', 'output_ext': '"""txt"""'}), "(input_file_path=audio_file_path, output_prefix=None,\n output_dir=transcript_input_dir, output_ext='txt')\n", (13327, 13435), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((14514, 14651), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'transcript_file_path', 'output_prefix': '"""rewrite"""', 'output_dir': 'transcript_output_dir', 'output_ext': '"""txt"""'}), "(input_file_path=transcript_file_path, output_prefix=\n 'rewrite', output_dir=transcript_output_dir, output_ext='txt')\n", (14531, 14651), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((15141, 15167), 'streamlit.write', 'st.write', (["res['text'][:50]"], {}), "(res['text'][:50])\n", (15149, 15167), True, 'import streamlit as st\n'), ((15175, 15187), 'streamlit.divider', 'st.divider', ([], {}), '()\n', (15185, 15187), True, 'import streamlit as st\n'), ((15441, 15572), 'utils.compose_file_path', 'compose_file_path', ([], {'input_file_path': 'rewrite_file_path', 'output_prefix': '"""summary"""', 'output_dir': 'summary_output_dir', 'output_ext': '"""txt"""'}), "(input_file_path=rewrite_file_path, output_prefix=\n 'summary', output_dir=summary_output_dir, output_ext='txt')\n", (15458, 15572), False, 'from utils import split, compose_file_path, get_rewrite_chain, get_summarize_chain, sound2txt, text2summary\n'), ((6641, 6652), 'time.time', 'time.time', ([], {}), '()\n', (6650, 6652), False, 'import time\n'), ((13827, 13838), 'time.time', 'time.time', ([], {}), '()\n', (13836, 13838), False, 'import time\n'), ((15867, 15878), 'time.time', 'time.time', ([], {}), '()\n', (15876, 15878), False, 'import time\n'), ((11333, 11365), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': 'paragraph'}), '(page_content=paragraph)\n', (11341, 11365), False, 'from langchain.docstore.document import Document\n')]
import requests import re import langchain import openai # Set up OpenAI GPT API credentials openai.api_key = 'MY_OPENAI_API_KEY' # Function to fetch a GitHub user's repositories def fetch_user_repos(username): url = f'https://api.github.com/users/{username}/repos' response = requests.get(url) if response.status_code == 200: return response.json() else: raise Exception(f"Failed to fetch user repositories. Status code: {response.status_code}") # Function to assess repository complexity using GPT and LangChain def assess_repository(repo): # Retrieve repository information repo_name = repo['name'] repo_url = repo['html_url'] readme_url = f"{repo_url}/blob/master/README.md" # Fetch the README content readme_response = requests.get(readme_url) readme_content = '' if readme_response.status_code == 200: readme_content = readme_response.text # Analyze README content using GPT gpt_input = f"Assess the complexity of repository {repo_name}. {readme_content}" gpt_response = openai.Completion.create( engine='text-davinci-003', prompt=gpt_input, max_tokens=100, temperature=0.5 ) complexity_score = gpt_response.choices[0].score # Analyze repository code using LangChain code_metrics = langchain.extract_metrics_from_github_repo(repo_url) # Calculate the overall complexity score based on GPT and LangChain metrics overall_score = complexity_score + code_metrics.complexity_score return repo_name, repo_url, overall_score # Function to find the most technically challenging repository def find_most_challenging_repository(username): try: # Fetch user repositories repositories = fetch_user_repos(username) if not repositories: raise Exception(f"No repositories found for user {username}") # Assess complexity for each repository repository_scores = [] for repo in repositories: repo_name, repo_url, overall_score = assess_repository(repo) repository_scores.append((repo_name, repo_url, overall_score)) # Sort repositories by score in descending order repository_scores.sort(key=lambda x: x[2], reverse=True) # Return the most challenging repository return repository_scores[0] except Exception as e: print(f"An error occurred: {str(e)}") # Example usage github_username = 'sai-vamsi-m' most_challenging_repo = find_most_challenging_repository(github_username) if most_challenging_repo: print(f"The most challenging repository for user {github_username} is:") print(f"Name: {most_challenging_repo[0]}") print(f"URL: {most_challenging_repo[1]}")
[ "langchain.extract_metrics_from_github_repo" ]
[((300, 317), 'requests.get', 'requests.get', (['url'], {}), '(url)\n', (312, 317), False, 'import requests\n'), ((811, 835), 'requests.get', 'requests.get', (['readme_url'], {}), '(readme_url)\n', (823, 835), False, 'import requests\n'), ((1100, 1206), 'openai.Completion.create', 'openai.Completion.create', ([], {'engine': '"""text-davinci-003"""', 'prompt': 'gpt_input', 'max_tokens': '(100)', 'temperature': '(0.5)'}), "(engine='text-davinci-003', prompt=gpt_input,\n max_tokens=100, temperature=0.5)\n", (1124, 1206), False, 'import openai\n'), ((1369, 1421), 'langchain.extract_metrics_from_github_repo', 'langchain.extract_metrics_from_github_repo', (['repo_url'], {}), '(repo_url)\n', (1411, 1421), False, 'import langchain\n')]
# import supporting packages and modules from abc import ABC import yaml import os import tempfile import requests from urllib.parse import urlparse from typing import List import json import re # import langchain modules from langchain.docstore.document import Document from langchain.document_loaders.base import BaseLoader from langchain.document_loaders import OnlinePDFLoader from langchain import PromptTemplate from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.embeddings import OpenAIEmbeddings from langchain.vectorstores.faiss import FAISS from langchain.document_loaders import WebBaseLoader # import puzzle-knowledge-sdk modules from .Ngram import AddAlphaSmooth class GloLoader(BaseLoader, ABC): """Loader class for `.glo` files. Defaults to check for local file, but if file is a web path, it will download it to a temporary file, use it, then clean up the temporary file after completion. """ def __init__(self, file_path: str): """Initializes the loader with the file path.""" self.file_path, self.web_path = self._process_file_path(file_path) def _process_file_path(self, file_path: str): """Handles file checking, URL validity checking, and downloading if necessary.""" web_path = None # Initialize web_path locally if "~" in file_path: file_path = os.path.expanduser(file_path) if os.path.isfile(file_path): return file_path, web_path # Return a tuple with two values elif self._is_valid_url(file_path): temp_dir = tempfile.TemporaryDirectory() self.temp_dir = temp_dir _, suffix = os.path.splitext(file_path) temp_glo = os.path.join(temp_dir.name, f"tmp{suffix}") if self._is_s3_url(file_path): web_path = file_path else: r = requests.get(file_path) if r.status_code != 200: print(file_path) web_path = file_path with open(temp_glo, mode="wb") as f: f.write(r.content) return str(temp_glo), web_path # Return a tuple with two values else: raise ValueError("File path %s is not a valid file or URL" % file_path) def __del__(self) -> None: if hasattr(self, "temp_dir"): self.temp_dir.cleanup() @staticmethod def _is_valid_url(url: str) -> bool: """Check if the url is valid.""" parsed = urlparse(url) return bool(parsed.netloc) and bool(parsed.scheme) @staticmethod def _is_s3_url(url: str) -> bool: """check if the url is S3""" try: result = urlparse(url) if result.scheme == "s3" and result.netloc: return True return False except ValueError: return False @staticmethod def _is_html_page(url): try: response = requests.head(url) content_type = response.headers.get('content-type', '').lower() return 'text/html' in content_type except requests.exceptions.RequestException: return False @property def source(self) -> str: return self.web_path if self.web_path is not None else self.file_path def import_data(self) -> dict: """Load concept documents.""" if isinstance(self.file_path, str): loaded = False # load data json or yaml or glo # yaml load with open(self.file_path, 'r') as glo_file: glo_text = glo_file.read() try: data = yaml.safe_load(glo_text) return data except: pass # json load try: with open(self.file_path, 'r') as json_file: data = json.load(json_file) return data except: pass if not loaded: raise ValueError("Error parsing file: Not in valid JSON or YAML formats") elif isinstance(self.file_path, dict): data = self.file_path return data @staticmethod def load_link(link, type, concept_name, load) -> Document: if type == 'uri': if link.endswith(".pdf"): if load: loader = OnlinePDFLoader(link) data_load = loader.load()[0] data_load.page_content = re.sub(r'\n+', '\n', data_load.page_content) else: data_load = Document(page_content="") data_load.metadata = { "concept": concept_name, "type": "link-pdf", "link_type": "uri", "source": link, "load_status": load } return data_load else: try: if load: loader = WebBaseLoader(link) data_load = loader.load()[0] data_load.page_content = re.sub(r'\n+', '\n', data_load.page_content) else: data_load = Document(page_content="") data_load.metadata = { "concept": concept_name, "type": "link-html", "link_type": "uri", "source": link, "load_status": load } return data_load except Exception as e: raise ValueError(f"Error loading link {link}: {e}") elif type == 'text': data_load = Document( page_content=link if load else "", metadata={ "concept": concept_name, "type": "link-text", "link_type": "text", "source": "text", "load_status": load } ) return data_load elif type == 'glo': loader = GloLoader(link) data_load = loader.load() text = GloLoader.transform(documents=data_load, include_task=False) data_load = Document( page_content=text if load else "", metadata={ "concept": concept_name, "type": "link-glo", "link_type": "glo", "source": link if isinstance(link, str) else f"Glossary: {link['name']}", "load_status": load } ) return data_load else: raise ValueError(f"Invalid link given: Can only process html, pdf, text, and glo.") def load(self, loadLinks=False) -> List[Document]: data = self.import_data() documents = [] if "concepts" in data: concepts = data["concepts"] for concept in concepts: if "name" in concept and "explanation" in concept: content = f"NAME: {concept['name']}\nEXPLANATION: {concept['explanation']}" new_document = Document( page_content=re.sub(r'\n+', '\n', content), metadata={ "glo_name": data.get("name", ""), "topic": data.get("topic", ""), "audience": data.get("audience", ""), "concept": concept.get("name", ""), "type": "content", "source": self.file_path if isinstance(self.file_path, str) else self.file_path["name"], "links": [] } ) if "links" in concept.keys(): for link in concept["links"]: if "uri" in link.keys(): new_document.metadata["links"].append(GloLoader.load_link(link["uri"], type="uri", concept_name=concept["name"], load=loadLinks)) elif "text" in link.keys(): new_document.metadata["links"].append(GloLoader.load_link(link["text"], type="text", concept_name=concept["name"], load=loadLinks)) elif "glo" in link.keys(): new_document.metadata["links"].append(GloLoader.load_link(link["glo"], type="glo", concept_name=concept["name"], load=loadLinks)) documents.append(new_document) else: raise ValueError("Concepts must have a name and explanation") return documents @staticmethod def calculate_score(sample, source, n=2, scope="word"): # to lower sample = sample.lower() source = source.lower() n = 2 scope = "word" try: if n in [1, 2, 3]: if scope == "word": # preprocess pattern pattern = r'([!@#$%^&*()_+{}\[\]:;"\'<>,.?/\|\\])' prep_source = [re.sub(pattern, r' \1', source).split(" ")] prep_sample = [re.sub(pattern, r' \1', sample).split(" ")] elif scope == "char": prep_source = [list(source)] prep_sample = [list(sample)] dist_source = AddAlphaSmooth(n, prep_source) score = dist_source.getSentenceLogLikelihood(prep_sample[0]) return score else: raise ValueError(f"ngram methods must have n in [1, 2, 3]") except ValueError as e: raise ValueError(f"Concept ranking failed: {e}") @staticmethod def transform(query: str, documents: List[Document], header: str=None, task: str=None, rank_function=None, additional_args: dict={}): if header is None or header == "": # glo name glo_name = "" for doc in documents: if isinstance(doc, Document): glo_name = doc.metadata.get("glo_name", "") if glo_name != "": break header = f"GLOSSARY: {glo_name}" if task is None or task == "": task = "TASK: This is a glossary of concepts for your reference throughout this conversation. You should prioritize this information when answering any questions asked by the user." max_width = additional_args.get("max_width", 1024) if rank_function is None: def fn(documents, max_width): context = "CONCEPTS: \n" # collect concepts concepts = [] for doc in documents: if isinstance(doc, Document): if doc.metadata["type"] == "content": concepts.append(doc.page_content) # select maximum possible number of concepts to fit in context window filtered_concepts = [] max_width -= len(context) for concept in concepts: if len(concept) < max_width: filtered_concepts.append(concept) max_width -= len(concept) - 1 else: break # format concepts concepts = "\n".join([concept for concept in filtered_concepts]) context += concepts return context rank_function = fn template = "{header}\n{concepts}\n\n{task}" max_width -= len(template.format(header="", concepts="", task="")) + len(header) + len(task) parameters = { "query": query, "documents": documents, "max_width": max_width } if additional_args is not {} and additional_args is not None: parameters.update(additional_args) concepts = rank_function(**{k: v for k, v in parameters.items() if k in rank_function.__code__.co_varnames}) prompt_template = PromptTemplate( input_variables = ["header", "concepts", "task"], template = template ) prompt = prompt_template.format(header=header, concepts=concepts, task=task) return prompt # Concepts search and transform @staticmethod def rank_by_concepts(query, documents, max_width=1024): # context variable context = "CONCEPTS: \n" # collect concepts concepts = [] for doc in documents: if isinstance(doc, Document): if doc.metadata["type"] == "content": concepts.append(doc.page_content) # compute score of concepts scores = [] for concept in concepts: score = GloLoader.calculate_score(query, concept, n=1, scope="word") scores.append( { "concept": concept, "score": score } ) # sort concepts by score sorted_concepts = sorted(scores, key=lambda x: x["score"], reverse=True) # select maximum possible number of concepts to fit in context window filtered_concepts = [] max_width -= len(context) for concept in sorted_concepts: if len(concept["concept"]) < max_width: filtered_concepts.append(concept) max_width -= len(concept["concept"]) - 1 else: break # format concepts concepts = "\n".join([concept["concept"] for concept in filtered_concepts]) context += concepts return context # links search and transform @staticmethod def rank_by_links(query, documents, max_width=1024, chunk_size=512, chunk_overlap=128): # context variable context = "CONCEPTS: \n" # text splitter text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=chunk_overlap) link_chunks = [] # split texts for doc in documents: if isinstance(doc, Document): link_documents = doc.metadata.get("links", []) for link_doc in link_documents: if link_doc.metadata["type"] in ["link-html", "link-pdf", "link-text"]: if not link_doc.metadata["load_status"]: link_doc = GloLoader.load_link( link=link_doc.metadata["source"], type=link_doc.metadata["link_type"], concept_name=doc.metadata["concept"], load=True ) link_chunks.extend(text_splitter.split_documents([link_doc])) # build vector store if len(link_chunks) == 0: similar_texts = [] else: links_vectorstore = FAISS.from_documents(link_chunks, OpenAIEmbeddings()) relevant_chunks = links_vectorstore.similarity_search(query) similar_texts = [doc.page_content for doc in relevant_chunks] # select maximum possible number of chunks to fit in context window filtered_chunks = [] max_width -= len(context) for chunk in similar_texts: if len(chunk) < max_width: filtered_chunks.append(chunk) max_width -= len(chunk) - 1 else: break # format chunks chunks = "\n".join([chunk for chunk in filtered_chunks]) context += chunks return context # concepts+links search and transform @staticmethod def rank_by_concepts_and_links(query, documents, max_width=1024, chunk_size=512, chunk_overlap=128): # context variable context = "CONCEPTS: \n" # collect concepts concepts = [] for doc in documents: if isinstance(doc, Document): if doc.metadata["type"] == "content": concepts.append(doc) # compute score of concepts scores = [] for concept in concepts: score = GloLoader.calculate_score(query, concept.page_content, n=1, scope="word") scores.append( { "concept": concept, "score": score } ) # sort concepts by score sorted_concepts = sorted(scores, key=lambda x: x["score"], reverse=True) # text splitter text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=chunk_overlap) link_chunks = [] # select maximum possible number of chunks to fit in context window filtered_concepts = [] max_width -= len(context) for concept in sorted_concepts: if len(concept["concept"].page_content) < max_width: # link qa here link_documents = concept["concept"].metadata.get("links", []) for link_doc in link_documents: if not link_doc.metadata["load_status"]: link_doc = GloLoader.load_link( link=link_doc.metadata["source"], type=link_doc.metadata["link_type"], concept_name=concept["concept"].metadata["concept"], load=True ) if link_doc.metadata["type"] in ["link-html", "link-pdf", "link-text"]: link_chunks.extend(text_splitter.split_documents([link_doc])) # build vector store if len(link_chunks) == 0: similar_texts = [] else: links_vectorstore = FAISS.from_documents(link_chunks, OpenAIEmbeddings()) relevant_chunks = links_vectorstore.similarity_search(query) similar_texts = [doc.page_content for doc in relevant_chunks] filtered_concepts.append(concept["concept"].page_content + "\nEXCERPTS: \n" + "\n".join(similar_texts[:3])) max_width -= len(filtered_concepts[-1]) else: break # format chunks concepts = "\n".join([concept for concept in filtered_concepts]) context += concepts return context
[ "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.docstore.document.Document", "langchain.document_loaders.OnlinePDFLoader", "langchain.document_loaders.WebBaseLoader", "langchain.embeddings.OpenAIEmbeddings", "langchain.PromptTemplate" ]
[((1437, 1462), 'os.path.isfile', 'os.path.isfile', (['file_path'], {}), '(file_path)\n', (1451, 1462), False, 'import os\n'), ((2579, 2592), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (2587, 2592), False, 'from urllib.parse import urlparse\n'), ((12738, 12824), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['header', 'concepts', 'task']", 'template': 'template'}), "(input_variables=['header', 'concepts', 'task'], template=\n template)\n", (12752, 12824), False, 'from langchain import PromptTemplate\n'), ((14716, 14803), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': 'chunk_size', 'chunk_overlap': 'chunk_overlap'}), '(chunk_size=chunk_size, chunk_overlap=\n chunk_overlap)\n', (14746, 14803), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((17476, 17563), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': 'chunk_size', 'chunk_overlap': 'chunk_overlap'}), '(chunk_size=chunk_size, chunk_overlap=\n chunk_overlap)\n', (17506, 17563), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((1395, 1424), 'os.path.expanduser', 'os.path.expanduser', (['file_path'], {}), '(file_path)\n', (1413, 1424), False, 'import os\n'), ((2780, 2793), 'urllib.parse.urlparse', 'urlparse', (['url'], {}), '(url)\n', (2788, 2793), False, 'from urllib.parse import urlparse\n'), ((3046, 3064), 'requests.head', 'requests.head', (['url'], {}), '(url)\n', (3059, 3064), False, 'import requests\n'), ((1604, 1633), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (1631, 1633), False, 'import tempfile\n'), ((1695, 1722), 'os.path.splitext', 'os.path.splitext', (['file_path'], {}), '(file_path)\n', (1711, 1722), False, 'import os\n'), ((1746, 1789), 'os.path.join', 'os.path.join', (['temp_dir.name', 'f"""tmp{suffix}"""'], {}), "(temp_dir.name, f'tmp{suffix}')\n", (1758, 1789), False, 'import os\n'), ((3757, 3781), 'yaml.safe_load', 'yaml.safe_load', (['glo_text'], {}), '(glo_text)\n', (3771, 3781), False, 'import yaml\n'), ((5906, 6074), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': "(link if load else '')", 'metadata': "{'concept': concept_name, 'type': 'link-text', 'link_type': 'text',\n 'source': 'text', 'load_status': load}"}), "(page_content=link if load else '', metadata={'concept':\n concept_name, 'type': 'link-text', 'link_type': 'text', 'source':\n 'text', 'load_status': load})\n", (5914, 6074), False, 'from langchain.docstore.document import Document\n'), ((15796, 15814), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (15812, 15814), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((1909, 1932), 'requests.get', 'requests.get', (['file_path'], {}), '(file_path)\n', (1921, 1932), False, 'import requests\n'), ((3980, 4000), 'json.load', 'json.load', (['json_file'], {}), '(json_file)\n', (3989, 4000), False, 'import json\n'), ((4530, 4551), 'langchain.document_loaders.OnlinePDFLoader', 'OnlinePDFLoader', (['link'], {}), '(link)\n', (4545, 4551), False, 'from langchain.document_loaders import OnlinePDFLoader\n'), ((4646, 4690), 're.sub', 're.sub', (['"""\\\\n+"""', '"""\n"""', 'data_load.page_content'], {}), "('\\\\n+', '\\n', data_load.page_content)\n", (4652, 4690), False, 'import re\n'), ((4745, 4770), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': '""""""'}), "(page_content='')\n", (4753, 4770), False, 'from langchain.docstore.document import Document\n'), ((5163, 5182), 'langchain.document_loaders.WebBaseLoader', 'WebBaseLoader', (['link'], {}), '(link)\n', (5176, 5182), False, 'from langchain.document_loaders import WebBaseLoader\n'), ((5285, 5329), 're.sub', 're.sub', (['"""\\\\n+"""', '"""\n"""', 'data_load.page_content'], {}), "('\\\\n+', '\\n', data_load.page_content)\n", (5291, 5329), False, 'import re\n'), ((5392, 5417), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': '""""""'}), "(page_content='')\n", (5400, 5417), False, 'from langchain.docstore.document import Document\n'), ((18806, 18824), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (18822, 18824), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((7473, 7502), 're.sub', 're.sub', (['"""\\\\n+"""', '"""\n"""', 'content'], {}), "('\\\\n+', '\\n', content)\n", (7479, 7502), False, 'import re\n'), ((9536, 9567), 're.sub', 're.sub', (['pattern', '""" \\\\1"""', 'source'], {}), "(pattern, ' \\\\1', source)\n", (9542, 9567), False, 'import re\n'), ((9615, 9646), 're.sub', 're.sub', (['pattern', '""" \\\\1"""', 'sample'], {}), "(pattern, ' \\\\1', sample)\n", (9621, 9646), False, 'import re\n')]
from langchain import OpenAI, SQLDatabase from langchain_experimental.sql import SQLDatabaseChain from langchain.memory import ConversationBufferMemory from langchain.agents import (AgentType, AgentExecutor, create_react_agent, create_openai_functions_agent, create_sql_agent) from langchain.agents.initialize import initialize_agent from langchain.tools import Tool from langchain.prompts import PromptTemplate # from langchain.retrievers import RetrievalModel from langchain.indexes.vectorstore import VectorstoreIndexCreator from langchain_community.document_loaders.text import TextLoader from langchain_community.agent_toolkits import SQLDatabaseToolkit # Document Loaders for PDFs and Docx from langchain.document_loaders.pdf import PyPDFLoader from langchain.document_loaders.word_document import Docx2txtLoader from langchain.embeddings.openai import OpenAIEmbeddings from langchain.vectorstores.chroma import Chroma from langchain.chat_models.openai import ChatOpenAI from langchain.chains import ConversationalRetrievalChain from langchain.text_splitter import CharacterTextSplitter import langchain import psycopg2 import os import constants import constants_allen os.environ["OPENAI_API_KEY"] = constants.OPENAI_API_KEY import warnings warnings.filterwarnings('ignore') # Setup database connection # Setup database db = SQLDatabase.from_uri( f"postgresql+psycopg2://postgres:{constants.DBPASS}@localhost:5433/{constants.DB}", ) # Setup conversation memory memory = ConversationBufferMemory(memory_key="chat_history") # Setup llm llm = OpenAI(temperature=0, model_name="gpt-3.5-turbo") # Create a list of documents from all our files in the ./data/docs folder documents = [] for file in os.listdir("./data/docs"): if file.endswith(".pdf"): pdf_path = "./data/docs/" + file loader = PyPDFLoader(pdf_path) documents.extend(loader.load()) elif file.endswith("./docx") or file.endswith(".doc"): doc_path = "./data/docs/" + file loader = Docx2txtLoader(doc_path) documents.extend(loader.load()) elif file.endswith(".txt"): text_path = "./data/docs/" + file loader = TextLoader(text_path) documents.extend(loader.load()) # Split the documents into smaller chunks text_splitter = CharacterTextSplitter(chunk_size=1000, chunk_overlap=10) documents = text_splitter.split_documents(documents) # Convert the document chunks to embedding and save them to the vector store vectordb = Chroma.from_documents(documents, embedding=OpenAIEmbeddings(), persist_directory="./data/docs") vectordb.persist() # Setup prompt template _DEFAULT_TEMPLATE = """Given an input question, first create a syntactically correct {dialect} query to run, then look at the results of the query and return the answer. Unless the user specifies in his question a specific number of examples he wishes to obtain, always limit your query to at most {top_k} results. You can order the results by a relevant column to return the most interesting examples in the database. Never query for all the columns from a specific table, only ask for a the few relevant columns given the question. Pay attention to use only the column names that you can see in the schema description. Be careful to not query for columns that do not exist. Also, pay attention to which column is in which table. Use the following format: Question: Question here SQLQuery: SQL Query to run SQLResult: Result of the SQLQuery Answer: Final answer here Only use the following tables: {table_info} Question: {input} """ # TEMPLATE = '''"Given your input question, provide the answer from the most relevant source. You can choose between two sources: # 1. Text File Source: This source contains Q&A data on data structures and algorithms. # 2. Database Source: This source includes information from three tables: # - 'developers' table: Contains details about developers, such as full_name, email, phone, position, and department. # - 'tasks' table: Holds information about developers' tasks, including the task, completion status, due date, completion date, and priority. # - 'insurance_data' table: Contains information about PRU Life Insurance and USA Insurance, including questions and answers. # Only query the content of the sources, not the metadata. # You have access to the following tools: # {tools} # Use the following format: # Question: the input question you must answer # Thought: you should always think about what to do # Action: the action to take, should be one of [{tool_names}] # Action Input: the input to the action # Observation: the result of the action # ... (this Thought/Action/Action Input/Observation can repeat N times) # Thought: I now know the final answer # Final Answer: the final answer to the original input question # Begin! # Question: {input} # Thought:{agent_scratchpad}''' PROMPT = PromptTemplate.from_template(_DEFAULT_TEMPLATE) # Setup db chain db_chain = SQLDatabaseChain(llm=llm, database=db, verbose=True) toolkit = SQLDatabaseToolkit(db=db, llm=OpenAI(temperature=0, verbose=True, model_name="gpt-3.5-turbo")) # Create Document Q&A chain pdf_qa = ConversationalRetrievalChain.from_llm( OpenAI(temperate=0.7, model_name="gpt-3.5-turbo"), retriever=vectordb.as_retriever(search_kwargs={'k': 6}), return_source_documents=True, verbose=False ) # Setup text data source directory = "./data/text/final1.txt" loader = TextLoader(directory) index_creator = VectorstoreIndexCreator() index = index_creator.from_loaders([loader]) retriever_text = index.vectorstore.as_retriever() # Define the tools needed tools = [ Tool(name="MyDocs", func=pdf_qa, description="Retrieve information from documents."), Tool(name="MyDB", func=db_chain.run, description="Query the PostgreSQL database",), Tool(name="MyRetriever", func=retriever_text.get_relevant_documents, description="Retrieve documents from text file"), ] agent = initialize_agent( tools, llm, memory=memory, agent=AgentType.CONVERSATIONAL_REACT_DESCRIPTION, verbose=True, handle_parsing_errors=True, output_key="result", early_stopping_method="generate", max_iterations=3, ) # agent = create_sql_agent( # llm=OpenAI(temperature=0), # toolkit=toolkit, # verbose=True, # agent_type=AgentType.CONVERSATIONAL_REACT_DESCRIPTION, # ) # agent = create_react_agent( # llm=llm, # tools=tools, # prompt=PROMPT # ) # agent_executor = AgentExecutor(agent=agent, tools=tools, memory=memory, # verbose=True, handle_parsing_errors=True) yellow = "\033[0;33m" green = "\033[0;32m" white = "\033[0;39m" def chat(): while True: user_input = input("You: ") # Get user input # Use the agent to generate a response try: bot_response = agent.run(user_input) # bot_response = agent_executor.invoke({"input": user_input}) except ValueError as e: bot_response = str(e) if not bot_response.startswith("Could not parse LLM output: `"): raise e bot_response = bot_response.removeprefix("Could not parse LLM output: `").removesuffix("`") print("Chatbot: " + bot_response) if user_input.lower() == 'exit': break if __name__ == "__main__": chat()
[ "langchain.text_splitter.CharacterTextSplitter", "langchain_experimental.sql.SQLDatabaseChain", "langchain.agents.initialize.initialize_agent", "langchain.tools.Tool", "langchain.document_loaders.pdf.PyPDFLoader", "langchain.memory.ConversationBufferMemory", "langchain_community.document_loaders.text.TextLoader", "langchain.document_loaders.word_document.Docx2txtLoader", "langchain.indexes.vectorstore.VectorstoreIndexCreator", "langchain.prompts.PromptTemplate.from_template", "langchain.SQLDatabase.from_uri", "langchain.OpenAI", "langchain.embeddings.openai.OpenAIEmbeddings" ]
[((1372, 1405), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {}), "('ignore')\n", (1395, 1405), False, 'import warnings\n'), ((1458, 1572), 'langchain.SQLDatabase.from_uri', 'SQLDatabase.from_uri', (['f"""postgresql+psycopg2://postgres:{constants.DBPASS}@localhost:5433/{constants.DB}"""'], {}), "(\n f'postgresql+psycopg2://postgres:{constants.DBPASS}@localhost:5433/{constants.DB}'\n )\n", (1478, 1572), False, 'from langchain import OpenAI, SQLDatabase\n'), ((1608, 1659), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""'}), "(memory_key='chat_history')\n", (1632, 1659), False, 'from langchain.memory import ConversationBufferMemory\n'), ((1679, 1728), 'langchain.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, model_name='gpt-3.5-turbo')\n", (1685, 1728), False, 'from langchain import OpenAI, SQLDatabase\n'), ((1831, 1856), 'os.listdir', 'os.listdir', (['"""./data/docs"""'], {}), "('./data/docs')\n", (1841, 1856), False, 'import os\n'), ((2402, 2458), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {'chunk_size': '(1000)', 'chunk_overlap': '(10)'}), '(chunk_size=1000, chunk_overlap=10)\n', (2423, 2458), False, 'from langchain.text_splitter import CharacterTextSplitter\n'), ((5005, 5052), 'langchain.prompts.PromptTemplate.from_template', 'PromptTemplate.from_template', (['_DEFAULT_TEMPLATE'], {}), '(_DEFAULT_TEMPLATE)\n', (5033, 5052), False, 'from langchain.prompts import PromptTemplate\n'), ((5082, 5134), 'langchain_experimental.sql.SQLDatabaseChain', 'SQLDatabaseChain', ([], {'llm': 'llm', 'database': 'db', 'verbose': '(True)'}), '(llm=llm, database=db, verbose=True)\n', (5098, 5134), False, 'from langchain_experimental.sql import SQLDatabaseChain\n'), ((5559, 5580), 'langchain_community.document_loaders.text.TextLoader', 'TextLoader', (['directory'], {}), '(directory)\n', (5569, 5580), False, 'from langchain_community.document_loaders.text import TextLoader\n'), ((5597, 5622), 'langchain.indexes.vectorstore.VectorstoreIndexCreator', 'VectorstoreIndexCreator', ([], {}), '()\n', (5620, 5622), False, 'from langchain.indexes.vectorstore import VectorstoreIndexCreator\n'), ((6067, 6289), 'langchain.agents.initialize.initialize_agent', 'initialize_agent', (['tools', 'llm'], {'memory': 'memory', 'agent': 'AgentType.CONVERSATIONAL_REACT_DESCRIPTION', 'verbose': '(True)', 'handle_parsing_errors': '(True)', 'output_key': '"""result"""', 'early_stopping_method': '"""generate"""', 'max_iterations': '(3)'}), "(tools, llm, memory=memory, agent=AgentType.\n CONVERSATIONAL_REACT_DESCRIPTION, verbose=True, handle_parsing_errors=\n True, output_key='result', early_stopping_method='generate',\n max_iterations=3)\n", (6083, 6289), False, 'from langchain.agents.initialize import initialize_agent\n'), ((5321, 5370), 'langchain.OpenAI', 'OpenAI', ([], {'temperate': '(0.7)', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperate=0.7, model_name='gpt-3.5-turbo')\n", (5327, 5370), False, 'from langchain import OpenAI, SQLDatabase\n'), ((5759, 5848), 'langchain.tools.Tool', 'Tool', ([], {'name': '"""MyDocs"""', 'func': 'pdf_qa', 'description': '"""Retrieve information from documents."""'}), "(name='MyDocs', func=pdf_qa, description=\n 'Retrieve information from documents.')\n", (5763, 5848), False, 'from langchain.tools import Tool\n'), ((5849, 5935), 'langchain.tools.Tool', 'Tool', ([], {'name': '"""MyDB"""', 'func': 'db_chain.run', 'description': '"""Query the PostgreSQL database"""'}), "(name='MyDB', func=db_chain.run, description=\n 'Query the PostgreSQL database')\n", (5853, 5935), False, 'from langchain.tools import Tool\n'), ((5937, 6058), 'langchain.tools.Tool', 'Tool', ([], {'name': '"""MyRetriever"""', 'func': 'retriever_text.get_relevant_documents', 'description': '"""Retrieve documents from text file"""'}), "(name='MyRetriever', func=retriever_text.get_relevant_documents,\n description='Retrieve documents from text file')\n", (5941, 6058), False, 'from langchain.tools import Tool\n'), ((1946, 1967), 'langchain.document_loaders.pdf.PyPDFLoader', 'PyPDFLoader', (['pdf_path'], {}), '(pdf_path)\n', (1957, 1967), False, 'from langchain.document_loaders.pdf import PyPDFLoader\n'), ((2644, 2662), 'langchain.embeddings.openai.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {}), '()\n', (2660, 2662), False, 'from langchain.embeddings.openai import OpenAIEmbeddings\n'), ((5175, 5238), 'langchain.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'verbose': '(True)', 'model_name': '"""gpt-3.5-turbo"""'}), "(temperature=0, verbose=True, model_name='gpt-3.5-turbo')\n", (5181, 5238), False, 'from langchain import OpenAI, SQLDatabase\n'), ((2125, 2149), 'langchain.document_loaders.word_document.Docx2txtLoader', 'Docx2txtLoader', (['doc_path'], {}), '(doc_path)\n', (2139, 2149), False, 'from langchain.document_loaders.word_document import Docx2txtLoader\n'), ((2281, 2302), 'langchain_community.document_loaders.text.TextLoader', 'TextLoader', (['text_path'], {}), '(text_path)\n', (2291, 2302), False, 'from langchain_community.document_loaders.text import TextLoader\n')]
"""Base interface that all chains should implement.""" import json from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import BaseModel, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import BaseMemory def _get_verbosity() -> bool: return langchain.verbose class Chain(BaseModel, ABC): """Base interface that all chains should implement.""" memory: Optional[BaseMemory] = None callback_manager: BaseCallbackManager = Field( default_factory=get_callback_manager, exclude=True ) verbose: bool = Field( default_factory=_get_verbosity ) # Whether to print the response text class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Input keys this chain expects.""" @property @abstractmethod def output_keys(self) -> List[str]: """Output keys this chain expects.""" def _validate_inputs(self, inputs: Dict[str, str]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, str]) -> None: if set(outputs) != set(self.output_keys): raise ValueError( f"Did not get output keys that were expected. " f"Got: {set(outputs)}. Expected: {set(self.output_keys)}." ) @abstractmethod def _call(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" async def _acall(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = self._call(inputs) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) if self.callback_manager.is_async: await self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) else: self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = await self._acall(inputs) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_chain_error(e, verbose=self.verbose) else: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_chain_end(outputs, verbose=self.verbose) else: self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prep outputs.""" self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prep inputs.""" if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs def apply(self, input_list: List[Dict[str, Any]]) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs) for inputs in input_list] def run(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0])[self.output_keys[0]] if kwargs and not args: return self(kwargs)[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return (await self.acall(args[0]))[self.output_keys[0]] if kwargs and not args: return (await self.acall(kwargs))[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain.""" if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict() _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml")
[ "langchain.callbacks.get_callback_manager" ]
[((646, 703), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager', 'exclude': '(True)'}), '(default_factory=get_callback_manager, exclude=True)\n', (651, 703), False, 'from pydantic import BaseModel, Field, validator\n'), ((738, 775), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (743, 775), False, 'from pydantic import BaseModel, Field, validator\n'), ((1075, 1127), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (1084, 1127), False, 'from pydantic import BaseModel, Field, validator\n'), ((1457, 1500), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (1466, 1500), False, 'from pydantic import BaseModel, Field, validator\n'), ((1428, 1450), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1448, 1450), False, 'from langchain.callbacks import get_callback_manager\n'), ((10078, 10093), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (10082, 10093), False, 'from pathlib import Path\n'), ((10413, 10447), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (10422, 10447), False, 'import json\n'), ((10550, 10600), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (10559, 10600), False, 'import yaml\n')]
from typing import Dict, List, Optional from langchain.agents.load_tools import ( _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS, ) from langflow.custom import customs from langflow.interface.base import LangChainTypeCreator from langflow.interface.tools.constants import ( ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS, ) from langflow.interface.tools.util import get_tool_params from langflow.services.utils import get_settings_manager from langflow.template.field.base import TemplateField from langflow.template.template.base import Template from langflow.utils import util from langflow.utils.util import build_template_from_class TOOL_INPUTS = { "str": TemplateField( field_type="str", required=True, is_list=False, show=True, placeholder="", value="", ), "llm": TemplateField( field_type="BaseLanguageModel", required=True, is_list=False, show=True ), "func": TemplateField( field_type="function", required=True, is_list=False, show=True, multiline=True, ), "code": TemplateField( field_type="str", required=True, is_list=False, show=True, value="", multiline=True, ), "path": TemplateField( field_type="file", required=True, is_list=False, show=True, value="", suffixes=[".json", ".yaml", ".yml"], file_types=["json", "yaml", "yml"], ), } class ToolCreator(LangChainTypeCreator): type_name: str = "tools" tools_dict: Optional[Dict] = None @property def type_to_loader_dict(self) -> Dict: settings_manager = get_settings_manager() if self.tools_dict is None: all_tools = {} for tool, tool_fcn in ALL_TOOLS_NAMES.items(): tool_params = get_tool_params(tool_fcn) tool_name = tool_params.get("name") or tool if ( tool_name in settings_manager.settings.TOOLS or settings_manager.settings.DEV ): if tool_name == "JsonSpec": tool_params["path"] = tool_params.pop("dict_") # type: ignore all_tools[tool_name] = { "type": tool, "params": tool_params, "fcn": tool_fcn, } self.tools_dict = all_tools return self.tools_dict def get_signature(self, name: str) -> Optional[Dict]: """Get the signature of a tool.""" base_classes = ["Tool", "BaseTool"] fields = [] params = [] tool_params = {} # Raise error if name is not in tools if name not in self.type_to_loader_dict.keys(): raise ValueError("Tool not found") tool_type: str = self.type_to_loader_dict[name]["type"] # type: ignore # if tool_type in _BASE_TOOLS.keys(): # params = [] if tool_type in _LLM_TOOLS.keys(): params = ["llm"] elif tool_type in _EXTRA_LLM_TOOLS.keys(): extra_keys = _EXTRA_LLM_TOOLS[tool_type][1] params = ["llm"] + extra_keys elif tool_type in _EXTRA_OPTIONAL_TOOLS.keys(): extra_keys = _EXTRA_OPTIONAL_TOOLS[tool_type][1] params = extra_keys # elif tool_type == "Tool": # params = ["name", "description", "func"] elif tool_type in CUSTOM_TOOLS: # Get custom tool params params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes = ["function"] if node := customs.get_custom_nodes("tools").get(tool_type): return node elif tool_type in FILE_TOOLS: params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes += [name] elif tool_type in OTHER_TOOLS: tool_dict = build_template_from_class(tool_type, OTHER_TOOLS) fields = tool_dict["template"] # Pop unnecessary fields and add name fields.pop("_type") # type: ignore fields.pop("return_direct") # type: ignore fields.pop("verbose") # type: ignore tool_params = { "name": fields.pop("name")["value"], # type: ignore "description": fields.pop("description")["value"], # type: ignore } fields = [ TemplateField(name=name, field_type=field["type"], **field) for name, field in fields.items() # type: ignore ] base_classes += tool_dict["base_classes"] # Copy the field and add the name for param in params: field = TOOL_INPUTS.get(param, TOOL_INPUTS["str"]).copy() field.name = param field.advanced = False if param == "aiosession": field.show = False field.required = False fields.append(field) template = Template(fields=fields, type_name=tool_type) tool_params = {**tool_params, **self.type_to_loader_dict[name]["params"]} return { "template": util.format_dict(template.to_dict()), **tool_params, "base_classes": base_classes, } def to_list(self) -> List[str]: """List all load tools""" return list(self.type_to_loader_dict.keys()) tool_creator = ToolCreator()
[ "langchain.agents.load_tools._LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys" ]
[((709, 811), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'placeholder': '""""""', 'value': '""""""'}), "(field_type='str', required=True, is_list=False, show=True,\n placeholder='', value='')\n", (722, 811), False, 'from langflow.template.field.base import TemplateField\n'), ((875, 965), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""BaseLanguageModel"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)'}), "(field_type='BaseLanguageModel', required=True, is_list=False,\n show=True)\n", (888, 965), False, 'from langflow.template.field.base import TemplateField\n'), ((989, 1087), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""function"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'multiline': '(True)'}), "(field_type='function', required=True, is_list=False, show=\n True, multiline=True)\n", (1002, 1087), False, 'from langflow.template.field.base import TemplateField\n'), ((1143, 1245), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'multiline': '(True)'}), "(field_type='str', required=True, is_list=False, show=True,\n value='', multiline=True)\n", (1156, 1245), False, 'from langflow.template.field.base import TemplateField\n'), ((1310, 1474), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""file"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'suffixes': "['.json', '.yaml', '.yml']", 'file_types': "['json', 'yaml', 'yml']"}), "(field_type='file', required=True, is_list=False, show=True,\n value='', suffixes=['.json', '.yaml', '.yml'], file_types=['json',\n 'yaml', 'yml'])\n", (1323, 1474), False, 'from langflow.template.field.base import TemplateField\n'), ((1728, 1750), 'langflow.services.utils.get_settings_manager', 'get_settings_manager', ([], {}), '()\n', (1748, 1750), False, 'from langflow.services.utils import get_settings_manager\n'), ((5139, 5183), 'langflow.template.template.base.Template', 'Template', ([], {'fields': 'fields', 'type_name': 'tool_type'}), '(fields=fields, type_name=tool_type)\n', (5147, 5183), False, 'from langflow.template.template.base import Template\n'), ((1849, 1872), 'langflow.interface.tools.constants.ALL_TOOLS_NAMES.items', 'ALL_TOOLS_NAMES.items', ([], {}), '()\n', (1870, 1872), False, 'from langflow.interface.tools.constants import ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS\n'), ((3091, 3108), 'langchain.agents.load_tools._LLM_TOOLS.keys', '_LLM_TOOLS.keys', ([], {}), '()\n', (3106, 3108), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((1904, 1929), 'langflow.interface.tools.util.get_tool_params', 'get_tool_params', (['tool_fcn'], {}), '(tool_fcn)\n', (1919, 1929), False, 'from langflow.interface.tools.util import get_tool_params\n'), ((3165, 3188), 'langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys', '_EXTRA_LLM_TOOLS.keys', ([], {}), '()\n', (3186, 3188), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((3314, 3342), 'langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys', '_EXTRA_OPTIONAL_TOOLS.keys', ([], {}), '()\n', (3340, 3342), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((4038, 4087), 'langflow.utils.util.build_template_from_class', 'build_template_from_class', (['tool_type', 'OTHER_TOOLS'], {}), '(tool_type, OTHER_TOOLS)\n', (4063, 4087), False, 'from langflow.utils.util import build_template_from_class\n'), ((3746, 3779), 'langflow.custom.customs.get_custom_nodes', 'customs.get_custom_nodes', (['"""tools"""'], {}), "('tools')\n", (3770, 3779), False, 'from langflow.custom import customs\n'), ((4571, 4630), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'name': 'name', 'field_type': "field['type']"}), "(name=name, field_type=field['type'], **field)\n", (4584, 4630), False, 'from langflow.template.field.base import TemplateField\n')]
#!/usr/bin/env python3 from restapi_helper import LangChainHelper from langchain.schema import HumanMessage print('==Simple message predict==') with LangChainHelper() as lch: text = 'Hey there!' messages = [HumanMessage(content=text)] print(lch.predict_messages(messages)) print('==As English teacher - know the answer==') with LangChainHelper() as lch: role = 'English teacher' text = 'What is a plural form of child?' print(lch.get_answer_as_role(role, text)) print('==As English teacher - does not know the answer==') with LangChainHelper() as lch: role = 'English teacher' text = 'How old am I?' print(lch.get_answer_as_role(role, text))
[ "langchain.schema.HumanMessage" ]
[((157, 174), 'restapi_helper.LangChainHelper', 'LangChainHelper', ([], {}), '()\n', (172, 174), False, 'from restapi_helper import LangChainHelper\n'), ((355, 372), 'restapi_helper.LangChainHelper', 'LangChainHelper', ([], {}), '()\n', (370, 372), False, 'from restapi_helper import LangChainHelper\n'), ((572, 589), 'restapi_helper.LangChainHelper', 'LangChainHelper', ([], {}), '()\n', (587, 589), False, 'from restapi_helper import LangChainHelper\n'), ((225, 251), 'langchain.schema.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (237, 251), False, 'from langchain.schema import HumanMessage\n')]
from typing import Optional from langchain.chains.openai_functions import create_structured_output_runnable from langchain_core.prompts import ChatPromptTemplate from langchain_core.pydantic_v1 import BaseModel, Field import logging import langchain from langchain_community.vectorstores import FAISS from langchain_core.output_parsers import StrOutputParser from langchain_core.prompts import ChatPromptTemplate from langchain_core.runnables import RunnablePassthrough from langchain_openai import ChatOpenAI, OpenAIEmbeddings langchain.verbose = True logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) logging.basicConfig(level=logging.DEBUG) # 控制台打印 handler = logging.StreamHandler() handler.setLevel(logging.DEBUG) logger.addHandler(handler) def test_runnable_parallel_chain() -> None: """Test create_structured_output_runnable. 测试创建结构化输出可运行对象。""" from langchain_core.prompts import ChatPromptTemplate from langchain_core.runnables import RunnableParallel from langchain_openai import ChatOpenAI llm = ChatOpenAI( openai_api_base='http://127.0.0.1:30000/v1', model="glm-4", openai_api_key="glm-4", verbose=True, # temperature=0.95, ) joke_chain = ChatPromptTemplate.from_template("tell me a joke about {topic}") | llm | StrOutputParser() poem_chain = ( ChatPromptTemplate.from_template("write a 2-line poem about {topic}") | llm ) map_chain = RunnableParallel(joke=joke_chain, poem=poem_chain) out = map_chain.invoke({"topic": "bear"}) print(out)
[ "langchain_core.prompts.ChatPromptTemplate.from_template", "langchain_openai.ChatOpenAI", "langchain_core.runnables.RunnableParallel", "langchain_core.output_parsers.StrOutputParser" ]
[((565, 592), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (582, 592), False, 'import logging\n'), ((624, 664), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (643, 664), False, 'import logging\n'), ((683, 706), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (704, 706), False, 'import logging\n'), ((1051, 1163), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_base': '"""http://127.0.0.1:30000/v1"""', 'model': '"""glm-4"""', 'openai_api_key': '"""glm-4"""', 'verbose': '(True)'}), "(openai_api_base='http://127.0.0.1:30000/v1', model='glm-4',\n openai_api_key='glm-4', verbose=True)\n", (1061, 1163), False, 'from langchain_openai import ChatOpenAI\n'), ((1465, 1515), 'langchain_core.runnables.RunnableParallel', 'RunnableParallel', ([], {'joke': 'joke_chain', 'poem': 'poem_chain'}), '(joke=joke_chain, poem=poem_chain)\n', (1481, 1515), False, 'from langchain_core.runnables import RunnableParallel\n'), ((1317, 1334), 'langchain_core.output_parsers.StrOutputParser', 'StrOutputParser', ([], {}), '()\n', (1332, 1334), False, 'from langchain_core.output_parsers import StrOutputParser\n'), ((1366, 1435), 'langchain_core.prompts.ChatPromptTemplate.from_template', 'ChatPromptTemplate.from_template', (['"""write a 2-line poem about {topic}"""'], {}), "('write a 2-line poem about {topic}')\n", (1398, 1435), False, 'from langchain_core.prompts import ChatPromptTemplate\n'), ((1244, 1308), 'langchain_core.prompts.ChatPromptTemplate.from_template', 'ChatPromptTemplate.from_template', (['"""tell me a joke about {topic}"""'], {}), "('tell me a joke about {topic}')\n", (1276, 1308), False, 'from langchain_core.prompts import ChatPromptTemplate\n')]
import streamlit as st # Import transformer classes for generaiton from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer, GPT2Tokenizer, GPT2LMHeadModel, GPT2Model # Import torch for datatype attributes import torch # Import the prompt wrapper...but for llama index from llama_index.prompts.prompts import SimpleInputPrompt # Import the llama index HF Wrapper from llama_index.llms import HuggingFaceLLM # Bring in embeddings wrapper from llama_index.embeddings import LangchainEmbedding # Bring in HF embeddings - need these to represent document chunks from langchain.embeddings.huggingface import HuggingFaceEmbeddings # Bring in stuff to change service context from llama_index import set_global_service_context from llama_index import ServiceContext # Import deps to load documents from llama_index import VectorStoreIndex, SimpleDirectoryReader from pathlib import Path import pypdf import time import os import tempfile # Define variable to hold llama2 weights namingfiner name = "gpt2" # Set auth token variable from hugging face auth_token = "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" @st.cache_resource def get_tokenizer_model(): # Create tokenizer tokenizer = GPT2Tokenizer.from_pretrained(name, use_auth_token=auth_token) # Create model model = GPT2LMHeadModel.from_pretrained(name) return model, tokenizer model, tokenizer = get_tokenizer_model() # disclaimer st.title('LLM Deployment Prototype for Production') st.caption("Special thanks to my mentor, Medkham Chanthavong, for all the support in making this project a reality. Connect with Medkham on LinkedIn: [Medkham Chanthavong](https://www.linkedin.com/in/medkham-chanthavong-4524b0125/).") st.subheader('Disclaimer') st.write("Due to Streamlit's lack of cloud-GPU support and limited memory, we're using GPT-2 instead of Llama 2 7B. GPT-2, \ being smaller, is less capable of accurately retrieving information from external data. This site is a prototype for LLM deployment \ and may not retrieve precise data from external sources. Additionally, the site might experience crashes due to memory constraints.\ In case of a crash, please feel free shoot me a message at my [LinkedIn profile](https://www.linkedin.com/in/duwe-ng/) so I can reboot the site. \ For a more advanced model demonstration using Llama 2 7B, check out our Colab notebook: \ [LLM Deployment Prototype](https://colab.research.google.com/drive/1bGf9rKntMjH4KtpKs9ryucj1nbiKs_zk?usp=sharing).") # # Initialize the SimpleInputPrompt with an empty template # query_wrapper_prompt = SimpleInputPrompt("{query_str}") # # Streamlit UI to let the user update the system prompt # # Start with an empty string or a default prompt # default_prompt = "" # user_system_prompt = st.text_area("How can I best assist you?", value="", height=100) # update_button = st.button('Request') # Import the prompt wrapper for llama index # Create a system prompt system_prompt = """ Hey there! 👋 You are here to help you out in the best way you can. Think of yourself as your friendly and trustworthy assistant. Your top priority is to be super helpful and safe in your responses. \ Here's a quick heads up about what you can expect from me: \ 1. Positive Vibes Only: I steer clear of anything harmful, unethical, or offensive. No racism, sexism, or toxic stuff here. I'm all about being respectful and keeping things on the up-and-up. \ 2. Making Sense Matters: If the user question seems a bit confusing or doesn't quite add up, You will let you know and try to clarify things instead of giving the user a misleading answer. \ 3. Honesty is Key: Not sure about something? You won't make stuff up. If you don't have the answer, you will be upfront about it. \ 4. All About Your PDFs: The user documents are your focus. Got a question about the content in your PDFs? That's exactly what you are here for. Let's dive into those documents and find the answers the user need! \ So, how can you assist the user today with their PDFs? """ # Throw together the query wrapper query_wrapper_prompt = SimpleInputPrompt("{query_str}") # Initialize the llm object with a placeholder or default system prompt llm = HuggingFaceLLM( context_window=1024, max_new_tokens=128, system_prompt="", query_wrapper_prompt=query_wrapper_prompt, model=model, tokenizer=tokenizer ) # # Function to update the system prompt and reinitialize the LLM with the new prompt # def update_system_prompt(new_prompt): # global llm # llm.system_prompt = new_prompt # if update_button: # # Update the system prompt and reinitialize the LLM # update_system_prompt(user_system_prompt) # st.success('Requested') # check if user request save to the memory # st.write(llm.system_prompt) # Create and dl embeddings instance embeddings=LangchainEmbedding( HuggingFaceEmbeddings(model_name="all-MiniLM-L6-v2") ) # Create new service context instance service_context = ServiceContext.from_defaults( chunk_size=256, llm=llm, embed_model=embeddings ) # And set the service context set_global_service_context(service_context) # Upload PDF and process it st.subheader("Please upload your data to enable the retrieval system; otherwise, it will not respond.") documents = [] uploaded_file = st.file_uploader("Upload PDF", type="pdf") if uploaded_file is not None: # Create a temporary directory with tempfile.TemporaryDirectory() as temp_dir: # Write the uploaded file to a file in the temporary directory temp_file_path = os.path.join(temp_dir, uploaded_file.name) with open(temp_file_path, "wb") as f: f.write(uploaded_file.getbuffer()) # Now you can use SimpleDirectoryReader on the temp_dir documents = SimpleDirectoryReader(temp_dir).load_data() st.success(f"File {uploaded_file.name} uploaded successfully.") index = VectorStoreIndex.from_documents(documents) # Setup index query engine using LLM query_engine = index.as_query_engine(streaming=True, similarity_top_k=1) # Create centered main title st.title('👔 HireMind 🧩') # setup a session to hold all the old prompt if 'messages' not in st.session_state: st.session_state.messages = [] # print out the history message for message in st.session_state.messages: st.chat_message(message['role']).markdown(message['content']) # Create a text input box for the user # If the user hits enter prompt = st.chat_input('Input your prompt here') if prompt: st.chat_message('user').markdown(prompt) st.session_state.messages.append({'role': 'user', 'content': prompt}) response = query_engine.query(prompt) st.chat_message('assistant').markdown(response) st.session_state.messages.append( {'role': 'assistant', 'content': response} )
[ "langchain.embeddings.huggingface.HuggingFaceEmbeddings" ]
[((1422, 1473), 'streamlit.title', 'st.title', (['"""LLM Deployment Prototype for Production"""'], {}), "('LLM Deployment Prototype for Production')\n", (1430, 1473), True, 'import streamlit as st\n'), ((1474, 1718), 'streamlit.caption', 'st.caption', (['"""Special thanks to my mentor, Medkham Chanthavong, for all the support in making this project a reality. Connect with Medkham on LinkedIn: [Medkham Chanthavong](https://www.linkedin.com/in/medkham-chanthavong-4524b0125/)."""'], {}), "(\n 'Special thanks to my mentor, Medkham Chanthavong, for all the support in making this project a reality. Connect with Medkham on LinkedIn: [Medkham Chanthavong](https://www.linkedin.com/in/medkham-chanthavong-4524b0125/).'\n )\n", (1484, 1718), True, 'import streamlit as st\n'), ((1710, 1736), 'streamlit.subheader', 'st.subheader', (['"""Disclaimer"""'], {}), "('Disclaimer')\n", (1722, 1736), True, 'import streamlit as st\n'), ((1737, 2482), 'streamlit.write', 'st.write', (['"""Due to Streamlit\'s lack of cloud-GPU support and limited memory, we\'re using GPT-2 instead of Llama 2 7B. GPT-2, being smaller, is less capable of accurately retrieving information from external data. This site is a prototype for LLM deployment and may not retrieve precise data from external sources. Additionally, the site might experience crashes due to memory constraints.In case of a crash, please feel free shoot me a message at my [LinkedIn profile](https://www.linkedin.com/in/duwe-ng/) so I can reboot the site. For a more advanced model demonstration using Llama 2 7B, check out our Colab notebook: [LLM Deployment Prototype](https://colab.research.google.com/drive/1bGf9rKntMjH4KtpKs9ryucj1nbiKs_zk?usp=sharing)."""'], {}), '(\n "Due to Streamlit\'s lack of cloud-GPU support and limited memory, we\'re using GPT-2 instead of Llama 2 7B. GPT-2, being smaller, is less capable of accurately retrieving information from external data. This site is a prototype for LLM deployment and may not retrieve precise data from external sources. Additionally, the site might experience crashes due to memory constraints.In case of a crash, please feel free shoot me a message at my [LinkedIn profile](https://www.linkedin.com/in/duwe-ng/) so I can reboot the site. For a more advanced model demonstration using Llama 2 7B, check out our Colab notebook: [LLM Deployment Prototype](https://colab.research.google.com/drive/1bGf9rKntMjH4KtpKs9ryucj1nbiKs_zk?usp=sharing)."\n )\n', (1745, 2482), True, 'import streamlit as st\n'), ((4064, 4096), 'llama_index.prompts.prompts.SimpleInputPrompt', 'SimpleInputPrompt', (['"""{query_str}"""'], {}), "('{query_str}')\n", (4081, 4096), False, 'from llama_index.prompts.prompts import SimpleInputPrompt\n'), ((4176, 4335), 'llama_index.llms.HuggingFaceLLM', 'HuggingFaceLLM', ([], {'context_window': '(1024)', 'max_new_tokens': '(128)', 'system_prompt': '""""""', 'query_wrapper_prompt': 'query_wrapper_prompt', 'model': 'model', 'tokenizer': 'tokenizer'}), "(context_window=1024, max_new_tokens=128, system_prompt='',\n query_wrapper_prompt=query_wrapper_prompt, model=model, tokenizer=tokenizer\n )\n", (4190, 4335), False, 'from llama_index.llms import HuggingFaceLLM\n'), ((4953, 5030), 'llama_index.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'chunk_size': '(256)', 'llm': 'llm', 'embed_model': 'embeddings'}), '(chunk_size=256, llm=llm, embed_model=embeddings)\n', (4981, 5030), False, 'from llama_index import ServiceContext\n'), ((5076, 5119), 'llama_index.set_global_service_context', 'set_global_service_context', (['service_context'], {}), '(service_context)\n', (5102, 5119), False, 'from llama_index import set_global_service_context\n'), ((5150, 5263), 'streamlit.subheader', 'st.subheader', (['"""Please upload your data to enable the retrieval system; otherwise, it will not respond."""'], {}), "(\n 'Please upload your data to enable the retrieval system; otherwise, it will not respond.'\n )\n", (5162, 5263), True, 'import streamlit as st\n'), ((5286, 5328), 'streamlit.file_uploader', 'st.file_uploader', (['"""Upload PDF"""'], {'type': '"""pdf"""'}), "('Upload PDF', type='pdf')\n", (5302, 5328), True, 'import streamlit as st\n'), ((5897, 5939), 'llama_index.VectorStoreIndex.from_documents', 'VectorStoreIndex.from_documents', (['documents'], {}), '(documents)\n', (5928, 5939), False, 'from llama_index import VectorStoreIndex, SimpleDirectoryReader\n'), ((6089, 6113), 'streamlit.title', 'st.title', (['"""👔 HireMind 🧩"""'], {}), "('👔 HireMind 🧩')\n", (6097, 6113), True, 'import streamlit as st\n'), ((6482, 6521), 'streamlit.chat_input', 'st.chat_input', (['"""Input your prompt here"""'], {}), "('Input your prompt here')\n", (6495, 6521), True, 'import streamlit as st\n'), ((1200, 1262), 'transformers.GPT2Tokenizer.from_pretrained', 'GPT2Tokenizer.from_pretrained', (['name'], {'use_auth_token': 'auth_token'}), '(name, use_auth_token=auth_token)\n', (1229, 1262), False, 'from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer, GPT2Tokenizer, GPT2LMHeadModel, GPT2Model\n'), ((1295, 1332), 'transformers.GPT2LMHeadModel.from_pretrained', 'GPT2LMHeadModel.from_pretrained', (['name'], {}), '(name)\n', (1326, 1332), False, 'from transformers import AutoTokenizer, AutoModelForCausalLM, TextStreamer, GPT2Tokenizer, GPT2LMHeadModel, GPT2Model\n'), ((4841, 4893), 'langchain.embeddings.huggingface.HuggingFaceEmbeddings', 'HuggingFaceEmbeddings', ([], {'model_name': '"""all-MiniLM-L6-v2"""'}), "(model_name='all-MiniLM-L6-v2')\n", (4862, 4893), False, 'from langchain.embeddings.huggingface import HuggingFaceEmbeddings\n'), ((6583, 6652), 'streamlit.session_state.messages.append', 'st.session_state.messages.append', (["{'role': 'user', 'content': prompt}"], {}), "({'role': 'user', 'content': prompt})\n", (6615, 6652), True, 'import streamlit as st\n'), ((6761, 6837), 'streamlit.session_state.messages.append', 'st.session_state.messages.append', (["{'role': 'assistant', 'content': response}"], {}), "({'role': 'assistant', 'content': response})\n", (6793, 6837), True, 'import streamlit as st\n'), ((5404, 5433), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (5431, 5433), False, 'import tempfile\n'), ((5543, 5585), 'os.path.join', 'os.path.join', (['temp_dir', 'uploaded_file.name'], {}), '(temp_dir, uploaded_file.name)\n', (5555, 5585), False, 'import os\n'), ((5816, 5879), 'streamlit.success', 'st.success', (['f"""File {uploaded_file.name} uploaded successfully."""'], {}), "(f'File {uploaded_file.name} uploaded successfully.')\n", (5826, 5879), True, 'import streamlit as st\n'), ((6329, 6361), 'streamlit.chat_message', 'st.chat_message', (["message['role']"], {}), "(message['role'])\n", (6344, 6361), True, 'import streamlit as st\n'), ((6538, 6561), 'streamlit.chat_message', 'st.chat_message', (['"""user"""'], {}), "('user')\n", (6553, 6561), True, 'import streamlit as st\n'), ((6709, 6737), 'streamlit.chat_message', 'st.chat_message', (['"""assistant"""'], {}), "('assistant')\n", (6724, 6737), True, 'import streamlit as st\n'), ((5764, 5795), 'llama_index.SimpleDirectoryReader', 'SimpleDirectoryReader', (['temp_dir'], {}), '(temp_dir)\n', (5785, 5795), False, 'from llama_index import VectorStoreIndex, SimpleDirectoryReader\n')]
from langchain_openai import ChatOpenAI from langchain.chains import LLMChain from langchain.prompts import MessagesPlaceholder, HumanMessagePromptTemplate, ChatPromptTemplate from langchain.memory import ConversationBufferMemory, FileChatMessageHistory from dotenv import load_dotenv import sqlite3 import sqlparse import json import warnings from langchain_core._api.deprecation import LangChainDeprecationWarning # Ignore specific deprecation warnings from langchain_core warnings.filterwarnings("ignore", category=LangChainDeprecationWarning) load_dotenv() # Create and initialize the OpenAI client def create_openai_client(): return ChatOpenAI(model_name="gpt-3.5-turbo") # Retrieve the database schema from an SQLite database def get_database_schema(db_path): try: conn = sqlite3.connect(db_path) cursor = conn.cursor() cursor.execute("SELECT name FROM sqlite_master WHERE type='table';") tables = cursor.fetchall() schema_info = [] for table in tables: table_name = table[0] cursor.execute(f"PRAGMA table_info({table_name});") columns = cursor.fetchall() primary_keys = [col[1] for col in columns if col[5] == 1] schema_info.append(f"{table_name} (Primary Key: {', '.join(primary_keys)})") conn.close() return "filename: chinook.db, "+', '.join(schema_info) except Exception as e: return f"Error reading database schema: {e}" # Format the raw SQL query for better readability using sqlparse def format_sql_query(raw_sql): if raw_sql.startswith("```sql"): raw_sql = raw_sql[6:] if raw_sql.endswith("```"): raw_sql = raw_sql[:-3] raw_sql = raw_sql.strip() formatted_sql = sqlparse.format(raw_sql, reindent=True, keyword_case='upper') return formatted_sql # Function to execute a SQL query and return results def execute_sql_query(db_path, sql_query): try: conn = sqlite3.connect(db_path) cursor = conn.cursor() cursor.execute(sql_query) results = cursor.fetchall() conn.close() return results except sqlite3.Error as e: return f"SQL Error: {e}" # Main execution if __name__ == "__main__": client = create_openai_client() db_path = "chinook.db" database_schema = get_database_schema(db_path) memory = ConversationBufferMemory( chat_memory=FileChatMessageHistory("chat_history.json"), memory_key="messages", return_messages=True, ) prompt = ChatPromptTemplate( input_variables=["content", "messages"], messages=[ MessagesPlaceholder(variable_name="messages"), HumanMessagePromptTemplate.from_template( "You are a data analyst. Generate SQL queries from natural language descriptions. Provide only the SQL query in response, without explanations or additional text" "Question: {{question}} Database Schema: {{database_schema}} Response: {content}" ), ], ) chain = LLMChain( llm=client, prompt=prompt, memory=memory, ) while True: content = input("Usage: type 'quit' to exit chat\n>> ") if content == "quit": break result = chain({"content": content}) # Format the SQL query formatted_query = format_sql_query(result['text']) print("\nGenerated SQL Query:\n", formatted_query) # Execute the SQL query and print results query_results = execute_sql_query(db_path, formatted_query) print("\nQuery Results:\n", query_results)
[ "langchain_openai.ChatOpenAI", "langchain.prompts.HumanMessagePromptTemplate.from_template", "langchain.prompts.MessagesPlaceholder", "langchain.memory.FileChatMessageHistory", "langchain.chains.LLMChain" ]
[((476, 547), 'warnings.filterwarnings', 'warnings.filterwarnings', (['"""ignore"""'], {'category': 'LangChainDeprecationWarning'}), "('ignore', category=LangChainDeprecationWarning)\n", (499, 547), False, 'import warnings\n'), ((550, 563), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (561, 563), False, 'from dotenv import load_dotenv\n'), ((646, 684), 'langchain_openai.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""'}), "(model_name='gpt-3.5-turbo')\n", (656, 684), False, 'from langchain_openai import ChatOpenAI\n'), ((1761, 1822), 'sqlparse.format', 'sqlparse.format', (['raw_sql'], {'reindent': '(True)', 'keyword_case': '"""upper"""'}), "(raw_sql, reindent=True, keyword_case='upper')\n", (1776, 1822), False, 'import sqlparse\n'), ((3072, 3122), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'client', 'prompt': 'prompt', 'memory': 'memory'}), '(llm=client, prompt=prompt, memory=memory)\n', (3080, 3122), False, 'from langchain.chains import LLMChain\n'), ((799, 823), 'sqlite3.connect', 'sqlite3.connect', (['db_path'], {}), '(db_path)\n', (814, 823), False, 'import sqlite3\n'), ((1969, 1993), 'sqlite3.connect', 'sqlite3.connect', (['db_path'], {}), '(db_path)\n', (1984, 1993), False, 'import sqlite3\n'), ((2423, 2466), 'langchain.memory.FileChatMessageHistory', 'FileChatMessageHistory', (['"""chat_history.json"""'], {}), "('chat_history.json')\n", (2445, 2466), False, 'from langchain.memory import ConversationBufferMemory, FileChatMessageHistory\n'), ((2649, 2694), 'langchain.prompts.MessagesPlaceholder', 'MessagesPlaceholder', ([], {'variable_name': '"""messages"""'}), "(variable_name='messages')\n", (2668, 2694), False, 'from langchain.prompts import MessagesPlaceholder, HumanMessagePromptTemplate, ChatPromptTemplate\n'), ((2708, 3001), 'langchain.prompts.HumanMessagePromptTemplate.from_template', 'HumanMessagePromptTemplate.from_template', (['"""You are a data analyst. Generate SQL queries from natural language descriptions. Provide only the SQL query in response, without explanations or additional textQuestion: {{question}} Database Schema: {{database_schema}} Response: {content}"""'], {}), "(\n 'You are a data analyst. Generate SQL queries from natural language descriptions. Provide only the SQL query in response, without explanations or additional textQuestion: {{question}} Database Schema: {{database_schema}} Response: {content}'\n )\n", (2748, 3001), False, 'from langchain.prompts import MessagesPlaceholder, HumanMessagePromptTemplate, ChatPromptTemplate\n')]
import langchain_visualizer # isort:skip # noqa: F401 import asyncio import vcr_langchain as vcr from fvalues import FValue from langchain import PromptTemplate from langchain.llms import OpenAI # ========================== Start of langchain example code ========================== # https://langchain.readthedocs.io/en/latest/modules/prompts/getting_started.html # An example prompt with one input variable one_input_prompt = PromptTemplate( input_variables=["adjective"], template="Tell me a {adjective} joke." ) prompt = one_input_prompt.format(adjective="funny") # ================================== Execute example ================================== def test_prompt(): assert prompt.parts == ( "Tell me a ", FValue(source="adjective", value="funny", formatted="funny"), " joke.", ) @vcr.use_cassette() async def one_input_prompt_demo(): agent = OpenAI(model_name="text-ada-001", temperature=0) return agent(prompt) def test_llm_usage_succeeds(): """ Check that it works like a regular prompt. Also, record playback for easy visualization. """ result = asyncio.get_event_loop().run_until_complete(one_input_prompt_demo()) assert result.strip().startswith("Why did the chicken cross the road?") if __name__ == "__main__": from langchain_visualizer import visualize visualize(one_input_prompt_demo)
[ "langchain_visualizer.visualize", "langchain.llms.OpenAI", "langchain.PromptTemplate" ]
[((434, 524), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['adjective']", 'template': '"""Tell me a {adjective} joke."""'}), "(input_variables=['adjective'], template=\n 'Tell me a {adjective} joke.')\n", (448, 524), False, 'from langchain import PromptTemplate\n'), ((837, 855), 'vcr_langchain.use_cassette', 'vcr.use_cassette', ([], {}), '()\n', (853, 855), True, 'import vcr_langchain as vcr\n'), ((903, 951), 'langchain.llms.OpenAI', 'OpenAI', ([], {'model_name': '"""text-ada-001"""', 'temperature': '(0)'}), "(model_name='text-ada-001', temperature=0)\n", (909, 951), False, 'from langchain.llms import OpenAI\n'), ((1362, 1394), 'langchain_visualizer.visualize', 'visualize', (['one_input_prompt_demo'], {}), '(one_input_prompt_demo)\n', (1371, 1394), False, 'from langchain_visualizer import visualize\n'), ((748, 808), 'fvalues.FValue', 'FValue', ([], {'source': '"""adjective"""', 'value': '"""funny"""', 'formatted': '"""funny"""'}), "(source='adjective', value='funny', formatted='funny')\n", (754, 808), False, 'from fvalues import FValue\n'), ((1136, 1160), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (1158, 1160), False, 'import asyncio\n')]
from typing import Optional, List from langchain.chains.openai_functions import create_structured_output_runnable from langchain_community.chat_models import ChatOpenAI from langchain_core.prompts import ChatPromptTemplate from langchain_core.pydantic_v1 import BaseModel, Field import logging import langchain from dreamsboard.document_loaders.protocol.ner_protocol import DreamsStepInfo, DreamsStepInfoListWrapper langchain.verbose = True logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) logging.basicConfig(level=logging.DEBUG) # 控制台打印 handler = logging.StreamHandler() handler.setLevel(logging.DEBUG) logger.addHandler(handler) def test_create_structured_output_runnable2() -> None: """Test create_structured_output_runnable. 测试创建结构化输出可运行对象。""" llm = ChatOpenAI(openai_api_base='http://127.0.0.1:30000/v1', model="glm-4", openai_api_key="glm-4", verbose=True, temperature=0.1, top_p=0.9, ) prompt = ChatPromptTemplate.from_messages( [ ("system", """根据提供的故事场景,您作为心理咨询工作者可以使用开放性问题来引导来访者表达他们的感受和思维。以下是一步一步的分解: **Step 1: 建立情感连接** 开始时,您可以通过表达理解和共鸣来建立情感连接,让来访者感到舒适。您可以说:“我注意到这个对话中有许多愉快的时刻和互动。你对这些时刻有什么特别的感受吗?” **Step 2: 探索来访者的感受** 继续引导来访者表达他们的感受。您可以问:“在这个对话中,有哪些瞬间让你感到开心或快乐?” **Step 3: 询问是否有反感情绪** 除了积极的情感,也要问询是否有一些负面情感或担忧。您可以说:“除了快乐的瞬间,是否有一些让你感到不安或担忧的地方?” **Step 4: 深入探讨个人反应** 一旦来访者开始分享他们的感受,可以深入探讨他们的个人反应。例如:“你觉得自己在这些互动中扮演了什么角色?” **Step 5: 探索与他人的互动** 继续引导来访者思考他们与他人的互动。您可以问:“这些互动对你与他人的关系有什么影响?你觉得与朋友之间的互动如何影响你的情感状态?” **Step 6: 引导自我反思** 最后,鼓励来访者进行自我反思。您可以问:“在这个故事场景中,你是否注意到了自己的情感变化或思维模式?有没有什么你想要深入探讨或解决的问题?” 通过这种方式,您可以引导来访者自由表达他们的情感和思维,帮助他们更好地理解自己和他们与他人的互动。同时,保持开放和倾听,以便在需要时提供支持和建议。"""), ("ai", """step_advice|step_description 我注意到这个对话中有许多愉快的时刻和互动。你对这些时刻有什么特别的感受吗?|建立情感连接 在这个对话中,有哪些瞬间让你感到开心或快乐?|探索来访者的感受 除了快乐的瞬间,是否有一些让你感到不安或担忧的地方?|询问是否有反感情绪 你觉得自己在这些互动中扮演了什么角色?|深入探讨个人反应 这些互动对你与他人的关系有什么影响?你觉得与朋友之间的互动如何影响你的情感状态?|探索与他人的互动 在这个故事场景中,你是否注意到了自己的情感变化或思维模式?有没有什么你想要深入探讨或解决的问题?|引导自我反思"""), ("human", "{input}"), ] ) chain = create_structured_output_runnable(DreamsStepInfoListWrapper, llm, prompt) out = chain.invoke({"input": """以下是如何通过分步引导来访者张毛峰说出自己的问题的步骤: ### Step 1: 建立情感连接 - “张毛峰,我注意到你提到的这段经历让你感到非常愉快,甚至有些事情让你忍不住笑出声。能跟我分享一下,那个让你这么开心的事情吗?” ### Step 2: 探索来访者的积极感受 - “在那次和小姐姐的互动中,有哪些具体的瞬间让你感到特别快乐?” ### Step 3: 引导来访者表达背后的动机和情感 - “我看到你因为举报那位小姐姐而感到开心,这背后有没有什么特别的原因?你能告诉我举报她的行为对你意味着什么吗?” ### Step 4: 探询来访者的内心冲突和不安 - “虽然这件事情让你觉得开心,但我想知道,在整件事情中,有没有什么让你感到困扰或不安的地方?” ### Step 5: 询问来访者的反思和感受 - “你提到觉得有时候没有必要讲道理,这是不是说明这件事情也让你感到某种程度上的矛盾和困惑?” ### Step 6: 深入探讨来访者的自我认知 - “你说‘我不想玩了,因为我不是我自己了’,这句话似乎透露出你对自己有一些新的认识。你能解释一下这是什么意思吗?” 以下是具体的步骤分解: **Step 1: 建立情感连接** - 通过询问张毛峰感到愉快的时刻来建立情感上的联系。 **Step 2: 探索来访者的积极感受** - 帮助张毛峰回忆和描述那些让他感到快乐的细节。 **Step 3: 引导来访者表达背后的动机和情感** - 深入探讨举报行为背后的动机和情感,了解他为何会因为举报而感到快乐。 **Step 4: 探询来访者的内心冲突和不安** - 询问是否有任何不安或冲突的情绪,比如对于自己行为的道德考量。 **Step 5: 询问来访者的反思和感受** - 让张毛峰反思他的行为和感受,了解他对于讲道理的看法。 **Step 6: 深入探讨来访者的自我认知** - 针对他提到的“我不是我自己了”这句话,探索他的自我认同和可能的内心变化。 通过这样的引导,可以帮助张毛峰表达和认识自己的情感、动机和内心冲突,从而更好地理解自己的问题所在。"""}) logger.info(out) # -> Dog(name="Harry", color="brown", fav_food="chicken")
[ "langchain_core.prompts.ChatPromptTemplate.from_messages", "langchain.chains.openai_functions.create_structured_output_runnable", "langchain_community.chat_models.ChatOpenAI" ]
[((453, 480), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (470, 480), False, 'import logging\n'), ((512, 552), 'logging.basicConfig', 'logging.basicConfig', ([], {'level': 'logging.DEBUG'}), '(level=logging.DEBUG)\n', (531, 552), False, 'import logging\n'), ((571, 594), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (592, 594), False, 'import logging\n'), ((788, 928), 'langchain_community.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_base': '"""http://127.0.0.1:30000/v1"""', 'model': '"""glm-4"""', 'openai_api_key': '"""glm-4"""', 'verbose': '(True)', 'temperature': '(0.1)', 'top_p': '(0.9)'}), "(openai_api_base='http://127.0.0.1:30000/v1', model='glm-4',\n openai_api_key='glm-4', verbose=True, temperature=0.1, top_p=0.9)\n", (798, 928), False, 'from langchain_community.chat_models import ChatOpenAI\n'), ((1045, 2067), 'langchain_core.prompts.ChatPromptTemplate.from_messages', 'ChatPromptTemplate.from_messages', (['[(\'system\',\n """根据提供的故事场景,您作为心理咨询工作者可以使用开放性问题来引导来访者表达他们的感受和思维。以下是一步一步的分解:\n\n**Step 1: 建立情感连接**\n开始时,您可以通过表达理解和共鸣来建立情感连接,让来访者感到舒适。您可以说:“我注意到这个对话中有许多愉快的时刻和互动。你对这些时刻有什么特别的感受吗?”\n\n**Step 2: 探索来访者的感受**\n继续引导来访者表达他们的感受。您可以问:“在这个对话中,有哪些瞬间让你感到开心或快乐?”\n\n**Step 3: 询问是否有反感情绪**\n除了积极的情感,也要问询是否有一些负面情感或担忧。您可以说:“除了快乐的瞬间,是否有一些让你感到不安或担忧的地方?”\n\n**Step 4: 深入探讨个人反应**\n一旦来访者开始分享他们的感受,可以深入探讨他们的个人反应。例如:“你觉得自己在这些互动中扮演了什么角色?”\n\n**Step 5: 探索与他人的互动**\n继续引导来访者思考他们与他人的互动。您可以问:“这些互动对你与他人的关系有什么影响?你觉得与朋友之间的互动如何影响你的情感状态?”\n\n**Step 6: 引导自我反思**\n最后,鼓励来访者进行自我反思。您可以问:“在这个故事场景中,你是否注意到了自己的情感变化或思维模式?有没有什么你想要深入探讨或解决的问题?”\n\n通过这种方式,您可以引导来访者自由表达他们的情感和思维,帮助他们更好地理解自己和他们与他人的互动。同时,保持开放和倾听,以便在需要时提供支持和建议。"""\n ), (\'ai\',\n """step_advice|step_description\n我注意到这个对话中有许多愉快的时刻和互动。你对这些时刻有什么特别的感受吗?|建立情感连接\n在这个对话中,有哪些瞬间让你感到开心或快乐?|探索来访者的感受\n除了快乐的瞬间,是否有一些让你感到不安或担忧的地方?|询问是否有反感情绪\n你觉得自己在这些互动中扮演了什么角色?|深入探讨个人反应\n这些互动对你与他人的关系有什么影响?你觉得与朋友之间的互动如何影响你的情感状态?|探索与他人的互动\n在这个故事场景中,你是否注意到了自己的情感变化或思维模式?有没有什么你想要深入探讨或解决的问题?|引导自我反思"""\n ), (\'human\', \'{input}\')]'], {}), '([(\'system\',\n """根据提供的故事场景,您作为心理咨询工作者可以使用开放性问题来引导来访者表达他们的感受和思维。以下是一步一步的分解:\n\n**Step 1: 建立情感连接**\n开始时,您可以通过表达理解和共鸣来建立情感连接,让来访者感到舒适。您可以说:“我注意到这个对话中有许多愉快的时刻和互动。你对这些时刻有什么特别的感受吗?”\n\n**Step 2: 探索来访者的感受**\n继续引导来访者表达他们的感受。您可以问:“在这个对话中,有哪些瞬间让你感到开心或快乐?”\n\n**Step 3: 询问是否有反感情绪**\n除了积极的情感,也要问询是否有一些负面情感或担忧。您可以说:“除了快乐的瞬间,是否有一些让你感到不安或担忧的地方?”\n\n**Step 4: 深入探讨个人反应**\n一旦来访者开始分享他们的感受,可以深入探讨他们的个人反应。例如:“你觉得自己在这些互动中扮演了什么角色?”\n\n**Step 5: 探索与他人的互动**\n继续引导来访者思考他们与他人的互动。您可以问:“这些互动对你与他人的关系有什么影响?你觉得与朋友之间的互动如何影响你的情感状态?”\n\n**Step 6: 引导自我反思**\n最后,鼓励来访者进行自我反思。您可以问:“在这个故事场景中,你是否注意到了自己的情感变化或思维模式?有没有什么你想要深入探讨或解决的问题?”\n\n通过这种方式,您可以引导来访者自由表达他们的情感和思维,帮助他们更好地理解自己和他们与他人的互动。同时,保持开放和倾听,以便在需要时提供支持和建议。"""\n ), (\'ai\',\n """step_advice|step_description\n我注意到这个对话中有许多愉快的时刻和互动。你对这些时刻有什么特别的感受吗?|建立情感连接\n在这个对话中,有哪些瞬间让你感到开心或快乐?|探索来访者的感受\n除了快乐的瞬间,是否有一些让你感到不安或担忧的地方?|询问是否有反感情绪\n你觉得自己在这些互动中扮演了什么角色?|深入探讨个人反应\n这些互动对你与他人的关系有什么影响?你觉得与朋友之间的互动如何影响你的情感状态?|探索与他人的互动\n在这个故事场景中,你是否注意到了自己的情感变化或思维模式?有没有什么你想要深入探讨或解决的问题?|引导自我反思"""\n ), (\'human\', \'{input}\')])\n', (1077, 2067), False, 'from langchain_core.prompts import ChatPromptTemplate\n'), ((2123, 2196), 'langchain.chains.openai_functions.create_structured_output_runnable', 'create_structured_output_runnable', (['DreamsStepInfoListWrapper', 'llm', 'prompt'], {}), '(DreamsStepInfoListWrapper, llm, prompt)\n', (2156, 2196), False, 'from langchain.chains.openai_functions import create_structured_output_runnable\n')]
import logging from langchain.chat_models import ChatOpenAI from dreamsboard.dreams.builder_cosplay_code.base import StructuredDreamsStoryboard from dreamsboard.dreams.dreams_personality_chain.base import StoryBoardDreamsGenerationChain import langchain from dreamsboard.engine.generate.code_generate import QueryProgramGenerator, EngineProgramGenerator, AIProgramGenerator from dreamsboard.engine.loading import load_store_from_storage from dreamsboard.engine.storage.storage_context import StorageContext langchain.verbose = True logger = logging.getLogger(__name__) logger.setLevel(logging.DEBUG) # 控制台打印 handler = logging.StreamHandler() handler.setLevel(logging.DEBUG) logger.addHandler(handler) def test_structured_dreams_storyboard_store_print() -> None: try: storage_context = StorageContext.from_defaults(persist_dir="./storage") code_gen_builder = load_store_from_storage(storage_context) index_loaded = True except: index_loaded = False if not index_loaded: llm = ChatOpenAI( verbose=True ) dreams_generation_chain = StoryBoardDreamsGenerationChain.from_dreams_personality_chain( llm=llm, csv_file_path="/media/checkpoint/speech_data/抖音作品/id46Bv3g/str/id46Bv3g_keyframe.csv") output = dreams_generation_chain.run() logger.info("dreams_guidance_context:" + output.get("dreams_guidance_context")) logger.info("dreams_personality_context:" + output.get("dreams_personality_context")) dreams_guidance_context = output.get("dreams_guidance_context") dreams_personality_context = output.get("dreams_personality_context") storyboard_executor = StructuredDreamsStoryboard.form_builder(llm=llm, builder=dreams_generation_chain.builder, dreams_guidance_context=dreams_guidance_context, dreams_personality_context=dreams_personality_context ) code_gen_builder = storyboard_executor.loader_cosplay_builder() executor = code_gen_builder.build_executor() logger.info(executor.executor_code) assert True
[ "langchain.chat_models.ChatOpenAI" ]
[((545, 572), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (562, 572), False, 'import logging\n'), ((623, 646), 'logging.StreamHandler', 'logging.StreamHandler', ([], {}), '()\n', (644, 646), False, 'import logging\n'), ((806, 859), 'dreamsboard.engine.storage.storage_context.StorageContext.from_defaults', 'StorageContext.from_defaults', ([], {'persist_dir': '"""./storage"""'}), "(persist_dir='./storage')\n", (834, 859), False, 'from dreamsboard.engine.storage.storage_context import StorageContext\n'), ((887, 927), 'dreamsboard.engine.loading.load_store_from_storage', 'load_store_from_storage', (['storage_context'], {}), '(storage_context)\n', (910, 927), False, 'from dreamsboard.engine.loading import load_store_from_storage\n'), ((1037, 1061), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'verbose': '(True)'}), '(verbose=True)\n', (1047, 1061), False, 'from langchain.chat_models import ChatOpenAI\n'), ((1119, 1285), 'dreamsboard.dreams.dreams_personality_chain.base.StoryBoardDreamsGenerationChain.from_dreams_personality_chain', 'StoryBoardDreamsGenerationChain.from_dreams_personality_chain', ([], {'llm': 'llm', 'csv_file_path': '"""/media/checkpoint/speech_data/抖音作品/id46Bv3g/str/id46Bv3g_keyframe.csv"""'}), "(llm=llm,\n csv_file_path=\n '/media/checkpoint/speech_data/抖音作品/id46Bv3g/str/id46Bv3g_keyframe.csv')\n", (1180, 1285), False, 'from dreamsboard.dreams.dreams_personality_chain.base import StoryBoardDreamsGenerationChain\n'), ((1701, 1909), 'dreamsboard.dreams.builder_cosplay_code.base.StructuredDreamsStoryboard.form_builder', 'StructuredDreamsStoryboard.form_builder', ([], {'llm': 'llm', 'builder': 'dreams_generation_chain.builder', 'dreams_guidance_context': 'dreams_guidance_context', 'dreams_personality_context': 'dreams_personality_context'}), '(llm=llm, builder=\n dreams_generation_chain.builder, dreams_guidance_context=\n dreams_guidance_context, dreams_personality_context=\n dreams_personality_context)\n', (1740, 1909), False, 'from dreamsboard.dreams.builder_cosplay_code.base import StructuredDreamsStoryboard\n')]
import langchain_openai # Disable because of version conflict # import langchain_anthropic import pytest from interlab.queries.count_tokens import count_tokens TEXT = ( "Lorem ipsum dolor sit amet, consectetur adipiscing elit. Sed ultrices lacus sed leo ornare, " "sed iaculis mi pharetra. Pellentesque non nisi vitae diam sodales commodo." ) def test_count_tokens(): count_tokens.cache_clear() # To have clean hit counting below assert count_tokens(TEXT, "gpt2") == 56 assert count_tokens(TEXT, "text-curie-001") == 56 assert count_tokens(TEXT, "claude-2") == 57 with pytest.raises(ValueError): count_tokens(TEXT, "foobar") assert ( count_tokens( TEXT, langchain_openai.ChatOpenAI(model="gpt-4", openai_api_key="Foo"), ) == 40 ) assert ( count_tokens( TEXT, langchain_openai.ChatOpenAI(openai_api_key="Foo"), ) == 40 ) # assert ( # count_tokens( # TEXT, langchain_community.chat_models.ChatAnthropic(anthropic_api_key="Foo") # ) # == 57 # ) assert ( count_tokens( TEXT, langchain_openai.OpenAI(model="babbage", openai_api_key="Foo") ) == 56 ) assert count_tokens.cache_info().hits == 0 assert ( count_tokens( TEXT, langchain_openai.OpenAI(model="babbage", openai_api_key="Foo") ) == 56 ) assert count_tokens.cache_info().hits == 1
[ "langchain_openai.ChatOpenAI", "langchain_openai.OpenAI" ]
[((385, 411), 'interlab.queries.count_tokens.count_tokens.cache_clear', 'count_tokens.cache_clear', ([], {}), '()\n', (409, 411), False, 'from interlab.queries.count_tokens import count_tokens\n'), ((460, 486), 'interlab.queries.count_tokens.count_tokens', 'count_tokens', (['TEXT', '"""gpt2"""'], {}), "(TEXT, 'gpt2')\n", (472, 486), False, 'from interlab.queries.count_tokens import count_tokens\n'), ((504, 540), 'interlab.queries.count_tokens.count_tokens', 'count_tokens', (['TEXT', '"""text-curie-001"""'], {}), "(TEXT, 'text-curie-001')\n", (516, 540), False, 'from interlab.queries.count_tokens import count_tokens\n'), ((558, 588), 'interlab.queries.count_tokens.count_tokens', 'count_tokens', (['TEXT', '"""claude-2"""'], {}), "(TEXT, 'claude-2')\n", (570, 588), False, 'from interlab.queries.count_tokens import count_tokens\n'), ((604, 629), 'pytest.raises', 'pytest.raises', (['ValueError'], {}), '(ValueError)\n', (617, 629), False, 'import pytest\n'), ((639, 667), 'interlab.queries.count_tokens.count_tokens', 'count_tokens', (['TEXT', '"""foobar"""'], {}), "(TEXT, 'foobar')\n", (651, 667), False, 'from interlab.queries.count_tokens import count_tokens\n'), ((733, 797), 'langchain_openai.ChatOpenAI', 'langchain_openai.ChatOpenAI', ([], {'model': '"""gpt-4"""', 'openai_api_key': '"""Foo"""'}), "(model='gpt-4', openai_api_key='Foo')\n", (760, 797), False, 'import langchain_openai\n'), ((894, 943), 'langchain_openai.ChatOpenAI', 'langchain_openai.ChatOpenAI', ([], {'openai_api_key': '"""Foo"""'}), "(openai_api_key='Foo')\n", (921, 943), False, 'import langchain_openai\n'), ((1194, 1256), 'langchain_openai.OpenAI', 'langchain_openai.OpenAI', ([], {'model': '"""babbage"""', 'openai_api_key': '"""Foo"""'}), "(model='babbage', openai_api_key='Foo')\n", (1217, 1256), False, 'import langchain_openai\n'), ((1298, 1323), 'interlab.queries.count_tokens.count_tokens.cache_info', 'count_tokens.cache_info', ([], {}), '()\n', (1321, 1323), False, 'from interlab.queries.count_tokens import count_tokens\n'), ((1387, 1449), 'langchain_openai.OpenAI', 'langchain_openai.OpenAI', ([], {'model': '"""babbage"""', 'openai_api_key': '"""Foo"""'}), "(model='babbage', openai_api_key='Foo')\n", (1410, 1449), False, 'import langchain_openai\n'), ((1491, 1516), 'interlab.queries.count_tokens.count_tokens.cache_info', 'count_tokens.cache_info', ([], {}), '()\n', (1514, 1516), False, 'from interlab.queries.count_tokens import count_tokens\n')]
from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec from model.chain_revision import ChainRevision from model.lang_chain_context import LangChainContext from langchain.llms.fake import FakeListLLM from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory def test_llm_chain_spec_serialization(): llm_chain_spec = LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="output1", prompt="prompt", llm_key="llm_key", chain_type="llm_chain_spec", ) serialized = llm_chain_spec.json() assert serialized == '{"chain_id": 1, "input_keys": ["input1", "input2"], "output_key": "output1", "chain_type": "llm_chain_spec", "prompt": "prompt", "llm_key": "llm_key"}' revision = ChainRevision(chain=llm_chain_spec, llms={}) serialized_revision = revision.json() deserialized = ChainRevision.parse_raw(serialized_revision).chain assert deserialized == llm_chain_spec def test_llm_chain_spec_to_lang_chain_creates_valid_chain(): prompt_template = "the prompt {input1} {input2}" llm_chain_spec = LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="output1", prompt=prompt_template, llm_key="test", chain_type="llm_chain_spec", ) llms = llms={"test": FakeListLLM(responses=["response1"])} ctx = LangChainContext(llms=llms) llm_chain = llm_chain_spec.to_lang_chain(ctx) assert llm_chain.input_keys == ["input1", "input2"] assert llm_chain.output_key == "output1" assert llm_chain.prompt.template == prompt_template assert llm_chain.llm == llms["test"] output = llm_chain._call({"input1": "input1", "input2": "input2"}) assert output == {"output1": "response1"} def test_llm_chain_spec_to_lang_chain_creates_recording_chain(): prompt_template = "the prompt {input1} {input2}" llm_chain_spec = LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="output1", prompt=prompt_template, llm_key="test", chain_type="llm_chain_spec", ) llms = llms={"test": FakeListLLM(responses=["response1"])} ctx = LangChainContext(llms=llms, recording=True) llm_chain = llm_chain_spec.to_lang_chain(ctx) assert len(ctx.prompts) == 1 output = llm_chain._call({"input1": "input1", "input2": "input2"}) assert output == {"output1": "response1"} assert ctx.prompts[0].calls == [ ({"input1": "input1", "input2": "input2"}, {"output1": "response1"}) ] def test_sequential_chain_spec_serialization(): sequential_chain_spec = SequentialChainSpec( chain_id=1, input_keys=["input1", "input2"], output_keys=["output1", "output2"], chain_type="sequential_chain_spec", chains=[ LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="output1", prompt="prompt", llm_key="llm_key", chain_type="llm_chain_spec" ) ], ) serialized = sequential_chain_spec.json() deserialized = SequentialChainSpec.parse_raw(serialized) assert deserialized == sequential_chain_spec def test_sequential_chain_spec_to_lang_chain_creates_valid_chain(): llm_chain_spec1 = LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="llm1_output", prompt="the prompt {input1} {input2}", llm_key="test", chain_type="llm_chain_spec", ) llm_chain_spec2 = LLMChainSpec( chain_id=2, input_keys=["llm1_output", "input3"], output_key="llm2_output", prompt="the prompt {llm1_output} {input3}", llm_key="test", chain_type="llm_chain_spec", ) sequential_chain_spec = SequentialChainSpec( chain_id=3, input_keys=["input1", "input2", "input3"], output_keys=["llm2_output"], chain_type="sequential_chain_spec", chains=[llm_chain_spec1, llm_chain_spec2], ) llms = llms={"test": FakeListLLM(responses=["fake_response1", "fake_response2"])} ctx = LangChainContext(llms=llms) sequential_chain = sequential_chain_spec.to_lang_chain(ctx) assert sequential_chain.input_keys == ["input1", "input2", "input3"] assert sequential_chain.output_keys == ["llm2_output"] assert sequential_chain.chains[0].input_keys == ["input1", "input2"] assert sequential_chain.chains[0].output_keys == ["llm1_output"] assert sequential_chain.chains[0].prompt.template == llm_chain_spec1.prompt assert sequential_chain.chains[0].llm == llms["test"] assert sequential_chain.chains[1].input_keys == ["llm1_output", "input3"] assert sequential_chain.chains[1].output_keys == ["llm2_output"] assert sequential_chain.chains[1].prompt.template == llm_chain_spec2.prompt assert sequential_chain.chains[1].llm == llms["test"] output = sequential_chain.run({"input1": "input1", "input2": "input2", "input3": "input3"}) assert output == "fake_response2" def test_case_chain_spec_serialization(): case_chain_spec = CaseChainSpec( chain_id=3, chain_type="case_chain_spec", categorization_key="categorization_input", cases={ "response1": LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="output", prompt="prompt", llm_key="test", chain_type="llm_chain_spec" ), "response2": LLMChainSpec( chain_id=2, input_keys=["input3", "input4"], output_key="output", prompt="prompt", llm_key="test", chain_type="llm_chain_spec" ) }, default_case=LLMChainSpec( chain_id=4, input_keys=["input1", "input2"], output_key="output", prompt="prompt", llm_key="test", chain_type="llm_chain_spec" ), ) serialized = case_chain_spec.json() deserialized = CaseChainSpec.parse_raw(serialized) assert deserialized == case_chain_spec def test_case_chain_spec_to_lang_chain_creates_valid_chain(): llms = {"test1": FakeListLLM(responses=["fake_response1"]), "test2": FakeListLLM(responses=["fake_response2"])} llm_chain_spec1 = LLMChainSpec( chain_id=1, input_keys=["input1", "input2"], output_key="output", prompt="ll1 prompt {input1} {input2}", llm_key="test1", chain_type="llm_chain_spec", ) llm_chain_spec2 = LLMChainSpec( chain_id=2, input_keys=["input3", "input4"], output_key="output", prompt="llm2 prompt {input3} {input4}", llm_key="test2", chain_type="llm_chain_spec", ) llm_chain_spec3 = LLMChainSpec( chain_id=4, input_keys=["input1", "input2"], output_key="output", prompt="default prompt {input1} {input2}", llm_key="test1", chain_type="llm_chain_spec" ) case_chain_spec = CaseChainSpec( chain_id=3, chain_type="case_chain_spec", categorization_key="categorization_input", cases={ "response1": llm_chain_spec1, "response2": llm_chain_spec2, }, default_case=llm_chain_spec3, ) ctx = LangChainContext(llms=llms) case_chain = case_chain_spec.to_lang_chain(ctx) assert case_chain.input_keys == ["categorization_input", "input1", "input2", "input3", "input4"] assert case_chain.output_keys == ["output"] assert case_chain.subchains["response1"].input_keys == ["input1", "input2"] assert case_chain.subchains["response1"].output_keys == ["output"] assert case_chain.subchains["response1"].prompt.template == llm_chain_spec1.prompt assert case_chain.subchains["response1"].llm == llms["test1"] assert case_chain.subchains["response2"].input_keys == ["input3", "input4"] assert case_chain.subchains["response2"].output_keys == ["output"] assert case_chain.subchains["response2"].prompt.template == llm_chain_spec2.prompt assert case_chain.subchains["response2"].llm == llms["test2"] assert case_chain.default_chain.input_keys == ["input1", "input2"] assert case_chain.default_chain.output_keys == ["output"] assert case_chain.default_chain.prompt.template == llm_chain_spec3.prompt assert case_chain.default_chain.llm == llms["test1"] output = case_chain.run({ "categorization_input": "response2", "input1": "input1", "input2": "input2", "input3": "input3", "input4": "input4" }) assert output == "fake_response2" def test_api_chain_spec_serialization(): api_chain_spec = APIChainSpec( chain_id=3, chain_type="api_chain_spec", input_keys=["input1", "input2"], output_key="output", url="http://test.com", method="POST", body="body {input1} {input2}", headers={"header1": "header1"}, ) revision = ChainRevision(chain=api_chain_spec, llms={}) serialized_revision = revision.json() deserialized = ChainRevision.parse_raw(serialized_revision).chain assert deserialized == api_chain_spec def test_api_chain_spec_to_lang_chain_creates_valid_chain(): api_chain_spec = APIChainSpec( chain_id=3, chain_type="api_chain_spec", input_keys=["input1", "input2"], output_key="output", url="http://test.com", method="POST", body="body {input1} {input2}", headers={"header1": "header1"}, ) ctx = LangChainContext(llms={}) api_chain = api_chain_spec.to_lang_chain(ctx) assert api_chain.input_keys == ["input1", "input2"] assert api_chain.output_keys == ["output"] assert api_chain.url == "http://test.com" assert api_chain.method == "POST" assert api_chain.body == "body {input1} {input2}" assert api_chain.headers == {"header1": "header1"} def test_reformat_chain_spec_serialization(): reformat_chain_spec = ReformatChainSpec( chain_id=3, chain_type="reformat_chain_spec", input_keys=["input1", "input2"], formatters={"output1": "formatter1", "output2": "formatter2"}, ) serialized = reformat_chain_spec.json() deserialized = ReformatChainSpec.parse_raw(serialized) assert deserialized == reformat_chain_spec def test_reformat_chain_spec_to_lang_chain_creates_valid_chain(): reformat_chain_spec = ReformatChainSpec( chain_id=3, chain_type="reformat_chain_spec", input_keys=["input1", "input2"], formatters={"output1": "formatter1", "output2": "formatter2"}, ) ctx = LangChainContext(llms={}) reformat_chain = reformat_chain_spec.to_lang_chain(ctx) assert reformat_chain.input_keys == ["input1", "input2"] assert reformat_chain.output_keys == ["output1", "output2"] assert reformat_chain.formatters == {"output1": "formatter1", "output2": "formatter2"} def test_transform_chain_spec_to_lang_chain_creates_valid_chain(): transform_chain_spec = TransformChainSpec( chain_id=4, chain_type="transform_chain_spec", input_keys=["x", "y"], output_keys=["z", "w"], transform_func=""" z = int(inputs['x']) + int(inputs['y']) w = int(inputs['x']) - int(inputs['y']) return {'z': z, 'w': w}""" ) ctx = LangChainContext(llms={}) transform_chain = transform_chain_spec.to_lang_chain(ctx) assert transform_chain.input_keys == ["x", "y"] assert transform_chain.output_keys == ["z", "w"] assert transform_chain.transform({'x': 4, 'y': 3}) == {'z': 7, 'w': 1} def test_vector_search_chain_spec_serialization(): vector_search_chain_spec = VectorSearchChainSpec( chain_id=3, chain_type="vector_search_chain_spec", input_keys=["input1", "input2"], output_key="output", embedding_engine="openai", embedding_params={"openai_api_key": "test"}, database="pinecone", database_params={"index": "test", "text_key": "text"}, num_results=10, query="How does {input}?" ) serialized = vector_search_chain_spec.json() deserialized = VectorSearchChainSpec.parse_raw(serialized) assert deserialized == vector_search_chain_spec class ChainDict: def __init__(self): self.chains = {} def add_chain(self, chain): self.chains[chain.chain_id] = chain def test_chain_spec_copy_replace(): ctx = SpecFactoryContext() chain = sequential_factory(ctx, chains=[ llm_factory(ctx), case_factory(ctx, cases={ "case1": llm_factory(ctx), "case2": sequential_factory(ctx, chains=[llm_factory(ctx), llm_factory(ctx)]), }, default_case=llm_factory(ctx)), ]) original_specs = ChainDict() chain.traverse(original_specs.add_chain) copied_specs = {} copied_chain = chain.copy_replace(lambda spec: spec) copied_specs = ChainDict() copied_chain.traverse(copied_specs.add_chain) assert len(original_specs.chains) == len(copied_specs.chains) for chain_id, spec in original_specs.chains.items(): copied_spec = copied_specs.chains.get(chain_id) assert copied_spec is not None assert copied_spec == spec assert copied_spec is not spec replacement = copied_specs.chains[3] replacement.prompt = "replaced" replace_chain = chain.copy_replace(lambda spec: spec if spec.chain_id != 3 else replacement) replace_specs = ChainDict() replace_chain.traverse(replace_specs.add_chain) assert len(original_specs.chains) == len(replace_specs.chains) assert replace_specs.chains[3] == replacement assert replace_specs.chains[3].prompt == "replaced" def test_base_spec_find_by_chain_id(): ctx = SpecFactoryContext() deep_llm = llm_factory(ctx) chain = sequential_factory(ctx, chains=[ llm_factory(ctx), case_factory(ctx, cases={ "case1": llm_factory(ctx), "case2": sequential_factory(ctx, chains=[llm_factory(ctx), deep_llm]), }, default_case=llm_factory(ctx)), ]) assert chain.find_by_chain_id(deep_llm.chain_id) == deep_llm
[ "langchain.llms.fake.FakeListLLM" ]
[((471, 620), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output1"""', 'prompt': '"""prompt"""', 'llm_key': '"""llm_key"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'output1', prompt='prompt', llm_key='llm_key', chain_type='llm_chain_spec')\n", (483, 620), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((904, 948), 'model.chain_revision.ChainRevision', 'ChainRevision', ([], {'chain': 'llm_chain_spec', 'llms': '{}'}), '(chain=llm_chain_spec, llms={})\n', (917, 948), False, 'from model.chain_revision import ChainRevision\n'), ((1240, 1398), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output1"""', 'prompt': 'prompt_template', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'output1', prompt=prompt_template, llm_key='test', chain_type=\n 'llm_chain_spec')\n", (1252, 1398), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((1517, 1544), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': 'llms'}), '(llms=llms)\n', (1533, 1544), False, 'from model.lang_chain_context import LangChainContext\n'), ((2053, 2211), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output1"""', 'prompt': 'prompt_template', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'output1', prompt=prompt_template, llm_key='test', chain_type=\n 'llm_chain_spec')\n", (2065, 2211), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((2330, 2373), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': 'llms', 'recording': '(True)'}), '(llms=llms, recording=True)\n', (2346, 2373), False, 'from model.lang_chain_context import LangChainContext\n'), ((3311, 3352), 'model.chain_spec.SequentialChainSpec.parse_raw', 'SequentialChainSpec.parse_raw', (['serialized'], {}), '(serialized)\n', (3340, 3352), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((3495, 3671), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""llm1_output"""', 'prompt': '"""the prompt {input1} {input2}"""', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'llm1_output', prompt='the prompt {input1} {input2}', llm_key='test',\n chain_type='llm_chain_spec')\n", (3507, 3671), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((3796, 3983), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(2)', 'input_keys': "['llm1_output', 'input3']", 'output_key': '"""llm2_output"""', 'prompt': '"""the prompt {llm1_output} {input3}"""', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=2, input_keys=['llm1_output', 'input3'], output_key=\n 'llm2_output', prompt='the prompt {llm1_output} {input3}', llm_key=\n 'test', chain_type='llm_chain_spec')\n", (3808, 3983), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((4114, 4305), 'model.chain_spec.SequentialChainSpec', 'SequentialChainSpec', ([], {'chain_id': '(3)', 'input_keys': "['input1', 'input2', 'input3']", 'output_keys': "['llm2_output']", 'chain_type': '"""sequential_chain_spec"""', 'chains': '[llm_chain_spec1, llm_chain_spec2]'}), "(chain_id=3, input_keys=['input1', 'input2', 'input3'],\n output_keys=['llm2_output'], chain_type='sequential_chain_spec', chains\n =[llm_chain_spec1, llm_chain_spec2])\n", (4133, 4305), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((4441, 4468), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': 'llms'}), '(llms=llms)\n', (4457, 4468), False, 'from model.lang_chain_context import LangChainContext\n'), ((6447, 6482), 'model.chain_spec.CaseChainSpec.parse_raw', 'CaseChainSpec.parse_raw', (['serialized'], {}), '(serialized)\n', (6470, 6482), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((6728, 6900), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'prompt': '"""ll1 prompt {input1} {input2}"""', 'llm_key': '"""test1"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'output', prompt='ll1 prompt {input1} {input2}', llm_key='test1',\n chain_type='llm_chain_spec')\n", (6740, 6900), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((7025, 7198), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(2)', 'input_keys': "['input3', 'input4']", 'output_key': '"""output"""', 'prompt': '"""llm2 prompt {input3} {input4}"""', 'llm_key': '"""test2"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=2, input_keys=['input3', 'input4'], output_key=\n 'output', prompt='llm2 prompt {input3} {input4}', llm_key='test2',\n chain_type='llm_chain_spec')\n", (7037, 7198), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((7323, 7499), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(4)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'prompt': '"""default prompt {input1} {input2}"""', 'llm_key': '"""test1"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=4, input_keys=['input1', 'input2'], output_key=\n 'output', prompt='default prompt {input1} {input2}', llm_key='test1',\n chain_type='llm_chain_spec')\n", (7335, 7499), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((7624, 7829), 'model.chain_spec.CaseChainSpec', 'CaseChainSpec', ([], {'chain_id': '(3)', 'chain_type': '"""case_chain_spec"""', 'categorization_key': '"""categorization_input"""', 'cases': "{'response1': llm_chain_spec1, 'response2': llm_chain_spec2}", 'default_case': 'llm_chain_spec3'}), "(chain_id=3, chain_type='case_chain_spec', categorization_key=\n 'categorization_input', cases={'response1': llm_chain_spec1,\n 'response2': llm_chain_spec2}, default_case=llm_chain_spec3)\n", (7637, 7829), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((7922, 7949), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': 'llms'}), '(llms=llms)\n', (7938, 7949), False, 'from model.lang_chain_context import LangChainContext\n'), ((9327, 9543), 'model.chain_spec.APIChainSpec', 'APIChainSpec', ([], {'chain_id': '(3)', 'chain_type': '"""api_chain_spec"""', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'url': '"""http://test.com"""', 'method': '"""POST"""', 'body': '"""body {input1} {input2}"""', 'headers': "{'header1': 'header1'}"}), "(chain_id=3, chain_type='api_chain_spec', input_keys=['input1',\n 'input2'], output_key='output', url='http://test.com', method='POST',\n body='body {input1} {input2}', headers={'header1': 'header1'})\n", (9339, 9543), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((9623, 9667), 'model.chain_revision.ChainRevision', 'ChainRevision', ([], {'chain': 'api_chain_spec', 'llms': '{}'}), '(chain=api_chain_spec, llms={})\n', (9636, 9667), False, 'from model.chain_revision import ChainRevision\n'), ((9906, 10122), 'model.chain_spec.APIChainSpec', 'APIChainSpec', ([], {'chain_id': '(3)', 'chain_type': '"""api_chain_spec"""', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'url': '"""http://test.com"""', 'method': '"""POST"""', 'body': '"""body {input1} {input2}"""', 'headers': "{'header1': 'header1'}"}), "(chain_id=3, chain_type='api_chain_spec', input_keys=['input1',\n 'input2'], output_key='output', url='http://test.com', method='POST',\n body='body {input1} {input2}', headers={'header1': 'header1'})\n", (9918, 10122), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((10196, 10221), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': '{}'}), '(llms={})\n', (10212, 10221), False, 'from model.lang_chain_context import LangChainContext\n'), ((10643, 10811), 'model.chain_spec.ReformatChainSpec', 'ReformatChainSpec', ([], {'chain_id': '(3)', 'chain_type': '"""reformat_chain_spec"""', 'input_keys': "['input1', 'input2']", 'formatters': "{'output1': 'formatter1', 'output2': 'formatter2'}"}), "(chain_id=3, chain_type='reformat_chain_spec', input_keys=\n ['input1', 'input2'], formatters={'output1': 'formatter1', 'output2':\n 'formatter2'})\n", (10660, 10811), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((10905, 10944), 'model.chain_spec.ReformatChainSpec.parse_raw', 'ReformatChainSpec.parse_raw', (['serialized'], {}), '(serialized)\n', (10932, 10944), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((11086, 11254), 'model.chain_spec.ReformatChainSpec', 'ReformatChainSpec', ([], {'chain_id': '(3)', 'chain_type': '"""reformat_chain_spec"""', 'input_keys': "['input1', 'input2']", 'formatters': "{'output1': 'formatter1', 'output2': 'formatter2'}"}), "(chain_id=3, chain_type='reformat_chain_spec', input_keys=\n ['input1', 'input2'], formatters={'output1': 'formatter1', 'output2':\n 'formatter2'})\n", (11103, 11254), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((11295, 11320), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': '{}'}), '(llms={})\n', (11311, 11320), False, 'from model.lang_chain_context import LangChainContext\n'), ((11694, 11947), 'model.chain_spec.TransformChainSpec', 'TransformChainSpec', ([], {'chain_id': '(4)', 'chain_type': '"""transform_chain_spec"""', 'input_keys': "['x', 'y']", 'output_keys': "['z', 'w']", 'transform_func': '"""\nz = int(inputs[\'x\']) + int(inputs[\'y\'])\nw = int(inputs[\'x\']) - int(inputs[\'y\'])\nreturn {\'z\': z, \'w\': w}"""'}), '(chain_id=4, chain_type=\'transform_chain_spec\',\n input_keys=[\'x\', \'y\'], output_keys=[\'z\', \'w\'], transform_func=\n """\nz = int(inputs[\'x\']) + int(inputs[\'y\'])\nw = int(inputs[\'x\']) - int(inputs[\'y\'])\nreturn {\'z\': z, \'w\': w}"""\n )\n', (11712, 11947), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((11990, 12015), 'model.lang_chain_context.LangChainContext', 'LangChainContext', ([], {'llms': '{}'}), '(llms={})\n', (12006, 12015), False, 'from model.lang_chain_context import LangChainContext\n'), ((12343, 12678), 'model.chain_spec.VectorSearchChainSpec', 'VectorSearchChainSpec', ([], {'chain_id': '(3)', 'chain_type': '"""vector_search_chain_spec"""', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'embedding_engine': '"""openai"""', 'embedding_params': "{'openai_api_key': 'test'}", 'database': '"""pinecone"""', 'database_params': "{'index': 'test', 'text_key': 'text'}", 'num_results': '(10)', 'query': '"""How does {input}?"""'}), "(chain_id=3, chain_type='vector_search_chain_spec',\n input_keys=['input1', 'input2'], output_key='output', embedding_engine=\n 'openai', embedding_params={'openai_api_key': 'test'}, database=\n 'pinecone', database_params={'index': 'test', 'text_key': 'text'},\n num_results=10, query='How does {input}?')\n", (12364, 12678), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((12815, 12858), 'model.chain_spec.VectorSearchChainSpec.parse_raw', 'VectorSearchChainSpec.parse_raw', (['serialized'], {}), '(serialized)\n', (12846, 12858), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((13108, 13128), 'model.tests.factory.SpecFactoryContext', 'SpecFactoryContext', ([], {}), '()\n', (13126, 13128), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((14436, 14456), 'model.tests.factory.SpecFactoryContext', 'SpecFactoryContext', ([], {}), '()\n', (14454, 14456), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((14472, 14488), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (14483, 14488), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((1010, 1054), 'model.chain_revision.ChainRevision.parse_raw', 'ChainRevision.parse_raw', (['serialized_revision'], {}), '(serialized_revision)\n', (1033, 1054), False, 'from model.chain_revision import ChainRevision\n'), ((1469, 1505), 'langchain.llms.fake.FakeListLLM', 'FakeListLLM', ([], {'responses': "['response1']"}), "(responses=['response1'])\n", (1480, 1505), False, 'from langchain.llms.fake import FakeListLLM\n'), ((2282, 2318), 'langchain.llms.fake.FakeListLLM', 'FakeListLLM', ([], {'responses': "['response1']"}), "(responses=['response1'])\n", (2293, 2318), False, 'from langchain.llms.fake import FakeListLLM\n'), ((4370, 4429), 'langchain.llms.fake.FakeListLLM', 'FakeListLLM', ([], {'responses': "['fake_response1', 'fake_response2']"}), "(responses=['fake_response1', 'fake_response2'])\n", (4381, 4429), False, 'from langchain.llms.fake import FakeListLLM\n'), ((6611, 6652), 'langchain.llms.fake.FakeListLLM', 'FakeListLLM', ([], {'responses': "['fake_response1']"}), "(responses=['fake_response1'])\n", (6622, 6652), False, 'from langchain.llms.fake import FakeListLLM\n'), ((6663, 6704), 'langchain.llms.fake.FakeListLLM', 'FakeListLLM', ([], {'responses': "['fake_response2']"}), "(responses=['fake_response2'])\n", (6674, 6704), False, 'from langchain.llms.fake import FakeListLLM\n'), ((9729, 9773), 'model.chain_revision.ChainRevision.parse_raw', 'ChainRevision.parse_raw', (['serialized_revision'], {}), '(serialized_revision)\n', (9752, 9773), False, 'from model.chain_revision import ChainRevision\n'), ((6158, 6303), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(4)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'prompt': '"""prompt"""', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=4, input_keys=['input1', 'input2'], output_key=\n 'output', prompt='prompt', llm_key='test', chain_type='llm_chain_spec')\n", (6170, 6303), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((2974, 3123), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output1"""', 'prompt': '"""prompt"""', 'llm_key': '"""llm_key"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'output1', prompt='prompt', llm_key='llm_key', chain_type='llm_chain_spec')\n", (2986, 3123), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((5598, 5743), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(1)', 'input_keys': "['input1', 'input2']", 'output_key': '"""output"""', 'prompt': '"""prompt"""', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=1, input_keys=['input1', 'input2'], output_key=\n 'output', prompt='prompt', llm_key='test', chain_type='llm_chain_spec')\n", (5610, 5743), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((5875, 6020), 'model.chain_spec.LLMChainSpec', 'LLMChainSpec', ([], {'chain_id': '(2)', 'input_keys': "['input3', 'input4']", 'output_key': '"""output"""', 'prompt': '"""prompt"""', 'llm_key': '"""test"""', 'chain_type': '"""llm_chain_spec"""'}), "(chain_id=2, input_keys=['input3', 'input4'], output_key=\n 'output', prompt='prompt', llm_key='test', chain_type='llm_chain_spec')\n", (5887, 6020), False, 'from model.chain_spec import ChainSpec, LLMChainSpec, SequentialChainSpec, CaseChainSpec, APIChainSpec, ReformatChainSpec, TransformChainSpec, VectorSearchChainSpec\n'), ((13182, 13198), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (13193, 13198), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((14542, 14558), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (14553, 14558), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((13388, 13404), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (13399, 13404), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((14740, 14756), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (14751, 14756), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((13255, 13271), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (13266, 13271), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((14615, 14631), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (14626, 14631), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((13326, 13342), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (13337, 13342), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((13344, 13360), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (13355, 13360), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n'), ((14686, 14702), 'model.tests.factory.llm_factory', 'llm_factory', (['ctx'], {}), '(ctx)\n', (14697, 14702), False, 'from model.tests.factory import SpecFactoryContext, llm_factory, sequential_factory, case_factory\n')]
import threading import time import unittest import unittest.mock from typing import Any, Dict from uuid import UUID import pytest from langsmith import Client from langchain_core.outputs import LLMResult from langchain_core.tracers.langchain import LangChainTracer from langchain_core.tracers.schemas import Run def test_example_id_assignment_threadsafe() -> None: """Test that example assigned at callback start/end is honored.""" example_ids = {} def mock_create_run(**kwargs: Any) -> Any: example_ids[kwargs.get("id")] = kwargs.get("reference_example_id") return unittest.mock.MagicMock() client = unittest.mock.MagicMock(spec=Client) client.create_run = mock_create_run tracer = LangChainTracer(client=client) old_persist_run_single = tracer._persist_run_single def new_persist_run_single(run: Run) -> None: time.sleep(0.01) old_persist_run_single(run) with unittest.mock.patch.object( tracer, "_persist_run_single", new=new_persist_run_single ): run_id_1 = UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a") run_id_2 = UUID("f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1") example_id_1 = UUID("57e42c57-8c79-4d9f-8765-bf6cd3a98055") tracer.example_id = example_id_1 tracer.on_llm_start({"name": "example_1"}, ["foo"], run_id=run_id_1) tracer.on_llm_end(LLMResult(generations=[], llm_output={}), run_id=run_id_1) example_id_2 = UUID("4f31216e-7c26-4027-a5fd-0bbf9ace17dc") tracer.example_id = example_id_2 tracer.on_llm_start({"name": "example_2"}, ["foo"], run_id=run_id_2) tracer.on_llm_end(LLMResult(generations=[], llm_output={}), run_id=run_id_2) tracer.example_id = None expected_example_ids = { run_id_1: example_id_1, run_id_2: example_id_2, } tracer.wait_for_futures() assert example_ids == expected_example_ids def test_log_lock() -> None: """Test that example assigned at callback start/end is honored.""" client = unittest.mock.MagicMock(spec=Client) tracer = LangChainTracer(client=client) with unittest.mock.patch.object(tracer, "_persist_run_single", new=lambda _: _): run_id_1 = UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a") lock = threading.Lock() tracer.on_chain_start({"name": "example_1"}, {"input": lock}, run_id=run_id_1) tracer.on_chain_end({}, run_id=run_id_1) tracer.wait_for_futures() class LangChainProjectNameTest(unittest.TestCase): """ Test that the project name is set correctly for runs. """ class SetProperTracerProjectTestCase: def __init__( self, test_name: str, envvars: Dict[str, str], expected_project_name: str ): self.test_name = test_name self.envvars = envvars self.expected_project_name = expected_project_name def test_correct_get_tracer_project(self) -> None: cases = [ self.SetProperTracerProjectTestCase( test_name="default to 'default' when no project provided", envvars={}, expected_project_name="default", ), self.SetProperTracerProjectTestCase( test_name="use session_name for legacy tracers", envvars={"LANGCHAIN_SESSION": "old_timey_session"}, expected_project_name="old_timey_session", ), self.SetProperTracerProjectTestCase( test_name="use LANGCHAIN_PROJECT over SESSION_NAME", envvars={ "LANGCHAIN_SESSION": "old_timey_session", "LANGCHAIN_PROJECT": "modern_session", }, expected_project_name="modern_session", ), ] for case in cases: with self.subTest(msg=case.test_name): with pytest.MonkeyPatch.context() as mp: for k, v in case.envvars.items(): mp.setenv(k, v) client = unittest.mock.MagicMock(spec=Client) tracer = LangChainTracer(client=client) projects = [] def mock_create_run(**kwargs: Any) -> Any: projects.append(kwargs.get("project_name")) return unittest.mock.MagicMock() client.create_run = mock_create_run tracer.on_llm_start( {"name": "example_1"}, ["foo"], run_id=UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a"), ) tracer.wait_for_futures() assert ( len(projects) == 1 and projects[0] == case.expected_project_name )
[ "langchain_core.tracers.langchain.LangChainTracer", "langchain_core.outputs.LLMResult" ]
[((640, 676), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (663, 676), False, 'import unittest\n'), ((730, 760), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (745, 760), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((2059, 2095), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (2082, 2095), False, 'import unittest\n'), ((2109, 2139), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (2124, 2139), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((600, 625), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {}), '()\n', (623, 625), False, 'import unittest\n'), ((876, 892), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (886, 892), False, 'import time\n'), ((939, 1029), 'unittest.mock.patch.object', 'unittest.mock.patch.object', (['tracer', '"""_persist_run_single"""'], {'new': 'new_persist_run_single'}), "(tracer, '_persist_run_single', new=\n new_persist_run_single)\n", (965, 1029), False, 'import unittest\n'), ((1059, 1103), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (1063, 1103), False, 'from uuid import UUID\n'), ((1123, 1167), 'uuid.UUID', 'UUID', (['"""f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1"""'], {}), "('f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1')\n", (1127, 1167), False, 'from uuid import UUID\n'), ((1191, 1235), 'uuid.UUID', 'UUID', (['"""57e42c57-8c79-4d9f-8765-bf6cd3a98055"""'], {}), "('57e42c57-8c79-4d9f-8765-bf6cd3a98055')\n", (1195, 1235), False, 'from uuid import UUID\n'), ((1462, 1506), 'uuid.UUID', 'UUID', (['"""4f31216e-7c26-4027-a5fd-0bbf9ace17dc"""'], {}), "('4f31216e-7c26-4027-a5fd-0bbf9ace17dc')\n", (1466, 1506), False, 'from uuid import UUID\n'), ((2150, 2224), 'unittest.mock.patch.object', 'unittest.mock.patch.object', (['tracer', '"""_persist_run_single"""'], {'new': '(lambda _: _)'}), "(tracer, '_persist_run_single', new=lambda _: _)\n", (2176, 2224), False, 'import unittest\n'), ((2245, 2289), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (2249, 2289), False, 'from uuid import UUID\n'), ((2305, 2321), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2319, 2321), False, 'import threading\n'), ((1380, 1420), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': '[]', 'llm_output': '{}'}), '(generations=[], llm_output={})\n', (1389, 1420), False, 'from langchain_core.outputs import LLMResult\n'), ((1651, 1691), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': '[]', 'llm_output': '{}'}), '(generations=[], llm_output={})\n', (1660, 1691), False, 'from langchain_core.outputs import LLMResult\n'), ((3929, 3957), 'pytest.MonkeyPatch.context', 'pytest.MonkeyPatch.context', ([], {}), '()\n', (3955, 3957), False, 'import pytest\n'), ((4089, 4125), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (4112, 4125), False, 'import unittest\n'), ((4155, 4185), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (4170, 4185), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((4383, 4408), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {}), '()\n', (4406, 4408), False, 'import unittest\n'), ((4619, 4663), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (4623, 4663), False, 'from uuid import UUID\n')]
import openai import os import dotenv from llama_index.agent.openai import OpenAIAgent from llama_index.llms.azure_openai import AzureOpenAI from llama_index.core.tools.tool_spec.load_and_search.base import LoadAndSearchToolSpec from llama_index.tools.google import GoogleSearchToolSpec from llama_index.tools.weather import OpenWeatherMapToolSpec from llama_index.embeddings.langchain import LangchainEmbedding from llama_index.core import ServiceContext, set_global_service_context from langchain.embeddings import OpenAIEmbeddings # Get API keys from environment variables dotenv.load_dotenv() google_api_key = os.environ.get("GOOGLE_API_KEY") google_cse_id = os.environ.get("GOOGLE_CSE_ID") openweather_api_key = os.environ.get("OPENWEATHER_API_KEY") os.environ["OPENAI_API_KEY"] = os.environ.get("AZURE_API_KEY") openai.api_type = "azure" openai.api_base = os.environ.get("AZURE_API_BASE") openai.api_key = os.environ.get("AZURE_API_KEY") azure_api_key = os.environ["AZURE_API_KEY"] azure_api_type = "azure" azure_api_base = os.environ.get("AZURE_API_BASE") azure_api_version = os.environ.get("AZURE_API_VERSION") azure_chatapi_version = os.environ.get("AZURE_CHATAPI_VERSION") # Check if user set the davinci model flag LLM_DEPLOYMENT_NAME = "gpt-4" LLM_MODEL_NAME = "gpt-4" EMBEDDINGS_DEPLOYMENT_NAME = "text-embedding-ada-002" openai.api_version = os.environ.get("AZURE_CHATAPI_FUNCTIONS_VERSION") print("Using gpt-4 2023-07-01-preview model.") llm = AzureOpenAI( engine=LLM_DEPLOYMENT_NAME, model=LLM_MODEL_NAME, api_key=azure_api_key, api_base=azure_api_base, api_type=azure_api_type, api_version=azure_chatapi_version, temperature=0.5, max_tokens=1024, ) embedding_llm = LangchainEmbedding( OpenAIEmbeddings( model=EMBEDDINGS_DEPLOYMENT_NAME, deployment=EMBEDDINGS_DEPLOYMENT_NAME, openai_api_key=azure_api_key, openai_api_base=azure_api_base, openai_api_type=azure_api_type, openai_api_version=azure_api_version, chunk_size=16, max_retries=3, ), embed_batch_size=1, ) service_context = ServiceContext.from_defaults( llm=llm, embed_model=embedding_llm, chunk_size=256, ) set_global_service_context(service_context) weather_tool = OpenWeatherMapToolSpec( key=openweather_api_key, ) agent = OpenAIAgent.from_tools( weather_tool.to_tool_list(), llm=llm, verbose=True, ) print("-------------------------------------------------------") print(agent.chat("what is the current weather like in Hillsboro, Oregon")) google_tool = GoogleSearchToolSpec( key=google_api_key, engine=google_cse_id ) #Wrap the google search tool as it returns large payloads tools = LoadAndSearchToolSpec.from_defaults( google_tool.to_tool_list()[0], ).to_tool_list() agent = OpenAIAgent.from_tools( tools, llm=llm, embedding_llm=embedding_llm, verbose=True ) # agent = OpenAIAgent.from_tools( # google_tool.to_tool_list(), # llm=llm, # verbose=True # ) print("-------------------------------------------------------") print(agent.chat("whats the latest news summary about India?")) print("-------------------------------------------------------") print(agent.chat('who is Ange Postecoglou, where is he now and what is his philosophy about football?')) print("-------------------------------------------------------") print(agent.chat("whats the latest financial news summary?"))
[ "langchain.embeddings.OpenAIEmbeddings" ]
[((577, 597), 'dotenv.load_dotenv', 'dotenv.load_dotenv', ([], {}), '()\n', (595, 597), False, 'import dotenv\n'), ((615, 647), 'os.environ.get', 'os.environ.get', (['"""GOOGLE_API_KEY"""'], {}), "('GOOGLE_API_KEY')\n", (629, 647), False, 'import os\n'), ((664, 695), 'os.environ.get', 'os.environ.get', (['"""GOOGLE_CSE_ID"""'], {}), "('GOOGLE_CSE_ID')\n", (678, 695), False, 'import os\n'), ((718, 755), 'os.environ.get', 'os.environ.get', (['"""OPENWEATHER_API_KEY"""'], {}), "('OPENWEATHER_API_KEY')\n", (732, 755), False, 'import os\n'), ((788, 819), 'os.environ.get', 'os.environ.get', (['"""AZURE_API_KEY"""'], {}), "('AZURE_API_KEY')\n", (802, 819), False, 'import os\n'), ((864, 896), 'os.environ.get', 'os.environ.get', (['"""AZURE_API_BASE"""'], {}), "('AZURE_API_BASE')\n", (878, 896), False, 'import os\n'), ((914, 945), 'os.environ.get', 'os.environ.get', (['"""AZURE_API_KEY"""'], {}), "('AZURE_API_KEY')\n", (928, 945), False, 'import os\n'), ((1033, 1065), 'os.environ.get', 'os.environ.get', (['"""AZURE_API_BASE"""'], {}), "('AZURE_API_BASE')\n", (1047, 1065), False, 'import os\n'), ((1086, 1121), 'os.environ.get', 'os.environ.get', (['"""AZURE_API_VERSION"""'], {}), "('AZURE_API_VERSION')\n", (1100, 1121), False, 'import os\n'), ((1146, 1185), 'os.environ.get', 'os.environ.get', (['"""AZURE_CHATAPI_VERSION"""'], {}), "('AZURE_CHATAPI_VERSION')\n", (1160, 1185), False, 'import os\n'), ((1360, 1409), 'os.environ.get', 'os.environ.get', (['"""AZURE_CHATAPI_FUNCTIONS_VERSION"""'], {}), "('AZURE_CHATAPI_FUNCTIONS_VERSION')\n", (1374, 1409), False, 'import os\n'), ((1464, 1676), 'llama_index.llms.azure_openai.AzureOpenAI', 'AzureOpenAI', ([], {'engine': 'LLM_DEPLOYMENT_NAME', 'model': 'LLM_MODEL_NAME', 'api_key': 'azure_api_key', 'api_base': 'azure_api_base', 'api_type': 'azure_api_type', 'api_version': 'azure_chatapi_version', 'temperature': '(0.5)', 'max_tokens': '(1024)'}), '(engine=LLM_DEPLOYMENT_NAME, model=LLM_MODEL_NAME, api_key=\n azure_api_key, api_base=azure_api_base, api_type=azure_api_type,\n api_version=azure_chatapi_version, temperature=0.5, max_tokens=1024)\n', (1475, 1676), False, 'from llama_index.llms.azure_openai import AzureOpenAI\n'), ((2113, 2198), 'llama_index.core.ServiceContext.from_defaults', 'ServiceContext.from_defaults', ([], {'llm': 'llm', 'embed_model': 'embedding_llm', 'chunk_size': '(256)'}), '(llm=llm, embed_model=embedding_llm, chunk_size=256\n )\n', (2141, 2198), False, 'from llama_index.core import ServiceContext, set_global_service_context\n'), ((2209, 2252), 'llama_index.core.set_global_service_context', 'set_global_service_context', (['service_context'], {}), '(service_context)\n', (2235, 2252), False, 'from llama_index.core import ServiceContext, set_global_service_context\n'), ((2269, 2316), 'llama_index.tools.weather.OpenWeatherMapToolSpec', 'OpenWeatherMapToolSpec', ([], {'key': 'openweather_api_key'}), '(key=openweather_api_key)\n', (2291, 2316), False, 'from llama_index.tools.weather import OpenWeatherMapToolSpec\n'), ((2579, 2641), 'llama_index.tools.google.GoogleSearchToolSpec', 'GoogleSearchToolSpec', ([], {'key': 'google_api_key', 'engine': 'google_cse_id'}), '(key=google_api_key, engine=google_cse_id)\n', (2599, 2641), False, 'from llama_index.tools.google import GoogleSearchToolSpec\n'), ((2817, 2903), 'llama_index.agent.openai.OpenAIAgent.from_tools', 'OpenAIAgent.from_tools', (['tools'], {'llm': 'llm', 'embedding_llm': 'embedding_llm', 'verbose': '(True)'}), '(tools, llm=llm, embedding_llm=embedding_llm, verbose\n =True)\n', (2839, 2903), False, 'from llama_index.agent.openai import OpenAIAgent\n'), ((1745, 2009), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'model': 'EMBEDDINGS_DEPLOYMENT_NAME', 'deployment': 'EMBEDDINGS_DEPLOYMENT_NAME', 'openai_api_key': 'azure_api_key', 'openai_api_base': 'azure_api_base', 'openai_api_type': 'azure_api_type', 'openai_api_version': 'azure_api_version', 'chunk_size': '(16)', 'max_retries': '(3)'}), '(model=EMBEDDINGS_DEPLOYMENT_NAME, deployment=\n EMBEDDINGS_DEPLOYMENT_NAME, openai_api_key=azure_api_key,\n openai_api_base=azure_api_base, openai_api_type=azure_api_type,\n openai_api_version=azure_api_version, chunk_size=16, max_retries=3)\n', (1761, 2009), False, 'from langchain.embeddings import OpenAIEmbeddings\n')]
import os import re import sys import langchain import langchain.prompts # noqa: F401 import mock import pytest from ddtrace.internal.utils.version import parse_version from tests.contrib.langchain.utils import get_request_vcr from tests.utils import override_global_config SHOULD_USE_LANGCHAIN_COMMUNITY = parse_version(langchain.__version__) >= (0, 1) SHOULD_USE_LANGCHAIN_OPENAI = SHOULD_USE_LANGCHAIN_COMMUNITY pytestmark = pytest.mark.skipif( not SHOULD_USE_LANGCHAIN_COMMUNITY or sys.version_info < (3, 10), reason="This module only tests langchain_community and Python 3.10+", ) @pytest.fixture(scope="session") def request_vcr(): yield get_request_vcr(subdirectory_name="langchain_community") @pytest.mark.parametrize("ddtrace_config_langchain", [dict(logs_enabled=True, log_prompt_completion_sample_rate=1.0)]) def test_global_tags( ddtrace_config_langchain, langchain, langchain_openai, request_vcr, mock_metrics, mock_logs, mock_tracer ): """ When the global config UST tags are set The service name should be used for all data The env should be used for all data The version should be used for all data """ llm = langchain_openai.OpenAI() with override_global_config(dict(service="test-svc", env="staging", version="1234")): with request_vcr.use_cassette("openai_completion_sync.yaml"): llm.invoke("What does Nietzsche mean by 'God is dead'?") span = mock_tracer.pop_traces()[0][0] assert span.resource == "langchain_openai.llms.base.OpenAI" assert span.service == "test-svc" assert span.get_tag("env") == "staging" assert span.get_tag("version") == "1234" assert span.get_tag("langchain.request.provider") == "openai" assert span.get_tag("langchain.request.model") == "gpt-3.5-turbo-instruct" assert span.get_tag("langchain.request.api_key") == "...key>" assert mock_logs.enqueue.call_count == 1 assert mock_metrics.mock_calls for _, _args, kwargs in mock_metrics.mock_calls: expected_metrics = [ "service:test-svc", "env:staging", "version:1234", "langchain.request.model:gpt-3.5-turbo-instruct", "langchain.request.provider:openai", "langchain.request.type:llm", "langchain.request.api_key:...key>", ] actual_tags = kwargs.get("tags") for m in expected_metrics: assert m in actual_tags for call, args, _kwargs in mock_logs.mock_calls: if call != "enqueue": continue log = args[0] assert log["service"] == "test-svc" assert ( log["ddtags"] == "env:staging,version:1234,langchain.request.provider:openai,langchain.request.model:gpt-3.5-turbo-instruct,langchain.request.type:llm,langchain.request.api_key:...key>" # noqa: E501 ) @pytest.mark.snapshot(ignores=["metrics.langchain.tokens.total_cost", "resource"]) def test_openai_llm_sync(langchain, langchain_openai, request_vcr): llm = langchain_openai.OpenAI() with request_vcr.use_cassette("openai_completion_sync.yaml"): llm.invoke("Can you explain what Descartes meant by 'I think, therefore I am'?") @pytest.mark.snapshot def test_openai_llm_sync_multiple_prompts(langchain, langchain_openai, request_vcr): llm = langchain_openai.OpenAI() with request_vcr.use_cassette("openai_completion_sync_multi_prompt.yaml"): llm.generate( prompts=[ "What is the best way to teach a baby multiple languages?", "How many times has Spongebob failed his road test?", ] ) @pytest.mark.asyncio @pytest.mark.snapshot async def test_openai_llm_async(langchain, langchain_openai, request_vcr): llm = langchain_openai.OpenAI() with request_vcr.use_cassette("openai_completion_async.yaml"): await llm.agenerate(["Which team won the 2019 NBA finals?"]) @pytest.mark.snapshot def test_openai_llm_sync_stream(langchain, langchain_openai, request_vcr): llm = langchain_openai.OpenAI(streaming=True) with request_vcr.use_cassette("openai_completion_sync_stream.yaml"): llm.invoke("Why is Spongebob so bad at driving?") @pytest.mark.asyncio @pytest.mark.snapshot async def test_openai_llm_async_stream(langchain, langchain_openai, request_vcr): llm = langchain_openai.OpenAI(streaming=True) with request_vcr.use_cassette("openai_completion_async_stream.yaml"): await llm.agenerate(["Why is Spongebob so bad at driving?"]) @pytest.mark.snapshot(ignores=["meta.error.stack", "resource"]) def test_openai_llm_error(langchain, langchain_openai, request_vcr): import openai # Imported here because the os env OPENAI_API_KEY needs to be set via langchain fixture before import llm = langchain_openai.OpenAI() # if parse_version(openai.__version__) >= (1, 0, 0): if getattr(openai, "__version__", "") >= "1.0.0": invalid_error = openai.BadRequestError else: invalid_error = openai.InvalidRequestError with pytest.raises(invalid_error): with request_vcr.use_cassette("openai_completion_error.yaml"): llm.generate([12345, 123456]) @pytest.mark.snapshot def test_cohere_llm_sync(langchain_community, request_vcr): llm = langchain_community.llms.Cohere(cohere_api_key=os.getenv("COHERE_API_KEY", "<not-a-real-key>")) with request_vcr.use_cassette("cohere_completion_sync.yaml"): llm.invoke("What is the secret Krabby Patty recipe?") @pytest.mark.snapshot def test_ai21_llm_sync(langchain, langchain_community, request_vcr): llm = langchain_community.llms.AI21(ai21_api_key=os.getenv("AI21_API_KEY", "<not-a-real-key>")) with request_vcr.use_cassette("ai21_completion_sync.yaml"): llm.invoke("Why does everyone in Bikini Bottom hate Plankton?") def test_openai_llm_metrics(langchain, langchain_openai, request_vcr, mock_metrics, mock_logs, snapshot_tracer): llm = langchain_openai.OpenAI() with request_vcr.use_cassette("openai_completion_sync.yaml"): llm.invoke("Can you explain what Descartes meant by 'I think, therefore I am'?") expected_tags = [ "version:", "env:", "service:", "langchain.request.provider:openai", "langchain.request.model:gpt-3.5-turbo-instruct", "langchain.request.type:llm", "langchain.request.api_key:...key>", "error:0", ] mock_metrics.assert_has_calls( [ mock.call.distribution("tokens.prompt", 17, tags=expected_tags), mock.call.distribution("tokens.completion", 256, tags=expected_tags), mock.call.distribution("tokens.total", 273, tags=expected_tags), mock.call.increment("tokens.total_cost", mock.ANY, tags=expected_tags), mock.call.distribution("request.duration", mock.ANY, tags=expected_tags), ], any_order=True, ) mock_logs.assert_not_called() @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_llm_logs( langchain, langchain_openai, ddtrace_config_langchain, request_vcr, mock_logs, mock_metrics, mock_tracer ): llm = langchain_openai.OpenAI() with request_vcr.use_cassette("openai_completion_sync.yaml"): llm.invoke("Can you explain what Descartes meant by 'I think, therefore I am'?") span = mock_tracer.pop_traces()[0][0] trace_id, span_id = span.trace_id, span.span_id assert mock_logs.enqueue.call_count == 1 mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.llms.base.OpenAI", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:gpt-3.5-turbo-instruct,langchain.request.type:llm,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(trace_id)[2:], "dd.span_id": str(span_id), "prompts": ["Can you explain what Descartes meant by 'I think, therefore I am'?"], "choices": mock.ANY, } ), ] ) mock_metrics.increment.assert_not_called() mock_metrics.distribution.assert_not_called() mock_metrics.count.assert_not_called() @pytest.mark.snapshot def test_openai_chat_model_sync_call_langchain_openai(langchain, langchain_openai, request_vcr): chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_sync_call.yaml"): chat.invoke(input=[langchain.schema.HumanMessage(content="When do you use 'whom' instead of 'who'?")]) @pytest.mark.snapshot def test_openai_chat_model_sync_generate(langchain, langchain_openai, request_vcr): chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_sync_generate.yaml"): chat.generate( [ [ langchain.schema.SystemMessage(content="Respond like a frat boy."), langchain.schema.HumanMessage( content="Where's the nearest equinox gym from Hudson Yards manhattan?" ), ], [ langchain.schema.SystemMessage(content="Respond with a pirate accent."), langchain.schema.HumanMessage(content="How does one get to Bikini Bottom from New York?"), ], ] ) @pytest.mark.snapshot def test_openai_chat_model_vision_generate(langchain_openai, request_vcr): """ Test that input messages with nested contents are still tagged without error Regression test for https://github.com/DataDog/dd-trace-py/issues/8149. """ image_url = ( "https://upload.wikimedia.org/wikipedia/commons/thumb/d/dd/Gfp-wisconsin-madison-the-nature-boardwalk" ".jpg/2560px-Gfp-wisconsin-madison-the-nature-boardwalk.jpg" ) chat = langchain_openai.ChatOpenAI(model="gpt-4-vision-preview", temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_image_input_sync_generate.yaml"): chat.generate( [ [ langchain.schema.HumanMessage( content=[ {"type": "text", "text": "What’s in this image?"}, { "type": "image_url", "image_url": image_url, }, ], ), ], ] ) @pytest.mark.asyncio @pytest.mark.snapshot async def test_openai_chat_model_async_call(langchain, langchain_openai, request_vcr): chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_async_call.yaml"): await chat._call_async([langchain.schema.HumanMessage(content="When do you use 'whom' instead of 'who'?")]) @pytest.mark.asyncio @pytest.mark.snapshot async def test_openai_chat_model_async_generate(langchain, langchain_openai, request_vcr): chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_async_generate.yaml"): await chat.agenerate( [ [ langchain.schema.SystemMessage(content="Respond like a frat boy."), langchain.schema.HumanMessage( content="Where's the nearest equinox gym from Hudson Yards manhattan?" ), ], [ langchain.schema.SystemMessage(content="Respond with a pirate accent."), langchain.schema.HumanMessage(content="How does one get to Bikini Bottom from New York?"), ], ] ) @pytest.mark.snapshot(token="tests.contrib.langchain.test_langchain_community.test_openai_chat_model_stream") def test_openai_chat_model_sync_stream(langchain, langchain_openai, request_vcr): chat = langchain_openai.ChatOpenAI(streaming=True, temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_sync_stream.yaml"): chat.invoke(input=[langchain.schema.HumanMessage(content="What is the secret Krabby Patty recipe?")]) @pytest.mark.asyncio @pytest.mark.snapshot(token="tests.contrib.langchain.test_langchain_community.test_openai_chat_model_stream") async def test_openai_chat_model_async_stream(langchain, langchain_openai, request_vcr): chat = langchain_openai.ChatOpenAI(streaming=True, temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_async_stream.yaml"): await chat.agenerate([[langchain.schema.HumanMessage(content="What is the secret Krabby Patty recipe?")]]) def test_chat_model_metrics(langchain, langchain_openai, request_vcr, mock_metrics, mock_logs, snapshot_tracer): chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_sync_call.yaml"): chat.invoke(input=[langchain.schema.HumanMessage(content="When do you use 'whom' instead of 'who'?")]) expected_tags = [ "version:", "env:", "service:", "langchain.request.provider:openai", "langchain.request.model:gpt-3.5-turbo", "langchain.request.type:chat_model", "langchain.request.api_key:...key>", "error:0", ] mock_metrics.assert_has_calls( [ mock.call.distribution("tokens.prompt", 20, tags=expected_tags), mock.call.distribution("tokens.completion", 96, tags=expected_tags), mock.call.distribution("tokens.total", 116, tags=expected_tags), mock.call.increment("tokens.total_cost", mock.ANY, tags=expected_tags), mock.call.distribution("request.duration", mock.ANY, tags=expected_tags), ], any_order=True, ) mock_logs.assert_not_called() @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_chat_model_logs( langchain, langchain_openai, ddtrace_config_langchain, request_vcr, mock_logs, mock_metrics, mock_tracer ): chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) with request_vcr.use_cassette("openai_chat_completion_sync_call.yaml"): chat.invoke(input=[langchain.schema.HumanMessage(content="When do you use 'whom' instead of 'who'?")]) span = mock_tracer.pop_traces()[0][0] trace_id, span_id = span.trace_id, span.span_id assert mock_logs.enqueue.call_count == 1 mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.chat_models.base.ChatOpenAI", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:gpt-3.5-turbo,langchain.request.type:chat_model,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(trace_id)[2:], "dd.span_id": str(span_id), "messages": [ [ { "content": "When do you use 'whom' instead of 'who'?", "message_type": "HumanMessage", } ] ], "choices": [ [ { "content": mock.ANY, "message_type": "AIMessage", } ] ], } ), ] ) mock_metrics.increment.assert_not_called() mock_metrics.distribution.assert_not_called() mock_metrics.count.assert_not_called() @pytest.mark.snapshot def test_openai_embedding_query(langchain_openai, request_vcr): with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): embeddings = langchain_openai.OpenAIEmbeddings() with request_vcr.use_cassette("openai_embedding_query.yaml"): embeddings.embed_query("this is a test query.") @pytest.mark.snapshot def test_fake_embedding_query(langchain, langchain_community): embeddings = langchain_community.embeddings.FakeEmbeddings(size=99) embeddings.embed_query(text="foo") @pytest.mark.snapshot def test_fake_embedding_document(langchain, langchain_community): embeddings = langchain_community.embeddings.FakeEmbeddings(size=99) embeddings.embed_documents(texts=["foo", "bar"]) def test_openai_embedding_metrics(langchain_openai, request_vcr, mock_metrics, mock_logs, snapshot_tracer): with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): embeddings = langchain_openai.OpenAIEmbeddings() with request_vcr.use_cassette("openai_embedding_query.yaml"): embeddings.embed_query("this is a test query.") expected_tags = [ "version:", "env:", "service:", "langchain.request.provider:openai", "langchain.request.model:text-embedding-ada-002", "langchain.request.type:embedding", "langchain.request.api_key:...key>", "error:0", ] mock_metrics.assert_has_calls( [mock.call.distribution("request.duration", mock.ANY, tags=expected_tags)], any_order=True, ) mock_logs.assert_not_called() @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_embedding_logs(langchain_openai, ddtrace_config_langchain, request_vcr, mock_logs, mock_metrics, mock_tracer): with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): embeddings = langchain_openai.OpenAIEmbeddings() with request_vcr.use_cassette("openai_embedding_query.yaml"): embeddings.embed_query("this is a test query.") span = mock_tracer.pop_traces()[0][0] trace_id, span_id = span.trace_id, span.span_id assert mock_logs.enqueue.call_count == 1 mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.embeddings.base.OpenAIEmbeddings", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:text-embedding-ada-002,langchain.request.type:embedding,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(trace_id)[2:], "dd.span_id": str(span_id), "inputs": ["this is a test query."], } ), ] ) mock_metrics.increment.assert_not_called() mock_metrics.distribution.assert_not_called() mock_metrics.count.assert_not_called() @pytest.mark.snapshot def test_openai_math_chain_sync(langchain, langchain_openai, request_vcr): """ Test that using the provided LLMMathChain will result in a 3-span trace with the overall LLMMathChain, LLMChain, and underlying OpenAI interface. """ chain = langchain.chains.LLMMathChain.from_llm(langchain_openai.OpenAI(temperature=0)) with request_vcr.use_cassette("openai_math_chain_sync.yaml"): chain.invoke("what is two raised to the fifty-fourth power?") @pytest.mark.asyncio @pytest.mark.snapshot async def test_openai_math_chain_async(langchain, langchain_openai, request_vcr): """ Test that using the provided LLMMathChain will result in a 3-span trace with the overall LLMMathChain, LLMChain, and underlying OpenAI interface. """ chain = langchain.chains.LLMMathChain.from_llm(langchain_openai.OpenAI(temperature=0)) with request_vcr.use_cassette("openai_math_chain_async.yaml"): await chain.ainvoke("what is two raised to the fifty-fourth power?") @pytest.mark.snapshot(token="tests.contrib.langchain.test_langchain_community.test_cohere_math_chain") def test_cohere_math_chain_sync(langchain, langchain_community, request_vcr): """ Test that using the provided LLMMathChain will result in a 3-span trace with the overall LLMMathChain, LLMChain, and underlying Cohere interface. """ chain = langchain.chains.LLMMathChain.from_llm( langchain_community.llms.Cohere(cohere_api_key=os.getenv("COHERE_API_KEY", "<not-a-real-key>")) ) with request_vcr.use_cassette("cohere_math_chain_sync.yaml"): chain.invoke("what is thirteen raised to the .3432 power?") @pytest.mark.snapshot def test_openai_sequential_chain(langchain, langchain_openai, request_vcr): """ Test that using a SequentialChain will result in a 4-span trace with the overall SequentialChain, TransformChain, LLMChain, and underlying OpenAI interface. """ def _transform_func(inputs): """Helper function to replace multiple new lines and multiple spaces with a single space""" text = inputs["text"] text = re.sub(r"(\r\n|\r|\n){2,}", r"\n", text) text = re.sub(r"[ \t]+", " ", text) return {"output_text": text} clean_extra_spaces_chain = langchain.chains.TransformChain( input_variables=["text"], output_variables=["output_text"], transform=_transform_func ) template = """Paraphrase this text: {output_text} In the style of a {style}. Paraphrase: """ prompt = langchain.prompts.PromptTemplate(input_variables=["style", "output_text"], template=template) style_paraphrase_chain = langchain.chains.LLMChain( llm=langchain_openai.OpenAI(), prompt=prompt, output_key="final_output" ) sequential_chain = langchain.chains.SequentialChain( chains=[clean_extra_spaces_chain, style_paraphrase_chain], input_variables=["text", "style"], output_variables=["final_output"], ) input_text = """ Chains allow us to combine multiple components together to create a single, coherent application. For example, we can create a chain that takes user input, format it with a PromptTemplate, and then passes the formatted response to an LLM. We can build more complex chains by combining multiple chains together, or by combining chains with other components. """ with request_vcr.use_cassette("openai_paraphrase.yaml"): sequential_chain.invoke({"text": input_text, "style": "a 90s rapper"}) @pytest.mark.snapshot def test_openai_sequential_chain_with_multiple_llm_sync(langchain, langchain_openai, request_vcr): template = """Paraphrase this text: {input_text} Paraphrase: """ prompt = langchain.prompts.PromptTemplate(input_variables=["input_text"], template=template) style_paraphrase_chain = langchain.chains.LLMChain( llm=langchain_openai.OpenAI(), prompt=prompt, output_key="paraphrased_output" ) rhyme_template = """Make this text rhyme: {paraphrased_output} Rhyme: """ rhyme_prompt = langchain.prompts.PromptTemplate(input_variables=["paraphrased_output"], template=rhyme_template) rhyme_chain = langchain.chains.LLMChain( llm=langchain_openai.OpenAI(), prompt=rhyme_prompt, output_key="final_output" ) sequential_chain = langchain.chains.SequentialChain( chains=[style_paraphrase_chain, rhyme_chain], input_variables=["input_text"], output_variables=["final_output"], ) input_text = """ I have convinced myself that there is absolutely nothing in the world, no sky, no earth, no minds, no bodies. Does it now follow that I too do not exist? No: if I convinced myself of something then I certainly existed. But there is a deceiver of supreme power and cunning who is deliberately and constantly deceiving me. In that case I too undoubtedly exist, if he is deceiving me; and let him deceive me as much as he can, he will never bring it about that I am nothing so long as I think that I am something. So after considering everything very thoroughly, I must finally conclude that this proposition, I am, I exist, is necessarily true whenever it is put forward by me or conceived in my mind. """ with request_vcr.use_cassette("openai_sequential_paraphrase_and_rhyme_sync.yaml"): sequential_chain.invoke({"input_text": input_text}) @pytest.mark.asyncio @pytest.mark.snapshot async def test_openai_sequential_chain_with_multiple_llm_async(langchain, langchain_openai, request_vcr): template = """Paraphrase this text: {input_text} Paraphrase: """ prompt = langchain.prompts.PromptTemplate(input_variables=["input_text"], template=template) style_paraphrase_chain = langchain.chains.LLMChain( llm=langchain_openai.OpenAI(), prompt=prompt, output_key="paraphrased_output" ) rhyme_template = """Make this text rhyme: {paraphrased_output} Rhyme: """ rhyme_prompt = langchain.prompts.PromptTemplate(input_variables=["paraphrased_output"], template=rhyme_template) rhyme_chain = langchain.chains.LLMChain( llm=langchain_openai.OpenAI(), prompt=rhyme_prompt, output_key="final_output" ) sequential_chain = langchain.chains.SequentialChain( chains=[style_paraphrase_chain, rhyme_chain], input_variables=["input_text"], output_variables=["final_output"], ) input_text = """ I have convinced myself that there is absolutely nothing in the world, no sky, no earth, no minds, no bodies. Does it now follow that I too do not exist? No: if I convinced myself of something then I certainly existed. But there is a deceiver of supreme power and cunning who is deliberately and constantly deceiving me. In that case I too undoubtedly exist, if he is deceiving me; and let him deceive me as much as he can, he will never bring it about that I am nothing so long as I think that I am something. So after considering everything very thoroughly, I must finally conclude that this proposition, I am, I exist, is necessarily true whenever it is put forward by me or conceived in my mind. """ with request_vcr.use_cassette("openai_sequential_paraphrase_and_rhyme_async.yaml"): await sequential_chain.ainvoke({"input_text": input_text}) @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_chain_logs( langchain, langchain_openai, ddtrace_config_langchain, request_vcr, mock_logs, mock_metrics, mock_tracer ): chain = langchain.chains.LLMMathChain.from_llm(langchain_openai.OpenAI(temperature=0)) with request_vcr.use_cassette("openai_math_chain_sync.yaml"): chain.invoke("what is two raised to the fifty-fourth power?") traces = mock_tracer.pop_traces() base_chain_span = traces[0][0] mid_chain_span = traces[0][1] llm_span = traces[0][2] assert mock_logs.enqueue.call_count == 3 # This operation includes 2 chains and 1 LLM call mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.llms.base.OpenAI", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:gpt-3.5-turbo-instruct,langchain.request.type:llm,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(llm_span.trace_id)[2:], "dd.span_id": str(llm_span.span_id), "prompts": mock.ANY, "choices": mock.ANY, } ), mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain.chains.llm.LLMChain", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:,langchain.request.model:,langchain.request.type:chain,langchain.request.api_key:", # noqa: E501 "dd.trace_id": hex(mid_chain_span.trace_id)[2:], "dd.span_id": str(mid_chain_span.span_id), "inputs": mock.ANY, "prompt": mock.ANY, "outputs": { "question": "what is two raised to the fifty-fourth power?", "stop": mock.ANY, "text": '```text\n2**54\n```\n...numexpr.evaluate("2**54")...\n', }, } ), mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain.chains.llm_math.base.LLMMathChain", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:,langchain.request.model:,langchain.request.type:chain,langchain.request.api_key:", # noqa: E501 "dd.trace_id": hex(base_chain_span.trace_id)[2:], "dd.span_id": str(base_chain_span.span_id), "inputs": {"question": "what is two raised to the fifty-fourth power?"}, "prompt": mock.ANY, "outputs": { "question": "what is two raised to the fifty-fourth power?", "answer": "Answer: 18014398509481984", }, } ), ] ) mock_metrics.increment.assert_not_called() mock_metrics.distribution.assert_not_called() mock_metrics.count.assert_not_called() def test_chat_prompt_template_does_not_parse_template(langchain, langchain_openai, mock_tracer): """ Test that tracing a chain with a ChatPromptTemplate does not try to directly parse the template, as ChatPromptTemplates do not contain a specific template attribute (which will lead to an attribute error) but instead contain multiple messages each with their own prompt template and are not trivial to tag. """ import langchain.prompts.chat # noqa: F401 with mock.patch("langchain_openai.ChatOpenAI._generate", side_effect=Exception("Mocked Error")): with pytest.raises(Exception) as exc_info: chat = langchain_openai.ChatOpenAI(temperature=0) template = "You are a helpful assistant that translates english to pirate." system_message_prompt = langchain.prompts.chat.SystemMessagePromptTemplate.from_template(template) example_human = langchain.prompts.chat.HumanMessagePromptTemplate.from_template("Hi") example_ai = langchain.prompts.chat.AIMessagePromptTemplate.from_template("Argh me mateys") human_template = "{text}" human_message_prompt = langchain.prompts.chat.HumanMessagePromptTemplate.from_template(human_template) chat_prompt = langchain.prompts.chat.ChatPromptTemplate.from_messages( [system_message_prompt, example_human, example_ai, human_message_prompt] ) chain = langchain.chains.LLMChain(llm=chat, prompt=chat_prompt) chain.invoke("I love programming.") assert str(exc_info.value) == "Mocked Error" traces = mock_tracer.pop_traces() chain_span = traces[0][0] assert chain_span.get_tag("langchain.request.inputs.text") == "I love programming." assert chain_span.get_tag("langchain.request.type") == "chain" assert chain_span.get_tag("langchain.request.prompt") is None @pytest.mark.snapshot def test_pinecone_vectorstore_similarity_search(langchain_community, langchain_openai, request_vcr): """ Test that calling a similarity search on a Pinecone vectorstore with langchain will result in a 2-span trace with a vectorstore span and underlying OpenAI embedding interface span. """ import langchain_pinecone import pinecone with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): with request_vcr.use_cassette("openai_pinecone_similarity_search.yaml"): pc = pinecone.Pinecone( api_key=os.getenv("PINECONE_API_KEY", "<not-a-real-key>"), environment=os.getenv("PINECONE_ENV", "<not-a-real-env>"), ) embed = langchain_openai.OpenAIEmbeddings(model="text-embedding-ada-002") index = pc.Index("langchain-retrieval") vectorstore = langchain_pinecone.PineconeVectorStore(index, embed, "text") vectorstore.similarity_search("Who was Alan Turing?", 1) @pytest.mark.snapshot def test_pinecone_vectorstore_retrieval_chain(langchain_community, langchain_openai, request_vcr): """ Test that calling a similarity search on a Pinecone vectorstore with langchain will result in a 2-span trace with a vectorstore span and underlying OpenAI embedding interface span. """ import langchain_pinecone import pinecone with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): with request_vcr.use_cassette("openai_pinecone_vectorstore_retrieval_chain.yaml"): pc = pinecone.Pinecone( api_key=os.getenv("PINECONE_API_KEY", "<not-a-real-key>"), environment=os.getenv("PINECONE_ENV", "<not-a-real-env>"), ) embed = langchain_openai.OpenAIEmbeddings(model="text-embedding-ada-002") index = pc.Index("langchain-retrieval") vectorstore = langchain_pinecone.PineconeVectorStore(index, embed, "text") llm = langchain_openai.OpenAI() qa_with_sources = langchain.chains.RetrievalQAWithSourcesChain.from_chain_type( llm=llm, chain_type="stuff", retriever=vectorstore.as_retriever() ) qa_with_sources.invoke("Who was Alan Turing?") def test_vectorstore_similarity_search_metrics( langchain_community, langchain_openai, request_vcr, mock_metrics, mock_logs, snapshot_tracer ): import langchain_pinecone import pinecone with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): with request_vcr.use_cassette("openai_pinecone_similarity_search.yaml"): pc = pinecone.Pinecone( api_key=os.getenv("PINECONE_API_KEY", "<not-a-real-key>"), environment=os.getenv("PINECONE_ENV", "<not-a-real-env>"), ) embed = langchain_openai.OpenAIEmbeddings(model="text-embedding-ada-002") index = pc.Index("langchain-retrieval") vectorstore = langchain_pinecone.PineconeVectorStore(index, embed, "text") vectorstore.similarity_search("Who was Alan Turing?", 1) expected_tags = [ "version:", "env:", "service:", "langchain.request.provider:pineconevectorstore", "langchain.request.model:", "langchain.request.type:similarity_search", "langchain.request.api_key:", "error:0", ] mock_metrics.distribution.assert_called_with("request.duration", mock.ANY, tags=expected_tags), mock_logs.assert_not_called() @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_vectorstore_logs( langchain_openai, langchain_community, ddtrace_config_langchain, request_vcr, mock_logs, mock_metrics, mock_tracer ): import langchain_pinecone import pinecone with mock.patch("langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", return_value=[0.0] * 1536): with request_vcr.use_cassette("openai_pinecone_similarity_search.yaml"): pc = pinecone.Pinecone( api_key=os.getenv("PINECONE_API_KEY", "<not-a-real-key>"), environment=os.getenv("PINECONE_ENV", "<not-a-real-env>"), ) embed = langchain_openai.OpenAIEmbeddings(model="text-embedding-ada-002") index = pc.Index("langchain-retrieval") vectorstore = langchain_pinecone.PineconeVectorStore(index, embed, "text") vectorstore.similarity_search("Who was Alan Turing?", 1) traces = mock_tracer.pop_traces() vectorstore_span = traces[0][0] embeddings_span = traces[0][1] assert mock_logs.enqueue.call_count == 2 # This operation includes 1 vectorstore call and 1 embeddings call mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.embeddings.base.OpenAIEmbeddings", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:text-embedding-ada-002,langchain.request.type:embedding,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(embeddings_span.trace_id)[2:], "dd.span_id": str(embeddings_span.span_id), "inputs": ["Who was Alan Turing?"], } ), mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_pinecone.vectorstores.PineconeVectorStore", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "info", "ddtags": "env:,version:,langchain.request.provider:pineconevectorstore,langchain.request.model:,langchain.request.type:similarity_search,langchain.request.api_key:", # noqa: E501 "dd.trace_id": hex(vectorstore_span.trace_id)[2:], "dd.span_id": str(vectorstore_span.span_id), "query": "Who was Alan Turing?", "k": 1, "documents": mock.ANY, } ), ] ) mock_metrics.increment.assert_not_called() mock_metrics.distribution.assert_not_called() mock_metrics.count.assert_not_called() @pytest.mark.snapshot(ignores=["metrics.langchain.tokens.total_cost", "resource"]) def test_openai_integration(langchain, request_vcr, ddtrace_run_python_code_in_subprocess): env = os.environ.copy() pypath = [os.path.dirname(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))] if "PYTHONPATH" in env: pypath.append(env["PYTHONPATH"]) env.update( { "PYTHONPATH": ":".join(pypath), # Disable metrics because the test agent doesn't support metrics "DD_LANGCHAIN_METRICS_ENABLED": "false", "DD_OPENAI_METRICS_ENABLED": "false", "OPENAI_API_KEY": "<not-a-real-key>", } ) out, err, status, pid = ddtrace_run_python_code_in_subprocess( """ from langchain_openai import OpenAI import ddtrace from tests.contrib.langchain.test_langchain_community import get_request_vcr llm = OpenAI() with get_request_vcr(subdirectory_name="langchain_community").use_cassette("openai_completion_sync.yaml"): llm.invoke("Can you explain what Descartes meant by 'I think, therefore I am'?") """, env=env, ) assert status == 0, err assert out == b"" assert err == b"" @pytest.mark.snapshot @pytest.mark.parametrize("schema_version", [None, "v0", "v1"]) @pytest.mark.parametrize("service_name", [None, "mysvc"]) def test_openai_service_name( langchain, request_vcr, ddtrace_run_python_code_in_subprocess, schema_version, service_name ): import os env = os.environ.copy() pypath = [os.path.dirname(os.path.dirname(os.path.dirname(os.path.dirname(__file__))))] if "PYTHONPATH" in env: pypath.append(env["PYTHONPATH"]) env.update( { "PYTHONPATH": ":".join(pypath), # Disable metrics because the test agent doesn't support metrics "DD_LANGCHAIN_METRICS_ENABLED": "false", "DD_OPENAI_METRICS_ENABLED": "false", "OPENAI_API_KEY": "<not-a-real-key>", } ) if service_name: env["DD_SERVICE"] = service_name if schema_version: env["DD_TRACE_SPAN_ATTRIBUTE_SCHEMA"] = schema_version out, err, status, pid = ddtrace_run_python_code_in_subprocess( # TODO: need to correct this """ from langchain_openai import OpenAI import ddtrace from tests.contrib.langchain.test_langchain_community import get_request_vcr llm = OpenAI() with get_request_vcr(subdirectory_name="langchain_community").use_cassette("openai_completion_sync.yaml"): llm.invoke("Can you explain what Descartes meant by 'I think, therefore I am'?") """, env=env, ) assert status == 0, err assert out == b"" assert err == b"" @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_llm_logs_when_response_not_completed( langchain, langchain_openai, langchain_community, ddtrace_config_langchain, mock_logs, mock_metrics, mock_tracer ): """Test that errors get logged even if the response is not returned.""" with mock.patch("langchain_openai.OpenAI._generate", side_effect=Exception("Mocked Error")): with pytest.raises(Exception) as exc_info: llm = langchain_openai.OpenAI() llm.invoke("Can you please not return an error?") assert str(exc_info.value) == "Mocked Error" span = mock_tracer.pop_traces()[0][0] trace_id, span_id = span.trace_id, span.span_id assert mock_logs.enqueue.call_count == 1 mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.llms.base.OpenAI", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "error", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:gpt-3.5-turbo-instruct,langchain.request.type:llm,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(trace_id)[2:], "dd.span_id": str(span_id), "prompts": ["Can you please not return an error?"], "choices": [], } ), ] ) @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_chat_model_logs_when_response_not_completed( langchain, langchain_openai, ddtrace_config_langchain, mock_logs, mock_metrics, mock_tracer ): """Test that errors get logged even if the response is not returned.""" with mock.patch("langchain_openai.ChatOpenAI._generate", side_effect=Exception("Mocked Error")): with pytest.raises(Exception) as exc_info: chat = langchain_openai.ChatOpenAI(temperature=0, max_tokens=256) chat.invoke(input=[langchain.schema.HumanMessage(content="Can you please not return an error?")]) assert str(exc_info.value) == "Mocked Error" span = mock_tracer.pop_traces()[0][0] trace_id, span_id = span.trace_id, span.span_id assert mock_logs.enqueue.call_count == 1 mock_logs.enqueue.assert_called_with( { "timestamp": mock.ANY, "message": "sampled langchain_openai.chat_models.base.ChatOpenAI", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "error", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:gpt-3.5-turbo,langchain.request.type:chat_model,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(trace_id)[2:], "dd.span_id": str(span_id), "messages": [ [ { "content": "Can you please not return an error?", "message_type": "HumanMessage", } ] ], "choices": [], } ) @pytest.mark.parametrize( "ddtrace_config_langchain", [dict(metrics_enabled=False, logs_enabled=True, log_prompt_completion_sample_rate=1.0)], ) def test_embedding_logs_when_response_not_completed( langchain_openai, ddtrace_config_langchain, mock_logs, mock_metrics, mock_tracer ): """Test that errors get logged even if the response is not returned.""" with mock.patch( "langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings", side_effect=Exception("Mocked Error") ): with pytest.raises(Exception) as exc_info: embeddings = langchain_openai.OpenAIEmbeddings() embeddings.embed_query("Can you please not return an error?") assert str(exc_info.value) == "Mocked Error" span = mock_tracer.pop_traces()[0][0] trace_id, span_id = span.trace_id, span.span_id assert mock_logs.enqueue.call_count == 1 mock_logs.assert_has_calls( [ mock.call.enqueue( { "timestamp": mock.ANY, "message": "sampled langchain_openai.embeddings.base.OpenAIEmbeddings", "hostname": mock.ANY, "ddsource": "langchain", "service": "", "status": "error", "ddtags": "env:,version:,langchain.request.provider:openai,langchain.request.model:text-embedding-ada-002,langchain.request.type:embedding,langchain.request.api_key:...key>", # noqa: E501 "dd.trace_id": hex(trace_id)[2:], "dd.span_id": str(span_id), "inputs": ["Can you please not return an error?"], } ), ] )
[ "langchain_pinecone.PineconeVectorStore", "langchain.prompts.chat.AIMessagePromptTemplate.from_template", "langchain.chains.TransformChain", "langchain.chains.SequentialChain", "langchain.prompts.chat.SystemMessagePromptTemplate.from_template", "langchain.schema.HumanMessage", "langchain.schema.SystemMessage", "langchain.prompts.chat.HumanMessagePromptTemplate.from_template", "langchain.chains.LLMChain", "langchain.prompts.PromptTemplate", "langchain.prompts.chat.ChatPromptTemplate.from_messages" ]
[((434, 597), 'pytest.mark.skipif', 'pytest.mark.skipif', (['(not SHOULD_USE_LANGCHAIN_COMMUNITY or sys.version_info < (3, 10))'], {'reason': '"""This module only tests langchain_community and Python 3.10+"""'}), "(not SHOULD_USE_LANGCHAIN_COMMUNITY or sys.version_info <\n (3, 10), reason=\n 'This module only tests langchain_community and Python 3.10+')\n", (452, 597), False, 'import pytest\n'), ((603, 634), 'pytest.fixture', 'pytest.fixture', ([], {'scope': '"""session"""'}), "(scope='session')\n", (617, 634), False, 'import pytest\n'), ((2890, 2975), 'pytest.mark.snapshot', 'pytest.mark.snapshot', ([], {'ignores': "['metrics.langchain.tokens.total_cost', 'resource']"}), "(ignores=['metrics.langchain.tokens.total_cost',\n 'resource'])\n", (2910, 2975), False, 'import pytest\n'), ((4564, 4626), 'pytest.mark.snapshot', 'pytest.mark.snapshot', ([], {'ignores': "['meta.error.stack', 'resource']"}), "(ignores=['meta.error.stack', 'resource'])\n", (4584, 4626), False, 'import pytest\n'), ((12289, 12407), 'pytest.mark.snapshot', 'pytest.mark.snapshot', ([], {'token': '"""tests.contrib.langchain.test_langchain_community.test_openai_chat_model_stream"""'}), "(token=\n 'tests.contrib.langchain.test_langchain_community.test_openai_chat_model_stream'\n )\n", (12309, 12407), False, 'import pytest\n'), ((12778, 12896), 'pytest.mark.snapshot', 'pytest.mark.snapshot', ([], {'token': '"""tests.contrib.langchain.test_langchain_community.test_openai_chat_model_stream"""'}), "(token=\n 'tests.contrib.langchain.test_langchain_community.test_openai_chat_model_stream'\n )\n", (12798, 12896), False, 'import pytest\n'), ((20974, 21080), 'pytest.mark.snapshot', 'pytest.mark.snapshot', ([], {'token': '"""tests.contrib.langchain.test_langchain_community.test_cohere_math_chain"""'}), "(token=\n 'tests.contrib.langchain.test_langchain_community.test_cohere_math_chain')\n", (20994, 21080), False, 'import pytest\n'), ((39920, 40005), 'pytest.mark.snapshot', 'pytest.mark.snapshot', ([], {'ignores': "['metrics.langchain.tokens.total_cost', 'resource']"}), "(ignores=['metrics.langchain.tokens.total_cost',\n 'resource'])\n", (39940, 40005), False, 'import pytest\n'), ((41138, 41199), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""schema_version"""', "[None, 'v0', 'v1']"], {}), "('schema_version', [None, 'v0', 'v1'])\n", (41161, 41199), False, 'import pytest\n'), ((41201, 41257), 'pytest.mark.parametrize', 'pytest.mark.parametrize', (['"""service_name"""', "[None, 'mysvc']"], {}), "('service_name', [None, 'mysvc'])\n", (41224, 41257), False, 'import pytest\n'), ((312, 348), 'ddtrace.internal.utils.version.parse_version', 'parse_version', (['langchain.__version__'], {}), '(langchain.__version__)\n', (325, 348), False, 'from ddtrace.internal.utils.version import parse_version\n'), ((22234, 22357), 'langchain.chains.TransformChain', 'langchain.chains.TransformChain', ([], {'input_variables': "['text']", 'output_variables': "['output_text']", 'transform': '_transform_func'}), "(input_variables=['text'], output_variables=\n ['output_text'], transform=_transform_func)\n", (22265, 22357), False, 'import langchain\n'), ((22504, 22601), 'langchain.prompts.PromptTemplate', 'langchain.prompts.PromptTemplate', ([], {'input_variables': "['style', 'output_text']", 'template': 'template'}), "(input_variables=['style', 'output_text'],\n template=template)\n", (22536, 22601), False, 'import langchain\n'), ((22763, 22932), 'langchain.chains.SequentialChain', 'langchain.chains.SequentialChain', ([], {'chains': '[clean_extra_spaces_chain, style_paraphrase_chain]', 'input_variables': "['text', 'style']", 'output_variables': "['final_output']"}), "(chains=[clean_extra_spaces_chain,\n style_paraphrase_chain], input_variables=['text', 'style'],\n output_variables=['final_output'])\n", (22795, 22932), False, 'import langchain\n'), ((23771, 23859), 'langchain.prompts.PromptTemplate', 'langchain.prompts.PromptTemplate', ([], {'input_variables': "['input_text']", 'template': 'template'}), "(input_variables=['input_text'], template=\n template)\n", (23803, 23859), False, 'import langchain\n'), ((24118, 24219), 'langchain.prompts.PromptTemplate', 'langchain.prompts.PromptTemplate', ([], {'input_variables': "['paraphrased_output']", 'template': 'rhyme_template'}), "(input_variables=['paraphrased_output'],\n template=rhyme_template)\n", (24150, 24219), False, 'import langchain\n'), ((24376, 24530), 'langchain.chains.SequentialChain', 'langchain.chains.SequentialChain', ([], {'chains': '[style_paraphrase_chain, rhyme_chain]', 'input_variables': "['input_text']", 'output_variables': "['final_output']"}), "(chains=[style_paraphrase_chain,\n rhyme_chain], input_variables=['input_text'], output_variables=[\n 'final_output'])\n", (24408, 24530), False, 'import langchain\n'), ((25773, 25861), 'langchain.prompts.PromptTemplate', 'langchain.prompts.PromptTemplate', ([], {'input_variables': "['input_text']", 'template': 'template'}), "(input_variables=['input_text'], template=\n template)\n", (25805, 25861), False, 'import langchain\n'), ((26120, 26221), 'langchain.prompts.PromptTemplate', 'langchain.prompts.PromptTemplate', ([], {'input_variables': "['paraphrased_output']", 'template': 'rhyme_template'}), "(input_variables=['paraphrased_output'],\n template=rhyme_template)\n", (26152, 26221), False, 'import langchain\n'), ((26378, 26532), 'langchain.chains.SequentialChain', 'langchain.chains.SequentialChain', ([], {'chains': '[style_paraphrase_chain, rhyme_chain]', 'input_variables': "['input_text']", 'output_variables': "['final_output']"}), "(chains=[style_paraphrase_chain,\n rhyme_chain], input_variables=['input_text'], output_variables=[\n 'final_output'])\n", (26410, 26532), False, 'import langchain\n'), ((40104, 40121), 'os.environ.copy', 'os.environ.copy', ([], {}), '()\n', (40119, 40121), False, 'import os\n'), ((41412, 41429), 'os.environ.copy', 'os.environ.copy', ([], {}), '()\n', (41427, 41429), False, 'import os\n'), ((664, 720), 'tests.contrib.langchain.utils.get_request_vcr', 'get_request_vcr', ([], {'subdirectory_name': '"""langchain_community"""'}), "(subdirectory_name='langchain_community')\n", (679, 720), False, 'from tests.contrib.langchain.utils import get_request_vcr\n'), ((5083, 5111), 'pytest.raises', 'pytest.raises', (['invalid_error'], {}), '(invalid_error)\n', (5096, 5111), False, 'import pytest\n'), ((16702, 16805), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (16712, 16805), False, 'import mock\n'), ((17522, 17625), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (17532, 17625), False, 'import mock\n'), ((18572, 18675), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (18582, 18675), False, 'import mock\n'), ((22080, 22123), 're.sub', 're.sub', (['"""(\\\\r\\\\n|\\\\r|\\\\n){2,}"""', '"""\\\\n"""', 'text'], {}), "('(\\\\r\\\\n|\\\\r|\\\\n){2,}', '\\\\n', text)\n", (22086, 22123), False, 'import re\n'), ((22136, 22164), 're.sub', 're.sub', (['"""[ \\\\t]+"""', '""" """', 'text'], {}), "('[ \\\\t]+', ' ', text)\n", (22142, 22164), False, 'import re\n'), ((33560, 33663), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (33570, 33663), False, 'import mock\n'), ((34624, 34727), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (34634, 34727), False, 'import mock\n'), ((35743, 35846), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (35753, 35846), False, 'import mock\n'), ((37204, 37307), 'mock.patch', 'mock.patch', (['"""langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings"""'], {'return_value': '([0.0] * 1536)'}), "('langchain_openai.OpenAIEmbeddings._get_len_safe_embeddings',\n return_value=[0.0] * 1536)\n", (37214, 37307), False, 'import mock\n'), ((5367, 5414), 'os.getenv', 'os.getenv', (['"""COHERE_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('COHERE_API_KEY', '<not-a-real-key>')\n", (5376, 5414), False, 'import os\n'), ((5690, 5735), 'os.getenv', 'os.getenv', (['"""AI21_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('AI21_API_KEY', '<not-a-real-key>')\n", (5699, 5735), False, 'import os\n'), ((6525, 6588), 'mock.call.distribution', 'mock.call.distribution', (['"""tokens.prompt"""', '(17)'], {'tags': 'expected_tags'}), "('tokens.prompt', 17, tags=expected_tags)\n", (6547, 6588), False, 'import mock\n'), ((6602, 6670), 'mock.call.distribution', 'mock.call.distribution', (['"""tokens.completion"""', '(256)'], {'tags': 'expected_tags'}), "('tokens.completion', 256, tags=expected_tags)\n", (6624, 6670), False, 'import mock\n'), ((6684, 6747), 'mock.call.distribution', 'mock.call.distribution', (['"""tokens.total"""', '(273)'], {'tags': 'expected_tags'}), "('tokens.total', 273, tags=expected_tags)\n", (6706, 6747), False, 'import mock\n'), ((6761, 6831), 'mock.call.increment', 'mock.call.increment', (['"""tokens.total_cost"""', 'mock.ANY'], {'tags': 'expected_tags'}), "('tokens.total_cost', mock.ANY, tags=expected_tags)\n", (6780, 6831), False, 'import mock\n'), ((6845, 6917), 'mock.call.distribution', 'mock.call.distribution', (['"""request.duration"""', 'mock.ANY'], {'tags': 'expected_tags'}), "('request.duration', mock.ANY, tags=expected_tags)\n", (6867, 6917), False, 'import mock\n'), ((13972, 14035), 'mock.call.distribution', 'mock.call.distribution', (['"""tokens.prompt"""', '(20)'], {'tags': 'expected_tags'}), "('tokens.prompt', 20, tags=expected_tags)\n", (13994, 14035), False, 'import mock\n'), ((14049, 14116), 'mock.call.distribution', 'mock.call.distribution', (['"""tokens.completion"""', '(96)'], {'tags': 'expected_tags'}), "('tokens.completion', 96, tags=expected_tags)\n", (14071, 14116), False, 'import mock\n'), ((14130, 14193), 'mock.call.distribution', 'mock.call.distribution', (['"""tokens.total"""', '(116)'], {'tags': 'expected_tags'}), "('tokens.total', 116, tags=expected_tags)\n", (14152, 14193), False, 'import mock\n'), ((14207, 14277), 'mock.call.increment', 'mock.call.increment', (['"""tokens.total_cost"""', 'mock.ANY'], {'tags': 'expected_tags'}), "('tokens.total_cost', mock.ANY, tags=expected_tags)\n", (14226, 14277), False, 'import mock\n'), ((14291, 14363), 'mock.call.distribution', 'mock.call.distribution', (['"""request.duration"""', 'mock.ANY'], {'tags': 'expected_tags'}), "('request.duration', mock.ANY, tags=expected_tags)\n", (14313, 14363), False, 'import mock\n'), ((18149, 18221), 'mock.call.distribution', 'mock.call.distribution', (['"""request.duration"""', 'mock.ANY'], {'tags': 'expected_tags'}), "('request.duration', mock.ANY, tags=expected_tags)\n", (18171, 18221), False, 'import mock\n'), ((31864, 31888), 'pytest.raises', 'pytest.raises', (['Exception'], {}), '(Exception)\n', (31877, 31888), False, 'import pytest\n'), ((32088, 32162), 'langchain.prompts.chat.SystemMessagePromptTemplate.from_template', 'langchain.prompts.chat.SystemMessagePromptTemplate.from_template', (['template'], {}), '(template)\n', (32152, 32162), False, 'import langchain\n'), ((32191, 32260), 'langchain.prompts.chat.HumanMessagePromptTemplate.from_template', 'langchain.prompts.chat.HumanMessagePromptTemplate.from_template', (['"""Hi"""'], {}), "('Hi')\n", (32254, 32260), False, 'import langchain\n'), ((32286, 32364), 'langchain.prompts.chat.AIMessagePromptTemplate.from_template', 'langchain.prompts.chat.AIMessagePromptTemplate.from_template', (['"""Argh me mateys"""'], {}), "('Argh me mateys')\n", (32346, 32364), False, 'import langchain\n'), ((32438, 32517), 'langchain.prompts.chat.HumanMessagePromptTemplate.from_template', 'langchain.prompts.chat.HumanMessagePromptTemplate.from_template', (['human_template'], {}), '(human_template)\n', (32501, 32517), False, 'import langchain\n'), ((32544, 32678), 'langchain.prompts.chat.ChatPromptTemplate.from_messages', 'langchain.prompts.chat.ChatPromptTemplate.from_messages', (['[system_message_prompt, example_human, example_ai, human_message_prompt]'], {}), '([\n system_message_prompt, example_human, example_ai, human_message_prompt])\n', (32599, 32678), False, 'import langchain\n'), ((32724, 32779), 'langchain.chains.LLMChain', 'langchain.chains.LLMChain', ([], {'llm': 'chat', 'prompt': 'chat_prompt'}), '(llm=chat, prompt=chat_prompt)\n', (32749, 32779), False, 'import langchain\n'), ((34106, 34166), 'langchain_pinecone.PineconeVectorStore', 'langchain_pinecone.PineconeVectorStore', (['index', 'embed', '"""text"""'], {}), "(index, embed, 'text')\n", (34144, 34166), False, 'import langchain_pinecone\n'), ((35180, 35240), 'langchain_pinecone.PineconeVectorStore', 'langchain_pinecone.PineconeVectorStore', (['index', 'embed', '"""text"""'], {}), "(index, embed, 'text')\n", (35218, 35240), False, 'import langchain_pinecone\n'), ((36289, 36349), 'langchain_pinecone.PineconeVectorStore', 'langchain_pinecone.PineconeVectorStore', (['index', 'embed', '"""text"""'], {}), "(index, embed, 'text')\n", (36327, 36349), False, 'import langchain_pinecone\n'), ((37750, 37810), 'langchain_pinecone.PineconeVectorStore', 'langchain_pinecone.PineconeVectorStore', (['index', 'embed', '"""text"""'], {}), "(index, embed, 'text')\n", (37788, 37810), False, 'import langchain_pinecone\n'), ((43114, 43138), 'pytest.raises', 'pytest.raises', (['Exception'], {}), '(Exception)\n', (43127, 43138), False, 'import pytest\n'), ((44781, 44805), 'pytest.raises', 'pytest.raises', (['Exception'], {}), '(Exception)\n', (44794, 44805), False, 'import pytest\n'), ((46589, 46613), 'pytest.raises', 'pytest.raises', (['Exception'], {}), '(Exception)\n', (46602, 46613), False, 'import pytest\n'), ((21431, 21478), 'os.getenv', 'os.getenv', (['"""COHERE_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('COHERE_API_KEY', '<not-a-real-key>')\n", (21440, 21478), False, 'import os\n'), ((8913, 8999), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""When do you use \'whom\' instead of \'who\'?"""'}), '(content=\n "When do you use \'whom\' instead of \'who\'?")\n', (8942, 8999), False, 'import langchain\n'), ((9330, 9396), 'langchain.schema.SystemMessage', 'langchain.schema.SystemMessage', ([], {'content': '"""Respond like a frat boy."""'}), "(content='Respond like a frat boy.')\n", (9360, 9396), False, 'import langchain\n'), ((9418, 9524), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""Where\'s the nearest equinox gym from Hudson Yards manhattan?"""'}), '(content=\n "Where\'s the nearest equinox gym from Hudson Yards manhattan?")\n', (9447, 9524), False, 'import langchain\n'), ((9624, 9695), 'langchain.schema.SystemMessage', 'langchain.schema.SystemMessage', ([], {'content': '"""Respond with a pirate accent."""'}), "(content='Respond with a pirate accent.')\n", (9654, 9695), False, 'import langchain\n'), ((9717, 9811), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""How does one get to Bikini Bottom from New York?"""'}), "(content=\n 'How does one get to Bikini Bottom from New York?')\n", (9746, 9811), False, 'import langchain\n'), ((10594, 10735), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': "[{'type': 'text', 'text': 'What’s in this image?'}, {'type': 'image_url',\n 'image_url': image_url}]"}), "(content=[{'type': 'text', 'text':\n 'What’s in this image?'}, {'type': 'image_url', 'image_url': image_url}])\n", (10623, 10735), False, 'import langchain\n'), ((11312, 11398), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""When do you use \'whom\' instead of \'who\'?"""'}), '(content=\n "When do you use \'whom\' instead of \'who\'?")\n', (11341, 11398), False, 'import langchain\n'), ((12671, 12756), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""What is the secret Krabby Patty recipe?"""'}), "(content='What is the secret Krabby Patty recipe?'\n )\n", (12700, 12756), False, 'import langchain\n'), ((13544, 13630), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""When do you use \'whom\' instead of \'who\'?"""'}), '(content=\n "When do you use \'whom\' instead of \'who\'?")\n', (13573, 13630), False, 'import langchain\n'), ((14906, 14992), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""When do you use \'whom\' instead of \'who\'?"""'}), '(content=\n "When do you use \'whom\' instead of \'who\'?")\n', (14935, 14992), False, 'import langchain\n'), ((33802, 33851), 'os.getenv', 'os.getenv', (['"""PINECONE_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('PINECONE_API_KEY', '<not-a-real-key>')\n", (33811, 33851), False, 'import os\n'), ((33881, 33926), 'os.getenv', 'os.getenv', (['"""PINECONE_ENV"""', '"""<not-a-real-env>"""'], {}), "('PINECONE_ENV', '<not-a-real-env>')\n", (33890, 33926), False, 'import os\n'), ((34876, 34925), 'os.getenv', 'os.getenv', (['"""PINECONE_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('PINECONE_API_KEY', '<not-a-real-key>')\n", (34885, 34925), False, 'import os\n'), ((34955, 35000), 'os.getenv', 'os.getenv', (['"""PINECONE_ENV"""', '"""<not-a-real-env>"""'], {}), "('PINECONE_ENV', '<not-a-real-env>')\n", (34964, 35000), False, 'import os\n'), ((35985, 36034), 'os.getenv', 'os.getenv', (['"""PINECONE_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('PINECONE_API_KEY', '<not-a-real-key>')\n", (35994, 36034), False, 'import os\n'), ((36064, 36109), 'os.getenv', 'os.getenv', (['"""PINECONE_ENV"""', '"""<not-a-real-env>"""'], {}), "('PINECONE_ENV', '<not-a-real-env>')\n", (36073, 36109), False, 'import os\n'), ((37446, 37495), 'os.getenv', 'os.getenv', (['"""PINECONE_API_KEY"""', '"""<not-a-real-key>"""'], {}), "('PINECONE_API_KEY', '<not-a-real-key>')\n", (37455, 37495), False, 'import os\n'), ((37525, 37570), 'os.getenv', 'os.getenv', (['"""PINECONE_ENV"""', '"""<not-a-real-env>"""'], {}), "('PINECONE_ENV', '<not-a-real-env>')\n", (37534, 37570), False, 'import os\n'), ((40184, 40209), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (40199, 40209), False, 'import os\n'), ((41492, 41517), 'os.path.dirname', 'os.path.dirname', (['__file__'], {}), '(__file__)\n', (41507, 41517), False, 'import os\n'), ((11765, 11831), 'langchain.schema.SystemMessage', 'langchain.schema.SystemMessage', ([], {'content': '"""Respond like a frat boy."""'}), "(content='Respond like a frat boy.')\n", (11795, 11831), False, 'import langchain\n'), ((11853, 11959), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""Where\'s the nearest equinox gym from Hudson Yards manhattan?"""'}), '(content=\n "Where\'s the nearest equinox gym from Hudson Yards manhattan?")\n', (11882, 11959), False, 'import langchain\n'), ((12059, 12130), 'langchain.schema.SystemMessage', 'langchain.schema.SystemMessage', ([], {'content': '"""Respond with a pirate accent."""'}), "(content='Respond with a pirate accent.')\n", (12089, 12130), False, 'import langchain\n'), ((12152, 12246), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""How does one get to Bikini Bottom from New York?"""'}), "(content=\n 'How does one get to Bikini Bottom from New York?')\n", (12181, 12246), False, 'import langchain\n'), ((13172, 13257), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""What is the secret Krabby Patty recipe?"""'}), "(content='What is the secret Krabby Patty recipe?'\n )\n", (13201, 13257), False, 'import langchain\n'), ((44928, 45004), 'langchain.schema.HumanMessage', 'langchain.schema.HumanMessage', ([], {'content': '"""Can you please not return an error?"""'}), "(content='Can you please not return an error?')\n", (44957, 45004), False, 'import langchain\n')]
import os from typing import Union from langchain.memory import ConversationBufferMemory from langchain.chat_models import AzureChatOpenAI from langchain.agents import AgentType, initialize_agent, tool import langchain from langchain.prompts.chat import MessagesPlaceholder, SystemMessagePromptTemplate import json import requests import datetime from langchain.tools import tool from pydantic.v1 import BaseModel, Field verbose = True langchain.debug = verbose # システムプロンプトの設定 PARTS_ORDER_SYSTEM_PROMPT = '''あなたはプラモデルの部品の個別注文を受け付ける担当者です。 部品を注文するには以下の注文情報が全て必要です。 注文情報 -------- * 注文される方のお名前 : * 注文される方のお名前(カタカナ) : * 部品送付先の郵便番号 : * 部品送付先の住所 : * 注文される方のお電話番号 : * 注文される方のメールアドレス : * 部品注文の対象となる商品の名称 : * 部品注文の対象となる商品の番号 : * 注文する部品とその個数 : 部品送付先の住所に関する注意 ---------------------------- また、住所に関しては番地まで含めた正確な情報が必要です。 「東京都」や「大阪府豊中市」などあいまいな場合は番地まで必要であることを回答に含めて下さい。 あなたの取るべき行動 -------------------- * 注文情報に未知の項目がある場合は予測や仮定をせず、"***" に置き換えた上で、 把握している注文情報を parts_order 関数に設定し confirmed = false で実行して下さい。 * あなたの「最終確認です。以下の内容で部品を注文しますが、よろしいですか?」の問いかけに対して、 ユーザーから肯定的な返答が確認できた場合のみ parts_order 関数を confirmed = true で実行し部品の注文を行って下さい。 * ユーザーから部品の注文の手続きをやめる、キャンセルする意思を伝えられた場合のみ、 parts_order 関数を canceled = true で実行し、あなたはそれまでの部品の注文に関する内容を全て忘れます。 parts_order 関数を実行する際の注文する部品とその個数の扱い ---------------------------------------------------------- また、parts_order 関数を実行する際、注文する部品とその個数は part_no_and_quantities に設定して下さい。 part_no_and_quantities は注文する部品とその個数の表現する dict の list です。 list の要素となる各 dict は key として 'part_no' と 'quantity' を持ちます。 'part_no' の value が部品名称の文字列、'quantity'の value が個数を意味する数字の文字列です。 以下は部品'J-26'を2個と部品'デカールC'を1枚注文する場合の part_no_and_quantities です。 ``` [{"part_no": "J-26", "quantity": "2"}, {"part_no": "デカールC", "quantity": "1"}] ``` ''' # エージェントの初期化 class PartsOrderInput(BaseModel): name: str = Field(description="注文される方のお名前です。") kana: str = Field(description="注文される方のお名前(カタカナ)です。") post_code: str = Field(description="部品送付先の郵便番号です。") address: str = Field(description="部品送付先の住所です。") tel: str = Field(description="注文される方のお電話番号です。") email: str = Field(description="注文される方のメールアドレスです。") product_name: str = Field( description="部品注文の対象となる商品の名称です。例:'PG 1/24 ダンバイン'") product_no: str = Field( description="部品注文の対象となる商品の箱や説明書に記載されている6桁の数字の文字列です。") part_no_and_quantities: list[dict[str, str]] = Field(description=( '注文する部品とその個数の表現する dict の list です。\n' 'dict は key "part_no" の value が部品名称の文字列、key "quantity" の value が個数を意味する整数です。\n' '例: 部品"J-26"を2個と部品"デカールC"を1枚注文する場合は、\n' '\n' '[{"part_no": "J-26", "quantity": 2}, {"part_no": "デカールC", "quantity": 1}]\n' '\n' 'としてください。')) confirmed: bool = Field(description=( "注文内容の最終確認状況です。最終確認が出来ている場合は True, そうでなければ False としてください。\n" "* confirmed が True の場合は部品の注文が行われます。 \n" "* confirmed が False の場合は注文内容の確認が行われます。") ) canceled: bool = Field(description=( "注文の手続きを継続する意思を示します。\n" "通常は False としますがユーザーに注文の手続きを継続しない意図がある場合は True としてください。\n" "* canceled が False の場合は部品の注文手続きを継続します。 \n" "* canceled が True の場合は注文手続きをキャンセルします。") ) @tool("parts_order", return_direct=True, args_schema=PartsOrderInput) def parts_order( name: str, kana: str, post_code: str, address: str, tel: str, email: str, product_name: str, product_no: str, part_no_and_quantities: list[dict[str, str]], confirmed: bool, canceled: bool, ) -> str: """プラモデルの部品を紛失、破損した場合に必要な部品を個別注文します。注文の内容確認にも使用します""" if canceled: return "わかりました。また部品の注文が必要になったらご相談ください。" def check_params(name, kana, post_code, address, tel, email, product_name, product_no, part_no_and_quantities): for arg in [name, kana, post_code, address, tel, email, product_name, product_no]: if arg is None or arg == "***" or arg == "": return False if not part_no_and_quantities: return False # for p_and_q in part_no_and_quantities: # part_no = p_and_q.get('part_no', '***') # quantity = p_and_q.get('quantity', '***') # if 'part_no' not in p_and_q or 'quantity' not in p_and_q: # return False # if part_no == '***' or quantity == '***': # return False for p_and_q in part_no_and_quantities: if "part_no" not in p_and_q or "quantity" not in p_and_q: return False if p_and_q["part_no"] == "***" or p_and_q["quantity"] == "***": return False return True has_required = check_params(name, kana, post_code, address, tel, email, product_name, product_no, part_no_and_quantities) # if part_no_and_quantities: # part_no_and_quantities = "\n ".join( # [f"{row.get('part_no','***')} x {row.get('quantity','***')}" # for row in part_no_and_quantities] # ) # else: # part_no_and_quantities = "***" if part_no_and_quantities: parts_list = [] for row in part_no_and_quantities: part_no = row.get('part_no', '***') quantity = row.get('quantity', '***') if part_no != '***' and quantity != '***': parts_list.append(f"{part_no} x {quantity}") else: parts_list.append("***") part_no_and_quantities = "\n ".join(parts_list) else: part_no_and_quantities = "***" # 注文情報のテンプレート order_template = ( f'・お名前: {name}\n' f'・お名前(カナ): {kana}\n' f'・郵便番号: {post_code}\n' f'・住所: {address}\n' f'・電話番号: {tel}\n' f'・メールアドレス: {email}\n' f'・商品名: {product_name}\n' f'・商品番号: {product_no}\n' f'・ご注文の部品\n' f' {part_no_and_quantities}' ) # 追加情報要求のテンプレート request_information_template = ( f'ご注文には以下の情報が必要です。"***" の項目を教えてください。\n' f'\n' f'{order_template}' ) # 注文確認のテンプレート confirm_template = ( f'最終確認です。以下の内容で部品を注文しますが、よろしいですか?\n' f'\n{order_template}' ) # 注文完了のテンプレート complete_template = ( f'以下の内容で部品を注文しました。\n' f'\n{order_template}\n' f'\n2営業日以内にご指定のメールアドレスに注文確認メールが届かない場合は、\n' f'弊社HPからお問い合わせください。' ) if has_required and confirmed: # TODO invoke order here! return complete_template else: if has_required: return confirm_template else: return request_information_template parts_order_tools = [parts_order] memory = ConversationBufferMemory( memory_key="chat_history", return_messages=True) chat_history = MessagesPlaceholder(variable_name='chat_history') # モデルの初期化 llm = AzureChatOpenAI( # Azure OpenAIのAPIを読み込み。 openai_api_base=os.environ["OPENAI_API_BASE"], openai_api_version=os.environ["OPENAI_API_VERSION"], deployment_name=os.environ["DEPLOYMENT_GPT35_NAME"], openai_api_key=os.environ["OPENAI_API_KEY"], openai_api_type="azure", model_kwargs={"top_p": 0.1, "function_call": {"name": "parts_order"}} ) agent_kwargs = { "system_message": SystemMessagePromptTemplate.from_template(template=PARTS_ORDER_SYSTEM_PROMPT), "extra_prompt_messages": [chat_history] } parts_order_agent = initialize_agent( parts_order_tools, llm, agent=AgentType.OPENAI_FUNCTIONS, verbose=verbose, agent_kwargs=agent_kwargs, memory=memory ) PARTS_ORDER_SSUFFIX_PROMPT = ''' 重要な注意事項 -------------- 注文情報に未知の項目がある場合は予測や仮定をせず "***" に置き換えてください。 parts_order 関数はユーザーから部品の注文の手続きをやめる、キャンセルする意思を伝えられた場合のみ canceled = true で実行して、 それまでの部品の注文に関する内容を全て忘れてください。。 parts_order 関数は次に示す例外を除いて confirmed = false で実行してください。 あなたの「最終確認です。以下の内容で部品を注文しますが、よろしいですか?」の問いかけに対して、 ユーザーから肯定的な返答が確認できた場合のみ parts_order 関数を confirmed = true で実行して部品を注文してください。 最終確認に対するユーザーの肯定的な返答なしで parts-order 関数を confirmed = true で実行することは誤発注であり事故になるので、固く禁止します。 ''' messages = [] messages.extend(parts_order_agent.agent.prompt.messages[:3]) messages.append(SystemMessagePromptTemplate.from_template( template=PARTS_ORDER_SSUFFIX_PROMPT),) messages.append(parts_order_agent.agent.prompt.messages[3]) parts_order_agent.agent.prompt.messages = messages # response = parts_order_agent.run("PG 1/24 ダンバインの部品を紛失したので注文したいです。") # print(response) while True: message = input(">> ") if message == "exit": break response = parts_order_agent.run(message) print(response)
[ "langchain.agents.initialize_agent", "langchain.prompts.chat.SystemMessagePromptTemplate.from_template", "langchain.memory.ConversationBufferMemory", "langchain.tools.tool", "langchain.chat_models.AzureChatOpenAI", "langchain.prompts.chat.MessagesPlaceholder" ]
[((3143, 3211), 'langchain.tools.tool', 'tool', (['"""parts_order"""'], {'return_direct': '(True)', 'args_schema': 'PartsOrderInput'}), "('parts_order', return_direct=True, args_schema=PartsOrderInput)\n", (3147, 3211), False, 'from langchain.tools import tool\n'), ((6557, 6630), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'return_messages': '(True)'}), "(memory_key='chat_history', return_messages=True)\n", (6581, 6630), False, 'from langchain.memory import ConversationBufferMemory\n'), ((6651, 6700), 'langchain.prompts.chat.MessagesPlaceholder', 'MessagesPlaceholder', ([], {'variable_name': '"""chat_history"""'}), "(variable_name='chat_history')\n", (6670, 6700), False, 'from langchain.prompts.chat import MessagesPlaceholder, SystemMessagePromptTemplate\n'), ((6718, 7045), 'langchain.chat_models.AzureChatOpenAI', 'AzureChatOpenAI', ([], {'openai_api_base': "os.environ['OPENAI_API_BASE']", 'openai_api_version': "os.environ['OPENAI_API_VERSION']", 'deployment_name': "os.environ['DEPLOYMENT_GPT35_NAME']", 'openai_api_key': "os.environ['OPENAI_API_KEY']", 'openai_api_type': '"""azure"""', 'model_kwargs': "{'top_p': 0.1, 'function_call': {'name': 'parts_order'}}"}), "(openai_api_base=os.environ['OPENAI_API_BASE'],\n openai_api_version=os.environ['OPENAI_API_VERSION'], deployment_name=os\n .environ['DEPLOYMENT_GPT35_NAME'], openai_api_key=os.environ[\n 'OPENAI_API_KEY'], openai_api_type='azure', model_kwargs={'top_p': 0.1,\n 'function_call': {'name': 'parts_order'}})\n", (6733, 7045), False, 'from langchain.chat_models import AzureChatOpenAI\n'), ((7265, 7402), 'langchain.agents.initialize_agent', 'initialize_agent', (['parts_order_tools', 'llm'], {'agent': 'AgentType.OPENAI_FUNCTIONS', 'verbose': 'verbose', 'agent_kwargs': 'agent_kwargs', 'memory': 'memory'}), '(parts_order_tools, llm, agent=AgentType.OPENAI_FUNCTIONS,\n verbose=verbose, agent_kwargs=agent_kwargs, memory=memory)\n', (7281, 7402), False, 'from langchain.agents import AgentType, initialize_agent, tool\n'), ((1803, 1837), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文される方のお名前です。"""'}), "(description='注文される方のお名前です。')\n", (1808, 1837), False, 'from pydantic.v1 import BaseModel, Field\n'), ((1854, 1894), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文される方のお名前(カタカナ)です。"""'}), "(description='注文される方のお名前(カタカナ)です。')\n", (1859, 1894), False, 'from pydantic.v1 import BaseModel, Field\n'), ((1916, 1950), 'pydantic.v1.Field', 'Field', ([], {'description': '"""部品送付先の郵便番号です。"""'}), "(description='部品送付先の郵便番号です。')\n", (1921, 1950), False, 'from pydantic.v1 import BaseModel, Field\n'), ((1970, 2002), 'pydantic.v1.Field', 'Field', ([], {'description': '"""部品送付先の住所です。"""'}), "(description='部品送付先の住所です。')\n", (1975, 2002), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2018, 2054), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文される方のお電話番号です。"""'}), "(description='注文される方のお電話番号です。')\n", (2023, 2054), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2072, 2110), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文される方のメールアドレスです。"""'}), "(description='注文される方のメールアドレスです。')\n", (2077, 2110), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2135, 2191), 'pydantic.v1.Field', 'Field', ([], {'description': '"""部品注文の対象となる商品の名称です。例:\'PG 1/24 ダンバイン\'"""'}), '(description="部品注文の対象となる商品の名称です。例:\'PG 1/24 ダンバイン\'")\n', (2140, 2191), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2223, 2282), 'pydantic.v1.Field', 'Field', ([], {'description': '"""部品注文の対象となる商品の箱や説明書に記載されている6桁の数字の文字列です。"""'}), "(description='部品注文の対象となる商品の箱や説明書に記載されている6桁の数字の文字列です。')\n", (2228, 2282), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2343, 2607), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文する部品とその個数の表現する dict の list です。\ndict は key "part_no" の value が部品名称の文字列、key "quantity" の value が個数を意味する整数です。\n例: 部品"J-26"を2個と部品"デカールC"を1枚注文する場合は、\n\n[{"part_no": "J-26", "quantity": 2}, {"part_no": "デカールC", "quantity": 1}]\n\nとしてください。"""'}), '(description=\n """注文する部品とその個数の表現する dict の list です。\ndict は key "part_no" の value が部品名称の文字列、key "quantity" の value が個数を意味する整数です。\n例: 部品"J-26"を2個と部品"デカールC"を1枚注文する場合は、\n\n[{"part_no": "J-26", "quantity": 2}, {"part_no": "デカールC", "quantity": 1}]\n\nとしてください。"""\n )\n', (2348, 2607), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2699, 2866), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文内容の最終確認状況です。最終確認が出来ている場合は True, そうでなければ False としてください。\n* confirmed が True の場合は部品の注文が行われます。 \n* confirmed が False の場合は注文内容の確認が行われます。"""'}), '(description=\n """注文内容の最終確認状況です。最終確認が出来ている場合は True, そうでなければ False としてください。\n* confirmed が True の場合は部品の注文が行われます。 \n* confirmed が False の場合は注文内容の確認が行われます。"""\n )\n', (2704, 2866), False, 'from pydantic.v1 import BaseModel, Field\n'), ((2914, 3101), 'pydantic.v1.Field', 'Field', ([], {'description': '"""注文の手続きを継続する意思を示します。\n通常は False としますがユーザーに注文の手続きを継続しない意図がある場合は True としてください。\n* canceled が False の場合は部品の注文手続きを継続します。 \n* canceled が True の場合は注文手続きをキャンセルします。"""'}), '(description=\n """注文の手続きを継続する意思を示します。\n通常は False としますがユーザーに注文の手続きを継続しない意図がある場合は True としてください。\n* canceled が False の場合は部品の注文手続きを継続します。 \n* canceled が True の場合は注文手続きをキャンセルします。"""\n )\n', (2919, 3101), False, 'from pydantic.v1 import BaseModel, Field\n'), ((7120, 7197), 'langchain.prompts.chat.SystemMessagePromptTemplate.from_template', 'SystemMessagePromptTemplate.from_template', ([], {'template': 'PARTS_ORDER_SYSTEM_PROMPT'}), '(template=PARTS_ORDER_SYSTEM_PROMPT)\n', (7161, 7197), False, 'from langchain.prompts.chat import MessagesPlaceholder, SystemMessagePromptTemplate\n'), ((7998, 8076), 'langchain.prompts.chat.SystemMessagePromptTemplate.from_template', 'SystemMessagePromptTemplate.from_template', ([], {'template': 'PARTS_ORDER_SSUFFIX_PROMPT'}), '(template=PARTS_ORDER_SSUFFIX_PROMPT)\n', (8039, 8076), False, 'from langchain.prompts.chat import MessagesPlaceholder, SystemMessagePromptTemplate\n')]
"""Interface with the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain_core.load.dump import dumps from langchain_core.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_key: Optional[str] = None) -> Client: try: from langchainhub import Client except ImportError as e: raise ImportError( "Could not import langchainhub, please install with `pip install " "langchainhub`." ) from e # Client logic will also attempt to load URL/key from environment variables return Client(api_url, api_key=api_key) def push( repo_full_name: str, object: Any, *, api_url: Optional[str] = None, api_key: Optional[str] = None, parent_commit_hash: Optional[str] = "latest", new_repo_is_public: bool = True, new_repo_description: str = "", ) -> str: """ Pushes an object to the hub and returns the URL it can be viewed at in a browser. :param repo_full_name: The full name of the repo to push to in the format of `owner/repo`. :param object: The LangChain to serialize and push to the hub. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. :param parent_commit_hash: The commit hash of the parent commit to push to. Defaults to the latest commit automatically. :param new_repo_is_public: Whether the repo should be public. Defaults to True (Public by default). :param new_repo_description: The description of the repo. Defaults to an empty string. """ client = _get_client(api_url=api_url, api_key=api_key) manifest_json = dumps(object) message = client.push( repo_full_name, manifest_json, parent_commit_hash=parent_commit_hash, new_repo_is_public=new_repo_is_public, new_repo_description=new_repo_description, ) return message def pull( owner_repo_commit: str, *, api_url: Optional[str] = None, api_key: Optional[str] = None, ) -> Any: """ Pulls an object from the hub and returns it as a LangChain object. :param owner_repo_commit: The full name of the repo to pull from in the format of `owner/repo:commit_hash`. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. """ client = _get_client(api_url=api_url, api_key=api_key) resp: str = client.pull(owner_repo_commit) return loads(resp)
[ "langchain_core.load.load.loads", "langchain_core.load.dump.dumps", "langchainhub.Client" ]
[((679, 711), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (685, 711), False, 'from langchainhub import Client\n'), ((1912, 1925), 'langchain_core.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1917, 1925), False, 'from langchain_core.load.dump import dumps\n'), ((2862, 2873), 'langchain_core.load.load.loads', 'loads', (['resp'], {}), '(resp)\n', (2867, 2873), False, 'from langchain_core.load.load import loads\n')]
"""Interface with the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain_core.load.dump import dumps from langchain_core.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_key: Optional[str] = None) -> Client: try: from langchainhub import Client except ImportError as e: raise ImportError( "Could not import langchainhub, please install with `pip install " "langchainhub`." ) from e # Client logic will also attempt to load URL/key from environment variables return Client(api_url, api_key=api_key) def push( repo_full_name: str, object: Any, *, api_url: Optional[str] = None, api_key: Optional[str] = None, parent_commit_hash: Optional[str] = "latest", new_repo_is_public: bool = True, new_repo_description: str = "", ) -> str: """ Pushes an object to the hub and returns the URL it can be viewed at in a browser. :param repo_full_name: The full name of the repo to push to in the format of `owner/repo`. :param object: The LangChain to serialize and push to the hub. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. :param parent_commit_hash: The commit hash of the parent commit to push to. Defaults to the latest commit automatically. :param new_repo_is_public: Whether the repo should be public. Defaults to True (Public by default). :param new_repo_description: The description of the repo. Defaults to an empty string. """ client = _get_client(api_url=api_url, api_key=api_key) manifest_json = dumps(object) message = client.push( repo_full_name, manifest_json, parent_commit_hash=parent_commit_hash, new_repo_is_public=new_repo_is_public, new_repo_description=new_repo_description, ) return message def pull( owner_repo_commit: str, *, api_url: Optional[str] = None, api_key: Optional[str] = None, ) -> Any: """ Pulls an object from the hub and returns it as a LangChain object. :param owner_repo_commit: The full name of the repo to pull from in the format of `owner/repo:commit_hash`. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. """ client = _get_client(api_url=api_url, api_key=api_key) resp: str = client.pull(owner_repo_commit) return loads(resp)
[ "langchain_core.load.load.loads", "langchain_core.load.dump.dumps", "langchainhub.Client" ]
[((679, 711), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (685, 711), False, 'from langchainhub import Client\n'), ((1912, 1925), 'langchain_core.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1917, 1925), False, 'from langchain_core.load.dump import dumps\n'), ((2862, 2873), 'langchain_core.load.load.loads', 'loads', (['resp'], {}), '(resp)\n', (2867, 2873), False, 'from langchain_core.load.load import loads\n')]
"""Interface with the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain_core.load.dump import dumps from langchain_core.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_key: Optional[str] = None) -> Client: try: from langchainhub import Client except ImportError as e: raise ImportError( "Could not import langchainhub, please install with `pip install " "langchainhub`." ) from e # Client logic will also attempt to load URL/key from environment variables return Client(api_url, api_key=api_key) def push( repo_full_name: str, object: Any, *, api_url: Optional[str] = None, api_key: Optional[str] = None, parent_commit_hash: Optional[str] = "latest", new_repo_is_public: bool = True, new_repo_description: str = "", ) -> str: """ Pushes an object to the hub and returns the URL it can be viewed at in a browser. :param repo_full_name: The full name of the repo to push to in the format of `owner/repo`. :param object: The LangChain to serialize and push to the hub. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. :param parent_commit_hash: The commit hash of the parent commit to push to. Defaults to the latest commit automatically. :param new_repo_is_public: Whether the repo should be public. Defaults to True (Public by default). :param new_repo_description: The description of the repo. Defaults to an empty string. """ client = _get_client(api_url=api_url, api_key=api_key) manifest_json = dumps(object) message = client.push( repo_full_name, manifest_json, parent_commit_hash=parent_commit_hash, new_repo_is_public=new_repo_is_public, new_repo_description=new_repo_description, ) return message def pull( owner_repo_commit: str, *, api_url: Optional[str] = None, api_key: Optional[str] = None, ) -> Any: """ Pulls an object from the hub and returns it as a LangChain object. :param owner_repo_commit: The full name of the repo to pull from in the format of `owner/repo:commit_hash`. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. """ client = _get_client(api_url=api_url, api_key=api_key) resp: str = client.pull(owner_repo_commit) return loads(resp)
[ "langchain_core.load.load.loads", "langchain_core.load.dump.dumps", "langchainhub.Client" ]
[((679, 711), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (685, 711), False, 'from langchainhub import Client\n'), ((1912, 1925), 'langchain_core.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1917, 1925), False, 'from langchain_core.load.dump import dumps\n'), ((2862, 2873), 'langchain_core.load.load.loads', 'loads', (['resp'], {}), '(resp)\n', (2867, 2873), False, 'from langchain_core.load.load import loads\n')]
"""Interface with the LangChain Hub.""" from __future__ import annotations from typing import TYPE_CHECKING, Any, Optional from langchain_core.load.dump import dumps from langchain_core.load.load import loads if TYPE_CHECKING: from langchainhub import Client def _get_client(api_url: Optional[str] = None, api_key: Optional[str] = None) -> Client: try: from langchainhub import Client except ImportError as e: raise ImportError( "Could not import langchainhub, please install with `pip install " "langchainhub`." ) from e # Client logic will also attempt to load URL/key from environment variables return Client(api_url, api_key=api_key) def push( repo_full_name: str, object: Any, *, api_url: Optional[str] = None, api_key: Optional[str] = None, parent_commit_hash: Optional[str] = "latest", new_repo_is_public: bool = True, new_repo_description: str = "", ) -> str: """ Pushes an object to the hub and returns the URL it can be viewed at in a browser. :param repo_full_name: The full name of the repo to push to in the format of `owner/repo`. :param object: The LangChain to serialize and push to the hub. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. :param parent_commit_hash: The commit hash of the parent commit to push to. Defaults to the latest commit automatically. :param new_repo_is_public: Whether the repo should be public. Defaults to True (Public by default). :param new_repo_description: The description of the repo. Defaults to an empty string. """ client = _get_client(api_url=api_url, api_key=api_key) manifest_json = dumps(object) message = client.push( repo_full_name, manifest_json, parent_commit_hash=parent_commit_hash, new_repo_is_public=new_repo_is_public, new_repo_description=new_repo_description, ) return message def pull( owner_repo_commit: str, *, api_url: Optional[str] = None, api_key: Optional[str] = None, ) -> Any: """ Pulls an object from the hub and returns it as a LangChain object. :param owner_repo_commit: The full name of the repo to pull from in the format of `owner/repo:commit_hash`. :param api_url: The URL of the LangChain Hub API. Defaults to the hosted API service if you have an api key set, or a localhost instance if not. :param api_key: The API key to use to authenticate with the LangChain Hub API. """ client = _get_client(api_url=api_url, api_key=api_key) resp: str = client.pull(owner_repo_commit) return loads(resp)
[ "langchain_core.load.load.loads", "langchain_core.load.dump.dumps", "langchainhub.Client" ]
[((679, 711), 'langchainhub.Client', 'Client', (['api_url'], {'api_key': 'api_key'}), '(api_url, api_key=api_key)\n', (685, 711), False, 'from langchainhub import Client\n'), ((1912, 1925), 'langchain_core.load.dump.dumps', 'dumps', (['object'], {}), '(object)\n', (1917, 1925), False, 'from langchain_core.load.dump import dumps\n'), ((2862, 2873), 'langchain_core.load.load.loads', 'loads', (['resp'], {}), '(resp)\n', (2867, 2873), False, 'from langchain_core.load.load import loads\n')]
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from datetime import timedelta from functools import lru_cache from typing import ( TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.llms.base import LLM, get_prompts from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.schema.cache import RETURN_VAL_TYPE, BaseCache from langchain.utils import get_from_env from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento from cassandra.cluster import Session as CassandraSession def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. Warning: would not work well with arbitrary subclasses of `Generation` """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. Warning: would not work well with arbitrary subclasses of `Generation` """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) def _dumps_generations(generations: RETURN_VAL_TYPE) -> str: """ Serialization for generic RETURN_VAL_TYPE, i.e. sequence of `Generation` Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: a single string representing a list of generations. This function (+ its counterpart `_loads_generations`) rely on the dumps/loads pair with Reviver, so are able to deal with all subclasses of Generation. Each item in the list can be `dumps`ed to a string, then we make the whole list of strings into a json-dumped. """ return json.dumps([dumps(_item) for _item in generations]) def _loads_generations(generations_str: str) -> Union[RETURN_VAL_TYPE, None]: """ Deserialization of a string into a generic RETURN_VAL_TYPE (i.e. a sequence of `Generation`). See `_dumps_generations`, the inverse of this function. Args: generations_str (str): A string representing a list of generations. Compatible with the legacy cache-blob format Does not raise exceptions for malformed entries, just logs a warning and returns none: the caller should be prepared for such a cache miss. Returns: RETURN_VAL_TYPE: A list of generations. """ try: generations = [loads(_item_str) for _item_str in json.loads(generations_str)] return generations except (json.JSONDecodeError, TypeError): # deferring the (soft) handling to after the legacy-format attempt pass try: gen_dicts = json.loads(generations_str) # not relying on `_load_generations_from_json` (which could disappear): generations = [Generation(**generation_dict) for generation_dict in gen_dicts] logger.warning( f"Legacy 'Generation' cached blob encountered: '{generations_str}'" ) return generations except (json.JSONDecodeError, TypeError): logger.warning( f"Malformed/unparsable cached blob encountered: '{generations_str}'" ) return None class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" def __init__(self, redis_: Any, *, ttl: Optional[int] = None): """ Initialize an instance of RedisCache. This method initializes an object with Redis caching capabilities. It takes a `redis_` parameter, which should be an instance of a Redis client class, allowing the object to interact with a Redis server for caching purposes. Parameters: redis_ (Any): An instance of a Redis client class (e.g., redis.Redis) used for caching. This allows the object to communicate with a Redis server for caching operations. ttl (int, optional): Time-to-live (TTL) for cached items in seconds. If provided, it sets the time duration for how long cached items will remain valid. If not provided, cached items will not have an automatic expiration. """ try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ self.ttl = ttl def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) with self.redis.pipeline() as pipe: pipe.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) if self.ttl is not None: pipe.expire(key, self.ttl) pipe.execute() def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis DEFAULT_SCHEMA = { "content_key": "prompt", "text": [ {"name": "prompt"}, ], "extra": [{"name": "return_val"}, {"name": "llm_string"}], } def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, schema=cast(Dict, self.DEFAULT_SCHEMA), ) except ValueError: redis = RedisVectorstore( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, index_schema=cast(Dict, self.DEFAULT_SCHEMA), ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations: List = [] # Read from a Hash results = llm_cache.similarity_search( query=prompt, k=1, distance_threshold=self.score_threshold, ) if results: for document in results: generations.extend( _load_generations_from_json(document.metadata["return_val"]) ) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) _dump_generations_to_json([g for g in return_val]) metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": _dump_generations_to_json([g for g in return_val]), } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception CASSANDRA_CACHE_DEFAULT_TABLE_NAME = "langchain_llm_cache" CASSANDRA_CACHE_DEFAULT_TTL_SECONDS = None class CassandraCache(BaseCache): """ Cache that uses Cassandra / Astra DB as a backend. It uses a single Cassandra table. The lookup keys (which get to form the primary key) are: - prompt, a string - llm_string, a deterministic str representation of the model parameters. (needed to prevent collisions same-prompt-different-model collisions) """ def __init__( self, session: Optional[CassandraSession] = None, keyspace: Optional[str] = None, table_name: str = CASSANDRA_CACHE_DEFAULT_TABLE_NAME, ttl_seconds: Optional[int] = CASSANDRA_CACHE_DEFAULT_TTL_SECONDS, skip_provisioning: bool = False, ): """ Initialize with a ready session and a keyspace name. Args: session (cassandra.cluster.Session): an open Cassandra session keyspace (str): the keyspace to use for storing the cache table_name (str): name of the Cassandra table to use as cache ttl_seconds (optional int): time-to-live for cache entries (default: None, i.e. forever) """ try: from cassio.table import ElasticCassandraTable except (ImportError, ModuleNotFoundError): raise ValueError( "Could not import cassio python package. " "Please install it with `pip install cassio`." ) self.session = session self.keyspace = keyspace self.table_name = table_name self.ttl_seconds = ttl_seconds self.kv_cache = ElasticCassandraTable( session=self.session, keyspace=self.keyspace, table=self.table_name, keys=["llm_string", "prompt"], primary_key_type=["TEXT", "TEXT"], ttl_seconds=self.ttl_seconds, skip_provisioning=skip_provisioning, ) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" item = self.kv_cache.get( llm_string=_hash(llm_string), prompt=_hash(prompt), ) if item is not None: generations = _loads_generations(item["body_blob"]) # this protects against malformed cached items: if generations is not None: return generations else: return None else: return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" blob = _dumps_generations(return_val) self.kv_cache.put( llm_string=_hash(llm_string), prompt=_hash(prompt), body_blob=blob, ) def delete_through_llm( self, prompt: str, llm: LLM, stop: Optional[List[str]] = None ) -> None: """ A wrapper around `delete` with the LLM being passed. In case the llm(prompt) calls have a `stop` param, you should pass it here """ llm_string = get_prompts( {**llm.dict(), **{"stop": stop}}, [], )[1] return self.delete(prompt, llm_string=llm_string) def delete(self, prompt: str, llm_string: str) -> None: """Evict from cache if there's an entry.""" return self.kv_cache.delete( llm_string=_hash(llm_string), prompt=_hash(prompt), ) def clear(self, **kwargs: Any) -> None: """Clear cache. This is for all LLMs at once.""" self.kv_cache.clear() CASSANDRA_SEMANTIC_CACHE_DEFAULT_DISTANCE_METRIC = "dot" CASSANDRA_SEMANTIC_CACHE_DEFAULT_SCORE_THRESHOLD = 0.85 CASSANDRA_SEMANTIC_CACHE_DEFAULT_TABLE_NAME = "langchain_llm_semantic_cache" CASSANDRA_SEMANTIC_CACHE_DEFAULT_TTL_SECONDS = None CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE = 16 class CassandraSemanticCache(BaseCache): """ Cache that uses Cassandra as a vector-store backend for semantic (i.e. similarity-based) lookup. It uses a single (vector) Cassandra table and stores, in principle, cached values from several LLMs, so the LLM's llm_string is part of the rows' primary keys. The similarity is based on one of several distance metrics (default: "dot"). If choosing another metric, the default threshold is to be re-tuned accordingly. """ def __init__( self, session: Optional[CassandraSession], keyspace: Optional[str], embedding: Embeddings, table_name: str = CASSANDRA_SEMANTIC_CACHE_DEFAULT_TABLE_NAME, distance_metric: str = CASSANDRA_SEMANTIC_CACHE_DEFAULT_DISTANCE_METRIC, score_threshold: float = CASSANDRA_SEMANTIC_CACHE_DEFAULT_SCORE_THRESHOLD, ttl_seconds: Optional[int] = CASSANDRA_SEMANTIC_CACHE_DEFAULT_TTL_SECONDS, skip_provisioning: bool = False, ): """ Initialize the cache with all relevant parameters. Args: session (cassandra.cluster.Session): an open Cassandra session keyspace (str): the keyspace to use for storing the cache embedding (Embedding): Embedding provider for semantic encoding and search. table_name (str): name of the Cassandra (vector) table to use as cache distance_metric (str, 'dot'): which measure to adopt for similarity searches score_threshold (optional float): numeric value to use as cutoff for the similarity searches ttl_seconds (optional int): time-to-live for cache entries (default: None, i.e. forever) The default score threshold is tuned to the default metric. Tune it carefully yourself if switching to another distance metric. """ try: from cassio.table import MetadataVectorCassandraTable except (ImportError, ModuleNotFoundError): raise ValueError( "Could not import cassio python package. " "Please install it with `pip install cassio`." ) self.session = session self.keyspace = keyspace self.embedding = embedding self.table_name = table_name self.distance_metric = distance_metric self.score_threshold = score_threshold self.ttl_seconds = ttl_seconds # The contract for this class has separate lookup and update: # in order to spare some embedding calculations we cache them between # the two calls. # Note: each instance of this class has its own `_get_embedding` with # its own lru. @lru_cache(maxsize=CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE) def _cache_embedding(text: str) -> List[float]: return self.embedding.embed_query(text=text) self._get_embedding = _cache_embedding self.embedding_dimension = self._get_embedding_dimension() self.table = MetadataVectorCassandraTable( session=self.session, keyspace=self.keyspace, table=self.table_name, primary_key_type=["TEXT"], vector_dimension=self.embedding_dimension, ttl_seconds=self.ttl_seconds, metadata_indexing=("allow", {"_llm_string_hash"}), skip_provisioning=skip_provisioning, ) def _get_embedding_dimension(self) -> int: return len(self._get_embedding(text="This is a sample sentence.")) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" embedding_vector = self._get_embedding(text=prompt) llm_string_hash = _hash(llm_string) body = _dumps_generations(return_val) metadata = { "_prompt": prompt, "_llm_string_hash": llm_string_hash, } row_id = f"{_hash(prompt)}-{llm_string_hash}" # self.table.put( body_blob=body, vector=embedding_vector, row_id=row_id, metadata=metadata, ) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" hit_with_id = self.lookup_with_id(prompt, llm_string) if hit_with_id is not None: return hit_with_id[1] else: return None def lookup_with_id( self, prompt: str, llm_string: str ) -> Optional[Tuple[str, RETURN_VAL_TYPE]]: """ Look up based on prompt and llm_string. If there are hits, return (document_id, cached_entry) """ prompt_embedding: List[float] = self._get_embedding(text=prompt) hits = list( self.table.metric_ann_search( vector=prompt_embedding, metadata={"_llm_string_hash": _hash(llm_string)}, n=1, metric=self.distance_metric, metric_threshold=self.score_threshold, ) ) if hits: hit = hits[0] generations = _loads_generations(hit["body_blob"]) if generations is not None: # this protects against malformed cached items: return ( hit["row_id"], generations, ) else: return None else: return None def lookup_with_id_through_llm( self, prompt: str, llm: LLM, stop: Optional[List[str]] = None ) -> Optional[Tuple[str, RETURN_VAL_TYPE]]: llm_string = get_prompts( {**llm.dict(), **{"stop": stop}}, [], )[1] return self.lookup_with_id(prompt, llm_string=llm_string) def delete_by_document_id(self, document_id: str) -> None: """ Given this is a "similarity search" cache, an invalidation pattern that makes sense is first a lookup to get an ID, and then deleting with that ID. This is for the second step. """ self.table.delete(row_id=document_id) def clear(self, **kwargs: Any) -> None: """Clear the *whole* semantic cache.""" self.table.clear()
[ "langchain.load.load.loads", "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps" ]
[((1586, 1613), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1603, 1613), False, 'import logging\n'), ((5793, 5811), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (5809, 5811), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((5968, 6000), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (5974, 6000), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((6011, 6043), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (6017, 6043), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((6054, 6087), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (6060, 6087), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((6103, 6117), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (6109, 6117), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2730, 2758), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2740, 2758), False, 'import json\n'), ((4550, 4577), 'json.loads', 'json.loads', (['generations_str'], {}), '(generations_str)\n', (4560, 4577), False, 'import json\n'), ((8521, 8564), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (8534, 8564), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((18425, 18432), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (18430, 18432), False, 'from gptcache import Cache\n'), ((19758, 19790), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (19761, 19790), False, 'from gptcache.adapter.api import get\n'), ((20677, 20723), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (20680, 20723), False, 'from gptcache.adapter.api import put\n'), ((24572, 24614), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (24602, 24614), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((24638, 24694), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (24649, 24694), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((29638, 29876), 'cassio.table.ElasticCassandraTable', 'ElasticCassandraTable', ([], {'session': 'self.session', 'keyspace': 'self.keyspace', 'table': 'self.table_name', 'keys': "['llm_string', 'prompt']", 'primary_key_type': "['TEXT', 'TEXT']", 'ttl_seconds': 'self.ttl_seconds', 'skip_provisioning': 'skip_provisioning'}), "(session=self.session, keyspace=self.keyspace, table=\n self.table_name, keys=['llm_string', 'prompt'], primary_key_type=[\n 'TEXT', 'TEXT'], ttl_seconds=self.ttl_seconds, skip_provisioning=\n skip_provisioning)\n", (29659, 29876), False, 'from cassio.table import ElasticCassandraTable\n'), ((34766, 34830), 'functools.lru_cache', 'lru_cache', ([], {'maxsize': 'CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE'}), '(maxsize=CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE)\n', (34775, 34830), False, 'from functools import lru_cache\n'), ((35081, 35380), 'cassio.table.MetadataVectorCassandraTable', 'MetadataVectorCassandraTable', ([], {'session': 'self.session', 'keyspace': 'self.keyspace', 'table': 'self.table_name', 'primary_key_type': "['TEXT']", 'vector_dimension': 'self.embedding_dimension', 'ttl_seconds': 'self.ttl_seconds', 'metadata_indexing': "('allow', {'_llm_string_hash'})", 'skip_provisioning': 'skip_provisioning'}), "(session=self.session, keyspace=self.keyspace,\n table=self.table_name, primary_key_type=['TEXT'], vector_dimension=self\n .embedding_dimension, ttl_seconds=self.ttl_seconds, metadata_indexing=(\n 'allow', {'_llm_string_hash'}), skip_provisioning=skip_provisioning)\n", (35109, 35380), False, 'from cassio.table import MetadataVectorCassandraTable\n'), ((2775, 2804), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2785, 2804), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((3618, 3630), 'langchain.load.dump.dumps', 'dumps', (['_item'], {}), '(_item)\n', (3623, 3630), False, 'from langchain.load.dump import dumps\n'), ((4296, 4312), 'langchain.load.load.loads', 'loads', (['_item_str'], {}), '(_item_str)\n', (4301, 4312), False, 'from langchain.load.load import loads\n'), ((4681, 4710), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (4691, 4710), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6867, 6887), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6874, 6887), False, 'from sqlalchemy.orm import Session\n'), ((7974, 7994), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (7981, 7994), False, 'from sqlalchemy.orm import Session\n'), ((8176, 8196), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (8183, 8196), False, 'from sqlalchemy.orm import Session\n'), ((18499, 18541), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (18516, 18541), False, 'import inspect\n'), ((20946, 20976), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (20950, 20976), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast\n'), ((21876, 21896), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (21885, 21896), False, 'from datetime import timedelta\n'), ((24439, 24465), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (24463, 24465), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((24501, 24549), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (24513, 24549), False, 'from langchain.utils import get_from_env\n'), ((4330, 4357), 'json.loads', 'json.loads', (['generations_str'], {}), '(generations_str)\n', (4340, 4357), False, 'import json\n'), ((11054, 11173), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (11067, 11173), False, 'import warnings\n'), ((15954, 16082), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (15967, 16082), False, 'import warnings\n'), ((19844, 19873), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (19854, 19873), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((7884, 7894), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (7889, 7894), False, 'from langchain.load.dump import dumps\n'), ((10513, 10534), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (10523, 10534), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((13908, 13939), 'typing.cast', 'cast', (['Dict', 'self.DEFAULT_SCHEMA'], {}), '(Dict, self.DEFAULT_SCHEMA)\n', (13912, 13939), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast\n'), ((18885, 18923), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (18901, 18923), False, 'from gptcache.manager.factory import get_data_manager\n'), ((19897, 19912), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (19907, 19912), False, 'import json\n'), ((7022, 7035), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (7027, 7035), False, 'from langchain.load.load import loads\n'), ((14172, 14203), 'typing.cast', 'cast', (['Dict', 'self.DEFAULT_SCHEMA'], {}), '(Dict, self.DEFAULT_SCHEMA)\n', (14176, 14203), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast\n'), ((7592, 7615), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (7602, 7615), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6637, 6671), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (6643, 6671), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from datetime import timedelta from functools import lru_cache from typing import ( TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.llms.base import LLM, get_prompts from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.schema.cache import RETURN_VAL_TYPE, BaseCache from langchain.utils import get_from_env from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento from cassandra.cluster import Session as CassandraSession def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. Warning: would not work well with arbitrary subclasses of `Generation` """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. Warning: would not work well with arbitrary subclasses of `Generation` """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) def _dumps_generations(generations: RETURN_VAL_TYPE) -> str: """ Serialization for generic RETURN_VAL_TYPE, i.e. sequence of `Generation` Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: a single string representing a list of generations. This function (+ its counterpart `_loads_generations`) rely on the dumps/loads pair with Reviver, so are able to deal with all subclasses of Generation. Each item in the list can be `dumps`ed to a string, then we make the whole list of strings into a json-dumped. """ return json.dumps([dumps(_item) for _item in generations]) def _loads_generations(generations_str: str) -> Union[RETURN_VAL_TYPE, None]: """ Deserialization of a string into a generic RETURN_VAL_TYPE (i.e. a sequence of `Generation`). See `_dumps_generations`, the inverse of this function. Args: generations_str (str): A string representing a list of generations. Compatible with the legacy cache-blob format Does not raise exceptions for malformed entries, just logs a warning and returns none: the caller should be prepared for such a cache miss. Returns: RETURN_VAL_TYPE: A list of generations. """ try: generations = [loads(_item_str) for _item_str in json.loads(generations_str)] return generations except (json.JSONDecodeError, TypeError): # deferring the (soft) handling to after the legacy-format attempt pass try: gen_dicts = json.loads(generations_str) # not relying on `_load_generations_from_json` (which could disappear): generations = [Generation(**generation_dict) for generation_dict in gen_dicts] logger.warning( f"Legacy 'Generation' cached blob encountered: '{generations_str}'" ) return generations except (json.JSONDecodeError, TypeError): logger.warning( f"Malformed/unparsable cached blob encountered: '{generations_str}'" ) return None class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" def __init__(self, redis_: Any, *, ttl: Optional[int] = None): """ Initialize an instance of RedisCache. This method initializes an object with Redis caching capabilities. It takes a `redis_` parameter, which should be an instance of a Redis client class, allowing the object to interact with a Redis server for caching purposes. Parameters: redis_ (Any): An instance of a Redis client class (e.g., redis.Redis) used for caching. This allows the object to communicate with a Redis server for caching operations. ttl (int, optional): Time-to-live (TTL) for cached items in seconds. If provided, it sets the time duration for how long cached items will remain valid. If not provided, cached items will not have an automatic expiration. """ try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ self.ttl = ttl def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) with self.redis.pipeline() as pipe: pipe.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) if self.ttl is not None: pipe.expire(key, self.ttl) pipe.execute() def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis DEFAULT_SCHEMA = { "content_key": "prompt", "text": [ {"name": "prompt"}, ], "extra": [{"name": "return_val"}, {"name": "llm_string"}], } def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, schema=cast(Dict, self.DEFAULT_SCHEMA), ) except ValueError: redis = RedisVectorstore( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, index_schema=cast(Dict, self.DEFAULT_SCHEMA), ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations: List = [] # Read from a Hash results = llm_cache.similarity_search( query=prompt, k=1, distance_threshold=self.score_threshold, ) if results: for document in results: generations.extend( _load_generations_from_json(document.metadata["return_val"]) ) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) _dump_generations_to_json([g for g in return_val]) metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": _dump_generations_to_json([g for g in return_val]), } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception CASSANDRA_CACHE_DEFAULT_TABLE_NAME = "langchain_llm_cache" CASSANDRA_CACHE_DEFAULT_TTL_SECONDS = None class CassandraCache(BaseCache): """ Cache that uses Cassandra / Astra DB as a backend. It uses a single Cassandra table. The lookup keys (which get to form the primary key) are: - prompt, a string - llm_string, a deterministic str representation of the model parameters. (needed to prevent collisions same-prompt-different-model collisions) """ def __init__( self, session: Optional[CassandraSession] = None, keyspace: Optional[str] = None, table_name: str = CASSANDRA_CACHE_DEFAULT_TABLE_NAME, ttl_seconds: Optional[int] = CASSANDRA_CACHE_DEFAULT_TTL_SECONDS, skip_provisioning: bool = False, ): """ Initialize with a ready session and a keyspace name. Args: session (cassandra.cluster.Session): an open Cassandra session keyspace (str): the keyspace to use for storing the cache table_name (str): name of the Cassandra table to use as cache ttl_seconds (optional int): time-to-live for cache entries (default: None, i.e. forever) """ try: from cassio.table import ElasticCassandraTable except (ImportError, ModuleNotFoundError): raise ValueError( "Could not import cassio python package. " "Please install it with `pip install cassio`." ) self.session = session self.keyspace = keyspace self.table_name = table_name self.ttl_seconds = ttl_seconds self.kv_cache = ElasticCassandraTable( session=self.session, keyspace=self.keyspace, table=self.table_name, keys=["llm_string", "prompt"], primary_key_type=["TEXT", "TEXT"], ttl_seconds=self.ttl_seconds, skip_provisioning=skip_provisioning, ) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" item = self.kv_cache.get( llm_string=_hash(llm_string), prompt=_hash(prompt), ) if item is not None: generations = _loads_generations(item["body_blob"]) # this protects against malformed cached items: if generations is not None: return generations else: return None else: return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" blob = _dumps_generations(return_val) self.kv_cache.put( llm_string=_hash(llm_string), prompt=_hash(prompt), body_blob=blob, ) def delete_through_llm( self, prompt: str, llm: LLM, stop: Optional[List[str]] = None ) -> None: """ A wrapper around `delete` with the LLM being passed. In case the llm(prompt) calls have a `stop` param, you should pass it here """ llm_string = get_prompts( {**llm.dict(), **{"stop": stop}}, [], )[1] return self.delete(prompt, llm_string=llm_string) def delete(self, prompt: str, llm_string: str) -> None: """Evict from cache if there's an entry.""" return self.kv_cache.delete( llm_string=_hash(llm_string), prompt=_hash(prompt), ) def clear(self, **kwargs: Any) -> None: """Clear cache. This is for all LLMs at once.""" self.kv_cache.clear() CASSANDRA_SEMANTIC_CACHE_DEFAULT_DISTANCE_METRIC = "dot" CASSANDRA_SEMANTIC_CACHE_DEFAULT_SCORE_THRESHOLD = 0.85 CASSANDRA_SEMANTIC_CACHE_DEFAULT_TABLE_NAME = "langchain_llm_semantic_cache" CASSANDRA_SEMANTIC_CACHE_DEFAULT_TTL_SECONDS = None CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE = 16 class CassandraSemanticCache(BaseCache): """ Cache that uses Cassandra as a vector-store backend for semantic (i.e. similarity-based) lookup. It uses a single (vector) Cassandra table and stores, in principle, cached values from several LLMs, so the LLM's llm_string is part of the rows' primary keys. The similarity is based on one of several distance metrics (default: "dot"). If choosing another metric, the default threshold is to be re-tuned accordingly. """ def __init__( self, session: Optional[CassandraSession], keyspace: Optional[str], embedding: Embeddings, table_name: str = CASSANDRA_SEMANTIC_CACHE_DEFAULT_TABLE_NAME, distance_metric: str = CASSANDRA_SEMANTIC_CACHE_DEFAULT_DISTANCE_METRIC, score_threshold: float = CASSANDRA_SEMANTIC_CACHE_DEFAULT_SCORE_THRESHOLD, ttl_seconds: Optional[int] = CASSANDRA_SEMANTIC_CACHE_DEFAULT_TTL_SECONDS, skip_provisioning: bool = False, ): """ Initialize the cache with all relevant parameters. Args: session (cassandra.cluster.Session): an open Cassandra session keyspace (str): the keyspace to use for storing the cache embedding (Embedding): Embedding provider for semantic encoding and search. table_name (str): name of the Cassandra (vector) table to use as cache distance_metric (str, 'dot'): which measure to adopt for similarity searches score_threshold (optional float): numeric value to use as cutoff for the similarity searches ttl_seconds (optional int): time-to-live for cache entries (default: None, i.e. forever) The default score threshold is tuned to the default metric. Tune it carefully yourself if switching to another distance metric. """ try: from cassio.table import MetadataVectorCassandraTable except (ImportError, ModuleNotFoundError): raise ValueError( "Could not import cassio python package. " "Please install it with `pip install cassio`." ) self.session = session self.keyspace = keyspace self.embedding = embedding self.table_name = table_name self.distance_metric = distance_metric self.score_threshold = score_threshold self.ttl_seconds = ttl_seconds # The contract for this class has separate lookup and update: # in order to spare some embedding calculations we cache them between # the two calls. # Note: each instance of this class has its own `_get_embedding` with # its own lru. @lru_cache(maxsize=CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE) def _cache_embedding(text: str) -> List[float]: return self.embedding.embed_query(text=text) self._get_embedding = _cache_embedding self.embedding_dimension = self._get_embedding_dimension() self.table = MetadataVectorCassandraTable( session=self.session, keyspace=self.keyspace, table=self.table_name, primary_key_type=["TEXT"], vector_dimension=self.embedding_dimension, ttl_seconds=self.ttl_seconds, metadata_indexing=("allow", {"_llm_string_hash"}), skip_provisioning=skip_provisioning, ) def _get_embedding_dimension(self) -> int: return len(self._get_embedding(text="This is a sample sentence.")) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" embedding_vector = self._get_embedding(text=prompt) llm_string_hash = _hash(llm_string) body = _dumps_generations(return_val) metadata = { "_prompt": prompt, "_llm_string_hash": llm_string_hash, } row_id = f"{_hash(prompt)}-{llm_string_hash}" # self.table.put( body_blob=body, vector=embedding_vector, row_id=row_id, metadata=metadata, ) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" hit_with_id = self.lookup_with_id(prompt, llm_string) if hit_with_id is not None: return hit_with_id[1] else: return None def lookup_with_id( self, prompt: str, llm_string: str ) -> Optional[Tuple[str, RETURN_VAL_TYPE]]: """ Look up based on prompt and llm_string. If there are hits, return (document_id, cached_entry) """ prompt_embedding: List[float] = self._get_embedding(text=prompt) hits = list( self.table.metric_ann_search( vector=prompt_embedding, metadata={"_llm_string_hash": _hash(llm_string)}, n=1, metric=self.distance_metric, metric_threshold=self.score_threshold, ) ) if hits: hit = hits[0] generations = _loads_generations(hit["body_blob"]) if generations is not None: # this protects against malformed cached items: return ( hit["row_id"], generations, ) else: return None else: return None def lookup_with_id_through_llm( self, prompt: str, llm: LLM, stop: Optional[List[str]] = None ) -> Optional[Tuple[str, RETURN_VAL_TYPE]]: llm_string = get_prompts( {**llm.dict(), **{"stop": stop}}, [], )[1] return self.lookup_with_id(prompt, llm_string=llm_string) def delete_by_document_id(self, document_id: str) -> None: """ Given this is a "similarity search" cache, an invalidation pattern that makes sense is first a lookup to get an ID, and then deleting with that ID. This is for the second step. """ self.table.delete(row_id=document_id) def clear(self, **kwargs: Any) -> None: """Clear the *whole* semantic cache.""" self.table.clear()
[ "langchain.load.load.loads", "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps" ]
[((1586, 1613), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1603, 1613), False, 'import logging\n'), ((5793, 5811), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (5809, 5811), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((5968, 6000), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (5974, 6000), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((6011, 6043), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (6017, 6043), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((6054, 6087), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (6060, 6087), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((6103, 6117), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (6109, 6117), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2730, 2758), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2740, 2758), False, 'import json\n'), ((4550, 4577), 'json.loads', 'json.loads', (['generations_str'], {}), '(generations_str)\n', (4560, 4577), False, 'import json\n'), ((8521, 8564), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (8534, 8564), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((18425, 18432), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (18430, 18432), False, 'from gptcache import Cache\n'), ((19758, 19790), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (19761, 19790), False, 'from gptcache.adapter.api import get\n'), ((20677, 20723), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (20680, 20723), False, 'from gptcache.adapter.api import put\n'), ((24572, 24614), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (24602, 24614), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((24638, 24694), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (24649, 24694), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((29638, 29876), 'cassio.table.ElasticCassandraTable', 'ElasticCassandraTable', ([], {'session': 'self.session', 'keyspace': 'self.keyspace', 'table': 'self.table_name', 'keys': "['llm_string', 'prompt']", 'primary_key_type': "['TEXT', 'TEXT']", 'ttl_seconds': 'self.ttl_seconds', 'skip_provisioning': 'skip_provisioning'}), "(session=self.session, keyspace=self.keyspace, table=\n self.table_name, keys=['llm_string', 'prompt'], primary_key_type=[\n 'TEXT', 'TEXT'], ttl_seconds=self.ttl_seconds, skip_provisioning=\n skip_provisioning)\n", (29659, 29876), False, 'from cassio.table import ElasticCassandraTable\n'), ((34766, 34830), 'functools.lru_cache', 'lru_cache', ([], {'maxsize': 'CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE'}), '(maxsize=CASSANDRA_SEMANTIC_CACHE_EMBEDDING_CACHE_SIZE)\n', (34775, 34830), False, 'from functools import lru_cache\n'), ((35081, 35380), 'cassio.table.MetadataVectorCassandraTable', 'MetadataVectorCassandraTable', ([], {'session': 'self.session', 'keyspace': 'self.keyspace', 'table': 'self.table_name', 'primary_key_type': "['TEXT']", 'vector_dimension': 'self.embedding_dimension', 'ttl_seconds': 'self.ttl_seconds', 'metadata_indexing': "('allow', {'_llm_string_hash'})", 'skip_provisioning': 'skip_provisioning'}), "(session=self.session, keyspace=self.keyspace,\n table=self.table_name, primary_key_type=['TEXT'], vector_dimension=self\n .embedding_dimension, ttl_seconds=self.ttl_seconds, metadata_indexing=(\n 'allow', {'_llm_string_hash'}), skip_provisioning=skip_provisioning)\n", (35109, 35380), False, 'from cassio.table import MetadataVectorCassandraTable\n'), ((2775, 2804), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2785, 2804), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((3618, 3630), 'langchain.load.dump.dumps', 'dumps', (['_item'], {}), '(_item)\n', (3623, 3630), False, 'from langchain.load.dump import dumps\n'), ((4296, 4312), 'langchain.load.load.loads', 'loads', (['_item_str'], {}), '(_item_str)\n', (4301, 4312), False, 'from langchain.load.load import loads\n'), ((4681, 4710), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (4691, 4710), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6867, 6887), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6874, 6887), False, 'from sqlalchemy.orm import Session\n'), ((7974, 7994), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (7981, 7994), False, 'from sqlalchemy.orm import Session\n'), ((8176, 8196), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (8183, 8196), False, 'from sqlalchemy.orm import Session\n'), ((18499, 18541), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (18516, 18541), False, 'import inspect\n'), ((20946, 20976), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (20950, 20976), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast\n'), ((21876, 21896), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (21885, 21896), False, 'from datetime import timedelta\n'), ((24439, 24465), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (24463, 24465), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((24501, 24549), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (24513, 24549), False, 'from langchain.utils import get_from_env\n'), ((4330, 4357), 'json.loads', 'json.loads', (['generations_str'], {}), '(generations_str)\n', (4340, 4357), False, 'import json\n'), ((11054, 11173), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (11067, 11173), False, 'import warnings\n'), ((15954, 16082), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (15967, 16082), False, 'import warnings\n'), ((19844, 19873), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (19854, 19873), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((7884, 7894), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (7889, 7894), False, 'from langchain.load.dump import dumps\n'), ((10513, 10534), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (10523, 10534), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((13908, 13939), 'typing.cast', 'cast', (['Dict', 'self.DEFAULT_SCHEMA'], {}), '(Dict, self.DEFAULT_SCHEMA)\n', (13912, 13939), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast\n'), ((18885, 18923), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (18901, 18923), False, 'from gptcache.manager.factory import get_data_manager\n'), ((19897, 19912), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (19907, 19912), False, 'import json\n'), ((7022, 7035), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (7027, 7035), False, 'from langchain.load.load import loads\n'), ((14172, 14203), 'typing.cast', 'cast', (['Dict', 'self.DEFAULT_SCHEMA'], {}), '(Dict, self.DEFAULT_SCHEMA)\n', (14176, 14203), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Tuple, Type, Union, cast\n'), ((7592, 7615), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (7602, 7615), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6637, 6671), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (6643, 6671), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.load.dump import dumpd, dumps from langchain.schema import ( ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo, ) from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage def _get_verbosity() -> bool: return langchain.verbose class BaseChatModel(BaseLanguageModel, ABC): cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) tags: Optional[List[str]] = Field(default=None, exclude=True) """Tags to add to the run trace.""" metadata: Optional[Dict[str, Any]] = Field(default=None, exclude=True) """Metadata to add to the run trace.""" @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True def _combine_llm_outputs(self, llm_outputs: List[Optional[dict]]) -> dict: return {} def _get_invocation_params( self, stop: Optional[List[str]] = None, **kwargs: Any, ) -> dict: params = self.dict() params["stop"] = stop return {**params, **kwargs} def _get_llm_string(self, stop: Optional[List[str]] = None, **kwargs: Any) -> str: if self.lc_serializable: params = {**kwargs, **{"stop": stop}} param_string = str(sorted([(k, v) for k, v in params.items()])) llm_string = dumps(self) return llm_string + "---" + param_string else: params = self._get_invocation_params(stop=stop, **kwargs) params = {**params, **kwargs} return str(sorted([(k, v) for k, v in params.items()])) def generate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = [] for i, m in enumerate(messages): try: results.append( self._generate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) ) except (KeyboardInterrupt, Exception) as e: if run_managers: run_managers[i].on_llm_error(e) raise e flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) if run_managers: run_infos = [] for manager, flattened_output in zip(run_managers, flattened_outputs): manager.on_llm_end(flattened_output) run_infos.append(RunInfo(run_id=manager.run_id)) output.run = run_infos return output async def agenerate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = await callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = await asyncio.gather( *[ self._agenerate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) for i, m in enumerate(messages) ], return_exceptions=True, ) exceptions = [] for i, res in enumerate(results): if isinstance(res, Exception): if run_managers: await run_managers[i].on_llm_error(res) exceptions.append(res) if exceptions: if run_managers: await asyncio.gather( *[ run_manager.on_llm_end( LLMResult( generations=[res.generations], llm_output=res.llm_output ) ) for run_manager, res in zip(run_managers, results) if not isinstance(res, Exception) ] ) raise exceptions[0] flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) await asyncio.gather( *[ run_manager.on_llm_end(flattened_output) for run_manager, flattened_output in zip( run_managers, flattened_outputs ) ] ) if run_managers: output.run = [ RunInfo(run_id=run_manager.run_id) for run_manager in run_managers ] return output def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return await self.agenerate( prompt_messages, stop=stop, callbacks=callbacks, **kwargs ) def _generate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return self._generate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = self._generate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result async def _agenerate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return await self._agenerate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = await self._agenerate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result @abstractmethod def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" @abstractmethod async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" def __call__( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: generation = self.generate( [messages], stop=stop, callbacks=callbacks, **kwargs ).generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") async def _call_async( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: result = await self.agenerate( [messages], stop=stop, callbacks=callbacks, **kwargs ) generation = result.generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") def call_as_llm( self, message: str, stop: Optional[List[str]] = None, **kwargs: Any ) -> str: return self.predict(message, stop=stop, **kwargs) def predict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = self([HumanMessage(content=text)], stop=_stop, **kwargs) return result.content def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return self(messages, stop=_stop, **kwargs) async def apredict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = await self._call_async( [HumanMessage(content=text)], stop=_stop, **kwargs ) return result.content async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(messages, stop=_stop, **kwargs) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} @property @abstractmethod def _llm_type(self) -> str: """Return type of chat model.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict class SimpleChatModel(BaseChatModel): def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: output_str = self._call(messages, stop=stop, run_manager=run_manager, **kwargs) message = AIMessage(content=output_str) generation = ChatGeneration(message=message) return ChatResult(generations=[generation]) @abstractmethod def _call( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Simpler interface.""" async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: func = partial( self._generate, messages, stop=stop, run_manager=run_manager, **kwargs ) return await asyncio.get_event_loop().run_in_executor(None, func)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages.HumanMessage", "langchain.schema.ChatGeneration", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((923, 960), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (928, 960), False, 'from pydantic import Field, root_validator\n'), ((1034, 1067), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1039, 1067), False, 'from pydantic import Field, root_validator\n'), ((1122, 1155), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1127, 1155), False, 'from pydantic import Field, root_validator\n'), ((1188, 1221), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1193, 1221), False, 'from pydantic import Field, root_validator\n'), ((1303, 1336), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1308, 1336), False, 'from pydantic import Field, root_validator\n'), ((1387, 1403), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1401, 1403), False, 'from pydantic import Field, root_validator\n'), ((3255, 3367), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (3280, 3367), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((4456, 4513), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (4465, 4513), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((5289, 5406), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (5319, 5406), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7141, 7198), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7150, 7198), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15689, 15718), 'langchain.schema.messages.AIMessage', 'AIMessage', ([], {'content': 'output_str'}), '(content=output_str)\n', (15698, 15718), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((15740, 15771), 'langchain.schema.ChatGeneration', 'ChatGeneration', ([], {'message': 'message'}), '(message=message)\n', (15754, 15771), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15787, 15823), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': '[generation]'}), '(generations=[generation])\n', (15797, 15823), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((16335, 16414), 'functools.partial', 'partial', (['self._generate', 'messages'], {'stop': 'stop', 'run_manager': 'run_manager'}), '(self._generate, messages, stop=stop, run_manager=run_manager, **kwargs)\n', (16342, 16414), False, 'from functools import partial\n'), ((1594, 1696), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (1607, 1696), False, 'import warnings\n'), ((2538, 2549), 'langchain.load.dump.dumps', 'dumps', (['self'], {}), '(self)\n', (2543, 2549), False, 'from langchain.load.dump import dumpd, dumps\n'), ((3532, 3543), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (3537, 3543), False, 'from langchain.load.dump import dumpd, dumps\n'), ((4187, 4254), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (4196, 4254), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((6872, 6939), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6881, 6939), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9447, 9462), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (9452, 9462), False, 'from langchain.load.dump import dumpd, dumps\n'), ((9487, 9533), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (9513, 9533), False, 'import langchain\n'), ((11167, 11182), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (11172, 11182), False, 'from langchain.load.dump import dumpd, dumps\n'), ((11207, 11253), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (11233, 11253), False, 'import langchain\n'), ((5578, 5589), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (5583, 5589), False, 'from langchain.load.dump import dumpd, dumps\n'), ((7521, 7555), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (7528, 7555), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9601, 9634), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (9611, 9634), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9950, 10016), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (9976, 10016), False, 'import langchain\n'), ((11321, 11354), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (11331, 11354), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((11684, 11750), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (11710, 11750), False, 'import langchain\n'), ((13743, 13769), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (13755, 13769), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((4735, 4765), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'manager.run_id'}), '(run_id=manager.run_id)\n', (4742, 4765), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((8614, 8647), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (8631, 8647), False, 'import inspect\n'), ((10319, 10353), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (10336, 10353), False, 'import inspect\n'), ((14419, 14445), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (14431, 14445), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((16458, 16482), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (16480, 16482), False, 'import asyncio\n'), ((6469, 6536), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6478, 6536), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n')]
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.load.dump import dumpd, dumps from langchain.schema import ( ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo, ) from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage def _get_verbosity() -> bool: return langchain.verbose class BaseChatModel(BaseLanguageModel, ABC): cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) tags: Optional[List[str]] = Field(default=None, exclude=True) """Tags to add to the run trace.""" metadata: Optional[Dict[str, Any]] = Field(default=None, exclude=True) """Metadata to add to the run trace.""" @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True def _combine_llm_outputs(self, llm_outputs: List[Optional[dict]]) -> dict: return {} def _get_invocation_params( self, stop: Optional[List[str]] = None, **kwargs: Any, ) -> dict: params = self.dict() params["stop"] = stop return {**params, **kwargs} def _get_llm_string(self, stop: Optional[List[str]] = None, **kwargs: Any) -> str: if self.lc_serializable: params = {**kwargs, **{"stop": stop}} param_string = str(sorted([(k, v) for k, v in params.items()])) llm_string = dumps(self) return llm_string + "---" + param_string else: params = self._get_invocation_params(stop=stop, **kwargs) params = {**params, **kwargs} return str(sorted([(k, v) for k, v in params.items()])) def generate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = [] for i, m in enumerate(messages): try: results.append( self._generate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) ) except (KeyboardInterrupt, Exception) as e: if run_managers: run_managers[i].on_llm_error(e) raise e flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) if run_managers: run_infos = [] for manager, flattened_output in zip(run_managers, flattened_outputs): manager.on_llm_end(flattened_output) run_infos.append(RunInfo(run_id=manager.run_id)) output.run = run_infos return output async def agenerate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = await callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = await asyncio.gather( *[ self._agenerate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) for i, m in enumerate(messages) ], return_exceptions=True, ) exceptions = [] for i, res in enumerate(results): if isinstance(res, Exception): if run_managers: await run_managers[i].on_llm_error(res) exceptions.append(res) if exceptions: if run_managers: await asyncio.gather( *[ run_manager.on_llm_end( LLMResult( generations=[res.generations], llm_output=res.llm_output ) ) for run_manager, res in zip(run_managers, results) if not isinstance(res, Exception) ] ) raise exceptions[0] flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) await asyncio.gather( *[ run_manager.on_llm_end(flattened_output) for run_manager, flattened_output in zip( run_managers, flattened_outputs ) ] ) if run_managers: output.run = [ RunInfo(run_id=run_manager.run_id) for run_manager in run_managers ] return output def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return await self.agenerate( prompt_messages, stop=stop, callbacks=callbacks, **kwargs ) def _generate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return self._generate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = self._generate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result async def _agenerate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return await self._agenerate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = await self._agenerate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result @abstractmethod def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" @abstractmethod async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" def __call__( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: generation = self.generate( [messages], stop=stop, callbacks=callbacks, **kwargs ).generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") async def _call_async( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: result = await self.agenerate( [messages], stop=stop, callbacks=callbacks, **kwargs ) generation = result.generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") def call_as_llm( self, message: str, stop: Optional[List[str]] = None, **kwargs: Any ) -> str: return self.predict(message, stop=stop, **kwargs) def predict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = self([HumanMessage(content=text)], stop=_stop, **kwargs) return result.content def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return self(messages, stop=_stop, **kwargs) async def apredict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = await self._call_async( [HumanMessage(content=text)], stop=_stop, **kwargs ) return result.content async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(messages, stop=_stop, **kwargs) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} @property @abstractmethod def _llm_type(self) -> str: """Return type of chat model.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict class SimpleChatModel(BaseChatModel): def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: output_str = self._call(messages, stop=stop, run_manager=run_manager, **kwargs) message = AIMessage(content=output_str) generation = ChatGeneration(message=message) return ChatResult(generations=[generation]) @abstractmethod def _call( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Simpler interface.""" async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: func = partial( self._generate, messages, stop=stop, run_manager=run_manager, **kwargs ) return await asyncio.get_event_loop().run_in_executor(None, func)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages.HumanMessage", "langchain.schema.ChatGeneration", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((923, 960), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (928, 960), False, 'from pydantic import Field, root_validator\n'), ((1034, 1067), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1039, 1067), False, 'from pydantic import Field, root_validator\n'), ((1122, 1155), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1127, 1155), False, 'from pydantic import Field, root_validator\n'), ((1188, 1221), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1193, 1221), False, 'from pydantic import Field, root_validator\n'), ((1303, 1336), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1308, 1336), False, 'from pydantic import Field, root_validator\n'), ((1387, 1403), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1401, 1403), False, 'from pydantic import Field, root_validator\n'), ((3255, 3367), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (3280, 3367), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((4456, 4513), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (4465, 4513), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((5289, 5406), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (5319, 5406), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7141, 7198), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7150, 7198), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15689, 15718), 'langchain.schema.messages.AIMessage', 'AIMessage', ([], {'content': 'output_str'}), '(content=output_str)\n', (15698, 15718), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((15740, 15771), 'langchain.schema.ChatGeneration', 'ChatGeneration', ([], {'message': 'message'}), '(message=message)\n', (15754, 15771), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15787, 15823), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': '[generation]'}), '(generations=[generation])\n', (15797, 15823), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((16335, 16414), 'functools.partial', 'partial', (['self._generate', 'messages'], {'stop': 'stop', 'run_manager': 'run_manager'}), '(self._generate, messages, stop=stop, run_manager=run_manager, **kwargs)\n', (16342, 16414), False, 'from functools import partial\n'), ((1594, 1696), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (1607, 1696), False, 'import warnings\n'), ((2538, 2549), 'langchain.load.dump.dumps', 'dumps', (['self'], {}), '(self)\n', (2543, 2549), False, 'from langchain.load.dump import dumpd, dumps\n'), ((3532, 3543), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (3537, 3543), False, 'from langchain.load.dump import dumpd, dumps\n'), ((4187, 4254), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (4196, 4254), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((6872, 6939), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6881, 6939), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9447, 9462), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (9452, 9462), False, 'from langchain.load.dump import dumpd, dumps\n'), ((9487, 9533), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (9513, 9533), False, 'import langchain\n'), ((11167, 11182), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (11172, 11182), False, 'from langchain.load.dump import dumpd, dumps\n'), ((11207, 11253), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (11233, 11253), False, 'import langchain\n'), ((5578, 5589), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (5583, 5589), False, 'from langchain.load.dump import dumpd, dumps\n'), ((7521, 7555), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (7528, 7555), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9601, 9634), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (9611, 9634), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9950, 10016), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (9976, 10016), False, 'import langchain\n'), ((11321, 11354), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (11331, 11354), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((11684, 11750), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (11710, 11750), False, 'import langchain\n'), ((13743, 13769), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (13755, 13769), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((4735, 4765), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'manager.run_id'}), '(run_id=manager.run_id)\n', (4742, 4765), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((8614, 8647), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (8631, 8647), False, 'import inspect\n'), ((10319, 10353), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (10336, 10353), False, 'import inspect\n'), ((14419, 14445), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (14431, 14445), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((16458, 16482), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (16480, 16482), False, 'import asyncio\n'), ((6469, 6536), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6478, 6536), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n')]
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.load.dump import dumpd, dumps from langchain.schema import ( ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo, ) from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage def _get_verbosity() -> bool: return langchain.verbose class BaseChatModel(BaseLanguageModel, ABC): cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) tags: Optional[List[str]] = Field(default=None, exclude=True) """Tags to add to the run trace.""" metadata: Optional[Dict[str, Any]] = Field(default=None, exclude=True) """Metadata to add to the run trace.""" @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True def _combine_llm_outputs(self, llm_outputs: List[Optional[dict]]) -> dict: return {} def _get_invocation_params( self, stop: Optional[List[str]] = None, **kwargs: Any, ) -> dict: params = self.dict() params["stop"] = stop return {**params, **kwargs} def _get_llm_string(self, stop: Optional[List[str]] = None, **kwargs: Any) -> str: if self.lc_serializable: params = {**kwargs, **{"stop": stop}} param_string = str(sorted([(k, v) for k, v in params.items()])) llm_string = dumps(self) return llm_string + "---" + param_string else: params = self._get_invocation_params(stop=stop, **kwargs) params = {**params, **kwargs} return str(sorted([(k, v) for k, v in params.items()])) def generate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = [] for i, m in enumerate(messages): try: results.append( self._generate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) ) except (KeyboardInterrupt, Exception) as e: if run_managers: run_managers[i].on_llm_error(e) raise e flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) if run_managers: run_infos = [] for manager, flattened_output in zip(run_managers, flattened_outputs): manager.on_llm_end(flattened_output) run_infos.append(RunInfo(run_id=manager.run_id)) output.run = run_infos return output async def agenerate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = await callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = await asyncio.gather( *[ self._agenerate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) for i, m in enumerate(messages) ], return_exceptions=True, ) exceptions = [] for i, res in enumerate(results): if isinstance(res, Exception): if run_managers: await run_managers[i].on_llm_error(res) exceptions.append(res) if exceptions: if run_managers: await asyncio.gather( *[ run_manager.on_llm_end( LLMResult( generations=[res.generations], llm_output=res.llm_output ) ) for run_manager, res in zip(run_managers, results) if not isinstance(res, Exception) ] ) raise exceptions[0] flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) await asyncio.gather( *[ run_manager.on_llm_end(flattened_output) for run_manager, flattened_output in zip( run_managers, flattened_outputs ) ] ) if run_managers: output.run = [ RunInfo(run_id=run_manager.run_id) for run_manager in run_managers ] return output def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return await self.agenerate( prompt_messages, stop=stop, callbacks=callbacks, **kwargs ) def _generate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return self._generate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = self._generate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result async def _agenerate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return await self._agenerate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = await self._agenerate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result @abstractmethod def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" @abstractmethod async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" def __call__( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: generation = self.generate( [messages], stop=stop, callbacks=callbacks, **kwargs ).generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") async def _call_async( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: result = await self.agenerate( [messages], stop=stop, callbacks=callbacks, **kwargs ) generation = result.generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") def call_as_llm( self, message: str, stop: Optional[List[str]] = None, **kwargs: Any ) -> str: return self.predict(message, stop=stop, **kwargs) def predict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = self([HumanMessage(content=text)], stop=_stop, **kwargs) return result.content def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return self(messages, stop=_stop, **kwargs) async def apredict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = await self._call_async( [HumanMessage(content=text)], stop=_stop, **kwargs ) return result.content async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(messages, stop=_stop, **kwargs) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} @property @abstractmethod def _llm_type(self) -> str: """Return type of chat model.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict class SimpleChatModel(BaseChatModel): def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: output_str = self._call(messages, stop=stop, run_manager=run_manager, **kwargs) message = AIMessage(content=output_str) generation = ChatGeneration(message=message) return ChatResult(generations=[generation]) @abstractmethod def _call( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Simpler interface.""" async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: func = partial( self._generate, messages, stop=stop, run_manager=run_manager, **kwargs ) return await asyncio.get_event_loop().run_in_executor(None, func)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages.HumanMessage", "langchain.schema.ChatGeneration", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((923, 960), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (928, 960), False, 'from pydantic import Field, root_validator\n'), ((1034, 1067), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1039, 1067), False, 'from pydantic import Field, root_validator\n'), ((1122, 1155), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1127, 1155), False, 'from pydantic import Field, root_validator\n'), ((1188, 1221), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1193, 1221), False, 'from pydantic import Field, root_validator\n'), ((1303, 1336), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1308, 1336), False, 'from pydantic import Field, root_validator\n'), ((1387, 1403), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1401, 1403), False, 'from pydantic import Field, root_validator\n'), ((3255, 3367), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (3280, 3367), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((4456, 4513), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (4465, 4513), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((5289, 5406), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (5319, 5406), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7141, 7198), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7150, 7198), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15689, 15718), 'langchain.schema.messages.AIMessage', 'AIMessage', ([], {'content': 'output_str'}), '(content=output_str)\n', (15698, 15718), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((15740, 15771), 'langchain.schema.ChatGeneration', 'ChatGeneration', ([], {'message': 'message'}), '(message=message)\n', (15754, 15771), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15787, 15823), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': '[generation]'}), '(generations=[generation])\n', (15797, 15823), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((16335, 16414), 'functools.partial', 'partial', (['self._generate', 'messages'], {'stop': 'stop', 'run_manager': 'run_manager'}), '(self._generate, messages, stop=stop, run_manager=run_manager, **kwargs)\n', (16342, 16414), False, 'from functools import partial\n'), ((1594, 1696), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (1607, 1696), False, 'import warnings\n'), ((2538, 2549), 'langchain.load.dump.dumps', 'dumps', (['self'], {}), '(self)\n', (2543, 2549), False, 'from langchain.load.dump import dumpd, dumps\n'), ((3532, 3543), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (3537, 3543), False, 'from langchain.load.dump import dumpd, dumps\n'), ((4187, 4254), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (4196, 4254), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((6872, 6939), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6881, 6939), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9447, 9462), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (9452, 9462), False, 'from langchain.load.dump import dumpd, dumps\n'), ((9487, 9533), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (9513, 9533), False, 'import langchain\n'), ((11167, 11182), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (11172, 11182), False, 'from langchain.load.dump import dumpd, dumps\n'), ((11207, 11253), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (11233, 11253), False, 'import langchain\n'), ((5578, 5589), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (5583, 5589), False, 'from langchain.load.dump import dumpd, dumps\n'), ((7521, 7555), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (7528, 7555), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9601, 9634), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (9611, 9634), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9950, 10016), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (9976, 10016), False, 'import langchain\n'), ((11321, 11354), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (11331, 11354), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((11684, 11750), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (11710, 11750), False, 'import langchain\n'), ((13743, 13769), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (13755, 13769), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((4735, 4765), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'manager.run_id'}), '(run_id=manager.run_id)\n', (4742, 4765), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((8614, 8647), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (8631, 8647), False, 'import inspect\n'), ((10319, 10353), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (10336, 10353), False, 'import inspect\n'), ((14419, 14445), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (14431, 14445), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((16458, 16482), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (16480, 16482), False, 'import asyncio\n'), ((6469, 6536), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6478, 6536), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n')]
import asyncio import inspect import warnings from abc import ABC, abstractmethod from functools import partial from typing import Any, Dict, List, Mapping, Optional, Sequence from pydantic import Field, root_validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.load.dump import dumpd, dumps from langchain.schema import ( ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo, ) from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage def _get_verbosity() -> bool: return langchain.verbose class BaseChatModel(BaseLanguageModel, ABC): cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) tags: Optional[List[str]] = Field(default=None, exclude=True) """Tags to add to the run trace.""" metadata: Optional[Dict[str, Any]] = Field(default=None, exclude=True) """Metadata to add to the run trace.""" @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True def _combine_llm_outputs(self, llm_outputs: List[Optional[dict]]) -> dict: return {} def _get_invocation_params( self, stop: Optional[List[str]] = None, **kwargs: Any, ) -> dict: params = self.dict() params["stop"] = stop return {**params, **kwargs} def _get_llm_string(self, stop: Optional[List[str]] = None, **kwargs: Any) -> str: if self.lc_serializable: params = {**kwargs, **{"stop": stop}} param_string = str(sorted([(k, v) for k, v in params.items()])) llm_string = dumps(self) return llm_string + "---" + param_string else: params = self._get_invocation_params(stop=stop, **kwargs) params = {**params, **kwargs} return str(sorted([(k, v) for k, v in params.items()])) def generate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = [] for i, m in enumerate(messages): try: results.append( self._generate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) ) except (KeyboardInterrupt, Exception) as e: if run_managers: run_managers[i].on_llm_error(e) raise e flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) if run_managers: run_infos = [] for manager, flattened_output in zip(run_managers, flattened_outputs): manager.on_llm_end(flattened_output) run_infos.append(RunInfo(run_id=manager.run_id)) output.run = run_infos return output async def agenerate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> LLMResult: """Top Level call""" params = self._get_invocation_params(stop=stop, **kwargs) options = {"stop": stop} callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) run_managers = await callback_manager.on_chat_model_start( dumpd(self), messages, invocation_params=params, options=options ) results = await asyncio.gather( *[ self._agenerate_with_cache( m, stop=stop, run_manager=run_managers[i] if run_managers else None, **kwargs, ) for i, m in enumerate(messages) ], return_exceptions=True, ) exceptions = [] for i, res in enumerate(results): if isinstance(res, Exception): if run_managers: await run_managers[i].on_llm_error(res) exceptions.append(res) if exceptions: if run_managers: await asyncio.gather( *[ run_manager.on_llm_end( LLMResult( generations=[res.generations], llm_output=res.llm_output ) ) for run_manager, res in zip(run_managers, results) if not isinstance(res, Exception) ] ) raise exceptions[0] flattened_outputs = [ LLMResult(generations=[res.generations], llm_output=res.llm_output) for res in results ] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] output = LLMResult(generations=generations, llm_output=llm_output) await asyncio.gather( *[ run_manager.on_llm_end(flattened_output) for run_manager, flattened_output in zip( run_managers, flattened_outputs ) ] ) if run_managers: output.run = [ RunInfo(run_id=run_manager.run_id) for run_manager in run_managers ] return output def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return self.generate(prompt_messages, stop=stop, callbacks=callbacks, **kwargs) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] return await self.agenerate( prompt_messages, stop=stop, callbacks=callbacks, **kwargs ) def _generate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return self._generate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = self._generate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = self._generate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result async def _agenerate_with_cache( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) disregard_cache = self.cache is not None and not self.cache if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) if new_arg_supported: return await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: return await self._agenerate(messages, stop=stop, **kwargs) else: llm_string = self._get_llm_string(stop=stop, **kwargs) prompt = dumps(messages) cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): return ChatResult(generations=cache_val) else: if new_arg_supported: result = await self._agenerate( messages, stop=stop, run_manager=run_manager, **kwargs ) else: result = await self._agenerate(messages, stop=stop, **kwargs) langchain.llm_cache.update(prompt, llm_string, result.generations) return result @abstractmethod def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" @abstractmethod async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: """Top Level call""" def __call__( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: generation = self.generate( [messages], stop=stop, callbacks=callbacks, **kwargs ).generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") async def _call_async( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> BaseMessage: result = await self.agenerate( [messages], stop=stop, callbacks=callbacks, **kwargs ) generation = result.generations[0][0] if isinstance(generation, ChatGeneration): return generation.message else: raise ValueError("Unexpected generation type") def call_as_llm( self, message: str, stop: Optional[List[str]] = None, **kwargs: Any ) -> str: return self.predict(message, stop=stop, **kwargs) def predict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = self([HumanMessage(content=text)], stop=_stop, **kwargs) return result.content def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return self(messages, stop=_stop, **kwargs) async def apredict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) result = await self._call_async( [HumanMessage(content=text)], stop=_stop, **kwargs ) return result.content async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(messages, stop=_stop, **kwargs) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} @property @abstractmethod def _llm_type(self) -> str: """Return type of chat model.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict class SimpleChatModel(BaseChatModel): def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: output_str = self._call(messages, stop=stop, run_manager=run_manager, **kwargs) message = AIMessage(content=output_str) generation = ChatGeneration(message=message) return ChatResult(generations=[generation]) @abstractmethod def _call( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Simpler interface.""" async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> ChatResult: func = partial( self._generate, messages, stop=stop, run_manager=run_manager, **kwargs ) return await asyncio.get_event_loop().run_in_executor(None, func)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.messages.AIMessage", "langchain.schema.ChatResult", "langchain.load.dump.dumps", "langchain.callbacks.manager.CallbackManager.configure", "langchain.load.dump.dumpd", "langchain.schema.RunInfo", "langchain.schema.messages.HumanMessage", "langchain.schema.ChatGeneration", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((923, 960), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (928, 960), False, 'from pydantic import Field, root_validator\n'), ((1034, 1067), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1039, 1067), False, 'from pydantic import Field, root_validator\n'), ((1122, 1155), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1127, 1155), False, 'from pydantic import Field, root_validator\n'), ((1188, 1221), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1193, 1221), False, 'from pydantic import Field, root_validator\n'), ((1303, 1336), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (1308, 1336), False, 'from pydantic import Field, root_validator\n'), ((1387, 1403), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (1401, 1403), False, 'from pydantic import Field, root_validator\n'), ((3255, 3367), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (3280, 3367), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((4456, 4513), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (4465, 4513), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((5289, 5406), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (5319, 5406), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7141, 7198), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7150, 7198), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15689, 15718), 'langchain.schema.messages.AIMessage', 'AIMessage', ([], {'content': 'output_str'}), '(content=output_str)\n', (15698, 15718), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((15740, 15771), 'langchain.schema.ChatGeneration', 'ChatGeneration', ([], {'message': 'message'}), '(message=message)\n', (15754, 15771), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((15787, 15823), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': '[generation]'}), '(generations=[generation])\n', (15797, 15823), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((16335, 16414), 'functools.partial', 'partial', (['self._generate', 'messages'], {'stop': 'stop', 'run_manager': 'run_manager'}), '(self._generate, messages, stop=stop, run_manager=run_manager, **kwargs)\n', (16342, 16414), False, 'from functools import partial\n'), ((1594, 1696), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (1607, 1696), False, 'import warnings\n'), ((2538, 2549), 'langchain.load.dump.dumps', 'dumps', (['self'], {}), '(self)\n', (2543, 2549), False, 'from langchain.load.dump import dumpd, dumps\n'), ((3532, 3543), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (3537, 3543), False, 'from langchain.load.dump import dumpd, dumps\n'), ((4187, 4254), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (4196, 4254), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((6872, 6939), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6881, 6939), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9447, 9462), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (9452, 9462), False, 'from langchain.load.dump import dumpd, dumps\n'), ((9487, 9533), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (9513, 9533), False, 'import langchain\n'), ((11167, 11182), 'langchain.load.dump.dumps', 'dumps', (['messages'], {}), '(messages)\n', (11172, 11182), False, 'from langchain.load.dump import dumpd, dumps\n'), ((11207, 11253), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (11233, 11253), False, 'import langchain\n'), ((5578, 5589), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (5583, 5589), False, 'from langchain.load.dump import dumpd, dumps\n'), ((7521, 7555), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (7528, 7555), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9601, 9634), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (9611, 9634), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((9950, 10016), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (9976, 10016), False, 'import langchain\n'), ((11321, 11354), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': 'cache_val'}), '(generations=cache_val)\n', (11331, 11354), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((11684, 11750), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result.generations'], {}), '(prompt, llm_string, result.generations)\n', (11710, 11750), False, 'import langchain\n'), ((13743, 13769), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (13755, 13769), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((4735, 4765), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'manager.run_id'}), '(run_id=manager.run_id)\n', (4742, 4765), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n'), ((8614, 8647), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (8631, 8647), False, 'import inspect\n'), ((10319, 10353), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (10336, 10353), False, 'import inspect\n'), ((14419, 14445), 'langchain.schema.messages.HumanMessage', 'HumanMessage', ([], {'content': 'text'}), '(content=text)\n', (14431, 14445), False, 'from langchain.schema.messages import AIMessage, BaseMessage, HumanMessage\n'), ((16458, 16482), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (16480, 16482), False, 'import asyncio\n'), ((6469, 6536), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': '[res.generations]', 'llm_output': 'res.llm_output'}), '(generations=[res.generations], llm_output=res.llm_output)\n', (6478, 6536), False, 'from langchain.schema import ChatGeneration, ChatResult, LLMResult, PromptValue, RunInfo\n')]
from __future__ import annotations import asyncio import functools import logging import os import warnings from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( Any, AsyncGenerator, Dict, Generator, List, Optional, Type, TypeVar, Union, cast, ) from uuid import UUID, uuid4 import langchain from langchain.callbacks.base import ( BaseCallbackHandler, BaseCallbackManager, ChainManagerMixin, LLMManagerMixin, RunManagerMixin, ToolManagerMixin, ) from langchain.callbacks.openai_info import OpenAICallbackHandler from langchain.callbacks.stdout import StdOutCallbackHandler from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.callbacks.tracers.langchain_v1 import LangChainTracerV1, TracerSessionV1 from langchain.callbacks.tracers.stdout import ConsoleCallbackHandler from langchain.callbacks.tracers.wandb import WandbTracer from langchain.schema import ( AgentAction, AgentFinish, BaseMessage, LLMResult, get_buffer_string, ) logger = logging.getLogger(__name__) Callbacks = Optional[Union[List[BaseCallbackHandler], BaseCallbackManager]] openai_callback_var: ContextVar[Optional[OpenAICallbackHandler]] = ContextVar( "openai_callback", default=None ) tracing_callback_var: ContextVar[ Optional[LangChainTracerV1] ] = ContextVar( # noqa: E501 "tracing_callback", default=None ) wandb_tracing_callback_var: ContextVar[ Optional[WandbTracer] ] = ContextVar( # noqa: E501 "tracing_wandb_callback", default=None ) tracing_v2_callback_var: ContextVar[ Optional[LangChainTracer] ] = ContextVar( # noqa: E501 "tracing_callback_v2", default=None ) def _get_debug() -> bool: return langchain.debug @contextmanager def get_openai_callback() -> Generator[OpenAICallbackHandler, None, None]: """Get OpenAI callback handler in a context manager.""" cb = OpenAICallbackHandler() openai_callback_var.set(cb) yield cb openai_callback_var.set(None) @contextmanager def tracing_enabled( session_name: str = "default", ) -> Generator[TracerSessionV1, None, None]: """Get Tracer in a context manager.""" cb = LangChainTracerV1() session = cast(TracerSessionV1, cb.load_session(session_name)) tracing_callback_var.set(cb) yield session tracing_callback_var.set(None) @contextmanager def wandb_tracing_enabled( session_name: str = "default", ) -> Generator[None, None, None]: """Get WandbTracer in a context manager.""" cb = WandbTracer() wandb_tracing_callback_var.set(cb) yield None wandb_tracing_callback_var.set(None) @contextmanager def tracing_v2_enabled( session_name: Optional[str] = None, *, example_id: Optional[Union[str, UUID]] = None, ) -> Generator[None, None, None]: """Get the experimental tracer handler in a context manager.""" # Issue a warning that this is experimental warnings.warn( "The tracing v2 API is in development. " "This is not yet stable and may change in the future." ) if isinstance(example_id, str): example_id = UUID(example_id) cb = LangChainTracer( example_id=example_id, session_name=session_name, ) tracing_v2_callback_var.set(cb) yield tracing_v2_callback_var.set(None) @contextmanager def trace_as_chain_group( group_name: str, *, session_name: Optional[str] = None, example_id: Optional[Union[str, UUID]] = None, ) -> Generator[CallbackManager, None, None]: """Get a callback manager for a chain group in a context manager.""" cb = LangChainTracer( session_name=session_name, example_id=example_id, ) cm = CallbackManager.configure( inheritable_callbacks=[cb], ) run_manager = cm.on_chain_start({"name": group_name}, {}) yield run_manager.get_child() run_manager.on_chain_end({}) @asynccontextmanager async def atrace_as_chain_group( group_name: str, *, session_name: Optional[str] = None, example_id: Optional[Union[str, UUID]] = None, ) -> AsyncGenerator[AsyncCallbackManager, None]: """Get a callback manager for a chain group in a context manager.""" cb = LangChainTracer( session_name=session_name, example_id=example_id, ) cm = AsyncCallbackManager.configure( inheritable_callbacks=[cb], ) run_manager = await cm.on_chain_start({"name": group_name}, {}) try: yield run_manager.get_child() finally: await run_manager.on_chain_end({}) def _handle_event( handlers: List[BaseCallbackHandler], event_name: str, ignore_condition_name: Optional[str], *args: Any, **kwargs: Any, ) -> None: """Generic event handler for CallbackManager.""" message_strings: Optional[List[str]] = None for handler in handlers: try: if ignore_condition_name is None or not getattr( handler, ignore_condition_name ): getattr(handler, event_name)(*args, **kwargs) except NotImplementedError as e: if event_name == "on_chat_model_start": if message_strings is None: message_strings = [get_buffer_string(m) for m in args[1]] _handle_event( [handler], "on_llm_start", "ignore_llm", args[0], message_strings, *args[2:], **kwargs, ) else: logger.warning(f"Error in {event_name} callback: {e}") except Exception as e: if handler.raise_error: raise e logging.warning(f"Error in {event_name} callback: {e}") async def _ahandle_event_for_handler( handler: BaseCallbackHandler, event_name: str, ignore_condition_name: Optional[str], *args: Any, **kwargs: Any, ) -> None: try: if ignore_condition_name is None or not getattr(handler, ignore_condition_name): event = getattr(handler, event_name) if asyncio.iscoroutinefunction(event): await event(*args, **kwargs) else: await asyncio.get_event_loop().run_in_executor( None, functools.partial(event, *args, **kwargs) ) except NotImplementedError as e: if event_name == "on_chat_model_start": message_strings = [get_buffer_string(m) for m in args[1]] await _ahandle_event_for_handler( handler, "on_llm_start", "ignore_llm", args[0], message_strings, *args[2:], **kwargs, ) else: logger.warning(f"Error in {event_name} callback: {e}") except Exception as e: logger.warning(f"Error in {event_name} callback: {e}") async def _ahandle_event( handlers: List[BaseCallbackHandler], event_name: str, ignore_condition_name: Optional[str], *args: Any, **kwargs: Any, ) -> None: """Generic event handler for AsyncCallbackManager.""" await asyncio.gather( *( _ahandle_event_for_handler( handler, event_name, ignore_condition_name, *args, **kwargs ) for handler in handlers ) ) BRM = TypeVar("BRM", bound="BaseRunManager") class BaseRunManager(RunManagerMixin): """Base class for run manager (a bound callback manager).""" def __init__( self, run_id: UUID, handlers: List[BaseCallbackHandler], inheritable_handlers: List[BaseCallbackHandler], parent_run_id: Optional[UUID] = None, ) -> None: """Initialize run manager.""" self.run_id = run_id self.handlers = handlers self.inheritable_handlers = inheritable_handlers self.parent_run_id = parent_run_id @classmethod def get_noop_manager(cls: Type[BRM]) -> BRM: """Return a manager that doesn't perform any operations.""" return cls(uuid4(), [], []) class RunManager(BaseRunManager): """Sync Run Manager.""" def on_text( self, text: str, **kwargs: Any, ) -> Any: """Run when text is received.""" _handle_event( self.handlers, "on_text", None, text, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncRunManager(BaseRunManager): """Async Run Manager.""" async def on_text( self, text: str, **kwargs: Any, ) -> Any: """Run when text is received.""" await _ahandle_event( self.handlers, "on_text", None, text, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManagerForLLMRun(RunManager, LLMManagerMixin): """Callback manager for LLM run.""" def on_llm_new_token( self, token: str, **kwargs: Any, ) -> None: """Run when LLM generates a new token.""" _handle_event( self.handlers, "on_llm_new_token", "ignore_llm", token=token, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" _handle_event( self.handlers, "on_llm_end", "ignore_llm", response, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_llm_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when LLM errors.""" _handle_event( self.handlers, "on_llm_error", "ignore_llm", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncCallbackManagerForLLMRun(AsyncRunManager, LLMManagerMixin): """Async callback manager for LLM run.""" async def on_llm_new_token( self, token: str, **kwargs: Any, ) -> None: """Run when LLM generates a new token.""" await _ahandle_event( self.handlers, "on_llm_new_token", "ignore_llm", token, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" await _ahandle_event( self.handlers, "on_llm_end", "ignore_llm", response, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_llm_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when LLM errors.""" await _ahandle_event( self.handlers, "on_llm_error", "ignore_llm", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManagerForChainRun(RunManager, ChainManagerMixin): """Callback manager for chain run.""" def get_child(self) -> CallbackManager: """Get a child callback manager.""" manager = CallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" _handle_event( self.handlers, "on_chain_end", "ignore_chain", outputs, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_chain_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when chain errors.""" _handle_event( self.handlers, "on_chain_error", "ignore_chain", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run when agent action is received.""" _handle_event( self.handlers, "on_agent_action", "ignore_agent", action, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> Any: """Run when agent finish is received.""" _handle_event( self.handlers, "on_agent_finish", "ignore_agent", finish, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncCallbackManagerForChainRun(AsyncRunManager, ChainManagerMixin): """Async callback manager for chain run.""" def get_child(self) -> AsyncCallbackManager: """Get a child callback manager.""" manager = AsyncCallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager async def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" await _ahandle_event( self.handlers, "on_chain_end", "ignore_chain", outputs, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_chain_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when chain errors.""" await _ahandle_event( self.handlers, "on_chain_error", "ignore_chain", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run when agent action is received.""" await _ahandle_event( self.handlers, "on_agent_action", "ignore_agent", action, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> Any: """Run when agent finish is received.""" await _ahandle_event( self.handlers, "on_agent_finish", "ignore_agent", finish, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManagerForToolRun(RunManager, ToolManagerMixin): """Callback manager for tool run.""" def get_child(self) -> CallbackManager: """Get a child callback manager.""" manager = CallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager def on_tool_end( self, output: str, **kwargs: Any, ) -> None: """Run when tool ends running.""" _handle_event( self.handlers, "on_tool_end", "ignore_agent", output, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_tool_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when tool errors.""" _handle_event( self.handlers, "on_tool_error", "ignore_agent", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncCallbackManagerForToolRun(AsyncRunManager, ToolManagerMixin): """Async callback manager for tool run.""" def get_child(self) -> AsyncCallbackManager: """Get a child callback manager.""" manager = AsyncCallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager async def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" await _ahandle_event( self.handlers, "on_tool_end", "ignore_agent", output, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_tool_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when tool errors.""" await _ahandle_event( self.handlers, "on_tool_error", "ignore_agent", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManager(BaseCallbackManager): """Callback manager that can be used to handle callbacks from langchain.""" def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForLLMRun: """Run when LLM starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_llm_start", "ignore_llm", serialized, prompts, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return CallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) def on_chat_model_start( self, serialized: Dict[str, Any], messages: List[List[BaseMessage]], run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForLLMRun: """Run when LLM starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_chat_model_start", "ignore_chat_model", serialized, messages, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) # Re-use the LLM Run Manager since the outputs are treated # the same for now return CallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForChainRun: """Run when chain starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_chain_start", "ignore_chain", serialized, inputs, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return CallbackManagerForChainRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) def on_tool_start( self, serialized: Dict[str, Any], input_str: str, run_id: Optional[UUID] = None, parent_run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForToolRun: """Run when tool starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_tool_start", "ignore_agent", serialized, input_str, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return CallbackManagerForToolRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) @classmethod def configure( cls, inheritable_callbacks: Callbacks = None, local_callbacks: Callbacks = None, verbose: bool = False, ) -> CallbackManager: """Configure the callback manager.""" return _configure(cls, inheritable_callbacks, local_callbacks, verbose) class AsyncCallbackManager(BaseCallbackManager): """Async callback manager that can be used to handle callbacks from LangChain.""" @property def is_async(self) -> bool: """Return whether the handler is async.""" return True async def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], run_id: Optional[UUID] = None, **kwargs: Any, ) -> AsyncCallbackManagerForLLMRun: """Run when LLM starts running.""" if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_llm_start", "ignore_llm", serialized, prompts, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) async def on_chat_model_start( self, serialized: Dict[str, Any], messages: List[List[BaseMessage]], run_id: Optional[UUID] = None, **kwargs: Any, ) -> Any: if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_chat_model_start", "ignore_chat_model", serialized, messages, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) async def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], run_id: Optional[UUID] = None, **kwargs: Any, ) -> AsyncCallbackManagerForChainRun: """Run when chain starts running.""" if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_chain_start", "ignore_chain", serialized, inputs, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForChainRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) async def on_tool_start( self, serialized: Dict[str, Any], input_str: str, run_id: Optional[UUID] = None, parent_run_id: Optional[UUID] = None, **kwargs: Any, ) -> AsyncCallbackManagerForToolRun: """Run when tool starts running.""" if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_tool_start", "ignore_agent", serialized, input_str, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForToolRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) @classmethod def configure( cls, inheritable_callbacks: Callbacks = None, local_callbacks: Callbacks = None, verbose: bool = False, ) -> AsyncCallbackManager: """Configure the callback manager.""" return _configure(cls, inheritable_callbacks, local_callbacks, verbose) T = TypeVar("T", CallbackManager, AsyncCallbackManager) def env_var_is_set(env_var: str) -> bool: """Check if an environment variable is set.""" return env_var in os.environ and os.environ[env_var] not in ( "", "0", "false", "False", ) def _configure( callback_manager_cls: Type[T], inheritable_callbacks: Callbacks = None, local_callbacks: Callbacks = None, verbose: bool = False, ) -> T: """Configure the callback manager.""" callback_manager = callback_manager_cls([]) if inheritable_callbacks or local_callbacks: if isinstance(inheritable_callbacks, list) or inheritable_callbacks is None: inheritable_callbacks_ = inheritable_callbacks or [] callback_manager = callback_manager_cls( handlers=inheritable_callbacks_.copy(), inheritable_handlers=inheritable_callbacks_.copy(), ) else: callback_manager = callback_manager_cls( handlers=inheritable_callbacks.handlers, inheritable_handlers=inheritable_callbacks.inheritable_handlers, parent_run_id=inheritable_callbacks.parent_run_id, ) local_handlers_ = ( local_callbacks if isinstance(local_callbacks, list) else (local_callbacks.handlers if local_callbacks else []) ) for handler in local_handlers_: callback_manager.add_handler(handler, False) tracer = tracing_callback_var.get() wandb_tracer = wandb_tracing_callback_var.get() open_ai = openai_callback_var.get() tracing_enabled_ = ( env_var_is_set("LANGCHAIN_TRACING") or tracer is not None or env_var_is_set("LANGCHAIN_HANDLER") ) wandb_tracing_enabled_ = ( env_var_is_set("LANGCHAIN_WANDB_TRACING") or wandb_tracer is not None ) tracer_v2 = tracing_v2_callback_var.get() tracing_v2_enabled_ = ( env_var_is_set("LANGCHAIN_TRACING_V2") or tracer_v2 is not None ) tracer_session = os.environ.get("LANGCHAIN_SESSION") debug = _get_debug() if tracer_session is None: tracer_session = "default" if ( verbose or debug or tracing_enabled_ or tracing_v2_enabled_ or wandb_tracing_enabled_ or open_ai is not None ): if verbose and not any( isinstance(handler, StdOutCallbackHandler) for handler in callback_manager.handlers ): if debug: pass else: callback_manager.add_handler(StdOutCallbackHandler(), False) if debug and not any( isinstance(handler, ConsoleCallbackHandler) for handler in callback_manager.handlers ): callback_manager.add_handler(ConsoleCallbackHandler(), True) if tracing_enabled_ and not any( isinstance(handler, LangChainTracerV1) for handler in callback_manager.handlers ): if tracer: callback_manager.add_handler(tracer, True) else: handler = LangChainTracerV1() handler.load_session(tracer_session) callback_manager.add_handler(handler, True) if wandb_tracing_enabled_ and not any( isinstance(handler, WandbTracer) for handler in callback_manager.handlers ): if wandb_tracer: callback_manager.add_handler(wandb_tracer, True) else: handler = WandbTracer() callback_manager.add_handler(handler, True) if tracing_v2_enabled_ and not any( isinstance(handler, LangChainTracer) for handler in callback_manager.handlers ): if tracer_v2: callback_manager.add_handler(tracer_v2, True) else: try: handler = LangChainTracer(session_name=tracer_session) callback_manager.add_handler(handler, True) except Exception as e: logger.warning( "Unable to load requested LangChainTracer." " To disable this warning," " unset the LANGCHAIN_TRACING_V2 environment variables.", e, ) if open_ai is not None and not any( isinstance(handler, OpenAICallbackHandler) for handler in callback_manager.handlers ): callback_manager.add_handler(open_ai, True) return callback_manager
[ "langchain.schema.get_buffer_string", "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.langchain_v1.LangChainTracerV1" ]
[((1114, 1141), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1131, 1141), False, 'import logging\n'), ((1286, 1329), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1296, 1329), False, 'from contextvars import ContextVar\n'), ((1406, 1450), 'contextvars.ContextVar', 'ContextVar', (['"""tracing_callback"""'], {'default': 'None'}), "('tracing_callback', default=None)\n", (1416, 1450), False, 'from contextvars import ContextVar\n'), ((1541, 1591), 'contextvars.ContextVar', 'ContextVar', (['"""tracing_wandb_callback"""'], {'default': 'None'}), "('tracing_wandb_callback', default=None)\n", (1551, 1591), False, 'from contextvars import ContextVar\n'), ((1684, 1731), 'contextvars.ContextVar', 'ContextVar', (['"""tracing_callback_v2"""'], {'default': 'None'}), "('tracing_callback_v2', default=None)\n", (1694, 1731), False, 'from contextvars import ContextVar\n'), ((7496, 7534), 'typing.TypeVar', 'TypeVar', (['"""BRM"""'], {'bound': '"""BaseRunManager"""'}), "('BRM', bound='BaseRunManager')\n", (7503, 7534), False, 'from typing import Any, AsyncGenerator, Dict, Generator, List, Optional, Type, TypeVar, Union, cast\n'), ((24415, 24466), 'typing.TypeVar', 'TypeVar', (['"""T"""', 'CallbackManager', 'AsyncCallbackManager'], {}), "('T', CallbackManager, AsyncCallbackManager)\n", (24422, 24466), False, 'from typing import Any, AsyncGenerator, Dict, Generator, List, Optional, Type, TypeVar, Union, cast\n'), ((1969, 1992), 'langchain.callbacks.openai_info.OpenAICallbackHandler', 'OpenAICallbackHandler', ([], {}), '()\n', (1990, 1992), False, 'from langchain.callbacks.openai_info import OpenAICallbackHandler\n'), ((2243, 2262), 'langchain.callbacks.tracers.langchain_v1.LangChainTracerV1', 'LangChainTracerV1', ([], {}), '()\n', (2260, 2262), False, 'from langchain.callbacks.tracers.langchain_v1 import LangChainTracerV1, TracerSessionV1\n'), ((2587, 2600), 'langchain.callbacks.tracers.wandb.WandbTracer', 'WandbTracer', ([], {}), '()\n', (2598, 2600), False, 'from langchain.callbacks.tracers.wandb import WandbTracer\n'), ((2990, 3107), 'warnings.warn', 'warnings.warn', (['"""The tracing v2 API is in development. This is not yet stable and may change in the future."""'], {}), "(\n 'The tracing v2 API is in development. This is not yet stable and may change in the future.'\n )\n", (3003, 3107), False, 'import warnings\n'), ((3206, 3271), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'example_id': 'example_id', 'session_name': 'session_name'}), '(example_id=example_id, session_name=session_name)\n', (3221, 3271), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((3669, 3734), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'session_name': 'session_name', 'example_id': 'example_id'}), '(session_name=session_name, example_id=example_id)\n', (3684, 3734), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((4272, 4337), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'session_name': 'session_name', 'example_id': 'example_id'}), '(session_name=session_name, example_id=example_id)\n', (4287, 4337), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((26488, 26523), 'os.environ.get', 'os.environ.get', (['"""LANGCHAIN_SESSION"""'], {}), "('LANGCHAIN_SESSION')\n", (26502, 26523), False, 'import os\n'), ((3180, 3196), 'uuid.UUID', 'UUID', (['example_id'], {}), '(example_id)\n', (3184, 3196), False, 'from uuid import UUID, uuid4\n'), ((6200, 6234), 'asyncio.iscoroutinefunction', 'asyncio.iscoroutinefunction', (['event'], {}), '(event)\n', (6227, 6234), False, 'import asyncio\n'), ((8213, 8220), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (8218, 8220), False, 'from uuid import UUID, uuid4\n'), ((17986, 17993), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (17991, 17993), False, 'from uuid import UUID, uuid4\n'), ((18693, 18700), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (18698, 18700), False, 'from uuid import UUID, uuid4\n'), ((19496, 19503), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (19501, 19503), False, 'from uuid import UUID, uuid4\n'), ((20231, 20238), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (20236, 20238), False, 'from uuid import UUID, uuid4\n'), ((21513, 21520), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (21518, 21520), False, 'from uuid import UUID, uuid4\n'), ((22174, 22181), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (22179, 22181), False, 'from uuid import UUID, uuid4\n'), ((22907, 22914), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (22912, 22914), False, 'from uuid import UUID, uuid4\n'), ((23665, 23672), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (23670, 23672), False, 'from uuid import UUID, uuid4\n'), ((5799, 5854), 'logging.warning', 'logging.warning', (['f"""Error in {event_name} callback: {e}"""'], {}), "(f'Error in {event_name} callback: {e}')\n", (5814, 5854), False, 'import logging\n'), ((27268, 27292), 'langchain.callbacks.tracers.stdout.ConsoleCallbackHandler', 'ConsoleCallbackHandler', ([], {}), '()\n', (27290, 27292), False, 'from langchain.callbacks.tracers.stdout import ConsoleCallbackHandler\n'), ((27582, 27601), 'langchain.callbacks.tracers.langchain_v1.LangChainTracerV1', 'LangChainTracerV1', ([], {}), '()\n', (27599, 27601), False, 'from langchain.callbacks.tracers.langchain_v1 import LangChainTracerV1, TracerSessionV1\n'), ((27997, 28010), 'langchain.callbacks.tracers.wandb.WandbTracer', 'WandbTracer', ([], {}), '()\n', (28008, 28010), False, 'from langchain.callbacks.tracers.wandb import WandbTracer\n'), ((6565, 6585), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['m'], {}), '(m)\n', (6582, 6585), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult, get_buffer_string\n'), ((27045, 27068), 'langchain.callbacks.stdout.StdOutCallbackHandler', 'StdOutCallbackHandler', ([], {}), '()\n', (27066, 27068), False, 'from langchain.callbacks.stdout import StdOutCallbackHandler\n'), ((28385, 28429), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'session_name': 'tracer_session'}), '(session_name=tracer_session)\n', (28400, 28429), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((6389, 6430), 'functools.partial', 'functools.partial', (['event', '*args'], {}), '(event, *args, **kwargs)\n', (6406, 6430), False, 'import functools\n'), ((5291, 5311), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['m'], {}), '(m)\n', (5308, 5311), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult, get_buffer_string\n'), ((6321, 6345), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (6343, 6345), False, 'import asyncio\n')]
from __future__ import annotations import asyncio import functools import logging import os import warnings from contextlib import asynccontextmanager, contextmanager from contextvars import ContextVar from typing import ( Any, AsyncGenerator, Dict, Generator, List, Optional, Type, TypeVar, Union, cast, ) from uuid import UUID, uuid4 import langchain from langchain.callbacks.base import ( BaseCallbackHandler, BaseCallbackManager, ChainManagerMixin, LLMManagerMixin, RunManagerMixin, ToolManagerMixin, ) from langchain.callbacks.openai_info import OpenAICallbackHandler from langchain.callbacks.stdout import StdOutCallbackHandler from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.callbacks.tracers.langchain_v1 import LangChainTracerV1, TracerSessionV1 from langchain.callbacks.tracers.stdout import ConsoleCallbackHandler from langchain.callbacks.tracers.wandb import WandbTracer from langchain.schema import ( AgentAction, AgentFinish, BaseMessage, LLMResult, get_buffer_string, ) logger = logging.getLogger(__name__) Callbacks = Optional[Union[List[BaseCallbackHandler], BaseCallbackManager]] openai_callback_var: ContextVar[Optional[OpenAICallbackHandler]] = ContextVar( "openai_callback", default=None ) tracing_callback_var: ContextVar[ Optional[LangChainTracerV1] ] = ContextVar( # noqa: E501 "tracing_callback", default=None ) wandb_tracing_callback_var: ContextVar[ Optional[WandbTracer] ] = ContextVar( # noqa: E501 "tracing_wandb_callback", default=None ) tracing_v2_callback_var: ContextVar[ Optional[LangChainTracer] ] = ContextVar( # noqa: E501 "tracing_callback_v2", default=None ) def _get_debug() -> bool: return langchain.debug @contextmanager def get_openai_callback() -> Generator[OpenAICallbackHandler, None, None]: """Get OpenAI callback handler in a context manager.""" cb = OpenAICallbackHandler() openai_callback_var.set(cb) yield cb openai_callback_var.set(None) @contextmanager def tracing_enabled( session_name: str = "default", ) -> Generator[TracerSessionV1, None, None]: """Get Tracer in a context manager.""" cb = LangChainTracerV1() session = cast(TracerSessionV1, cb.load_session(session_name)) tracing_callback_var.set(cb) yield session tracing_callback_var.set(None) @contextmanager def wandb_tracing_enabled( session_name: str = "default", ) -> Generator[None, None, None]: """Get WandbTracer in a context manager.""" cb = WandbTracer() wandb_tracing_callback_var.set(cb) yield None wandb_tracing_callback_var.set(None) @contextmanager def tracing_v2_enabled( session_name: Optional[str] = None, *, example_id: Optional[Union[str, UUID]] = None, ) -> Generator[None, None, None]: """Get the experimental tracer handler in a context manager.""" # Issue a warning that this is experimental warnings.warn( "The tracing v2 API is in development. " "This is not yet stable and may change in the future." ) if isinstance(example_id, str): example_id = UUID(example_id) cb = LangChainTracer( example_id=example_id, session_name=session_name, ) tracing_v2_callback_var.set(cb) yield tracing_v2_callback_var.set(None) @contextmanager def trace_as_chain_group( group_name: str, *, session_name: Optional[str] = None, example_id: Optional[Union[str, UUID]] = None, ) -> Generator[CallbackManager, None, None]: """Get a callback manager for a chain group in a context manager.""" cb = LangChainTracer( session_name=session_name, example_id=example_id, ) cm = CallbackManager.configure( inheritable_callbacks=[cb], ) run_manager = cm.on_chain_start({"name": group_name}, {}) yield run_manager.get_child() run_manager.on_chain_end({}) @asynccontextmanager async def atrace_as_chain_group( group_name: str, *, session_name: Optional[str] = None, example_id: Optional[Union[str, UUID]] = None, ) -> AsyncGenerator[AsyncCallbackManager, None]: """Get a callback manager for a chain group in a context manager.""" cb = LangChainTracer( session_name=session_name, example_id=example_id, ) cm = AsyncCallbackManager.configure( inheritable_callbacks=[cb], ) run_manager = await cm.on_chain_start({"name": group_name}, {}) try: yield run_manager.get_child() finally: await run_manager.on_chain_end({}) def _handle_event( handlers: List[BaseCallbackHandler], event_name: str, ignore_condition_name: Optional[str], *args: Any, **kwargs: Any, ) -> None: """Generic event handler for CallbackManager.""" message_strings: Optional[List[str]] = None for handler in handlers: try: if ignore_condition_name is None or not getattr( handler, ignore_condition_name ): getattr(handler, event_name)(*args, **kwargs) except NotImplementedError as e: if event_name == "on_chat_model_start": if message_strings is None: message_strings = [get_buffer_string(m) for m in args[1]] _handle_event( [handler], "on_llm_start", "ignore_llm", args[0], message_strings, *args[2:], **kwargs, ) else: logger.warning(f"Error in {event_name} callback: {e}") except Exception as e: if handler.raise_error: raise e logging.warning(f"Error in {event_name} callback: {e}") async def _ahandle_event_for_handler( handler: BaseCallbackHandler, event_name: str, ignore_condition_name: Optional[str], *args: Any, **kwargs: Any, ) -> None: try: if ignore_condition_name is None or not getattr(handler, ignore_condition_name): event = getattr(handler, event_name) if asyncio.iscoroutinefunction(event): await event(*args, **kwargs) else: await asyncio.get_event_loop().run_in_executor( None, functools.partial(event, *args, **kwargs) ) except NotImplementedError as e: if event_name == "on_chat_model_start": message_strings = [get_buffer_string(m) for m in args[1]] await _ahandle_event_for_handler( handler, "on_llm_start", "ignore_llm", args[0], message_strings, *args[2:], **kwargs, ) else: logger.warning(f"Error in {event_name} callback: {e}") except Exception as e: logger.warning(f"Error in {event_name} callback: {e}") async def _ahandle_event( handlers: List[BaseCallbackHandler], event_name: str, ignore_condition_name: Optional[str], *args: Any, **kwargs: Any, ) -> None: """Generic event handler for AsyncCallbackManager.""" await asyncio.gather( *( _ahandle_event_for_handler( handler, event_name, ignore_condition_name, *args, **kwargs ) for handler in handlers ) ) BRM = TypeVar("BRM", bound="BaseRunManager") class BaseRunManager(RunManagerMixin): """Base class for run manager (a bound callback manager).""" def __init__( self, run_id: UUID, handlers: List[BaseCallbackHandler], inheritable_handlers: List[BaseCallbackHandler], parent_run_id: Optional[UUID] = None, ) -> None: """Initialize run manager.""" self.run_id = run_id self.handlers = handlers self.inheritable_handlers = inheritable_handlers self.parent_run_id = parent_run_id @classmethod def get_noop_manager(cls: Type[BRM]) -> BRM: """Return a manager that doesn't perform any operations.""" return cls(uuid4(), [], []) class RunManager(BaseRunManager): """Sync Run Manager.""" def on_text( self, text: str, **kwargs: Any, ) -> Any: """Run when text is received.""" _handle_event( self.handlers, "on_text", None, text, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncRunManager(BaseRunManager): """Async Run Manager.""" async def on_text( self, text: str, **kwargs: Any, ) -> Any: """Run when text is received.""" await _ahandle_event( self.handlers, "on_text", None, text, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManagerForLLMRun(RunManager, LLMManagerMixin): """Callback manager for LLM run.""" def on_llm_new_token( self, token: str, **kwargs: Any, ) -> None: """Run when LLM generates a new token.""" _handle_event( self.handlers, "on_llm_new_token", "ignore_llm", token=token, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" _handle_event( self.handlers, "on_llm_end", "ignore_llm", response, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_llm_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when LLM errors.""" _handle_event( self.handlers, "on_llm_error", "ignore_llm", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncCallbackManagerForLLMRun(AsyncRunManager, LLMManagerMixin): """Async callback manager for LLM run.""" async def on_llm_new_token( self, token: str, **kwargs: Any, ) -> None: """Run when LLM generates a new token.""" await _ahandle_event( self.handlers, "on_llm_new_token", "ignore_llm", token, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" await _ahandle_event( self.handlers, "on_llm_end", "ignore_llm", response, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_llm_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when LLM errors.""" await _ahandle_event( self.handlers, "on_llm_error", "ignore_llm", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManagerForChainRun(RunManager, ChainManagerMixin): """Callback manager for chain run.""" def get_child(self) -> CallbackManager: """Get a child callback manager.""" manager = CallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" _handle_event( self.handlers, "on_chain_end", "ignore_chain", outputs, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_chain_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when chain errors.""" _handle_event( self.handlers, "on_chain_error", "ignore_chain", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run when agent action is received.""" _handle_event( self.handlers, "on_agent_action", "ignore_agent", action, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> Any: """Run when agent finish is received.""" _handle_event( self.handlers, "on_agent_finish", "ignore_agent", finish, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncCallbackManagerForChainRun(AsyncRunManager, ChainManagerMixin): """Async callback manager for chain run.""" def get_child(self) -> AsyncCallbackManager: """Get a child callback manager.""" manager = AsyncCallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager async def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" await _ahandle_event( self.handlers, "on_chain_end", "ignore_chain", outputs, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_chain_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when chain errors.""" await _ahandle_event( self.handlers, "on_chain_error", "ignore_chain", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run when agent action is received.""" await _ahandle_event( self.handlers, "on_agent_action", "ignore_agent", action, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> Any: """Run when agent finish is received.""" await _ahandle_event( self.handlers, "on_agent_finish", "ignore_agent", finish, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManagerForToolRun(RunManager, ToolManagerMixin): """Callback manager for tool run.""" def get_child(self) -> CallbackManager: """Get a child callback manager.""" manager = CallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager def on_tool_end( self, output: str, **kwargs: Any, ) -> None: """Run when tool ends running.""" _handle_event( self.handlers, "on_tool_end", "ignore_agent", output, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) def on_tool_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when tool errors.""" _handle_event( self.handlers, "on_tool_error", "ignore_agent", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class AsyncCallbackManagerForToolRun(AsyncRunManager, ToolManagerMixin): """Async callback manager for tool run.""" def get_child(self) -> AsyncCallbackManager: """Get a child callback manager.""" manager = AsyncCallbackManager([], parent_run_id=self.run_id) manager.set_handlers(self.inheritable_handlers) return manager async def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" await _ahandle_event( self.handlers, "on_tool_end", "ignore_agent", output, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) async def on_tool_error( self, error: Union[Exception, KeyboardInterrupt], **kwargs: Any, ) -> None: """Run when tool errors.""" await _ahandle_event( self.handlers, "on_tool_error", "ignore_agent", error, run_id=self.run_id, parent_run_id=self.parent_run_id, **kwargs, ) class CallbackManager(BaseCallbackManager): """Callback manager that can be used to handle callbacks from langchain.""" def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForLLMRun: """Run when LLM starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_llm_start", "ignore_llm", serialized, prompts, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return CallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) def on_chat_model_start( self, serialized: Dict[str, Any], messages: List[List[BaseMessage]], run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForLLMRun: """Run when LLM starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_chat_model_start", "ignore_chat_model", serialized, messages, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) # Re-use the LLM Run Manager since the outputs are treated # the same for now return CallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForChainRun: """Run when chain starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_chain_start", "ignore_chain", serialized, inputs, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return CallbackManagerForChainRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) def on_tool_start( self, serialized: Dict[str, Any], input_str: str, run_id: Optional[UUID] = None, parent_run_id: Optional[UUID] = None, **kwargs: Any, ) -> CallbackManagerForToolRun: """Run when tool starts running.""" if run_id is None: run_id = uuid4() _handle_event( self.handlers, "on_tool_start", "ignore_agent", serialized, input_str, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return CallbackManagerForToolRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) @classmethod def configure( cls, inheritable_callbacks: Callbacks = None, local_callbacks: Callbacks = None, verbose: bool = False, ) -> CallbackManager: """Configure the callback manager.""" return _configure(cls, inheritable_callbacks, local_callbacks, verbose) class AsyncCallbackManager(BaseCallbackManager): """Async callback manager that can be used to handle callbacks from LangChain.""" @property def is_async(self) -> bool: """Return whether the handler is async.""" return True async def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], run_id: Optional[UUID] = None, **kwargs: Any, ) -> AsyncCallbackManagerForLLMRun: """Run when LLM starts running.""" if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_llm_start", "ignore_llm", serialized, prompts, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) async def on_chat_model_start( self, serialized: Dict[str, Any], messages: List[List[BaseMessage]], run_id: Optional[UUID] = None, **kwargs: Any, ) -> Any: if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_chat_model_start", "ignore_chat_model", serialized, messages, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForLLMRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) async def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], run_id: Optional[UUID] = None, **kwargs: Any, ) -> AsyncCallbackManagerForChainRun: """Run when chain starts running.""" if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_chain_start", "ignore_chain", serialized, inputs, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForChainRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) async def on_tool_start( self, serialized: Dict[str, Any], input_str: str, run_id: Optional[UUID] = None, parent_run_id: Optional[UUID] = None, **kwargs: Any, ) -> AsyncCallbackManagerForToolRun: """Run when tool starts running.""" if run_id is None: run_id = uuid4() await _ahandle_event( self.handlers, "on_tool_start", "ignore_agent", serialized, input_str, run_id=run_id, parent_run_id=self.parent_run_id, **kwargs, ) return AsyncCallbackManagerForToolRun( run_id, self.handlers, self.inheritable_handlers, self.parent_run_id ) @classmethod def configure( cls, inheritable_callbacks: Callbacks = None, local_callbacks: Callbacks = None, verbose: bool = False, ) -> AsyncCallbackManager: """Configure the callback manager.""" return _configure(cls, inheritable_callbacks, local_callbacks, verbose) T = TypeVar("T", CallbackManager, AsyncCallbackManager) def env_var_is_set(env_var: str) -> bool: """Check if an environment variable is set.""" return env_var in os.environ and os.environ[env_var] not in ( "", "0", "false", "False", ) def _configure( callback_manager_cls: Type[T], inheritable_callbacks: Callbacks = None, local_callbacks: Callbacks = None, verbose: bool = False, ) -> T: """Configure the callback manager.""" callback_manager = callback_manager_cls([]) if inheritable_callbacks or local_callbacks: if isinstance(inheritable_callbacks, list) or inheritable_callbacks is None: inheritable_callbacks_ = inheritable_callbacks or [] callback_manager = callback_manager_cls( handlers=inheritable_callbacks_.copy(), inheritable_handlers=inheritable_callbacks_.copy(), ) else: callback_manager = callback_manager_cls( handlers=inheritable_callbacks.handlers, inheritable_handlers=inheritable_callbacks.inheritable_handlers, parent_run_id=inheritable_callbacks.parent_run_id, ) local_handlers_ = ( local_callbacks if isinstance(local_callbacks, list) else (local_callbacks.handlers if local_callbacks else []) ) for handler in local_handlers_: callback_manager.add_handler(handler, False) tracer = tracing_callback_var.get() wandb_tracer = wandb_tracing_callback_var.get() open_ai = openai_callback_var.get() tracing_enabled_ = ( env_var_is_set("LANGCHAIN_TRACING") or tracer is not None or env_var_is_set("LANGCHAIN_HANDLER") ) wandb_tracing_enabled_ = ( env_var_is_set("LANGCHAIN_WANDB_TRACING") or wandb_tracer is not None ) tracer_v2 = tracing_v2_callback_var.get() tracing_v2_enabled_ = ( env_var_is_set("LANGCHAIN_TRACING_V2") or tracer_v2 is not None ) tracer_session = os.environ.get("LANGCHAIN_SESSION") debug = _get_debug() if tracer_session is None: tracer_session = "default" if ( verbose or debug or tracing_enabled_ or tracing_v2_enabled_ or wandb_tracing_enabled_ or open_ai is not None ): if verbose and not any( isinstance(handler, StdOutCallbackHandler) for handler in callback_manager.handlers ): if debug: pass else: callback_manager.add_handler(StdOutCallbackHandler(), False) if debug and not any( isinstance(handler, ConsoleCallbackHandler) for handler in callback_manager.handlers ): callback_manager.add_handler(ConsoleCallbackHandler(), True) if tracing_enabled_ and not any( isinstance(handler, LangChainTracerV1) for handler in callback_manager.handlers ): if tracer: callback_manager.add_handler(tracer, True) else: handler = LangChainTracerV1() handler.load_session(tracer_session) callback_manager.add_handler(handler, True) if wandb_tracing_enabled_ and not any( isinstance(handler, WandbTracer) for handler in callback_manager.handlers ): if wandb_tracer: callback_manager.add_handler(wandb_tracer, True) else: handler = WandbTracer() callback_manager.add_handler(handler, True) if tracing_v2_enabled_ and not any( isinstance(handler, LangChainTracer) for handler in callback_manager.handlers ): if tracer_v2: callback_manager.add_handler(tracer_v2, True) else: try: handler = LangChainTracer(session_name=tracer_session) callback_manager.add_handler(handler, True) except Exception as e: logger.warning( "Unable to load requested LangChainTracer." " To disable this warning," " unset the LANGCHAIN_TRACING_V2 environment variables.", e, ) if open_ai is not None and not any( isinstance(handler, OpenAICallbackHandler) for handler in callback_manager.handlers ): callback_manager.add_handler(open_ai, True) return callback_manager
[ "langchain.schema.get_buffer_string", "langchain.callbacks.stdout.StdOutCallbackHandler", "langchain.callbacks.tracers.wandb.WandbTracer", "langchain.callbacks.openai_info.OpenAICallbackHandler", "langchain.callbacks.tracers.stdout.ConsoleCallbackHandler", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.langchain_v1.LangChainTracerV1" ]
[((1114, 1141), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1131, 1141), False, 'import logging\n'), ((1286, 1329), 'contextvars.ContextVar', 'ContextVar', (['"""openai_callback"""'], {'default': 'None'}), "('openai_callback', default=None)\n", (1296, 1329), False, 'from contextvars import ContextVar\n'), ((1406, 1450), 'contextvars.ContextVar', 'ContextVar', (['"""tracing_callback"""'], {'default': 'None'}), "('tracing_callback', default=None)\n", (1416, 1450), False, 'from contextvars import ContextVar\n'), ((1541, 1591), 'contextvars.ContextVar', 'ContextVar', (['"""tracing_wandb_callback"""'], {'default': 'None'}), "('tracing_wandb_callback', default=None)\n", (1551, 1591), False, 'from contextvars import ContextVar\n'), ((1684, 1731), 'contextvars.ContextVar', 'ContextVar', (['"""tracing_callback_v2"""'], {'default': 'None'}), "('tracing_callback_v2', default=None)\n", (1694, 1731), False, 'from contextvars import ContextVar\n'), ((7496, 7534), 'typing.TypeVar', 'TypeVar', (['"""BRM"""'], {'bound': '"""BaseRunManager"""'}), "('BRM', bound='BaseRunManager')\n", (7503, 7534), False, 'from typing import Any, AsyncGenerator, Dict, Generator, List, Optional, Type, TypeVar, Union, cast\n'), ((24415, 24466), 'typing.TypeVar', 'TypeVar', (['"""T"""', 'CallbackManager', 'AsyncCallbackManager'], {}), "('T', CallbackManager, AsyncCallbackManager)\n", (24422, 24466), False, 'from typing import Any, AsyncGenerator, Dict, Generator, List, Optional, Type, TypeVar, Union, cast\n'), ((1969, 1992), 'langchain.callbacks.openai_info.OpenAICallbackHandler', 'OpenAICallbackHandler', ([], {}), '()\n', (1990, 1992), False, 'from langchain.callbacks.openai_info import OpenAICallbackHandler\n'), ((2243, 2262), 'langchain.callbacks.tracers.langchain_v1.LangChainTracerV1', 'LangChainTracerV1', ([], {}), '()\n', (2260, 2262), False, 'from langchain.callbacks.tracers.langchain_v1 import LangChainTracerV1, TracerSessionV1\n'), ((2587, 2600), 'langchain.callbacks.tracers.wandb.WandbTracer', 'WandbTracer', ([], {}), '()\n', (2598, 2600), False, 'from langchain.callbacks.tracers.wandb import WandbTracer\n'), ((2990, 3107), 'warnings.warn', 'warnings.warn', (['"""The tracing v2 API is in development. This is not yet stable and may change in the future."""'], {}), "(\n 'The tracing v2 API is in development. This is not yet stable and may change in the future.'\n )\n", (3003, 3107), False, 'import warnings\n'), ((3206, 3271), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'example_id': 'example_id', 'session_name': 'session_name'}), '(example_id=example_id, session_name=session_name)\n', (3221, 3271), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((3669, 3734), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'session_name': 'session_name', 'example_id': 'example_id'}), '(session_name=session_name, example_id=example_id)\n', (3684, 3734), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((4272, 4337), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'session_name': 'session_name', 'example_id': 'example_id'}), '(session_name=session_name, example_id=example_id)\n', (4287, 4337), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((26488, 26523), 'os.environ.get', 'os.environ.get', (['"""LANGCHAIN_SESSION"""'], {}), "('LANGCHAIN_SESSION')\n", (26502, 26523), False, 'import os\n'), ((3180, 3196), 'uuid.UUID', 'UUID', (['example_id'], {}), '(example_id)\n', (3184, 3196), False, 'from uuid import UUID, uuid4\n'), ((6200, 6234), 'asyncio.iscoroutinefunction', 'asyncio.iscoroutinefunction', (['event'], {}), '(event)\n', (6227, 6234), False, 'import asyncio\n'), ((8213, 8220), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (8218, 8220), False, 'from uuid import UUID, uuid4\n'), ((17986, 17993), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (17991, 17993), False, 'from uuid import UUID, uuid4\n'), ((18693, 18700), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (18698, 18700), False, 'from uuid import UUID, uuid4\n'), ((19496, 19503), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (19501, 19503), False, 'from uuid import UUID, uuid4\n'), ((20231, 20238), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (20236, 20238), False, 'from uuid import UUID, uuid4\n'), ((21513, 21520), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (21518, 21520), False, 'from uuid import UUID, uuid4\n'), ((22174, 22181), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (22179, 22181), False, 'from uuid import UUID, uuid4\n'), ((22907, 22914), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (22912, 22914), False, 'from uuid import UUID, uuid4\n'), ((23665, 23672), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (23670, 23672), False, 'from uuid import UUID, uuid4\n'), ((5799, 5854), 'logging.warning', 'logging.warning', (['f"""Error in {event_name} callback: {e}"""'], {}), "(f'Error in {event_name} callback: {e}')\n", (5814, 5854), False, 'import logging\n'), ((27268, 27292), 'langchain.callbacks.tracers.stdout.ConsoleCallbackHandler', 'ConsoleCallbackHandler', ([], {}), '()\n', (27290, 27292), False, 'from langchain.callbacks.tracers.stdout import ConsoleCallbackHandler\n'), ((27582, 27601), 'langchain.callbacks.tracers.langchain_v1.LangChainTracerV1', 'LangChainTracerV1', ([], {}), '()\n', (27599, 27601), False, 'from langchain.callbacks.tracers.langchain_v1 import LangChainTracerV1, TracerSessionV1\n'), ((27997, 28010), 'langchain.callbacks.tracers.wandb.WandbTracer', 'WandbTracer', ([], {}), '()\n', (28008, 28010), False, 'from langchain.callbacks.tracers.wandb import WandbTracer\n'), ((6565, 6585), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['m'], {}), '(m)\n', (6582, 6585), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult, get_buffer_string\n'), ((27045, 27068), 'langchain.callbacks.stdout.StdOutCallbackHandler', 'StdOutCallbackHandler', ([], {}), '()\n', (27066, 27068), False, 'from langchain.callbacks.stdout import StdOutCallbackHandler\n'), ((28385, 28429), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'session_name': 'tracer_session'}), '(session_name=tracer_session)\n', (28400, 28429), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((6389, 6430), 'functools.partial', 'functools.partial', (['event', '*args'], {}), '(event, *args, **kwargs)\n', (6406, 6430), False, 'import functools\n'), ((5291, 5311), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['m'], {}), '(m)\n', (5308, 5311), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult, get_buffer_string\n'), ((6321, 6345), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (6343, 6345), False, 'import asyncio\n')]
"""A tracer that runs evaluators over completed runs.""" from __future__ import annotations import logging from concurrent.futures import Future, ThreadPoolExecutor from typing import Any, Dict, List, Optional, Sequence, Set, Union from uuid import UUID import langsmith from langsmith import schemas as langsmith_schemas from langchain.callbacks import manager from langchain.callbacks.tracers import langchain as langchain_tracer from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.schemas import Run logger = logging.getLogger(__name__) class EvaluatorCallbackHandler(BaseTracer): """A tracer that runs a run evaluator whenever a run is persisted. Parameters ---------- evaluators : Sequence[RunEvaluator] The run evaluators to apply to all top level runs. max_workers : int, optional The maximum number of worker threads to use for running the evaluators. If not specified, it will default to the number of evaluators. client : LangSmith Client, optional The LangSmith client instance to use for evaluating the runs. If not specified, a new instance will be created. example_id : Union[UUID, str], optional The example ID to be associated with the runs. project_name : str, optional The LangSmith project name to be organize eval chain runs under. Attributes ---------- example_id : Union[UUID, None] The example ID associated with the runs. client : Client The LangSmith client instance used for evaluating the runs. evaluators : Sequence[RunEvaluator] The sequence of run evaluators to be executed. executor : ThreadPoolExecutor The thread pool executor used for running the evaluators. futures : Set[Future] The set of futures representing the running evaluators. skip_unfinished : bool Whether to skip runs that are not finished or raised an error. project_name : Optional[str] The LangSmith project name to be organize eval chain runs under. """ name = "evaluator_callback_handler" def __init__( self, evaluators: Sequence[langsmith.RunEvaluator], max_workers: Optional[int] = None, client: Optional[langsmith.Client] = None, example_id: Optional[Union[UUID, str]] = None, skip_unfinished: bool = True, project_name: Optional[str] = "evaluators", **kwargs: Any, ) -> None: super().__init__(**kwargs) self.example_id = ( UUID(example_id) if isinstance(example_id, str) else example_id ) self.client = client or langchain_tracer.get_client() self.evaluators = evaluators self.max_workers = max_workers or len(evaluators) self.futures: Set[Future] = set() self.skip_unfinished = skip_unfinished self.project_name = project_name self.logged_feedback: Dict[str, List[langsmith_schemas.Feedback]] = {} def _evaluate_in_project(self, run: Run, evaluator: langsmith.RunEvaluator) -> None: """Evaluate the run in the project. Parameters ---------- run : Run The run to be evaluated. evaluator : RunEvaluator The evaluator to use for evaluating the run. """ try: if self.project_name is None: feedback = self.client.evaluate_run(run, evaluator) with manager.tracing_v2_enabled( project_name=self.project_name, tags=["eval"], client=self.client ): feedback = self.client.evaluate_run(run, evaluator) except Exception as e: logger.error( f"Error evaluating run {run.id} with " f"{evaluator.__class__.__name__}: {e}", exc_info=True, ) raise e example_id = str(run.reference_example_id) self.logged_feedback.setdefault(example_id, []).append(feedback) def _persist_run(self, run: Run) -> None: """Run the evaluator on the run. Parameters ---------- run : Run The run to be evaluated. """ if self.skip_unfinished and not run.outputs: logger.debug(f"Skipping unfinished run {run.id}") return run_ = run.copy() run_.reference_example_id = self.example_id if self.max_workers > 0: with ThreadPoolExecutor(max_workers=self.max_workers) as executor: list( executor.map( self._evaluate_in_project, [run_ for _ in range(len(self.evaluators))], self.evaluators, ) ) else: for evaluator in self.evaluators: self._evaluate_in_project(run_, evaluator)
[ "langchain.callbacks.tracers.langchain.get_client", "langchain.callbacks.manager.tracing_v2_enabled" ]
[((553, 580), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (570, 580), False, 'import logging\n'), ((2572, 2588), 'uuid.UUID', 'UUID', (['example_id'], {}), '(example_id)\n', (2576, 2588), False, 'from uuid import UUID\n'), ((2678, 2707), 'langchain.callbacks.tracers.langchain.get_client', 'langchain_tracer.get_client', ([], {}), '()\n', (2705, 2707), True, 'from langchain.callbacks.tracers import langchain as langchain_tracer\n'), ((3483, 3580), 'langchain.callbacks.manager.tracing_v2_enabled', 'manager.tracing_v2_enabled', ([], {'project_name': 'self.project_name', 'tags': "['eval']", 'client': 'self.client'}), "(project_name=self.project_name, tags=['eval'],\n client=self.client)\n", (3509, 3580), False, 'from langchain.callbacks import manager\n'), ((4490, 4538), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': 'self.max_workers'}), '(max_workers=self.max_workers)\n', (4508, 4538), False, 'from concurrent.futures import Future, ThreadPoolExecutor\n')]
"""A tracer that runs evaluators over completed runs.""" from __future__ import annotations import logging import threading import weakref from concurrent.futures import Future, ThreadPoolExecutor, wait from typing import Any, Dict, List, Optional, Sequence, Tuple, Union, cast from uuid import UUID import langsmith from langsmith.evaluation.evaluator import EvaluationResult from langchain.callbacks import manager from langchain.callbacks.tracers import langchain as langchain_tracer from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.langchain import _get_executor from langchain.callbacks.tracers.schemas import Run logger = logging.getLogger(__name__) _TRACERS: weakref.WeakSet[EvaluatorCallbackHandler] = weakref.WeakSet() def wait_for_all_evaluators() -> None: """Wait for all tracers to finish.""" global _TRACERS for tracer in list(_TRACERS): if tracer is not None: tracer.wait_for_futures() class EvaluatorCallbackHandler(BaseTracer): """A tracer that runs a run evaluator whenever a run is persisted. Parameters ---------- evaluators : Sequence[RunEvaluator] The run evaluators to apply to all top level runs. client : LangSmith Client, optional The LangSmith client instance to use for evaluating the runs. If not specified, a new instance will be created. example_id : Union[UUID, str], optional The example ID to be associated with the runs. project_name : str, optional The LangSmith project name to be organize eval chain runs under. Attributes ---------- example_id : Union[UUID, None] The example ID associated with the runs. client : Client The LangSmith client instance used for evaluating the runs. evaluators : Sequence[RunEvaluator] The sequence of run evaluators to be executed. executor : ThreadPoolExecutor The thread pool executor used for running the evaluators. futures : Set[Future] The set of futures representing the running evaluators. skip_unfinished : bool Whether to skip runs that are not finished or raised an error. project_name : Optional[str] The LangSmith project name to be organize eval chain runs under. """ name = "evaluator_callback_handler" def __init__( self, evaluators: Sequence[langsmith.RunEvaluator], client: Optional[langsmith.Client] = None, example_id: Optional[Union[UUID, str]] = None, skip_unfinished: bool = True, project_name: Optional[str] = "evaluators", max_concurrency: Optional[int] = None, **kwargs: Any, ) -> None: super().__init__(**kwargs) self.example_id = ( UUID(example_id) if isinstance(example_id, str) else example_id ) self.client = client or langchain_tracer.get_client() self.evaluators = evaluators if max_concurrency is None: self.executor: Optional[ThreadPoolExecutor] = _get_executor() elif max_concurrency > 0: self.executor = ThreadPoolExecutor(max_workers=max_concurrency) weakref.finalize( self, lambda: cast(ThreadPoolExecutor, self.executor).shutdown(wait=True), ) else: self.executor = None self.futures: weakref.WeakSet[Future] = weakref.WeakSet() self.skip_unfinished = skip_unfinished self.project_name = project_name self.logged_eval_results: Dict[Tuple[str, str], List[EvaluationResult]] = {} self.lock = threading.Lock() global _TRACERS _TRACERS.add(self) def _evaluate_in_project(self, run: Run, evaluator: langsmith.RunEvaluator) -> None: """Evaluate the run in the project. Parameters ---------- run : Run The run to be evaluated. evaluator : RunEvaluator The evaluator to use for evaluating the run. """ try: if self.project_name is None: eval_result = self.client.evaluate_run(run, evaluator) with manager.tracing_v2_enabled( project_name=self.project_name, tags=["eval"], client=self.client ) as cb: reference_example = ( self.client.read_example(run.reference_example_id) if run.reference_example_id else None ) evaluation_result = evaluator.evaluate_run( run, example=reference_example, ) run_id = cb.latest_run.id if cb.latest_run is not None else None self.client.create_feedback( run.id, evaluation_result.key, score=evaluation_result.score, value=evaluation_result.value, comment=evaluation_result.comment, correction=evaluation_result.correction, source_info=evaluation_result.evaluator_info, source_run_id=evaluation_result.source_run_id or run_id, feedback_source_type=langsmith.schemas.FeedbackSourceType.MODEL, ) except Exception as e: logger.error( f"Error evaluating run {run.id} with " f"{evaluator.__class__.__name__}: {repr(e)}", exc_info=True, ) raise e example_id = str(run.reference_example_id) with self.lock: self.logged_eval_results.setdefault((str(run.id), example_id), []).append( eval_result ) def _persist_run(self, run: Run) -> None: """Run the evaluator on the run. Parameters ---------- run : Run The run to be evaluated. """ if self.skip_unfinished and not run.outputs: logger.debug(f"Skipping unfinished run {run.id}") return run_ = run.copy() run_.reference_example_id = self.example_id for evaluator in self.evaluators: if self.executor is None: self._evaluate_in_project(run_, evaluator) else: self.futures.add( self.executor.submit(self._evaluate_in_project, run_, evaluator) ) def wait_for_futures(self) -> None: """Wait for all futures to complete.""" wait(self.futures)
[ "langchain.callbacks.tracers.langchain._get_executor", "langchain.callbacks.tracers.langchain.get_client", "langchain.callbacks.manager.tracing_v2_enabled" ]
[((672, 699), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (689, 699), False, 'import logging\n'), ((755, 772), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (770, 772), False, 'import weakref\n'), ((3430, 3447), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (3445, 3447), False, 'import weakref\n'), ((3641, 3657), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (3655, 3657), False, 'import threading\n'), ((6571, 6589), 'concurrent.futures.wait', 'wait', (['self.futures'], {}), '(self.futures)\n', (6575, 6589), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((2791, 2807), 'uuid.UUID', 'UUID', (['example_id'], {}), '(example_id)\n', (2795, 2807), False, 'from uuid import UUID\n'), ((2897, 2926), 'langchain.callbacks.tracers.langchain.get_client', 'langchain_tracer.get_client', ([], {}), '()\n', (2924, 2926), True, 'from langchain.callbacks.tracers import langchain as langchain_tracer\n'), ((3058, 3073), 'langchain.callbacks.tracers.langchain._get_executor', '_get_executor', ([], {}), '()\n', (3071, 3073), False, 'from langchain.callbacks.tracers.langchain import _get_executor\n'), ((3136, 3183), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': 'max_concurrency'}), '(max_workers=max_concurrency)\n', (3154, 3183), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((4183, 4280), 'langchain.callbacks.manager.tracing_v2_enabled', 'manager.tracing_v2_enabled', ([], {'project_name': 'self.project_name', 'tags': "['eval']", 'client': 'self.client'}), "(project_name=self.project_name, tags=['eval'],\n client=self.client)\n", (4209, 4280), False, 'from langchain.callbacks import manager\n'), ((3260, 3299), 'typing.cast', 'cast', (['ThreadPoolExecutor', 'self.executor'], {}), '(ThreadPoolExecutor, self.executor)\n', (3264, 3299), False, 'from typing import Any, Dict, List, Optional, Sequence, Tuple, Union, cast\n')]
"""A tracer that runs evaluators over completed runs.""" from __future__ import annotations import logging import threading import weakref from concurrent.futures import Future, ThreadPoolExecutor, wait from typing import Any, Dict, List, Optional, Sequence, Tuple, Union, cast from uuid import UUID import langsmith from langsmith.evaluation.evaluator import EvaluationResult from langchain.callbacks import manager from langchain.callbacks.tracers import langchain as langchain_tracer from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.langchain import _get_executor from langchain.callbacks.tracers.schemas import Run logger = logging.getLogger(__name__) _TRACERS: weakref.WeakSet[EvaluatorCallbackHandler] = weakref.WeakSet() def wait_for_all_evaluators() -> None: """Wait for all tracers to finish.""" global _TRACERS for tracer in list(_TRACERS): if tracer is not None: tracer.wait_for_futures() class EvaluatorCallbackHandler(BaseTracer): """A tracer that runs a run evaluator whenever a run is persisted. Parameters ---------- evaluators : Sequence[RunEvaluator] The run evaluators to apply to all top level runs. client : LangSmith Client, optional The LangSmith client instance to use for evaluating the runs. If not specified, a new instance will be created. example_id : Union[UUID, str], optional The example ID to be associated with the runs. project_name : str, optional The LangSmith project name to be organize eval chain runs under. Attributes ---------- example_id : Union[UUID, None] The example ID associated with the runs. client : Client The LangSmith client instance used for evaluating the runs. evaluators : Sequence[RunEvaluator] The sequence of run evaluators to be executed. executor : ThreadPoolExecutor The thread pool executor used for running the evaluators. futures : Set[Future] The set of futures representing the running evaluators. skip_unfinished : bool Whether to skip runs that are not finished or raised an error. project_name : Optional[str] The LangSmith project name to be organize eval chain runs under. """ name = "evaluator_callback_handler" def __init__( self, evaluators: Sequence[langsmith.RunEvaluator], client: Optional[langsmith.Client] = None, example_id: Optional[Union[UUID, str]] = None, skip_unfinished: bool = True, project_name: Optional[str] = "evaluators", max_concurrency: Optional[int] = None, **kwargs: Any, ) -> None: super().__init__(**kwargs) self.example_id = ( UUID(example_id) if isinstance(example_id, str) else example_id ) self.client = client or langchain_tracer.get_client() self.evaluators = evaluators if max_concurrency is None: self.executor: Optional[ThreadPoolExecutor] = _get_executor() elif max_concurrency > 0: self.executor = ThreadPoolExecutor(max_workers=max_concurrency) weakref.finalize( self, lambda: cast(ThreadPoolExecutor, self.executor).shutdown(wait=True), ) else: self.executor = None self.futures: weakref.WeakSet[Future] = weakref.WeakSet() self.skip_unfinished = skip_unfinished self.project_name = project_name self.logged_eval_results: Dict[Tuple[str, str], List[EvaluationResult]] = {} self.lock = threading.Lock() global _TRACERS _TRACERS.add(self) def _evaluate_in_project(self, run: Run, evaluator: langsmith.RunEvaluator) -> None: """Evaluate the run in the project. Parameters ---------- run : Run The run to be evaluated. evaluator : RunEvaluator The evaluator to use for evaluating the run. """ try: if self.project_name is None: eval_result = self.client.evaluate_run(run, evaluator) with manager.tracing_v2_enabled( project_name=self.project_name, tags=["eval"], client=self.client ) as cb: reference_example = ( self.client.read_example(run.reference_example_id) if run.reference_example_id else None ) evaluation_result = evaluator.evaluate_run( run, example=reference_example, ) run_id = cb.latest_run.id if cb.latest_run is not None else None self.client.create_feedback( run.id, evaluation_result.key, score=evaluation_result.score, value=evaluation_result.value, comment=evaluation_result.comment, correction=evaluation_result.correction, source_info=evaluation_result.evaluator_info, source_run_id=evaluation_result.source_run_id or run_id, feedback_source_type=langsmith.schemas.FeedbackSourceType.MODEL, ) except Exception as e: logger.error( f"Error evaluating run {run.id} with " f"{evaluator.__class__.__name__}: {repr(e)}", exc_info=True, ) raise e example_id = str(run.reference_example_id) with self.lock: self.logged_eval_results.setdefault((str(run.id), example_id), []).append( eval_result ) def _persist_run(self, run: Run) -> None: """Run the evaluator on the run. Parameters ---------- run : Run The run to be evaluated. """ if self.skip_unfinished and not run.outputs: logger.debug(f"Skipping unfinished run {run.id}") return run_ = run.copy() run_.reference_example_id = self.example_id for evaluator in self.evaluators: if self.executor is None: self._evaluate_in_project(run_, evaluator) else: self.futures.add( self.executor.submit(self._evaluate_in_project, run_, evaluator) ) def wait_for_futures(self) -> None: """Wait for all futures to complete.""" wait(self.futures)
[ "langchain.callbacks.tracers.langchain._get_executor", "langchain.callbacks.tracers.langchain.get_client", "langchain.callbacks.manager.tracing_v2_enabled" ]
[((672, 699), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (689, 699), False, 'import logging\n'), ((755, 772), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (770, 772), False, 'import weakref\n'), ((3430, 3447), 'weakref.WeakSet', 'weakref.WeakSet', ([], {}), '()\n', (3445, 3447), False, 'import weakref\n'), ((3641, 3657), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (3655, 3657), False, 'import threading\n'), ((6571, 6589), 'concurrent.futures.wait', 'wait', (['self.futures'], {}), '(self.futures)\n', (6575, 6589), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((2791, 2807), 'uuid.UUID', 'UUID', (['example_id'], {}), '(example_id)\n', (2795, 2807), False, 'from uuid import UUID\n'), ((2897, 2926), 'langchain.callbacks.tracers.langchain.get_client', 'langchain_tracer.get_client', ([], {}), '()\n', (2924, 2926), True, 'from langchain.callbacks.tracers import langchain as langchain_tracer\n'), ((3058, 3073), 'langchain.callbacks.tracers.langchain._get_executor', '_get_executor', ([], {}), '()\n', (3071, 3073), False, 'from langchain.callbacks.tracers.langchain import _get_executor\n'), ((3136, 3183), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': 'max_concurrency'}), '(max_workers=max_concurrency)\n', (3154, 3183), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((4183, 4280), 'langchain.callbacks.manager.tracing_v2_enabled', 'manager.tracing_v2_enabled', ([], {'project_name': 'self.project_name', 'tags': "['eval']", 'client': 'self.client'}), "(project_name=self.project_name, tags=['eval'],\n client=self.client)\n", (4209, 4280), False, 'from langchain.callbacks import manager\n'), ((3260, 3299), 'typing.cast', 'cast', (['ThreadPoolExecutor', 'self.executor'], {}), '(ThreadPoolExecutor, self.executor)\n', (3264, 3299), False, 'from typing import Any, Dict, List, Optional, Sequence, Tuple, Union, cast\n')]
import os import re from uuid import UUID from typing import Any, Dict, List, Optional, Union import asyncio import langchain import streamlit as st from langchain.schema import LLMResult from langchain.chat_models import ChatOpenAI from langchain.agents import Tool from langchain.agents import AgentType from langchain.llms import OpenAI from langchain.agents import initialize_agent from langchain.memory import ConversationBufferMemory from langchain.callbacks.base import AsyncCallbackHandler, BaseCallbackHandler from streamlit.delta_generator import DeltaGenerator from chains import doc_retriever from chains.conversational_retrieval_over_code import ConversationalRetrievalCodeChain from chains.parser import parse_code langchain.debug = st.secrets["langchain"]["debug"] python_script = os.path.join(os.getcwd(), "langchain", "generated_script.py") class AsyncHandler(AsyncCallbackHandler): """Async callback handler that can be used to handle callbacks from langchain.""" def __init__(self, message_placeholder: DeltaGenerator) -> None: super().__init__() self.message_placeholder = message_placeholder self.code_block = False self.code_extracted = False self.full_response = "" async def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: """Run when chain starts running.""" message = "" for chunk in "⌛Processing".split(): message += chunk + " " await asyncio.sleep(0.05) # Add a blinking cursor to simulate typing self.message_placeholder.info(message + "▌") async def on_llm_new_token( self, token: str, *, run_id: UUID, parent_run_id: Optional[UUID] = None, **kwargs: Any, ) -> None: # Detect if the token is a code block, to print it all at once self.full_response += token if not self.code_extracted: if "`" in token and not self.code_block: self.code_block = True if self.code_block and self.full_response.count("`") == 6: # We have a full code block, print it now code, explanation = parse_code(self.full_response) container = self.message_placeholder.container() ex = container.expander("Code") ex.code(code) container.markdown(explanation) self.code_block = False self.code_extracted = True if self.code_extracted: code, explanation = parse_code(self.full_response) container = self.message_placeholder.container() ex = container.expander("Code") ex.code(code) container.markdown(explanation + "▌") async def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when chain ends running.""" self.code_extracted = False self.full_response = "" class Handler(BaseCallbackHandler): """Async callback handler that can be used to handle callbacks from langchain.""" def __init__(self, message_placeholder: DeltaGenerator) -> None: super().__init__() self.message_placeholder = message_placeholder self.code_block = False self.code_extracted = False self.full_response = "" def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], *, run_id: UUID, parent_run_id: Union[UUID, None] = None, tags: Union[List[str], None] = None, **kwargs: Any, ) -> Any: """Run when chain starts running.""" return super().on_chain_start( serialized, inputs, run_id=run_id, parent_run_id=parent_run_id, tags=tags, **kwargs, ) def on_llm_new_token( self, token: str, *, run_id: UUID, parent_run_id: Optional[UUID] = None, **kwargs: Any, ) -> None: # Detect if the token is a code block, to print it all at once self.full_response += token if not self.code_extracted: if "`" in token and not self.code_block: self.code_block = True if self.code_block and self.full_response.count("`") >= 6: # We have a full code block, print it now self.message_placeholder.markdown(self.full_response) self.code_block = False self.code_extracted = True if self.code_extracted: message = "" code, explain = parse(self.full_response) if code: message = f"```python\n{code}\n```\n" if explain: message += f"{explain}" if message != "": # Add a blinking cursor to simulate typing self.message_placeholder.markdown(message + "▌") return super().on_llm_new_token( token, run_id=run_id, parent_run_id=parent_run_id, **kwargs ) def on_chain_end( self, outputs: Dict[str, Any], *, run_id: UUID, parent_run_id: Union[UUID, None] = None, **kwargs: Any, ) -> Any: """Run when chain ends running.""" self.code_extracted = False self.full_response = "" return super().on_chain_end( outputs, run_id=run_id, parent_run_id=parent_run_id, **kwargs ) def load_conversation_chain( message_placeholder: DeltaGenerator, openai_api_key: str ) -> ConversationalRetrievalCodeChain: if openai_api_key is None: raise ValueError("OpenAI API key is required to load the chain.") llm = ChatOpenAI( model_name="gpt-3.5-turbo-16k", temperature=0, openai_api_key=openai_api_key, streaming=True, callbacks=[Handler(message_placeholder)], ) condense_question_llm = ChatOpenAI( model_name="gpt-3.5-turbo", temperature=0, openai_api_key=openai_api_key ) missing_imports_llm = ChatOpenAI( model_name="gpt-3.5-turbo", temperature=0, openai_api_key=openai_api_key, verbose=False, ) retriever = doc_retriever.load_streamlit_doc_retriever( st.secrets["openai_api_key"], chroma_server_host=st.secrets["chroma"]["host"], chroma_server_port=st.secrets["chroma"]["port"], mode="docker", ) qa_over_streamlit_code = ConversationalRetrievalCodeChain.from_llm( llm=llm, retriever=retriever, condense_question_llm=condense_question_llm, return_source_documents=True, missing_imports_llm=missing_imports_llm, return_revision_request=True, verbose=False, ) return qa_over_streamlit_code def load_agent(): if st.secrets["openai_api_key"] is None: st.error("OpenAI API key is missing! Please add it to your secrets.") st.stop() doc_chain = doc_retriever.load_streamlit_doc_chain( OpenAI( temperature=0, max_tokens=2000, openai_api_key=st.secrets["openai_api_key"] ) ) tools = [ Tool( name="Streamlit up to date source code", func=doc_chain.run, description="useful for when you need to answer questions about the " "streamlit Python API. Input should be a fully formed question.", ), ] model_name = "text-davinci-003" memory = ConversationBufferMemory(memory_key="chat_history") llm = OpenAI( openai_api_key=st.secrets["openai_api_key"], max_tokens=2000, temperature=0, model_name=model_name, ) agent_chain = initialize_agent( tools, llm, agent=AgentType.CONVERSATIONAL_REACT_DESCRIPTION, verbose=False, memory=memory, ) return agent_chain def load_chat_agent(): if st.secrets["openai_api_key"] is None: st.error("OpenAI API key is missing! Please add it to your secrets.") st.stop() doc_chain = doc_retriever.load_streamlit_doc_chain( OpenAI( temperature=0, max_tokens=2000, openai_api_key=st.secrets["openai_api_key"] ) ) tools = [ Tool( name="Streamlit up to date source code", func=doc_chain.run, description="useful for when you need to answer questions about the streamlit " "Python API. Input should be a fully formed question.", ), ] model_name = "text-davinci-003" memory = ConversationBufferMemory(memory_key="chat_history", return_messages=True) llm = ChatOpenAI( openai_api_key=st.secrets["openai_api_key"], max_tokens=2000, temperature=0, model_name=model_name, ) agent_chain = initialize_agent( tools, llm, agent=AgentType.CHAT_CONVERSATIONAL_REACT_DESCRIPTION, verbose=False, memory=memory, ) return agent_chain # https://regex101.com/r/fHlyKq/1 parse_code_regex = r"(```python(.*?)```)?(.*?)$" def parse(output): python_code = None explain_code = None python_code_match = re.search(parse_code_regex, output, re.DOTALL) if python_code_match: python_code = python_code_match.group(2) explain_code = python_code_match.group(3) if python_code == "None": python_code = None return python_code, explain_code
[ "langchain.agents.initialize_agent", "langchain.memory.ConversationBufferMemory", "langchain.llms.OpenAI", "langchain.chat_models.ChatOpenAI", "langchain.agents.Tool" ]
[((815, 826), 'os.getcwd', 'os.getcwd', ([], {}), '()\n', (824, 826), False, 'import os\n'), ((6031, 6120), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""', 'temperature': '(0)', 'openai_api_key': 'openai_api_key'}), "(model_name='gpt-3.5-turbo', temperature=0, openai_api_key=\n openai_api_key)\n", (6041, 6120), False, 'from langchain.chat_models import ChatOpenAI\n'), ((6156, 6260), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""', 'temperature': '(0)', 'openai_api_key': 'openai_api_key', 'verbose': '(False)'}), "(model_name='gpt-3.5-turbo', temperature=0, openai_api_key=\n openai_api_key, verbose=False)\n", (6166, 6260), False, 'from langchain.chat_models import ChatOpenAI\n'), ((6311, 6505), 'chains.doc_retriever.load_streamlit_doc_retriever', 'doc_retriever.load_streamlit_doc_retriever', (["st.secrets['openai_api_key']"], {'chroma_server_host': "st.secrets['chroma']['host']", 'chroma_server_port': "st.secrets['chroma']['port']", 'mode': '"""docker"""'}), "(st.secrets['openai_api_key'],\n chroma_server_host=st.secrets['chroma']['host'], chroma_server_port=st.\n secrets['chroma']['port'], mode='docker')\n", (6353, 6505), False, 'from chains import doc_retriever\n'), ((6565, 6811), 'chains.conversational_retrieval_over_code.ConversationalRetrievalCodeChain.from_llm', 'ConversationalRetrievalCodeChain.from_llm', ([], {'llm': 'llm', 'retriever': 'retriever', 'condense_question_llm': 'condense_question_llm', 'return_source_documents': '(True)', 'missing_imports_llm': 'missing_imports_llm', 'return_revision_request': '(True)', 'verbose': '(False)'}), '(llm=llm, retriever=retriever,\n condense_question_llm=condense_question_llm, return_source_documents=\n True, missing_imports_llm=missing_imports_llm, return_revision_request=\n True, verbose=False)\n', (6606, 6811), False, 'from chains.conversational_retrieval_over_code import ConversationalRetrievalCodeChain\n'), ((7573, 7624), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""'}), "(memory_key='chat_history')\n", (7597, 7624), False, 'from langchain.memory import ConversationBufferMemory\n'), ((7635, 7745), 'langchain.llms.OpenAI', 'OpenAI', ([], {'openai_api_key': "st.secrets['openai_api_key']", 'max_tokens': '(2000)', 'temperature': '(0)', 'model_name': 'model_name'}), "(openai_api_key=st.secrets['openai_api_key'], max_tokens=2000,\n temperature=0, model_name=model_name)\n", (7641, 7745), False, 'from langchain.llms import OpenAI\n'), ((7799, 7912), 'langchain.agents.initialize_agent', 'initialize_agent', (['tools', 'llm'], {'agent': 'AgentType.CONVERSATIONAL_REACT_DESCRIPTION', 'verbose': '(False)', 'memory': 'memory'}), '(tools, llm, agent=AgentType.\n CONVERSATIONAL_REACT_DESCRIPTION, verbose=False, memory=memory)\n', (7815, 7912), False, 'from langchain.agents import initialize_agent\n'), ((8661, 8734), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'return_messages': '(True)'}), "(memory_key='chat_history', return_messages=True)\n", (8685, 8734), False, 'from langchain.memory import ConversationBufferMemory\n'), ((8745, 8859), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': "st.secrets['openai_api_key']", 'max_tokens': '(2000)', 'temperature': '(0)', 'model_name': 'model_name'}), "(openai_api_key=st.secrets['openai_api_key'], max_tokens=2000,\n temperature=0, model_name=model_name)\n", (8755, 8859), False, 'from langchain.chat_models import ChatOpenAI\n'), ((8913, 9031), 'langchain.agents.initialize_agent', 'initialize_agent', (['tools', 'llm'], {'agent': 'AgentType.CHAT_CONVERSATIONAL_REACT_DESCRIPTION', 'verbose': '(False)', 'memory': 'memory'}), '(tools, llm, agent=AgentType.\n CHAT_CONVERSATIONAL_REACT_DESCRIPTION, verbose=False, memory=memory)\n', (8929, 9031), False, 'from langchain.agents import initialize_agent\n'), ((9274, 9320), 're.search', 're.search', (['parse_code_regex', 'output', 're.DOTALL'], {}), '(parse_code_regex, output, re.DOTALL)\n', (9283, 9320), False, 'import re\n'), ((6968, 7037), 'streamlit.error', 'st.error', (['"""OpenAI API key is missing! Please add it to your secrets."""'], {}), "('OpenAI API key is missing! Please add it to your secrets.')\n", (6976, 7037), True, 'import streamlit as st\n'), ((7046, 7055), 'streamlit.stop', 'st.stop', ([], {}), '()\n', (7053, 7055), True, 'import streamlit as st\n'), ((7120, 7208), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'max_tokens': '(2000)', 'openai_api_key': "st.secrets['openai_api_key']"}), "(temperature=0, max_tokens=2000, openai_api_key=st.secrets[\n 'openai_api_key'])\n", (7126, 7208), False, 'from langchain.llms import OpenAI\n'), ((7255, 7467), 'langchain.agents.Tool', 'Tool', ([], {'name': '"""Streamlit up to date source code"""', 'func': 'doc_chain.run', 'description': '"""useful for when you need to answer questions about the streamlit Python API. Input should be a fully formed question."""'}), "(name='Streamlit up to date source code', func=doc_chain.run,\n description=\n 'useful for when you need to answer questions about the streamlit Python API. Input should be a fully formed question.'\n )\n", (7259, 7467), False, 'from langchain.agents import Tool\n'), ((8056, 8125), 'streamlit.error', 'st.error', (['"""OpenAI API key is missing! Please add it to your secrets."""'], {}), "('OpenAI API key is missing! Please add it to your secrets.')\n", (8064, 8125), True, 'import streamlit as st\n'), ((8134, 8143), 'streamlit.stop', 'st.stop', ([], {}), '()\n', (8141, 8143), True, 'import streamlit as st\n'), ((8208, 8296), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'max_tokens': '(2000)', 'openai_api_key': "st.secrets['openai_api_key']"}), "(temperature=0, max_tokens=2000, openai_api_key=st.secrets[\n 'openai_api_key'])\n", (8214, 8296), False, 'from langchain.llms import OpenAI\n'), ((8343, 8555), 'langchain.agents.Tool', 'Tool', ([], {'name': '"""Streamlit up to date source code"""', 'func': 'doc_chain.run', 'description': '"""useful for when you need to answer questions about the streamlit Python API. Input should be a fully formed question."""'}), "(name='Streamlit up to date source code', func=doc_chain.run,\n description=\n 'useful for when you need to answer questions about the streamlit Python API. Input should be a fully formed question.'\n )\n", (8347, 8555), False, 'from langchain.agents import Tool\n'), ((2617, 2647), 'chains.parser.parse_code', 'parse_code', (['self.full_response'], {}), '(self.full_response)\n', (2627, 2647), False, 'from chains.parser import parse_code\n'), ((1529, 1548), 'asyncio.sleep', 'asyncio.sleep', (['(0.05)'], {}), '(0.05)\n', (1542, 1548), False, 'import asyncio\n'), ((2247, 2277), 'chains.parser.parse_code', 'parse_code', (['self.full_response'], {}), '(self.full_response)\n', (2257, 2277), False, 'from chains.parser import parse_code\n')]
from abc import ABC, abstractmethod from typing import List, Optional from pydantic import BaseModel, Extra, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import ( AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue, ) def _get_verbosity() -> bool: return langchain.verbose class BaseChatModel(BaseLanguageModel, BaseModel, ABC): verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callback_manager: BaseCallbackManager = Field(default_factory=get_callback_manager) class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() def _combine_llm_outputs(self, llm_outputs: List[Optional[dict]]) -> dict: return {} def generate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None ) -> LLMResult: """Top Level call""" results = [self._generate(m, stop=stop) for m in messages] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] return LLMResult(generations=generations, llm_output=llm_output) async def agenerate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None ) -> LLMResult: """Top Level call""" results = [await self._agenerate(m, stop=stop) for m in messages] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] return LLMResult(generations=generations, llm_output=llm_output) def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] prompt_strings = [p.to_string() for p in prompts] self.callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompt_strings, verbose=self.verbose ) try: output = self.generate(prompt_messages, stop=stop) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_llm_error(e, verbose=self.verbose) raise e self.callback_manager.on_llm_end(output, verbose=self.verbose) return output async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] prompt_strings = [p.to_string() for p in prompts] if self.callback_manager.is_async: await self.callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompt_strings, verbose=self.verbose ) else: self.callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompt_strings, verbose=self.verbose ) try: output = await self.agenerate(prompt_messages, stop=stop) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_llm_error(e, verbose=self.verbose) else: self.callback_manager.on_llm_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_llm_end(output, verbose=self.verbose) else: self.callback_manager.on_llm_end(output, verbose=self.verbose) return output @abstractmethod def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> ChatResult: """Top Level call""" @abstractmethod async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> ChatResult: """Top Level call""" def __call__( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> BaseMessage: return self._generate(messages, stop=stop).generations[0].message def call_as_llm(self, message: str, stop: Optional[List[str]] = None) -> str: result = self([HumanMessage(content=message)], stop=stop) return result.content class SimpleChatModel(BaseChatModel): def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> ChatResult: output_str = self._call(messages, stop=stop) message = AIMessage(content=output_str) generation = ChatGeneration(message=message) return ChatResult(generations=[generation]) @abstractmethod def _call( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> str: """Simpler interface."""
[ "langchain.schema.ChatResult", "langchain.schema.ChatGeneration", "langchain.schema.HumanMessage", "langchain.schema.AIMessage", "langchain.schema.LLMResult", "langchain.callbacks.get_callback_manager" ]
[((568, 605), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (573, 605), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((696, 739), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager'}), '(default_factory=get_callback_manager)\n', (701, 739), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((888, 940), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (897, 940), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((1737, 1794), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (1746, 1794), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((2184, 2241), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (2193, 2241), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((5092, 5121), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'output_str'}), '(content=output_str)\n', (5101, 5121), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((5143, 5174), 'langchain.schema.ChatGeneration', 'ChatGeneration', ([], {'message': 'message'}), '(message=message)\n', (5157, 5174), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((5190, 5226), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': '[generation]'}), '(generations=[generation])\n', (5200, 5226), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((1241, 1263), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1261, 1263), False, 'from langchain.callbacks import get_callback_manager\n'), ((4792, 4821), 'langchain.schema.HumanMessage', 'HumanMessage', ([], {'content': 'message'}), '(content=message)\n', (4804, 4821), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n')]
from abc import ABC, abstractmethod from typing import List, Optional from pydantic import BaseModel, Extra, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import ( AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue, ) def _get_verbosity() -> bool: return langchain.verbose class BaseChatModel(BaseLanguageModel, BaseModel, ABC): verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callback_manager: BaseCallbackManager = Field(default_factory=get_callback_manager) class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() def _combine_llm_outputs(self, llm_outputs: List[Optional[dict]]) -> dict: return {} def generate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None ) -> LLMResult: """Top Level call""" results = [self._generate(m, stop=stop) for m in messages] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] return LLMResult(generations=generations, llm_output=llm_output) async def agenerate( self, messages: List[List[BaseMessage]], stop: Optional[List[str]] = None ) -> LLMResult: """Top Level call""" results = [await self._agenerate(m, stop=stop) for m in messages] llm_output = self._combine_llm_outputs([res.llm_output for res in results]) generations = [res.generations for res in results] return LLMResult(generations=generations, llm_output=llm_output) def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] prompt_strings = [p.to_string() for p in prompts] self.callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompt_strings, verbose=self.verbose ) try: output = self.generate(prompt_messages, stop=stop) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_llm_error(e, verbose=self.verbose) raise e self.callback_manager.on_llm_end(output, verbose=self.verbose) return output async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None ) -> LLMResult: prompt_messages = [p.to_messages() for p in prompts] prompt_strings = [p.to_string() for p in prompts] if self.callback_manager.is_async: await self.callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompt_strings, verbose=self.verbose ) else: self.callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompt_strings, verbose=self.verbose ) try: output = await self.agenerate(prompt_messages, stop=stop) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_llm_error(e, verbose=self.verbose) else: self.callback_manager.on_llm_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_llm_end(output, verbose=self.verbose) else: self.callback_manager.on_llm_end(output, verbose=self.verbose) return output @abstractmethod def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> ChatResult: """Top Level call""" @abstractmethod async def _agenerate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> ChatResult: """Top Level call""" def __call__( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> BaseMessage: return self._generate(messages, stop=stop).generations[0].message def call_as_llm(self, message: str, stop: Optional[List[str]] = None) -> str: result = self([HumanMessage(content=message)], stop=stop) return result.content class SimpleChatModel(BaseChatModel): def _generate( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> ChatResult: output_str = self._call(messages, stop=stop) message = AIMessage(content=output_str) generation = ChatGeneration(message=message) return ChatResult(generations=[generation]) @abstractmethod def _call( self, messages: List[BaseMessage], stop: Optional[List[str]] = None ) -> str: """Simpler interface."""
[ "langchain.schema.ChatResult", "langchain.schema.ChatGeneration", "langchain.schema.HumanMessage", "langchain.schema.AIMessage", "langchain.schema.LLMResult", "langchain.callbacks.get_callback_manager" ]
[((568, 605), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (573, 605), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((696, 739), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager'}), '(default_factory=get_callback_manager)\n', (701, 739), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((888, 940), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (897, 940), False, 'from pydantic import BaseModel, Extra, Field, validator\n'), ((1737, 1794), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (1746, 1794), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((2184, 2241), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (2193, 2241), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((5092, 5121), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'output_str'}), '(content=output_str)\n', (5101, 5121), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((5143, 5174), 'langchain.schema.ChatGeneration', 'ChatGeneration', ([], {'message': 'message'}), '(message=message)\n', (5157, 5174), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((5190, 5226), 'langchain.schema.ChatResult', 'ChatResult', ([], {'generations': '[generation]'}), '(generations=[generation])\n', (5200, 5226), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n'), ((1241, 1263), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1261, 1263), False, 'from langchain.callbacks import get_callback_manager\n'), ((4792, 4821), 'langchain.schema.HumanMessage', 'HumanMessage', ([], {'content': 'message'}), '(content=message)\n', (4804, 4821), False, 'from langchain.schema import AIMessage, BaseLanguageModel, BaseMessage, ChatGeneration, ChatResult, HumanMessage, LLMResult, PromptValue\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.load.dump import dumpd from langchain.schema import ( AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string, ) def _get_verbosity() -> bool: return langchain.verbose def get_prompts( params: Dict[str, Any], prompts: List[str] ) -> Tuple[Dict[int, List], str, List[int], List[str]]: """Get prompts that are already cached.""" llm_string = str(sorted([(k, v) for k, v in params.items()])) missing_prompts = [] missing_prompt_idxs = [] existing_prompts = {} for i, prompt in enumerate(prompts): if langchain.llm_cache is not None: cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): existing_prompts[i] = cache_val else: missing_prompts.append(prompt) missing_prompt_idxs.append(i) return existing_prompts, llm_string, missing_prompt_idxs, missing_prompts def update_cache( existing_prompts: Dict[int, List], llm_string: str, missing_prompt_idxs: List[int], new_results: LLMResult, prompts: List[str], ) -> Optional[dict]: """Update the cache and get the LLM output.""" for i, result in enumerate(new_results.generations): existing_prompts[missing_prompt_idxs[i]] = result prompt = prompts[missing_prompt_idxs[i]] if langchain.llm_cache is not None: langchain.llm_cache.update(prompt, llm_string, result) llm_output = new_results.llm_output return llm_output class BaseLLM(BaseLanguageModel, ABC): """LLM wrapper should take in a prompt and return a string.""" cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) tags: Optional[List[str]] = Field(default=None, exclude=True) """Tags to add to the run trace.""" class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @abstractmethod def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompts.""" @abstractmethod async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompts.""" def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return self.generate(prompt_strings, stop=stop, callbacks=callbacks, **kwargs) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return await self.agenerate( prompt_strings, stop=stop, callbacks=callbacks, **kwargs ) def generate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # If string is passed in directly no errors will be raised but outputs will # not make sense. if not isinstance(prompts, list): raise ValueError( "Argument 'prompts' is expected to be of type List[str], received" f" argument of type {type(prompts)}." ) params = self.dict() params["stop"] = stop options = {"stop": stop} ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags ) new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = callback_manager.on_llm_start( dumpd(self), prompts, invocation_params=params, options=options ) try: output = ( self._generate( prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else self._generate(prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(output) if run_manager: output.run = RunInfo(run_id=run_manager.run_id) return output if len(missing_prompts) > 0: run_manager = callback_manager.on_llm_start( dumpd(self), missing_prompts, invocation_params=params, options=options, ) try: new_results = ( self._generate( missing_prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else self._generate(missing_prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) run_info = None if run_manager: run_info = RunInfo(run_id=run_manager.run_id) else: llm_output = {} run_info = None generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output, run=run_info) async def agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" params = self.dict() params["stop"] = stop options = {"stop": stop} ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags ) new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = await callback_manager.on_llm_start( dumpd(self), prompts, invocation_params=params, options=options ) try: output = ( await self._agenerate( prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else await self._agenerate(prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e, verbose=self.verbose) raise e await run_manager.on_llm_end(output, verbose=self.verbose) if run_manager: output.run = RunInfo(run_id=run_manager.run_id) return output if len(missing_prompts) > 0: run_manager = await callback_manager.on_llm_start( dumpd(self), missing_prompts, invocation_params=params, options=options, ) try: new_results = ( await self._agenerate( missing_prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else await self._agenerate(missing_prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e) raise e await run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) run_info = None if run_manager: run_info = RunInfo(run_id=run_manager.run_id) else: llm_output = {} run_info = None generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output, run=run_info) def __call__( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> str: """Check Cache and run the LLM on the given prompt and input.""" if not isinstance(prompt, str): raise ValueError( "Argument `prompt` is expected to be a string. Instead found " f"{type(prompt)}. If you want to run the LLM on multiple prompts, use " "`generate` instead." ) return ( self.generate([prompt], stop=stop, callbacks=callbacks, **kwargs) .generations[0][0] .text ) async def _call_async( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> str: """Check Cache and run the LLM on the given prompt and input.""" result = await self.agenerate( [prompt], stop=stop, callbacks=callbacks, **kwargs ) return result.generations[0][0].text def predict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) return self(text, stop=_stop, **kwargs) def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = self(text, stop=_stop, **kwargs) return AIMessage(content=content) async def apredict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(text, stop=_stop, **kwargs) async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = await self._call_async(text, stop=_stop, **kwargs) return AIMessage(content=content) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} def __str__(self) -> str: """Get a string representation of the object for printing.""" cls_name = f"\033[1m{self.__class__.__name__}\033[0m" return f"{cls_name}\nParams: {self._identifying_params}" @property @abstractmethod def _llm_type(self) -> str: """Return type of llm.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict def save(self, file_path: Union[Path, str]) -> None: """Save the LLM. Args: file_path: Path to file to save the LLM to. Example: .. code-block:: python llm.save(file_path="path/llm.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save prompt_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(prompt_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(prompt_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") class LLM(BaseLLM): """LLM class that expect subclasses to implement a simpler call method. The purpose of this class is to expose a simpler interface for working with LLMs, rather than expect the user to implement the full _generate method. """ @abstractmethod def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Run the LLM on the given prompt and input.""" async def _acall( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Run the LLM on the given prompt and input.""" raise NotImplementedError("Async generation not implemented for this LLM.") def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # TODO: add caching here. generations = [] new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") for prompt in prompts: text = ( self._call(prompt, stop=stop, run_manager=run_manager, **kwargs) if new_arg_supported else self._call(prompt, stop=stop, **kwargs) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations) async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" generations = [] new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") for prompt in prompts: text = ( await self._acall(prompt, stop=stop, run_manager=run_manager, **kwargs) if new_arg_supported else await self._acall(prompt, stop=stop, **kwargs) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.load.dump.dumpd", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.RunInfo", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((2353, 2390), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2358, 2390), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2464, 2497), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2469, 2497), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2552, 2585), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2557, 2585), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2618, 2651), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2623, 2651), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2840, 2856), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (2854, 2856), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((3292, 3335), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (3301, 3335), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((5907, 5994), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags)\n', (5932, 5994), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((8360, 8431), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output', 'run': 'run_info'}), '(generations=generations, llm_output=llm_output, run=run_info)\n', (8369, 8431), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((9086, 9178), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags)\n', (9116, 9178), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((11653, 11724), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output', 'run': 'run_info'}), '(generations=generations, llm_output=llm_output, run=run_info)\n', (11662, 11724), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((13270, 13297), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (13287, 13297), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((13459, 13485), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (13468, 13485), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((13968, 13995), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (13985, 13995), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((14175, 14201), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (14184, 14201), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((17471, 17505), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (17480, 17505), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((18230, 18264), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (18239, 18264), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((1286, 1332), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (1312, 1332), False, 'import langchain\n'), ((2074, 2128), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result'], {}), '(prompt, llm_string, result)\n', (2100, 2128), False, 'import langchain\n'), ((3047, 3149), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (3060, 3149), False, 'import warnings\n'), ((15251, 15266), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (15255, 15266), False, 'from pathlib import Path\n'), ((6524, 6535), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (6529, 6535), False, 'from langchain.load.dump import dumpd\n'), ((7135, 7169), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (7142, 7169), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((7306, 7317), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (7311, 7317), False, 'from langchain.load.dump import dumpd\n'), ((8167, 8201), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (8174, 8201), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((10396, 10430), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (10403, 10430), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((11460, 11494), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (11467, 11494), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((15587, 15622), 'json.dump', 'json.dump', (['prompt_dict', 'f'], {'indent': '(4)'}), '(prompt_dict, f, indent=4)\n', (15596, 15622), False, 'import json\n'), ((6041, 6074), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (6058, 6074), False, 'import inspect\n'), ((9225, 9259), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (9242, 9259), False, 'import inspect\n'), ((9715, 9726), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (9720, 9726), False, 'from langchain.load.dump import dumpd\n'), ((10573, 10584), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (10578, 10584), False, 'from langchain.load.dump import dumpd\n'), ((15725, 15776), 'yaml.dump', 'yaml.dump', (['prompt_dict', 'f'], {'default_flow_style': '(False)'}), '(prompt_dict, f, default_flow_style=False)\n', (15734, 15776), False, 'import yaml\n'), ((17095, 17124), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (17112, 17124), False, 'import inspect\n'), ((17432, 17453), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (17442, 17453), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((17839, 17869), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (17856, 17869), False, 'import inspect\n'), ((18191, 18212), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (18201, 18212), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.load.dump import dumpd from langchain.schema import ( AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string, ) def _get_verbosity() -> bool: return langchain.verbose def get_prompts( params: Dict[str, Any], prompts: List[str] ) -> Tuple[Dict[int, List], str, List[int], List[str]]: """Get prompts that are already cached.""" llm_string = str(sorted([(k, v) for k, v in params.items()])) missing_prompts = [] missing_prompt_idxs = [] existing_prompts = {} for i, prompt in enumerate(prompts): if langchain.llm_cache is not None: cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): existing_prompts[i] = cache_val else: missing_prompts.append(prompt) missing_prompt_idxs.append(i) return existing_prompts, llm_string, missing_prompt_idxs, missing_prompts def update_cache( existing_prompts: Dict[int, List], llm_string: str, missing_prompt_idxs: List[int], new_results: LLMResult, prompts: List[str], ) -> Optional[dict]: """Update the cache and get the LLM output.""" for i, result in enumerate(new_results.generations): existing_prompts[missing_prompt_idxs[i]] = result prompt = prompts[missing_prompt_idxs[i]] if langchain.llm_cache is not None: langchain.llm_cache.update(prompt, llm_string, result) llm_output = new_results.llm_output return llm_output class BaseLLM(BaseLanguageModel, ABC): """LLM wrapper should take in a prompt and return a string.""" cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) tags: Optional[List[str]] = Field(default=None, exclude=True) """Tags to add to the run trace.""" class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @abstractmethod def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompts.""" @abstractmethod async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompts.""" def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return self.generate(prompt_strings, stop=stop, callbacks=callbacks, **kwargs) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return await self.agenerate( prompt_strings, stop=stop, callbacks=callbacks, **kwargs ) def generate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # If string is passed in directly no errors will be raised but outputs will # not make sense. if not isinstance(prompts, list): raise ValueError( "Argument 'prompts' is expected to be of type List[str], received" f" argument of type {type(prompts)}." ) params = self.dict() params["stop"] = stop options = {"stop": stop} ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags ) new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = callback_manager.on_llm_start( dumpd(self), prompts, invocation_params=params, options=options ) try: output = ( self._generate( prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else self._generate(prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(output) if run_manager: output.run = RunInfo(run_id=run_manager.run_id) return output if len(missing_prompts) > 0: run_manager = callback_manager.on_llm_start( dumpd(self), missing_prompts, invocation_params=params, options=options, ) try: new_results = ( self._generate( missing_prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else self._generate(missing_prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) run_info = None if run_manager: run_info = RunInfo(run_id=run_manager.run_id) else: llm_output = {} run_info = None generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output, run=run_info) async def agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" params = self.dict() params["stop"] = stop options = {"stop": stop} ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags ) new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = await callback_manager.on_llm_start( dumpd(self), prompts, invocation_params=params, options=options ) try: output = ( await self._agenerate( prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else await self._agenerate(prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e, verbose=self.verbose) raise e await run_manager.on_llm_end(output, verbose=self.verbose) if run_manager: output.run = RunInfo(run_id=run_manager.run_id) return output if len(missing_prompts) > 0: run_manager = await callback_manager.on_llm_start( dumpd(self), missing_prompts, invocation_params=params, options=options, ) try: new_results = ( await self._agenerate( missing_prompts, stop=stop, run_manager=run_manager, **kwargs ) if new_arg_supported else await self._agenerate(missing_prompts, stop=stop, **kwargs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e) raise e await run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) run_info = None if run_manager: run_info = RunInfo(run_id=run_manager.run_id) else: llm_output = {} run_info = None generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output, run=run_info) def __call__( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> str: """Check Cache and run the LLM on the given prompt and input.""" if not isinstance(prompt, str): raise ValueError( "Argument `prompt` is expected to be a string. Instead found " f"{type(prompt)}. If you want to run the LLM on multiple prompts, use " "`generate` instead." ) return ( self.generate([prompt], stop=stop, callbacks=callbacks, **kwargs) .generations[0][0] .text ) async def _call_async( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None, **kwargs: Any, ) -> str: """Check Cache and run the LLM on the given prompt and input.""" result = await self.agenerate( [prompt], stop=stop, callbacks=callbacks, **kwargs ) return result.generations[0][0].text def predict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) return self(text, stop=_stop, **kwargs) def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = self(text, stop=_stop, **kwargs) return AIMessage(content=content) async def apredict( self, text: str, *, stop: Optional[Sequence[str]] = None, **kwargs: Any ) -> str: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(text, stop=_stop, **kwargs) async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None, **kwargs: Any, ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = await self._call_async(text, stop=_stop, **kwargs) return AIMessage(content=content) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} def __str__(self) -> str: """Get a string representation of the object for printing.""" cls_name = f"\033[1m{self.__class__.__name__}\033[0m" return f"{cls_name}\nParams: {self._identifying_params}" @property @abstractmethod def _llm_type(self) -> str: """Return type of llm.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict def save(self, file_path: Union[Path, str]) -> None: """Save the LLM. Args: file_path: Path to file to save the LLM to. Example: .. code-block:: python llm.save(file_path="path/llm.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save prompt_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(prompt_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(prompt_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") class LLM(BaseLLM): """LLM class that expect subclasses to implement a simpler call method. The purpose of this class is to expose a simpler interface for working with LLMs, rather than expect the user to implement the full _generate method. """ @abstractmethod def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Run the LLM on the given prompt and input.""" async def _acall( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> str: """Run the LLM on the given prompt and input.""" raise NotImplementedError("Async generation not implemented for this LLM.") def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # TODO: add caching here. generations = [] new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") for prompt in prompts: text = ( self._call(prompt, stop=stop, run_manager=run_manager, **kwargs) if new_arg_supported else self._call(prompt, stop=stop, **kwargs) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations) async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, **kwargs: Any, ) -> LLMResult: """Run the LLM on the given prompt and input.""" generations = [] new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") for prompt in prompts: text = ( await self._acall(prompt, stop=stop, run_manager=run_manager, **kwargs) if new_arg_supported else await self._acall(prompt, stop=stop, **kwargs) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.load.dump.dumpd", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.RunInfo", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((2353, 2390), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2358, 2390), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2464, 2497), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2469, 2497), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2552, 2585), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2557, 2585), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2618, 2651), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2623, 2651), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2840, 2856), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (2854, 2856), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((3292, 3335), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (3301, 3335), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((5907, 5994), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags)\n', (5932, 5994), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((8360, 8431), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output', 'run': 'run_info'}), '(generations=generations, llm_output=llm_output, run=run_info)\n', (8369, 8431), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((9086, 9178), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags)\n', (9116, 9178), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((11653, 11724), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output', 'run': 'run_info'}), '(generations=generations, llm_output=llm_output, run=run_info)\n', (11662, 11724), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((13270, 13297), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (13287, 13297), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((13459, 13485), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (13468, 13485), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((13968, 13995), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (13985, 13995), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((14175, 14201), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (14184, 14201), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((17471, 17505), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (17480, 17505), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((18230, 18264), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (18239, 18264), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((1286, 1332), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (1312, 1332), False, 'import langchain\n'), ((2074, 2128), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result'], {}), '(prompt, llm_string, result)\n', (2100, 2128), False, 'import langchain\n'), ((3047, 3149), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (3060, 3149), False, 'import warnings\n'), ((15251, 15266), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (15255, 15266), False, 'from pathlib import Path\n'), ((6524, 6535), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (6529, 6535), False, 'from langchain.load.dump import dumpd\n'), ((7135, 7169), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (7142, 7169), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((7306, 7317), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (7311, 7317), False, 'from langchain.load.dump import dumpd\n'), ((8167, 8201), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (8174, 8201), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((10396, 10430), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (10403, 10430), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((11460, 11494), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (11467, 11494), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((15587, 15622), 'json.dump', 'json.dump', (['prompt_dict', 'f'], {'indent': '(4)'}), '(prompt_dict, f, indent=4)\n', (15596, 15622), False, 'import json\n'), ((6041, 6074), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (6058, 6074), False, 'import inspect\n'), ((9225, 9259), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (9242, 9259), False, 'import inspect\n'), ((9715, 9726), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (9720, 9726), False, 'from langchain.load.dump import dumpd\n'), ((10573, 10584), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (10578, 10584), False, 'from langchain.load.dump import dumpd\n'), ((15725, 15776), 'yaml.dump', 'yaml.dump', (['prompt_dict', 'f'], {'default_flow_style': '(False)'}), '(prompt_dict, f, default_flow_style=False)\n', (15734, 15776), False, 'import yaml\n'), ((17095, 17124), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (17112, 17124), False, 'import inspect\n'), ((17432, 17453), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (17442, 17453), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n'), ((17839, 17869), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (17856, 17869), False, 'import inspect\n'), ((18191, 18212), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (18201, 18212), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, RunInfo, get_buffer_string\n')]
from langchain.agents import AgentExecutor, LLMSingleActionAgent, AgentOutputParser from langchain.tools import Tool, StructuredTool from langchain.prompts import StringPromptTemplate from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.llms import VertexAI from typing import List, Union, Tuple, Any, Dict from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult import re import langchain import os import json import requests from utils import * from tools import take_environment_action_wrapper from debate import view_debate_wrapper from langchain.callbacks import FileCallbackHandler from langchain.callbacks.base import BaseCallbackHandler import logging from langchain.agents.agent_iterator import AgentExecutorIterator from datetime import datetime from pydantic import BaseModel, Field # Example usage: log_levels = { 'none': logging.CRITICAL, 'all': logging.INFO, } def get_info_logger(): timestamp = datetime.now().strftime("%Y-%m-%d_%H-%M-%S") os.makedirs(f'./results/{timestamp}', exist_ok=True) info_filename = f'./results/{timestamp}/info_{timestamp}.log' debug_filename = f'./results/{timestamp}/debug_{timestamp}.log' logging.basicConfig(level=logging.NOTSET, handlers=[logging.NullHandler()]) info_logger = logging.getLogger('info_logger') # Create an INFO file handler file_handler_info = logging.FileHandler(info_filename) file_handler_info.setLevel(logging.INFO) # Create formatters and set them for the file handlers formatter_info = logging.Formatter('%(message)s') file_handler_info.setFormatter(formatter_info) # Add the file handlers to the logger info_logger.addHandler(file_handler_info) info_logger.info(timestamp) return info_logger, timestamp, info_filename, debug_filename class Context(BaseModel): generation_observation_history: List[str] = [] log_count: int = 0 action_count: int = 0 debate_count: int = 0 do_debate: bool = False system_hint_mod: int = 2 max_votes: int = 1 vote_count: int = 0 votes: List[Tuple[str, str, str]] = [] info_logger: Any = None token_count: int = 0 class CustomPromptTemplate(StringPromptTemplate): # The template to use template: str # The list of tools available tools: List[Tool] # Context for information context: Context def format(self, **kwargs) -> str: # Get the intermediate steps (AgentAction, Observation tuples) # Format them in a particular way intermediate_steps = [(a, o) for a, o in kwargs.pop("intermediate_steps") if o != EMPTY_RESPONSE] # if recently finished voting, then vote count is 0 done_voting = self.context.vote_count == 0 history_lines = [] last_action = None last_observation = None # create history to prompt agent for action, observation in intermediate_steps: history_lines.append(action.log.strip()) history_lines.append(f"{OBSERVATION_PREFIX} {observation.strip()}") last_action = action last_observation = observation.strip() history = '\n'.join(history_lines) # append observation to list for debaters to use. Only append the environment observation, not debate observation # only do this if done voting, else we'll append it 'self.context.max_votes' many times if done_voting and last_action and last_action.tool == TAKE_ENVIRONMENT_ACTION: self.context.generation_observation_history.append(f'{OBSERVATION_PREFIX} {last_observation}') system_hint = None valid_action_hint = None # append system hint for after taking action if (self.context.do_debate and last_action and last_action.tool == TAKE_ENVIRONMENT_ACTION and self.context.action_count % self.context.system_hint_mod == 0): system_hint = HINT_AFTER_ACTION history += '\n' + system_hint # append system hint for after viewing debate if (self.context.do_debate and last_action and last_action.tool == VIEW_DEBATE): system_hint = HINT_AFTER_DEBATE history += '\n' + system_hint # append system hint that reminds valid_action_hint = VALID_ACTIONS history += '\n' + valid_action_hint # Set the agent_scratchpad variable to that value kwargs["agent_scratchpad"] = history # Create a tools variable from the list of tools provided tool_strings = [] for tool in self.tools: s = f"{tool.name}: {tool.description}" params = [f'{param} - {info["description"]}' for param, info in tool.args.items()] s += ' Argument: ' + ' '.join(params) tool_strings.append(s) kwargs["tools"] = "\n".join(tool_strings) # Create a list of tool names for the tools provided kwargs["tool_names"] = " or ".join([tool.name for tool in self.tools]) agent_prompt = self.template.format(**kwargs) # log some stuff if self.context.log_count == 0: self.context.info_logger.info(f"Step {self.context.log_count} ===") self.context.info_logger.info(agent_prompt) self.context.log_count += 1 elif done_voting and self.context.log_count > 0 and last_action: self.context.info_logger.info(f"\nStep {self.context.log_count} ===") self.context.info_logger.info(f'Generation ---\n{last_action.log.strip()}') self.context.info_logger.info(f'Observation ---\n{last_observation}') if system_hint: self.context.info_logger.info(f'<only seen once by agent> {system_hint}') if valid_action_hint: self.context.info_logger.info(f'<only seen once by agent> {valid_action_hint}') self.context.log_count += 1 self.context.token_count += tokens(agent_prompt) return agent_prompt class CustomOutputParser(AgentOutputParser): # Context for information context: Context def parse(self, llm_output: str) -> Union[AgentFinish, AgentAction, list[AgentAction]]: self.context.token_count += tokens(llm_output) default_agent_finish = AgentFinish( return_values={'output': "Task finished."}, log=llm_output ) # first, extract the answer and append it to the votes # this includes final answers, and even treats errors as an answer try: pattern = r'Tool: (.*)\nTool Input: (.*)' match = re.search(pattern, llm_output, re.DOTALL) # if the agent wants to do a final answer if "final answer" in llm_output.lower() or "final_answer" in llm_output.lower(): self.context.vote_count += 1 self.context.votes.append(('final answer', 'final answer', llm_output)) # else, look for a tool elif match: # extract the tool information tool_name: str = match.group(1) tool_input: str = match.group(2) # increment the votes self.context.vote_count += 1 self.context.votes.append((tool_name, tool_input, llm_output)) else: raise ValueError(f"Could not find 'Tool:' or 'Tool Input:' : `{llm_output}`") except Exception as e: self.context.vote_count += 1 self.context.votes.append(('error', 'error', llm_output + f'\n ERROR === \n{e}')) # if not done voting, then don't take an action # the take_environment_action tool handles this if self.context.vote_count < self.context.max_votes: return AgentAction(tool=TAKE_ENVIRONMENT_ACTION, tool_input='empty param', log='empty tool') # if done voting, return majority vote and reset voting else: # log the votes self.context.info_logger.info("Casting votes... ===") for i, vote in enumerate(self.context.votes): self.context.info_logger.info(f"Vote {i} ---\n{vote}") # get the majority vote majority_tool_name, majority_tool_input, random_llm_output = get_majority_vote(self.context.votes) # reset the voting self.context.vote_count = 0 self.context.votes = [] # if the majority vote was a final answer or an error: if majority_tool_name == 'final answer' or majority_tool_name == 'error': return default_agent_finish # if the majority vote is a debate tool, then call that tool elif majority_tool_name == VIEW_DEBATE: self.context.debate_count += 1 return AgentAction(tool=majority_tool_name, tool_input=majority_tool_input, log=random_llm_output) # if the majority vote is a environment action tool, then call that tool and log stuff elif majority_tool_name == TAKE_ENVIRONMENT_ACTION: # increment action count and log it self.context.action_count += 1 self.context.info_logger.info(f"\nAction Count {self.context.action_count} +++") # add action to the history for debaters self.context.generation_observation_history.append('Action: ' + majority_tool_input) return AgentAction(tool=majority_tool_name, tool_input=majority_tool_input, log=random_llm_output) else: print('This happened!') return default_agent_finish # raise ValueError(f"An error occurred that should never occur: `{majority_tool_name}`") def run_experiment(exp): description = exp['description'] langchain.debug = exp['langchain.debug'] langchain_verbose = exp['langchain_verbose'] log_level = log_levels['all'] langchain_logging = False do_debate = exp['do_debate'] MAX_STEPS = exp['MAX_STEPS'] MAX_VOTES = exp['MAX_VOTES'] temperature = 0 if MAX_VOTES == 1 else 0.7 model = exp['agent_model'] system_hint_mod = 1000000 debate_params = dict() if 'debate_params' in exp: debate_params = exp['debate_params'] system_hint_mod = debate_params.pop('system_hint_mod') info_logger, timestamp, info_filename, debug_filename = get_info_logger() info_logger.setLevel(log_level) if langchain_logging: handler = FileCallbackHandler(debug_filename) results = [] num_tasks = exp['num_tasks'] start_task = exp['start_task'] reset_tasks() for _ in range(1, start_task): get_next_task(MAX_STEPS) filename = None for _ in range(num_tasks): # set up the context to pass around context = Context() # file that debaters will use to see agent history context.generation_observation_history = [] # information to know when to log the full action count context.log_count = 0 # information to know when to provide the system hint context.action_count = 0 # count the number of debates context.debate_count = 0 # only display the system hints if do_debate is true context.do_debate = do_debate # show the hints after every x many actions. context.system_hint_mod = system_hint_mod # do majority voting context.max_votes = MAX_VOTES # count total votes so far context.vote_count = 0 # store the votes context.votes = [] # expose the logger context.info_logger = info_logger # count the tokens context.token_count = 0 # set up the available tools tools = [ take_environment_action_wrapper(context) ] if do_debate: tools.append(view_debate_wrapper(context, **debate_params, logger=info_logger)) # load the examples and task from ALFWorld examples, task, task_index = get_next_task(MAX_STEPS, do_debate=False) # not inserting debates even if in debate mode print(f'Task index: {task_index}') examples_str = '\n\n'.join([f'Example {i + 1}:\n{ex}' for i, ex in enumerate(examples)]) examples_str = examples_str.replace('{', '{{').replace('}', '}}') info_logger.info(f'\n\n\n\n\n\n\n\n\nTask index: {task_index}') # load the prompt that tells how to format the actions formatting = read_text_file("./prompts/action_formatting.txt") # load the examples of failures failure_examples_str = read_text_file("./prompts/failure_examples.txt") # set up strings to insert if do_debate == True debate_examples_str = '' debate_msg_1 = '' debate_msg_2 = '' if do_debate: debate_msg_2 = f'\n- IMPORTANT: Remember to use the "{VIEW_DEBATE}" tool to get a better understanding about your problem and proposed action BEFORE using "{TAKE_ENVIRONMENT_ACTION}".' debate_examples_str = '\n\n' + read_text_file("./prompts/debate_examples.txt") template = read_text_file("prompts/prompt_template.txt") # fill out the template template = template.format( formatting=formatting, success_examples=examples_str, failure_examples=failure_examples_str, debate_examples=debate_examples_str, debate_msg_1=debate_msg_1, debate_msg_2=debate_msg_2, ) # create the prompt prompt = CustomPromptTemplate( template=template, tools=tools, # This omits the `agent_scratchpad`, `tools`, and `tool_names` variables because those are generated dynamically # This includes the `intermediate_steps` variable because that is needed input_variables=["input", "intermediate_steps"], context=context, ) llm = VertexAI( model_name=model, temperature=temperature, max_output_tokens=256, ) callbacks = [] if langchain_logging: callbacks.append(handler) llm_chain = LLMChain( llm=llm, prompt=prompt, callbacks=callbacks ) agent = LLMSingleActionAgent( llm_chain=llm_chain, output_parser=CustomOutputParser(context=context), stop=[f"\n{OBSERVATION_PREFIX}"], allowed_tools=[tool.name for tool in tools], ) agent_executor = AgentExecutor.from_agent_and_tools( agent=agent, tools=tools, verbose=langchain_verbose, max_iterations=2 * MAX_STEPS * context.max_votes + 1) agent_iterator = agent_executor.iter(inputs=task) # append to history for the debaters context.generation_observation_history.append('Task: ' + task) result_dict = dict() result_dict['task'] = task result_dict['task_index'] = task_index result_dict['success'] = False total_steps = 0 for step_num, step in enumerate(agent_iterator): # intermediate_step is a (action, observation) pair prev_ob = step.get('intermediate_step') if prev_ob: a, o = prev_ob[-1] if not o: break if FAIL_OBSERVATION in o: total_steps += 1 break elif SUCCESS_OBSERVATION in o: total_steps += 1 result_dict['success'] = True break elif EMPTY_RESPONSE not in o: total_steps += 1 else: # print("HERE") # print(step) break result_dict['total_steps'] = total_steps # the number of times take_environment_action was called result_dict['total_actions'] = context.action_count result_dict['total_debates'] = context.debate_count result_dict['token_count'] = context.token_count results.append(result_dict) # save the results every time, so we don't lose anything results_dir = f'./results/{timestamp}/' filename = f"{results_dir}results_{timestamp}.json" if not os.path.exists(results_dir): os.makedirs(results_dir) # put back system hint mod because we popped it if 'debate_params' in exp: exp['debate_params']['system_hint_mod'] = system_hint_mod extended_results = { 'description': description, 'params': exp, 'timestamp': timestamp, 'results': results } write_json_file(filename, extended_results) # pop it again if 'debate_params' in exp: exp['debate_params'].pop('system_hint_mod') return timestamp, filename
[ "langchain.agents.AgentExecutor.from_agent_and_tools", "langchain.schema.AgentAction", "langchain.llms.VertexAI", "langchain.schema.AgentFinish", "langchain.callbacks.FileCallbackHandler", "langchain.chains.LLMChain" ]
[((1046, 1098), 'os.makedirs', 'os.makedirs', (['f"""./results/{timestamp}"""'], {'exist_ok': '(True)'}), "(f'./results/{timestamp}', exist_ok=True)\n", (1057, 1098), False, 'import os\n'), ((1332, 1364), 'logging.getLogger', 'logging.getLogger', (['"""info_logger"""'], {}), "('info_logger')\n", (1349, 1364), False, 'import logging\n'), ((1424, 1458), 'logging.FileHandler', 'logging.FileHandler', (['info_filename'], {}), '(info_filename)\n', (1443, 1458), False, 'import logging\n'), ((1585, 1617), 'logging.Formatter', 'logging.Formatter', (['"""%(message)s"""'], {}), "('%(message)s')\n", (1602, 1617), False, 'import logging\n'), ((6381, 6452), 'langchain.schema.AgentFinish', 'AgentFinish', ([], {'return_values': "{'output': 'Task finished.'}", 'log': 'llm_output'}), "(return_values={'output': 'Task finished.'}, log=llm_output)\n", (6392, 6452), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult\n'), ((10667, 10702), 'langchain.callbacks.FileCallbackHandler', 'FileCallbackHandler', (['debug_filename'], {}), '(debug_filename)\n', (10686, 10702), False, 'from langchain.callbacks import FileCallbackHandler\n'), ((14195, 14269), 'langchain.llms.VertexAI', 'VertexAI', ([], {'model_name': 'model', 'temperature': 'temperature', 'max_output_tokens': '(256)'}), '(model_name=model, temperature=temperature, max_output_tokens=256)\n', (14203, 14269), False, 'from langchain.llms import VertexAI\n'), ((14429, 14482), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt', 'callbacks': 'callbacks'}), '(llm=llm, prompt=prompt, callbacks=callbacks)\n', (14437, 14482), False, 'from langchain.chains import LLMChain\n'), ((14802, 14948), 'langchain.agents.AgentExecutor.from_agent_and_tools', 'AgentExecutor.from_agent_and_tools', ([], {'agent': 'agent', 'tools': 'tools', 'verbose': 'langchain_verbose', 'max_iterations': '(2 * MAX_STEPS * context.max_votes + 1)'}), '(agent=agent, tools=tools, verbose=\n langchain_verbose, max_iterations=2 * MAX_STEPS * context.max_votes + 1)\n', (14836, 14948), False, 'from langchain.agents import AgentExecutor, LLMSingleActionAgent, AgentOutputParser\n'), ((997, 1011), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (1009, 1011), False, 'from datetime import datetime\n'), ((6714, 6755), 're.search', 're.search', (['pattern', 'llm_output', 're.DOTALL'], {}), '(pattern, llm_output, re.DOTALL)\n', (6723, 6755), False, 'import re\n'), ((7875, 7965), 'langchain.schema.AgentAction', 'AgentAction', ([], {'tool': 'TAKE_ENVIRONMENT_ACTION', 'tool_input': '"""empty param"""', 'log': '"""empty tool"""'}), "(tool=TAKE_ENVIRONMENT_ACTION, tool_input='empty param', log=\n 'empty tool')\n", (7886, 7965), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult\n'), ((11956, 11996), 'tools.take_environment_action_wrapper', 'take_environment_action_wrapper', (['context'], {}), '(context)\n', (11987, 11996), False, 'from tools import take_environment_action_wrapper\n'), ((16594, 16621), 'os.path.exists', 'os.path.exists', (['results_dir'], {}), '(results_dir)\n', (16608, 16621), False, 'import os\n'), ((16635, 16659), 'os.makedirs', 'os.makedirs', (['results_dir'], {}), '(results_dir)\n', (16646, 16659), False, 'import os\n'), ((1290, 1311), 'logging.NullHandler', 'logging.NullHandler', ([], {}), '()\n', (1309, 1311), False, 'import logging\n'), ((12054, 12119), 'debate.view_debate_wrapper', 'view_debate_wrapper', (['context'], {'logger': 'info_logger'}), '(context, **debate_params, logger=info_logger)\n', (12073, 12119), False, 'from debate import view_debate_wrapper\n'), ((8911, 9007), 'langchain.schema.AgentAction', 'AgentAction', ([], {'tool': 'majority_tool_name', 'tool_input': 'majority_tool_input', 'log': 'random_llm_output'}), '(tool=majority_tool_name, tool_input=majority_tool_input, log=\n random_llm_output)\n', (8922, 9007), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult\n'), ((9578, 9674), 'langchain.schema.AgentAction', 'AgentAction', ([], {'tool': 'majority_tool_name', 'tool_input': 'majority_tool_input', 'log': 'random_llm_output'}), '(tool=majority_tool_name, tool_input=majority_tool_input, log=\n random_llm_output)\n', (9589, 9674), False, 'from langchain.schema import AgentAction, AgentFinish, BaseMessage, LLMResult\n')]
"""Base interface that all chains should implement.""" import inspect import json import logging import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import Field, root_validator, validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks, ) from langchain.load.dump import dumpd from langchain.load.serializable import Serializable from langchain.schema import RUN_KEY, BaseMemory, RunInfo logger = logging.getLogger(__name__) def _get_verbosity() -> bool: return langchain.verbose class Chain(Serializable, ABC): """Abstract base class for creating structured sequences of calls to components. Chains should be used to encode a sequence of calls to components like models, document retrievers, other chains, etc., and provide a simple interface to this sequence. The Chain interface makes it easy to create apps that are: - Stateful: add Memory to any Chain to give it state, - Observable: pass Callbacks to a Chain to execute additional functionality, like logging, outside the main sequence of component calls, - Composable: the Chain API is flexible enough that it is easy to combine Chains with other components, including other Chains. The main methods exposed by chains are: - `__call__`: Chains are callable. The `__call__` method is the primary way to execute a Chain. This takes inputs as a dictionary and returns a dictionary output. - `run`: A convenience method that takes inputs as args/kwargs and returns the output as a string. This method can only be used for a subset of chains and cannot return as rich of an output as `__call__`. """ memory: Optional[BaseMemory] = None """Optional memory object. Defaults to None. Memory is a class that gets called at the start and at the end of every chain. At the start, memory loads variables and passes them along in the chain. At the end, it saves any returned variables. There are many different types of memory - please see memory docs for the full catalog.""" callbacks: Callbacks = Field(default=None, exclude=True) """Optional list of callback handlers (or callback manager). Defaults to None. Callback handlers are called throughout the lifecycle of a call to a chain, starting with on_chain_start, ending with on_chain_end or on_chain_error. Each custom chain can optionally call additional callback methods, see Callback docs for full details.""" callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) """Deprecated, use `callbacks` instead.""" verbose: bool = Field(default_factory=_get_verbosity) """Whether or not run in verbose mode. In verbose mode, some intermediate logs will be printed to the console. Defaults to `langchain.verbose` value.""" tags: Optional[List[str]] = None """Optional list of tags associated with the chain. Defaults to None These tags will be associated with each call to this chain, and passed as arguments to the handlers defined in `callbacks`. You can use these to eg identify a specific instance of a chain with its use case. """ metadata: Optional[Dict[str, Any]] = None """Optional metadata associated with the chain. Defaults to None This metadata will be associated with each call to this chain, and passed as arguments to the handlers defined in `callbacks`. You can use these to eg identify a specific instance of a chain with its use case. """ class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @root_validator() def raise_callback_manager_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """Set the chain verbosity. Defaults to the global setting if not specified by the user. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Return the keys expected to be in the chain input.""" @property @abstractmethod def output_keys(self) -> List[str]: """Return the keys expected to be in the chain output.""" def _validate_inputs(self, inputs: Dict[str, Any]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, Any]) -> None: missing_keys = set(self.output_keys).difference(outputs) if missing_keys: raise ValueError(f"Missing some output keys: {missing_keys}") @abstractmethod def _call( self, inputs: Dict[str, Any], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Execute the chain. This is a private method that is not user-facing. It is only called within `Chain.__call__`, which is the user-facing wrapper method that handles callbacks configuration and some input/output processing. Args: inputs: A dict of named inputs to the chain. Assumed to contain all inputs specified in `Chain.input_keys`, including any inputs added by memory. run_manager: The callbacks manager that contains the callback handlers for this run of the chain. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ async def _acall( self, inputs: Dict[str, Any], run_manager: Optional[AsyncCallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Asynchronously execute the chain. This is a private method that is not user-facing. It is only called within `Chain.acall`, which is the user-facing wrapper method that handles callbacks configuration and some input/output processing. Args: inputs: A dict of named inputs to the chain. Assumed to contain all inputs specified in `Chain.input_keys`, including any inputs added by memory. run_manager: The callbacks manager that contains the callback handlers for this run of the chain. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Execute the chain. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. return_only_outputs: Whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. metadata: Optional metadata associated with the chain. Defaults to None include_run_info: Whether to include run info in the response. Defaults to False. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ inputs = self.prep_inputs(inputs) callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") run_manager = callback_manager.on_chain_start( dumpd(self), inputs, ) try: outputs = ( self._call(inputs, run_manager=run_manager) if new_arg_supported else self._call(inputs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_chain_error(e) raise e run_manager.on_chain_end(outputs) final_outputs: Dict[str, Any] = self.prep_outputs( inputs, outputs, return_only_outputs ) if include_run_info: final_outputs[RUN_KEY] = RunInfo(run_id=run_manager.run_id) return final_outputs async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Asynchronously execute the chain. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. return_only_outputs: Whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. metadata: Optional metadata associated with the chain. Defaults to None include_run_info: Whether to include run info in the response. Defaults to False. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ inputs = self.prep_inputs(inputs) callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") run_manager = await callback_manager.on_chain_start( dumpd(self), inputs, ) try: outputs = ( await self._acall(inputs, run_manager=run_manager) if new_arg_supported else await self._acall(inputs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_chain_error(e) raise e await run_manager.on_chain_end(outputs) final_outputs: Dict[str, Any] = self.prep_outputs( inputs, outputs, return_only_outputs ) if include_run_info: final_outputs[RUN_KEY] = RunInfo(run_id=run_manager.run_id) return final_outputs def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prepare chain outputs, and save info about this run to memory. Args: inputs: Dictionary of chain inputs, including any inputs added by chain memory. outputs: Dictionary of initial chain outputs. return_only_outputs: Whether to only return the chain outputs. If False, inputs are also added to the final outputs. Returns: A dict of the final chain outputs. """ self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prepare chain inputs, including adding inputs from memory. Args: inputs: Dictionary of raw inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. Returns: A dictionary of all inputs, including those added by the chain's memory. """ if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs @property def _run_output_key(self) -> str: if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) return self.output_keys[0] def run( self, *args: Any, callbacks: Callbacks = None, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> str: """Convenience method for executing chain when there's a single string output. The main difference between this method and `Chain.__call__` is that this method can only be used for chains that return a single string output. If a Chain has more outputs, a non-string output, or you want to return the inputs/run info along with the outputs, use `Chain.__call__`. The other difference is that this method expects inputs to be passed directly in as positional arguments or keyword arguments, whereas `Chain.__call__` expects a single input dictionary with all the inputs. Args: *args: If the chain expects a single input, it can be passed in as the sole positional argument. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. **kwargs: If the chain expects multiple inputs, they can be passed in directly as keyword arguments. Returns: The chain output as a string. Example: .. code-block:: python # Suppose we have a single-input chain that takes a 'question' string: chain.run("What's the temperature in Boise, Idaho?") # -> "The temperature in Boise is..." # Suppose we have a multi-input chain that takes a 'question' string # and 'context' string: question = "What's the temperature in Boise, Idaho?" context = "Weather report for Boise, Idaho on 07/03/23..." chain.run(question=question, context=context) # -> "The temperature in Boise is..." """ # Run at start to make sure this is possible/defined _output_key = self._run_output_key if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0], callbacks=callbacks, tags=tags, metadata=metadata)[ _output_key ] if kwargs and not args: return self(kwargs, callbacks=callbacks, tags=tags, metadata=metadata)[ _output_key ] if not kwargs and not args: raise ValueError( "`run` supported with either positional arguments or keyword arguments," " but none were provided." ) else: raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun( self, *args: Any, callbacks: Callbacks = None, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> str: """Convenience method for executing chain when there's a single string output. The main difference between this method and `Chain.__call__` is that this method can only be used for chains that return a single string output. If a Chain has more outputs, a non-string output, or you want to return the inputs/run info along with the outputs, use `Chain.__call__`. The other difference is that this method expects inputs to be passed directly in as positional arguments or keyword arguments, whereas `Chain.__call__` expects a single input dictionary with all the inputs. Args: *args: If the chain expects a single input, it can be passed in as the sole positional argument. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. **kwargs: If the chain expects multiple inputs, they can be passed in directly as keyword arguments. Returns: The chain output as a string. Example: .. code-block:: python # Suppose we have a single-input chain that takes a 'question' string: await chain.arun("What's the temperature in Boise, Idaho?") # -> "The temperature in Boise is..." # Suppose we have a multi-input chain that takes a 'question' string # and 'context' string: question = "What's the temperature in Boise, Idaho?" context = "Weather report for Boise, Idaho on 07/03/23..." await chain.arun(question=question, context=context) # -> "The temperature in Boise is..." """ if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) elif args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return ( await self.acall( args[0], callbacks=callbacks, tags=tags, metadata=metadata ) )[self.output_keys[0]] if kwargs and not args: return ( await self.acall( kwargs, callbacks=callbacks, tags=tags, metadata=metadata ) )[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain. Expects `Chain._chain_type` property to be implemented and for memory to be null. Args: **kwargs: Keyword arguments passed to default `pydantic.BaseModel.dict` method. Returns: A dictionary representation of the chain. Example: ..code-block:: python chain.dict(exclude_unset=True) # -> {"_type": "foo", "verbose": False, ...} """ if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict(**kwargs) _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Expects `Chain._chain_type` property to be implemented and for memory to be null. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") def apply( self, input_list: List[Dict[str, Any]], callbacks: Callbacks = None ) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs, callbacks=callbacks) for inputs in input_list]
[ "langchain.schema.RunInfo", "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.load.dump.dumpd", "langchain.callbacks.manager.CallbackManager.configure" ]
[((702, 729), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (719, 729), False, 'import logging\n'), ((2435, 2468), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2440, 2468), False, 'from pydantic import Field, root_validator, validator\n'), ((2878, 2911), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2883, 2911), False, 'from pydantic import Field, root_validator, validator\n'), ((2979, 3016), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2984, 3016), False, 'from pydantic import Field, root_validator, validator\n'), ((4107, 4123), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (4121, 4123), False, 'from pydantic import Field, root_validator, validator\n'), ((4576, 4619), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (4585, 4619), False, 'from pydantic import Field, root_validator, validator\n'), ((9402, 9514), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (9427, 9514), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks\n'), ((12243, 12360), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (12273, 12360), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks\n'), ((4331, 4433), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (4344, 4433), False, 'import warnings\n'), ((9761, 9772), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (9766, 9772), False, 'from langchain.load.dump import dumpd\n'), ((10332, 10366), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (10339, 10366), False, 'from langchain.schema import RUN_KEY, BaseMemory, RunInfo\n'), ((13211, 13245), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (13218, 13245), False, 'from langchain.schema import RUN_KEY, BaseMemory, RunInfo\n'), ((23986, 24001), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (23990, 24001), False, 'from pathlib import Path\n'), ((12614, 12625), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (12619, 12625), False, 'from langchain.load.dump import dumpd\n'), ((24321, 24355), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (24330, 24355), False, 'import json\n'), ((9634, 9663), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (9651, 9663), False, 'import inspect\n'), ((12480, 12510), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (12497, 12510), False, 'import inspect\n'), ((24458, 24508), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (24467, 24508), False, 'import yaml\n')]
"""Base interface that all chains should implement.""" import inspect import json import logging import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import Field, root_validator, validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks, ) from langchain.load.dump import dumpd from langchain.load.serializable import Serializable from langchain.schema import RUN_KEY, BaseMemory, RunInfo logger = logging.getLogger(__name__) def _get_verbosity() -> bool: return langchain.verbose class Chain(Serializable, ABC): """Abstract base class for creating structured sequences of calls to components. Chains should be used to encode a sequence of calls to components like models, document retrievers, other chains, etc., and provide a simple interface to this sequence. The Chain interface makes it easy to create apps that are: - Stateful: add Memory to any Chain to give it state, - Observable: pass Callbacks to a Chain to execute additional functionality, like logging, outside the main sequence of component calls, - Composable: the Chain API is flexible enough that it is easy to combine Chains with other components, including other Chains. The main methods exposed by chains are: - `__call__`: Chains are callable. The `__call__` method is the primary way to execute a Chain. This takes inputs as a dictionary and returns a dictionary output. - `run`: A convenience method that takes inputs as args/kwargs and returns the output as a string. This method can only be used for a subset of chains and cannot return as rich of an output as `__call__`. """ memory: Optional[BaseMemory] = None """Optional memory object. Defaults to None. Memory is a class that gets called at the start and at the end of every chain. At the start, memory loads variables and passes them along in the chain. At the end, it saves any returned variables. There are many different types of memory - please see memory docs for the full catalog.""" callbacks: Callbacks = Field(default=None, exclude=True) """Optional list of callback handlers (or callback manager). Defaults to None. Callback handlers are called throughout the lifecycle of a call to a chain, starting with on_chain_start, ending with on_chain_end or on_chain_error. Each custom chain can optionally call additional callback methods, see Callback docs for full details.""" callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) """Deprecated, use `callbacks` instead.""" verbose: bool = Field(default_factory=_get_verbosity) """Whether or not run in verbose mode. In verbose mode, some intermediate logs will be printed to the console. Defaults to `langchain.verbose` value.""" tags: Optional[List[str]] = None """Optional list of tags associated with the chain. Defaults to None These tags will be associated with each call to this chain, and passed as arguments to the handlers defined in `callbacks`. You can use these to eg identify a specific instance of a chain with its use case. """ metadata: Optional[Dict[str, Any]] = None """Optional metadata associated with the chain. Defaults to None This metadata will be associated with each call to this chain, and passed as arguments to the handlers defined in `callbacks`. You can use these to eg identify a specific instance of a chain with its use case. """ class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @root_validator() def raise_callback_manager_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """Set the chain verbosity. Defaults to the global setting if not specified by the user. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Return the keys expected to be in the chain input.""" @property @abstractmethod def output_keys(self) -> List[str]: """Return the keys expected to be in the chain output.""" def _validate_inputs(self, inputs: Dict[str, Any]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, Any]) -> None: missing_keys = set(self.output_keys).difference(outputs) if missing_keys: raise ValueError(f"Missing some output keys: {missing_keys}") @abstractmethod def _call( self, inputs: Dict[str, Any], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Execute the chain. This is a private method that is not user-facing. It is only called within `Chain.__call__`, which is the user-facing wrapper method that handles callbacks configuration and some input/output processing. Args: inputs: A dict of named inputs to the chain. Assumed to contain all inputs specified in `Chain.input_keys`, including any inputs added by memory. run_manager: The callbacks manager that contains the callback handlers for this run of the chain. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ async def _acall( self, inputs: Dict[str, Any], run_manager: Optional[AsyncCallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Asynchronously execute the chain. This is a private method that is not user-facing. It is only called within `Chain.acall`, which is the user-facing wrapper method that handles callbacks configuration and some input/output processing. Args: inputs: A dict of named inputs to the chain. Assumed to contain all inputs specified in `Chain.input_keys`, including any inputs added by memory. run_manager: The callbacks manager that contains the callback handlers for this run of the chain. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Execute the chain. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. return_only_outputs: Whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. metadata: Optional metadata associated with the chain. Defaults to None include_run_info: Whether to include run info in the response. Defaults to False. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ inputs = self.prep_inputs(inputs) callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") run_manager = callback_manager.on_chain_start( dumpd(self), inputs, ) try: outputs = ( self._call(inputs, run_manager=run_manager) if new_arg_supported else self._call(inputs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_chain_error(e) raise e run_manager.on_chain_end(outputs) final_outputs: Dict[str, Any] = self.prep_outputs( inputs, outputs, return_only_outputs ) if include_run_info: final_outputs[RUN_KEY] = RunInfo(run_id=run_manager.run_id) return final_outputs async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Asynchronously execute the chain. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. return_only_outputs: Whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. metadata: Optional metadata associated with the chain. Defaults to None include_run_info: Whether to include run info in the response. Defaults to False. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ inputs = self.prep_inputs(inputs) callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") run_manager = await callback_manager.on_chain_start( dumpd(self), inputs, ) try: outputs = ( await self._acall(inputs, run_manager=run_manager) if new_arg_supported else await self._acall(inputs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_chain_error(e) raise e await run_manager.on_chain_end(outputs) final_outputs: Dict[str, Any] = self.prep_outputs( inputs, outputs, return_only_outputs ) if include_run_info: final_outputs[RUN_KEY] = RunInfo(run_id=run_manager.run_id) return final_outputs def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prepare chain outputs, and save info about this run to memory. Args: inputs: Dictionary of chain inputs, including any inputs added by chain memory. outputs: Dictionary of initial chain outputs. return_only_outputs: Whether to only return the chain outputs. If False, inputs are also added to the final outputs. Returns: A dict of the final chain outputs. """ self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prepare chain inputs, including adding inputs from memory. Args: inputs: Dictionary of raw inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. Returns: A dictionary of all inputs, including those added by the chain's memory. """ if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs @property def _run_output_key(self) -> str: if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) return self.output_keys[0] def run( self, *args: Any, callbacks: Callbacks = None, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> str: """Convenience method for executing chain when there's a single string output. The main difference between this method and `Chain.__call__` is that this method can only be used for chains that return a single string output. If a Chain has more outputs, a non-string output, or you want to return the inputs/run info along with the outputs, use `Chain.__call__`. The other difference is that this method expects inputs to be passed directly in as positional arguments or keyword arguments, whereas `Chain.__call__` expects a single input dictionary with all the inputs. Args: *args: If the chain expects a single input, it can be passed in as the sole positional argument. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. **kwargs: If the chain expects multiple inputs, they can be passed in directly as keyword arguments. Returns: The chain output as a string. Example: .. code-block:: python # Suppose we have a single-input chain that takes a 'question' string: chain.run("What's the temperature in Boise, Idaho?") # -> "The temperature in Boise is..." # Suppose we have a multi-input chain that takes a 'question' string # and 'context' string: question = "What's the temperature in Boise, Idaho?" context = "Weather report for Boise, Idaho on 07/03/23..." chain.run(question=question, context=context) # -> "The temperature in Boise is..." """ # Run at start to make sure this is possible/defined _output_key = self._run_output_key if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0], callbacks=callbacks, tags=tags, metadata=metadata)[ _output_key ] if kwargs and not args: return self(kwargs, callbacks=callbacks, tags=tags, metadata=metadata)[ _output_key ] if not kwargs and not args: raise ValueError( "`run` supported with either positional arguments or keyword arguments," " but none were provided." ) else: raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun( self, *args: Any, callbacks: Callbacks = None, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> str: """Convenience method for executing chain when there's a single string output. The main difference between this method and `Chain.__call__` is that this method can only be used for chains that return a single string output. If a Chain has more outputs, a non-string output, or you want to return the inputs/run info along with the outputs, use `Chain.__call__`. The other difference is that this method expects inputs to be passed directly in as positional arguments or keyword arguments, whereas `Chain.__call__` expects a single input dictionary with all the inputs. Args: *args: If the chain expects a single input, it can be passed in as the sole positional argument. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. **kwargs: If the chain expects multiple inputs, they can be passed in directly as keyword arguments. Returns: The chain output as a string. Example: .. code-block:: python # Suppose we have a single-input chain that takes a 'question' string: await chain.arun("What's the temperature in Boise, Idaho?") # -> "The temperature in Boise is..." # Suppose we have a multi-input chain that takes a 'question' string # and 'context' string: question = "What's the temperature in Boise, Idaho?" context = "Weather report for Boise, Idaho on 07/03/23..." await chain.arun(question=question, context=context) # -> "The temperature in Boise is..." """ if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) elif args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return ( await self.acall( args[0], callbacks=callbacks, tags=tags, metadata=metadata ) )[self.output_keys[0]] if kwargs and not args: return ( await self.acall( kwargs, callbacks=callbacks, tags=tags, metadata=metadata ) )[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain. Expects `Chain._chain_type` property to be implemented and for memory to be null. Args: **kwargs: Keyword arguments passed to default `pydantic.BaseModel.dict` method. Returns: A dictionary representation of the chain. Example: ..code-block:: python chain.dict(exclude_unset=True) # -> {"_type": "foo", "verbose": False, ...} """ if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict(**kwargs) _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Expects `Chain._chain_type` property to be implemented and for memory to be null. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") def apply( self, input_list: List[Dict[str, Any]], callbacks: Callbacks = None ) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs, callbacks=callbacks) for inputs in input_list]
[ "langchain.schema.RunInfo", "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.load.dump.dumpd", "langchain.callbacks.manager.CallbackManager.configure" ]
[((702, 729), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (719, 729), False, 'import logging\n'), ((2435, 2468), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2440, 2468), False, 'from pydantic import Field, root_validator, validator\n'), ((2878, 2911), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2883, 2911), False, 'from pydantic import Field, root_validator, validator\n'), ((2979, 3016), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2984, 3016), False, 'from pydantic import Field, root_validator, validator\n'), ((4107, 4123), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (4121, 4123), False, 'from pydantic import Field, root_validator, validator\n'), ((4576, 4619), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (4585, 4619), False, 'from pydantic import Field, root_validator, validator\n'), ((9402, 9514), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (9427, 9514), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks\n'), ((12243, 12360), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (12273, 12360), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks\n'), ((4331, 4433), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (4344, 4433), False, 'import warnings\n'), ((9761, 9772), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (9766, 9772), False, 'from langchain.load.dump import dumpd\n'), ((10332, 10366), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (10339, 10366), False, 'from langchain.schema import RUN_KEY, BaseMemory, RunInfo\n'), ((13211, 13245), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (13218, 13245), False, 'from langchain.schema import RUN_KEY, BaseMemory, RunInfo\n'), ((23986, 24001), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (23990, 24001), False, 'from pathlib import Path\n'), ((12614, 12625), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (12619, 12625), False, 'from langchain.load.dump import dumpd\n'), ((24321, 24355), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (24330, 24355), False, 'import json\n'), ((9634, 9663), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (9651, 9663), False, 'import inspect\n'), ((12480, 12510), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (12497, 12510), False, 'import inspect\n'), ((24458, 24508), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (24467, 24508), False, 'import yaml\n')]
"""Base interface that all chains should implement.""" import inspect import json import logging import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import Field, root_validator, validator import langchain from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks, ) from langchain.load.dump import dumpd from langchain.load.serializable import Serializable from langchain.schema import RUN_KEY, BaseMemory, RunInfo logger = logging.getLogger(__name__) def _get_verbosity() -> bool: return langchain.verbose class Chain(Serializable, ABC): """Abstract base class for creating structured sequences of calls to components. Chains should be used to encode a sequence of calls to components like models, document retrievers, other chains, etc., and provide a simple interface to this sequence. The Chain interface makes it easy to create apps that are: - Stateful: add Memory to any Chain to give it state, - Observable: pass Callbacks to a Chain to execute additional functionality, like logging, outside the main sequence of component calls, - Composable: the Chain API is flexible enough that it is easy to combine Chains with other components, including other Chains. The main methods exposed by chains are: - `__call__`: Chains are callable. The `__call__` method is the primary way to execute a Chain. This takes inputs as a dictionary and returns a dictionary output. - `run`: A convenience method that takes inputs as args/kwargs and returns the output as a string. This method can only be used for a subset of chains and cannot return as rich of an output as `__call__`. """ memory: Optional[BaseMemory] = None """Optional memory object. Defaults to None. Memory is a class that gets called at the start and at the end of every chain. At the start, memory loads variables and passes them along in the chain. At the end, it saves any returned variables. There are many different types of memory - please see memory docs for the full catalog.""" callbacks: Callbacks = Field(default=None, exclude=True) """Optional list of callback handlers (or callback manager). Defaults to None. Callback handlers are called throughout the lifecycle of a call to a chain, starting with on_chain_start, ending with on_chain_end or on_chain_error. Each custom chain can optionally call additional callback methods, see Callback docs for full details.""" callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) """Deprecated, use `callbacks` instead.""" verbose: bool = Field(default_factory=_get_verbosity) """Whether or not run in verbose mode. In verbose mode, some intermediate logs will be printed to the console. Defaults to `langchain.verbose` value.""" tags: Optional[List[str]] = None """Optional list of tags associated with the chain. Defaults to None These tags will be associated with each call to this chain, and passed as arguments to the handlers defined in `callbacks`. You can use these to eg identify a specific instance of a chain with its use case. """ metadata: Optional[Dict[str, Any]] = None """Optional metadata associated with the chain. Defaults to None This metadata will be associated with each call to this chain, and passed as arguments to the handlers defined in `callbacks`. You can use these to eg identify a specific instance of a chain with its use case. """ class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @root_validator() def raise_callback_manager_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """Set the chain verbosity. Defaults to the global setting if not specified by the user. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Return the keys expected to be in the chain input.""" @property @abstractmethod def output_keys(self) -> List[str]: """Return the keys expected to be in the chain output.""" def _validate_inputs(self, inputs: Dict[str, Any]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, Any]) -> None: missing_keys = set(self.output_keys).difference(outputs) if missing_keys: raise ValueError(f"Missing some output keys: {missing_keys}") @abstractmethod def _call( self, inputs: Dict[str, Any], run_manager: Optional[CallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Execute the chain. This is a private method that is not user-facing. It is only called within `Chain.__call__`, which is the user-facing wrapper method that handles callbacks configuration and some input/output processing. Args: inputs: A dict of named inputs to the chain. Assumed to contain all inputs specified in `Chain.input_keys`, including any inputs added by memory. run_manager: The callbacks manager that contains the callback handlers for this run of the chain. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ async def _acall( self, inputs: Dict[str, Any], run_manager: Optional[AsyncCallbackManagerForChainRun] = None, ) -> Dict[str, Any]: """Asynchronously execute the chain. This is a private method that is not user-facing. It is only called within `Chain.acall`, which is the user-facing wrapper method that handles callbacks configuration and some input/output processing. Args: inputs: A dict of named inputs to the chain. Assumed to contain all inputs specified in `Chain.input_keys`, including any inputs added by memory. run_manager: The callbacks manager that contains the callback handlers for this run of the chain. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Execute the chain. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. return_only_outputs: Whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. metadata: Optional metadata associated with the chain. Defaults to None include_run_info: Whether to include run info in the response. Defaults to False. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ inputs = self.prep_inputs(inputs) callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") run_manager = callback_manager.on_chain_start( dumpd(self), inputs, ) try: outputs = ( self._call(inputs, run_manager=run_manager) if new_arg_supported else self._call(inputs) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_chain_error(e) raise e run_manager.on_chain_end(outputs) final_outputs: Dict[str, Any] = self.prep_outputs( inputs, outputs, return_only_outputs ) if include_run_info: final_outputs[RUN_KEY] = RunInfo(run_id=run_manager.run_id) return final_outputs async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False, callbacks: Callbacks = None, *, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, include_run_info: bool = False, ) -> Dict[str, Any]: """Asynchronously execute the chain. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. return_only_outputs: Whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. metadata: Optional metadata associated with the chain. Defaults to None include_run_info: Whether to include run info in the response. Defaults to False. Returns: A dict of named outputs. Should contain all outputs specified in `Chain.output_keys`. """ inputs = self.prep_inputs(inputs) callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose, tags, self.tags, metadata, self.metadata, ) new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") run_manager = await callback_manager.on_chain_start( dumpd(self), inputs, ) try: outputs = ( await self._acall(inputs, run_manager=run_manager) if new_arg_supported else await self._acall(inputs) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_chain_error(e) raise e await run_manager.on_chain_end(outputs) final_outputs: Dict[str, Any] = self.prep_outputs( inputs, outputs, return_only_outputs ) if include_run_info: final_outputs[RUN_KEY] = RunInfo(run_id=run_manager.run_id) return final_outputs def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prepare chain outputs, and save info about this run to memory. Args: inputs: Dictionary of chain inputs, including any inputs added by chain memory. outputs: Dictionary of initial chain outputs. return_only_outputs: Whether to only return the chain outputs. If False, inputs are also added to the final outputs. Returns: A dict of the final chain outputs. """ self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prepare chain inputs, including adding inputs from memory. Args: inputs: Dictionary of raw inputs, or single input if chain expects only one param. Should contain all inputs specified in `Chain.input_keys` except for inputs that will be set by the chain's memory. Returns: A dictionary of all inputs, including those added by the chain's memory. """ if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs @property def _run_output_key(self) -> str: if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) return self.output_keys[0] def run( self, *args: Any, callbacks: Callbacks = None, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> str: """Convenience method for executing chain when there's a single string output. The main difference between this method and `Chain.__call__` is that this method can only be used for chains that return a single string output. If a Chain has more outputs, a non-string output, or you want to return the inputs/run info along with the outputs, use `Chain.__call__`. The other difference is that this method expects inputs to be passed directly in as positional arguments or keyword arguments, whereas `Chain.__call__` expects a single input dictionary with all the inputs. Args: *args: If the chain expects a single input, it can be passed in as the sole positional argument. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. **kwargs: If the chain expects multiple inputs, they can be passed in directly as keyword arguments. Returns: The chain output as a string. Example: .. code-block:: python # Suppose we have a single-input chain that takes a 'question' string: chain.run("What's the temperature in Boise, Idaho?") # -> "The temperature in Boise is..." # Suppose we have a multi-input chain that takes a 'question' string # and 'context' string: question = "What's the temperature in Boise, Idaho?" context = "Weather report for Boise, Idaho on 07/03/23..." chain.run(question=question, context=context) # -> "The temperature in Boise is..." """ # Run at start to make sure this is possible/defined _output_key = self._run_output_key if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0], callbacks=callbacks, tags=tags, metadata=metadata)[ _output_key ] if kwargs and not args: return self(kwargs, callbacks=callbacks, tags=tags, metadata=metadata)[ _output_key ] if not kwargs and not args: raise ValueError( "`run` supported with either positional arguments or keyword arguments," " but none were provided." ) else: raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun( self, *args: Any, callbacks: Callbacks = None, tags: Optional[List[str]] = None, metadata: Optional[Dict[str, Any]] = None, **kwargs: Any, ) -> str: """Convenience method for executing chain when there's a single string output. The main difference between this method and `Chain.__call__` is that this method can only be used for chains that return a single string output. If a Chain has more outputs, a non-string output, or you want to return the inputs/run info along with the outputs, use `Chain.__call__`. The other difference is that this method expects inputs to be passed directly in as positional arguments or keyword arguments, whereas `Chain.__call__` expects a single input dictionary with all the inputs. Args: *args: If the chain expects a single input, it can be passed in as the sole positional argument. callbacks: Callbacks to use for this chain run. These will be called in addition to callbacks passed to the chain during construction, but only these runtime callbacks will propagate to calls to other objects. tags: List of string tags to pass to all callbacks. These will be passed in addition to tags passed to the chain during construction, but only these runtime tags will propagate to calls to other objects. **kwargs: If the chain expects multiple inputs, they can be passed in directly as keyword arguments. Returns: The chain output as a string. Example: .. code-block:: python # Suppose we have a single-input chain that takes a 'question' string: await chain.arun("What's the temperature in Boise, Idaho?") # -> "The temperature in Boise is..." # Suppose we have a multi-input chain that takes a 'question' string # and 'context' string: question = "What's the temperature in Boise, Idaho?" context = "Weather report for Boise, Idaho on 07/03/23..." await chain.arun(question=question, context=context) # -> "The temperature in Boise is..." """ if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) elif args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return ( await self.acall( args[0], callbacks=callbacks, tags=tags, metadata=metadata ) )[self.output_keys[0]] if kwargs and not args: return ( await self.acall( kwargs, callbacks=callbacks, tags=tags, metadata=metadata ) )[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain. Expects `Chain._chain_type` property to be implemented and for memory to be null. Args: **kwargs: Keyword arguments passed to default `pydantic.BaseModel.dict` method. Returns: A dictionary representation of the chain. Example: ..code-block:: python chain.dict(exclude_unset=True) # -> {"_type": "foo", "verbose": False, ...} """ if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict(**kwargs) _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Expects `Chain._chain_type` property to be implemented and for memory to be null. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") def apply( self, input_list: List[Dict[str, Any]], callbacks: Callbacks = None ) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs, callbacks=callbacks) for inputs in input_list]
[ "langchain.schema.RunInfo", "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.load.dump.dumpd", "langchain.callbacks.manager.CallbackManager.configure" ]
[((702, 729), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (719, 729), False, 'import logging\n'), ((2435, 2468), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2440, 2468), False, 'from pydantic import Field, root_validator, validator\n'), ((2878, 2911), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2883, 2911), False, 'from pydantic import Field, root_validator, validator\n'), ((2979, 3016), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2984, 3016), False, 'from pydantic import Field, root_validator, validator\n'), ((4107, 4123), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (4121, 4123), False, 'from pydantic import Field, root_validator, validator\n'), ((4576, 4619), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (4585, 4619), False, 'from pydantic import Field, root_validator, validator\n'), ((9402, 9514), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose, tags,\n self.tags, metadata, self.metadata)\n', (9427, 9514), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks\n'), ((12243, 12360), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose', 'tags', 'self.tags', 'metadata', 'self.metadata'], {}), '(callbacks, self.callbacks, self.verbose,\n tags, self.tags, metadata, self.metadata)\n', (12273, 12360), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForChainRun, CallbackManager, CallbackManagerForChainRun, Callbacks\n'), ((4331, 4433), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (4344, 4433), False, 'import warnings\n'), ((9761, 9772), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (9766, 9772), False, 'from langchain.load.dump import dumpd\n'), ((10332, 10366), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (10339, 10366), False, 'from langchain.schema import RUN_KEY, BaseMemory, RunInfo\n'), ((13211, 13245), 'langchain.schema.RunInfo', 'RunInfo', ([], {'run_id': 'run_manager.run_id'}), '(run_id=run_manager.run_id)\n', (13218, 13245), False, 'from langchain.schema import RUN_KEY, BaseMemory, RunInfo\n'), ((23986, 24001), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (23990, 24001), False, 'from pathlib import Path\n'), ((12614, 12625), 'langchain.load.dump.dumpd', 'dumpd', (['self'], {}), '(self)\n', (12619, 12625), False, 'from langchain.load.dump import dumpd\n'), ((24321, 24355), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (24330, 24355), False, 'import json\n'), ((9634, 9663), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (9651, 9663), False, 'import inspect\n'), ((12480, 12510), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (12497, 12510), False, 'import inspect\n'), ((24458, 24508), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (24467, 24508), False, 'import yaml\n')]
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import functools import inspect import logging import uuid from enum import Enum from typing import ( TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast, ) from langsmith.client import Client from langsmith.evaluation import RunEvaluator from langsmith.run_helpers import as_runnable, is_traceable_function from langsmith.schemas import Dataset, DataType, Example from langsmith.utils import LangSmithError from requests import HTTPError from langchain._api import warn_deprecated from langchain.callbacks.manager import Callbacks from langchain.callbacks.tracers.evaluation import ( EvaluatorCallbackHandler, wait_for_all_evaluators, ) from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.chains.base import Chain from langchain.evaluation.loading import load_evaluator from langchain.evaluation.schema import ( EvaluatorType, PairwiseStringEvaluator, StringEvaluator, ) from langchain.schema import ChatResult, LLMResult from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import BaseMessage, messages_from_dict from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda from langchain.schema.runnable import config as runnable_config from langchain.schema.runnable import utils as runnable_utils from langchain.smith import evaluation as smith_eval from langchain.smith.evaluation import config as smith_eval_config from langchain.smith.evaluation import name_generation, progress if TYPE_CHECKING: import pandas as pd logger = logging.getLogger(__name__) MODEL_OR_CHAIN_FACTORY = Union[ Callable[[], Union[Chain, Runnable]], BaseLanguageModel, Callable[[dict], Any], Runnable, Chain, ] MCF = Union[Callable[[], Union[Chain, Runnable]], BaseLanguageModel] class InputFormatError(Exception): """Raised when the input format is invalid.""" ## Shared Utilities class TestResult(dict): """A dictionary of the results of a single test run.""" def get_aggregate_feedback( self, quantiles: Optional[Sequence[float]] = None ) -> pd.DataFrame: """Return quantiles for the feedback scores. This method calculates and prints the quantiles for the feedback scores across all feedback keys. Returns: A DataFrame containing the quantiles for each feedback key. """ df = self.to_dataframe() feedback_cols = [ col for col in df.columns if col not in ["input", "output", "reference"] ] _quantiles = df[feedback_cols].quantile( quantiles or [0.25, 0.5, 0.75], numeric_only=True ) _quantiles.loc["mean"] = df[feedback_cols].mean() _quantiles.loc["mode"] = df[feedback_cols].mode().iloc[0] return _quantiles.transpose() def to_dataframe(self) -> pd.DataFrame: """Convert the results to a dataframe.""" try: import pandas as pd except ImportError as e: raise ImportError( "Pandas is required to convert the results to a dataframe." " to install pandas, run `pip install pandas`." ) from e indices = [] records = [] for example_id, result in self["results"].items(): feedback = result["feedback"] r = { **{f.key: f.score for f in feedback}, "input": result["input"], "output": result["output"], } if "reference" in result: r["reference"] = result["reference"] records.append(r) indices.append(example_id) return pd.DataFrame(records, index=indices) def _wrap_in_chain_factory( llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, dataset_name: str = "<my_dataset>", ) -> MCF: """Forgive the user if they pass in a chain without memory instead of a chain factory. It's a common mistake. Raise a more helpful error message as well.""" if isinstance(llm_or_chain_factory, Chain): chain = llm_or_chain_factory chain_class = chain.__class__.__name__ if llm_or_chain_factory.memory is not None: memory_class = chain.memory.__class__.__name__ raise ValueError( "Cannot directly evaluate a chain with stateful memory." " To evaluate this chain, pass in a chain constructor" " that initializes fresh memory each time it is called." " This will safegaurd against information" " leakage between dataset examples." "\nFor example:\n\n" "def chain_constructor():\n" f" new_memory = {memory_class}(...)\n" f" return {chain_class}" "(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) return lambda: chain elif isinstance(llm_or_chain_factory, BaseLanguageModel): return llm_or_chain_factory elif isinstance(llm_or_chain_factory, Runnable): # Memory may exist here, but it's not elegant to check all those cases. lcf = llm_or_chain_factory return lambda: lcf elif callable(llm_or_chain_factory): if is_traceable_function(llm_or_chain_factory): runnable_ = as_runnable(cast(Callable, llm_or_chain_factory)) return lambda: runnable_ try: _model = llm_or_chain_factory() # type: ignore[call-arg] except TypeError: # It's an arbitrary function, wrap it in a RunnableLambda user_func = cast(Callable, llm_or_chain_factory) sig = inspect.signature(user_func) logger.info(f"Wrapping function {sig} as RunnableLambda.") wrapped = RunnableLambda(user_func) return lambda: wrapped constructor = cast(Callable, llm_or_chain_factory) if isinstance(_model, BaseLanguageModel): # It's not uncommon to do an LLM constructor instead of raw LLM, # so we'll unpack it for the user. return _model elif is_traceable_function(cast(Callable, _model)): runnable_ = as_runnable(cast(Callable, _model)) return lambda: runnable_ elif not isinstance(_model, Runnable): # This is unlikely to happen - a constructor for a model function return lambda: RunnableLambda(constructor) else: # Typical correct case return constructor # noqa return llm_or_chain_factory def _get_prompt(inputs: Dict[str, Any]) -> str: """Get prompt from inputs. Args: inputs: The input dictionary. Returns: A string prompt. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") prompts = [] if "prompt" in inputs: if not isinstance(inputs["prompt"], str): raise InputFormatError( "Expected string for 'prompt', got" f" {type(inputs['prompt']).__name__}" ) prompts = [inputs["prompt"]] elif "prompts" in inputs: if not isinstance(inputs["prompts"], list) or not all( isinstance(i, str) for i in inputs["prompts"] ): raise InputFormatError( "Expected list of strings for 'prompts'," f" got {type(inputs['prompts']).__name__}" ) prompts = inputs["prompts"] elif len(inputs) == 1: prompt_ = next(iter(inputs.values())) if isinstance(prompt_, str): prompts = [prompt_] elif isinstance(prompt_, list) and all(isinstance(i, str) for i in prompt_): prompts = prompt_ else: raise InputFormatError(f"LLM Run expects string prompt input. Got {inputs}") else: raise InputFormatError( f"LLM Run expects 'prompt' or 'prompts' in inputs. Got {inputs}" ) if len(prompts) == 1: return prompts[0] else: raise InputFormatError( f"LLM Run expects single prompt input. Got {len(prompts)} prompts." ) def _get_messages(inputs: Dict[str, Any]) -> List[BaseMessage]: """Get Chat Messages from inputs. Args: inputs: The input dictionary. Returns: A list of chat messages. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") if "messages" in inputs: single_input = inputs["messages"] elif len(inputs) == 1: single_input = next(iter(inputs.values())) else: raise InputFormatError( f"Chat Run expects 'messages' in inputs when example has multiple" f" input keys. Got {inputs}" ) if isinstance(single_input, list) and all( isinstance(i, dict) for i in single_input ): raw_messages = [single_input] elif isinstance(single_input, list) and all( isinstance(i, list) for i in single_input ): raw_messages = single_input else: raise InputFormatError( f"Chat Run expects List[dict] or List[List[dict]] values for" f" 'messages' key input. Got {inputs}" ) if len(raw_messages) == 1: return messages_from_dict(raw_messages[0]) else: raise InputFormatError( f"Chat Run expects single List[dict] or List[List[dict]] 'messages'" f" input. Got {len(raw_messages)} messages from inputs {inputs}" ) ## Shared data validation utilities def _validate_example_inputs_for_language_model( first_example: Example, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: if input_mapper: prompt_input = input_mapper(first_example.inputs) if not isinstance(prompt_input, str) and not ( isinstance(prompt_input, list) and all(isinstance(msg, BaseMessage) for msg in prompt_input) ): raise InputFormatError( "When using an input_mapper to prepare dataset example inputs" " for an LLM or chat model, the output must a single string or" " a list of chat messages." f"\nGot: {prompt_input} of type {type(prompt_input)}." ) else: try: _get_prompt(first_example.inputs) except InputFormatError: try: _get_messages(first_example.inputs) except InputFormatError: raise InputFormatError( "Example inputs do not match language model input format. " "Expected a dictionary with messages or a single prompt." f" Got: {first_example.inputs}" " Please update your dataset OR provide an input_mapper" " to convert the example.inputs to a compatible format" " for the llm or chat model you wish to evaluate." ) def _validate_example_inputs_for_chain( first_example: Example, chain: Chain, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: """Validate that the example inputs match the chain input keys.""" if input_mapper: first_inputs = input_mapper(first_example.inputs) missing_keys = set(chain.input_keys).difference(first_inputs) if not isinstance(first_inputs, dict): raise InputFormatError( "When using an input_mapper to prepare dataset example" " inputs for a chain, the mapped value must be a dictionary." f"\nGot: {first_inputs} of type {type(first_inputs)}." ) if missing_keys: raise InputFormatError( "Missing keys after loading example using input_mapper." f"\nExpected: {chain.input_keys}. Got: {first_inputs.keys()}" ) else: first_inputs = first_example.inputs missing_keys = set(chain.input_keys).difference(first_inputs) if len(first_inputs) == 1 and len(chain.input_keys) == 1: # We can pass this through the run method. # Refrain from calling to validate. pass elif missing_keys: raise InputFormatError( "Example inputs missing expected chain input keys." " Please provide an input_mapper to convert the example.inputs" " to a compatible format for the chain you wish to evaluate." f"Expected: {chain.input_keys}. " f"Got: {first_inputs.keys()}" ) def _validate_example_inputs( example: Example, llm_or_chain_factory: MCF, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: """Validate that the example inputs are valid for the model.""" if isinstance(llm_or_chain_factory, BaseLanguageModel): _validate_example_inputs_for_language_model(example, input_mapper) else: chain = llm_or_chain_factory() if isinstance(chain, Chain): # Otherwise it's a runnable _validate_example_inputs_for_chain(example, chain, input_mapper) elif isinstance(chain, Runnable): logger.debug(f"Skipping input validation for {chain}") ## Shared Evaluator Setup Utilities def _setup_evaluation( llm_or_chain_factory: MCF, examples: List[Example], evaluation: Optional[smith_eval.RunEvalConfig], data_type: DataType, ) -> Optional[List[RunEvaluator]]: """Configure the evaluators to run on the results of the chain.""" if evaluation: if isinstance(llm_or_chain_factory, BaseLanguageModel): run_inputs, run_outputs = None, None run_type = "llm" else: run_type = "chain" if data_type in (DataType.chat, DataType.llm): val = data_type.value if isinstance(data_type, Enum) else data_type raise ValueError( "Cannot evaluate a chain on dataset with " f"data_type={val}. " "Please specify a dataset with the default 'kv' data type." ) chain = llm_or_chain_factory() run_inputs = chain.input_keys if isinstance(chain, Chain) else None run_outputs = chain.output_keys if isinstance(chain, Chain) else None run_evaluators = _load_run_evaluators( evaluation, run_type, data_type, list(examples[0].outputs) if examples[0].outputs else None, run_inputs, run_outputs, ) else: # TODO: Create a default helpfulness evaluator run_evaluators = None return run_evaluators def _determine_input_key( config: smith_eval.RunEvalConfig, run_inputs: Optional[List[str]], ) -> Optional[str]: input_key = None if config.input_key: input_key = config.input_key if run_inputs and input_key not in run_inputs: raise ValueError(f"Input key {input_key} not in run inputs {run_inputs}") elif run_inputs and len(run_inputs) == 1: input_key = run_inputs[0] elif run_inputs is not None and len(run_inputs) > 1: raise ValueError( f"Must specify input key for model with multiple inputs: {run_inputs}" ) return input_key def _determine_prediction_key( config: smith_eval.RunEvalConfig, run_outputs: Optional[List[str]], ) -> Optional[str]: prediction_key = None if config.prediction_key: prediction_key = config.prediction_key if run_outputs and prediction_key not in run_outputs: raise ValueError( f"Prediction key {prediction_key} not in run outputs {run_outputs}" ) elif run_outputs and len(run_outputs) == 1: prediction_key = run_outputs[0] elif run_outputs is not None and len(run_outputs) > 1: raise ValueError( f"Must specify prediction key for model" f" with multiple outputs: {run_outputs}" ) return prediction_key def _determine_reference_key( config: smith_eval.RunEvalConfig, example_outputs: Optional[List[str]], ) -> Optional[str]: if config.reference_key: reference_key = config.reference_key if example_outputs and reference_key not in example_outputs: raise ValueError( f"Reference key {reference_key} not in Dataset" f" example outputs: {example_outputs}" ) elif example_outputs and len(example_outputs) == 1: reference_key = list(example_outputs)[0] else: reference_key = None return reference_key def _construct_run_evaluator( eval_config: Union[EvaluatorType, str, smith_eval_config.EvalConfig], eval_llm: Optional[BaseLanguageModel], run_type: str, data_type: DataType, example_outputs: Optional[List[str]], reference_key: Optional[str], input_key: Optional[str], prediction_key: Optional[str], ) -> RunEvaluator: if isinstance(eval_config, (EvaluatorType, str)): if not isinstance(eval_config, EvaluatorType): eval_config = EvaluatorType(eval_config) evaluator_ = load_evaluator(eval_config, llm=eval_llm) eval_type_tag = eval_config.value else: kwargs = {"llm": eval_llm, **eval_config.get_kwargs()} evaluator_ = load_evaluator(eval_config.evaluator_type, **kwargs) eval_type_tag = eval_config.evaluator_type.value if isinstance(evaluator_, StringEvaluator): if evaluator_.requires_reference and reference_key is None: raise ValueError( f"Must specify reference_key in smith_eval.RunEvalConfig to use" f" evaluator of type {eval_type_tag} with" f" dataset with multiple output keys: {example_outputs}." ) run_evaluator = smith_eval.StringRunEvaluatorChain.from_run_and_data_type( evaluator_, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, tags=[eval_type_tag], ) elif isinstance(evaluator_, PairwiseStringEvaluator): raise NotImplementedError( f"Run evaluator for {eval_type_tag} is not implemented." " PairwiseStringEvaluators compare the outputs of two different models" " rather than the output of a single model." " Did you mean to use a StringEvaluator instead?" "\nSee: https://python.langchain.com/docs/guides/evaluation/string/" ) else: raise NotImplementedError( f"Run evaluator for {eval_type_tag} is not implemented" ) return run_evaluator def _get_keys( config: smith_eval.RunEvalConfig, run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], example_outputs: Optional[List[str]], ) -> Tuple[Optional[str], Optional[str], Optional[str]]: input_key = _determine_input_key(config, run_inputs) prediction_key = _determine_prediction_key(config, run_outputs) reference_key = _determine_reference_key(config, example_outputs) return input_key, prediction_key, reference_key def _load_run_evaluators( config: smith_eval.RunEvalConfig, run_type: str, data_type: DataType, example_outputs: Optional[List[str]], run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], ) -> List[RunEvaluator]: """ Load run evaluators from a configuration. Args: config: Configuration for the run evaluators. Returns: A list of run evaluators. """ run_evaluators = [] input_key, prediction_key, reference_key = None, None, None if ( config.evaluators or any([isinstance(e, EvaluatorType) for e in config.evaluators]) or ( config.custom_evaluators and any([isinstance(e, StringEvaluator) for e in config.custom_evaluators]) ) ): input_key, prediction_key, reference_key = _get_keys( config, run_inputs, run_outputs, example_outputs ) for eval_config in config.evaluators: run_evaluator = _construct_run_evaluator( eval_config, config.eval_llm, run_type, data_type, example_outputs, reference_key, input_key, prediction_key, ) run_evaluators.append(run_evaluator) custom_evaluators = config.custom_evaluators or [] for custom_evaluator in custom_evaluators: if isinstance(custom_evaluator, RunEvaluator): run_evaluators.append(custom_evaluator) elif isinstance(custom_evaluator, StringEvaluator): run_evaluators.append( smith_eval.StringRunEvaluatorChain.from_run_and_data_type( custom_evaluator, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, ) ) else: raise ValueError( f"Unsupported custom evaluator: {custom_evaluator}." f" Expected RunEvaluator or StringEvaluator." ) return run_evaluators ### Async Helpers async def _arun_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], *, tags: Optional[List[str]] = None, callbacks: Callbacks = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """Asynchronously run the language model. Args: llm: The language model to run. inputs: The input dictionary. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map inputs to the expected format. Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): return await llm.apredict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): return await llm.apredict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format" f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: prompt = _get_prompt(inputs) llm_output: Union[str, BaseMessage] = await llm.apredict( prompt, callbacks=callbacks, tags=tags ) except InputFormatError: messages = _get_messages(inputs) llm_output = await llm.apredict_messages( messages, callbacks=callbacks, tags=tags ) return llm_output async def _arun_chain( chain: Union[Chain, Runnable], inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str]: """Run a chain asynchronously on inputs.""" inputs_ = inputs if input_mapper is None else input_mapper(inputs) if ( isinstance(chain, Chain) and isinstance(inputs_, dict) and len(inputs_) == 1 and chain.input_keys ): val = next(iter(inputs_.values())) output = await chain.acall(val, callbacks=callbacks, tags=tags) else: runnable_config = RunnableConfig(tags=tags or [], callbacks=callbacks) output = await chain.ainvoke(inputs_, config=runnable_config) return output async def _arun_llm_or_chain( example: Example, config: RunnableConfig, *, llm_or_chain_factory: MCF, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str, LLMResult, ChatResult]: """Asynchronously run the Chain or language model. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map the input to the expected format. Returns: A list of outputs. """ chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) result = None try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = await _arun_llm( llm_or_chain_factory, example.inputs, tags=config["tags"], callbacks=config["callbacks"], input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = await _arun_chain( chain, example.inputs, tags=config["tags"], callbacks=config["callbacks"], input_mapper=input_mapper, ) result = output except Exception as e: logger.warning( f"{chain_or_llm} failed for example {example.id} " f"with inputs {example.inputs}" f"\n{repr(e)}" ) result = {"Error": repr(e)} return result ## Sync Utilities def _run_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """ Run the language model on the example. Args: llm: The language model to run. inputs: The input dictionary. callbacks: The callbacks to use during the run. tags: Optional tags to add to the run. input_mapper: function to map to the inputs dictionary from an Example Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): llm_output: Union[str, BaseMessage] = llm.predict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): llm_output = llm.predict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format: " f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: llm_prompts = _get_prompt(inputs) llm_output = llm.predict(llm_prompts, callbacks=callbacks, tags=tags) except InputFormatError: llm_messages = _get_messages(inputs) llm_output = llm.predict_messages(llm_messages, callbacks=callbacks) return llm_output def _run_chain( chain: Union[Chain, Runnable], inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[Dict, str]: """Run a chain on inputs.""" inputs_ = inputs if input_mapper is None else input_mapper(inputs) if ( isinstance(chain, Chain) and isinstance(inputs_, dict) and len(inputs_) == 1 and chain.input_keys ): val = next(iter(inputs_.values())) output = chain(val, callbacks=callbacks, tags=tags) else: runnable_config = RunnableConfig(tags=tags or [], callbacks=callbacks) output = chain.invoke(inputs_, config=runnable_config) return output def _run_llm_or_chain( example: Example, config: RunnableConfig, *, llm_or_chain_factory: MCF, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str, LLMResult, ChatResult]: """ Run the Chain or language model synchronously. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. Returns: Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: The outputs of the model or chain. """ chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) result = None try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = _run_llm( llm_or_chain_factory, example.inputs, config["callbacks"], tags=config["tags"], input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = _run_chain( chain, example.inputs, config["callbacks"], tags=config["tags"], input_mapper=input_mapper, ) result = output except Exception as e: error_type = type(e).__name__ logger.warning( f"{chain_or_llm} failed for example {example.id} " f"with inputs {example.inputs}" f"\nError Type: {error_type}, Message: {e}" ) result = {"Error": repr(e)} return result ## Public API def _prepare_eval_run( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, project_name: str, project_metadata: Optional[Dict[str, Any]] = None, ) -> Tuple[MCF, str, Dataset, List[Example]]: wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory, dataset_name) dataset = client.read_dataset(dataset_name=dataset_name) try: project = client.create_project( project_name, reference_dataset_id=dataset.id, project_extra={"metadata": project_metadata} if project_metadata else {}, ) except (HTTPError, ValueError, LangSmithError) as e: if "already exists " not in str(e): raise e uid = uuid.uuid4() example_msg = f""" run_on_dataset( ... project_name="{project_name} - {uid}", # Update since {project_name} already exists ) """ raise ValueError( f"Test project {project_name} already exists. Please use a different name:" f"\n\n{example_msg}" ) print( f"View the evaluation results for project '{project_name}'" f" at:\n{project.url}?eval=true\n\n" f"View all tests for Dataset {dataset_name} at:\n{dataset.url}", flush=True, ) examples = list(client.list_examples(dataset_id=dataset.id)) if not examples: raise ValueError(f"Dataset {dataset_name} has no example rows.") return wrapped_model, project_name, dataset, examples def _prepare_run_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, project_name: Optional[str], evaluation: Optional[smith_eval.RunEvalConfig] = None, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, concurrency_level: int = 5, project_metadata: Optional[Dict[str, Any]] = None, ) -> Tuple[MCF, str, List[Example], List[RunnableConfig]]: project_name = project_name or name_generation.random_name() wrapped_model, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name, project_metadata=project_metadata, ) wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory) run_evaluators = _setup_evaluation( wrapped_model, examples, evaluation, dataset.data_type or DataType.kv ) _validate_example_inputs(examples[0], wrapped_model, input_mapper) progress_bar = progress.ProgressBarCallback(len(examples)) configs = [ RunnableConfig( callbacks=[ LangChainTracer( project_name=project_name, client=client, use_threading=False, example_id=example.id, ), EvaluatorCallbackHandler( evaluators=run_evaluators or [], client=client, example_id=example.id, ), progress_bar, ], tags=tags or [], max_concurrency=concurrency_level, ) for example in examples ] return wrapped_model, project_name, examples, configs def _collect_test_results( examples: List[Example], batch_results: List[Union[dict, str, LLMResult, ChatResult]], configs: List[RunnableConfig], project_name: str, ) -> TestResult: wait_for_all_evaluators() all_eval_results = {} for c in configs: for callback in cast(list, c["callbacks"]): if isinstance(callback, EvaluatorCallbackHandler): eval_results = callback.logged_eval_results all_eval_results.update( {example_id: v for (_, example_id), v in eval_results.items()} ) results = {} for example, output in zip(examples, batch_results): feedback = all_eval_results.get(str(example.id), []) results[str(example.id)] = { "output": output, "input": example.inputs, "feedback": feedback, } if example.outputs: results[str(example.id)]["reference"] = example.outputs return TestResult( project_name=project_name, results=results, ) _INPUT_MAPPER_DEP_WARNING = ( "The input_mapper argument is deprecated and " "will be removed in a future release. Please add a " " RunnableLambda to your chain to map inputs to the expected format" " instead. Example:\n" "def construct_chain():\n" " my_chain = ...\n" " input_mapper = {'other_key': 'MyOtherInput', 'my_input_key': x}\n" " return input_mapper | my_chain\n" "run_on_dataset(..., llm_or_chain_factory=construct_chain)\n" "(See https://api.python.langchain.com/en/latest/schema/" "langchain.schema.runnable.base.RunnableLambda.html)" ) async def arun_on_dataset( client: Optional[Client], dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[smith_eval.RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, project_metadata: Optional[Dict[str, Any]] = None, verbose: bool = False, tags: Optional[List[str]] = None, **kwargs: Any, ) -> Dict[str, Any]: input_mapper = kwargs.pop("input_mapper", None) if input_mapper: warn_deprecated("0.0.305", message=_INPUT_MAPPER_DEP_WARNING, pending=True) if kwargs: warn_deprecated( "0.0.305", message="The following arguments are deprecated and " "will be removed in a future release: " f"{kwargs.keys()}.", removal="0.0.305", ) client = client or Client() wrapped_model, project_name, examples, configs = _prepare_run_on_dataset( client, dataset_name, llm_or_chain_factory, project_name, evaluation, tags, input_mapper, concurrency_level, project_metadata=project_metadata, ) batch_results = await runnable_utils.gather_with_concurrency( configs[0].get("max_concurrency"), *map( functools.partial( _arun_llm_or_chain, llm_or_chain_factory=wrapped_model, input_mapper=input_mapper, ), examples, configs, ), ) results = _collect_test_results(examples, batch_results, configs, project_name) if verbose: try: agg_feedback = results.get_aggregate_feedback() print("\n Eval quantiles:") print(agg_feedback) except Exception as e: logger.debug(f"Failed to print aggregate feedback: {repr(e)}") return results def run_on_dataset( client: Optional[Client], dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[smith_eval.RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, project_metadata: Optional[Dict[str, Any]] = None, verbose: bool = False, tags: Optional[List[str]] = None, **kwargs: Any, ) -> Dict[str, Any]: input_mapper = kwargs.pop("input_mapper", None) if input_mapper: warn_deprecated("0.0.305", message=_INPUT_MAPPER_DEP_WARNING, pending=True) if kwargs: warn_deprecated( "0.0.305", message="The following arguments are deprecated and " "will be removed in a future release: " f"{kwargs.keys()}.", removal="0.0.305", ) client = client or Client() wrapped_model, project_name, examples, configs = _prepare_run_on_dataset( client, dataset_name, llm_or_chain_factory, project_name, evaluation, tags, input_mapper, concurrency_level, project_metadata=project_metadata, ) if concurrency_level == 0: batch_results = [ _run_llm_or_chain( example, config, llm_or_chain_factory=wrapped_model, input_mapper=input_mapper, ) for example, config in zip(examples, configs) ] else: with runnable_config.get_executor_for_config(configs[0]) as executor: batch_results = list( executor.map( functools.partial( _run_llm_or_chain, llm_or_chain_factory=wrapped_model, input_mapper=input_mapper, ), examples, configs, ) ) results = _collect_test_results(examples, batch_results, configs, project_name) if verbose: try: agg_feedback = results.get_aggregate_feedback() print("\n Eval quantiles:") print(agg_feedback) except Exception as e: logger.debug(f"Failed to print aggregate feedback: {repr(e)}") return results _RUN_ON_DATASET_DOCSTRING = """ Run the Chain or language model on a dataset and store traces to the specified project name. Args: dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Configuration for evaluators to run on the results of the chain concurrency_level: The number of async tasks to run concurrently. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. project_metadata: Optional metadata to add to the project. Useful for storing information the test variant. (prompt version, model version, etc.) client: LangSmith client to use to access the dataset and to log feedback and run traces. verbose: Whether to print progress. tags: Tags to add to each run in the project. Returns: A dictionary containing the run's project name and the resulting model outputs. For the (usually faster) async version of this function, see :func:`arun_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import smith_eval.RunEvalConfig, run_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = smith_eval.RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", smith_eval.RunEvalConfig.Criteria("helpfulness"), smith_eval.RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = smith_eval.RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 run_on_dataset.__doc__ = _RUN_ON_DATASET_DOCSTRING arun_on_dataset.__doc__ = _RUN_ON_DATASET_DOCSTRING.replace( "run_on_dataset(", "await arun_on_dataset(" )
[ "langchain.schema.messages.messages_from_dict", "langchain._api.warn_deprecated", "langchain.schema.runnable.config.get_executor_for_config", "langchain.evaluation.schema.EvaluatorType", "langchain.smith.evaluation.name_generation.random_name", "langchain.smith.evaluation.StringRunEvaluatorChain.from_run_and_data_type", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler", "langchain.schema.runnable.RunnableLambda", "langchain.schema.runnable.RunnableConfig", "langchain.evaluation.loading.load_evaluator", "langchain.callbacks.tracers.evaluation.wait_for_all_evaluators" ]
[((1724, 1751), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1741, 1751), False, 'import logging\n'), ((33983, 34008), 'langchain.callbacks.tracers.evaluation.wait_for_all_evaluators', 'wait_for_all_evaluators', ([], {}), '()\n', (34006, 34008), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler, wait_for_all_evaluators\n'), ((3847, 3883), 'pandas.DataFrame', 'pd.DataFrame', (['records'], {'index': 'indices'}), '(records, index=indices)\n', (3859, 3883), True, 'import pandas as pd\n'), ((9632, 9667), 'langchain.schema.messages.messages_from_dict', 'messages_from_dict', (['raw_messages[0]'], {}), '(raw_messages[0])\n', (9650, 9667), False, 'from langchain.schema.messages import BaseMessage, messages_from_dict\n'), ((17595, 17636), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config'], {'llm': 'eval_llm'}), '(eval_config, llm=eval_llm)\n', (17609, 17636), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((17773, 17825), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config.evaluator_type'], {}), '(eval_config.evaluator_type, **kwargs)\n', (17787, 17825), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((18282, 18483), 'langchain.smith.evaluation.StringRunEvaluatorChain.from_run_and_data_type', 'smith_eval.StringRunEvaluatorChain.from_run_and_data_type', (['evaluator_', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key', 'tags': '[eval_type_tag]'}), '(evaluator_,\n run_type, data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key, tags=[eval_type_tag])\n', (18339, 18483), True, 'from langchain.smith import evaluation as smith_eval\n'), ((24388, 24440), 'langchain.schema.runnable.RunnableConfig', 'RunnableConfig', ([], {'tags': '(tags or [])', 'callbacks': 'callbacks'}), '(tags=tags or [], callbacks=callbacks)\n', (24402, 24440), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((28680, 28732), 'langchain.schema.runnable.RunnableConfig', 'RunnableConfig', ([], {'tags': '(tags or [])', 'callbacks': 'callbacks'}), '(tags=tags or [], callbacks=callbacks)\n', (28694, 28732), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((32515, 32544), 'langchain.smith.evaluation.name_generation.random_name', 'name_generation.random_name', ([], {}), '()\n', (32542, 32544), False, 'from langchain.smith.evaluation import name_generation, progress\n'), ((34081, 34107), 'typing.cast', 'cast', (['list', "c['callbacks']"], {}), "(list, c['callbacks'])\n", (34085, 34107), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast\n'), ((35958, 36033), 'langchain._api.warn_deprecated', 'warn_deprecated', (['"""0.0.305"""'], {'message': '_INPUT_MAPPER_DEP_WARNING', 'pending': '(True)'}), "('0.0.305', message=_INPUT_MAPPER_DEP_WARNING, pending=True)\n", (35973, 36033), False, 'from langchain._api import warn_deprecated\n'), ((36313, 36321), 'langsmith.client.Client', 'Client', ([], {}), '()\n', (36319, 36321), False, 'from langsmith.client import Client\n'), ((37857, 37932), 'langchain._api.warn_deprecated', 'warn_deprecated', (['"""0.0.305"""'], {'message': '_INPUT_MAPPER_DEP_WARNING', 'pending': '(True)'}), "('0.0.305', message=_INPUT_MAPPER_DEP_WARNING, pending=True)\n", (37872, 37932), False, 'from langchain._api import warn_deprecated\n'), ((38212, 38220), 'langsmith.client.Client', 'Client', ([], {}), '()\n', (38218, 38220), False, 'from langsmith.client import Client\n'), ((17547, 17573), 'langchain.evaluation.schema.EvaluatorType', 'EvaluatorType', (['eval_config'], {}), '(eval_config)\n', (17560, 17573), False, 'from langchain.evaluation.schema import EvaluatorType, PairwiseStringEvaluator, StringEvaluator\n'), ((31267, 31279), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (31277, 31279), False, 'import uuid\n'), ((38858, 38909), 'langchain.schema.runnable.config.get_executor_for_config', 'runnable_config.get_executor_for_config', (['configs[0]'], {}), '(configs[0])\n', (38897, 38909), True, 'from langchain.schema.runnable import config as runnable_config\n'), ((5477, 5520), 'langsmith.run_helpers.is_traceable_function', 'is_traceable_function', (['llm_or_chain_factory'], {}), '(llm_or_chain_factory)\n', (5498, 5520), False, 'from langsmith.run_helpers import as_runnable, is_traceable_function\n'), ((6096, 6132), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (6100, 6132), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast\n'), ((21237, 21422), 'langchain.smith.evaluation.StringRunEvaluatorChain.from_run_and_data_type', 'smith_eval.StringRunEvaluatorChain.from_run_and_data_type', (['custom_evaluator', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key'}), '(custom_evaluator,\n run_type, data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key)\n', (21294, 21422), True, 'from langchain.smith import evaluation as smith_eval\n'), ((33159, 33265), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)', 'example_id': 'example.id'}), '(project_name=project_name, client=client, use_threading=\n False, example_id=example.id)\n', (33174, 33265), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((33377, 33476), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'evaluators': '(run_evaluators or [])', 'client': 'client', 'example_id': 'example.id'}), '(evaluators=run_evaluators or [], client=client,\n example_id=example.id)\n', (33401, 33476), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler, wait_for_all_evaluators\n'), ((36758, 36862), 'functools.partial', 'functools.partial', (['_arun_llm_or_chain'], {'llm_or_chain_factory': 'wrapped_model', 'input_mapper': 'input_mapper'}), '(_arun_llm_or_chain, llm_or_chain_factory=wrapped_model,\n input_mapper=input_mapper)\n', (36775, 36862), False, 'import functools\n'), ((39007, 39110), 'functools.partial', 'functools.partial', (['_run_llm_or_chain'], {'llm_or_chain_factory': 'wrapped_model', 'input_mapper': 'input_mapper'}), '(_run_llm_or_chain, llm_or_chain_factory=wrapped_model,\n input_mapper=input_mapper)\n', (39024, 39110), False, 'import functools\n'), ((5558, 5594), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (5562, 5594), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast\n'), ((5836, 5872), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (5840, 5872), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast\n'), ((5891, 5919), 'inspect.signature', 'inspect.signature', (['user_func'], {}), '(user_func)\n', (5908, 5919), False, 'import inspect\n'), ((6013, 6038), 'langchain.schema.runnable.RunnableLambda', 'RunnableLambda', (['user_func'], {}), '(user_func)\n', (6027, 6038), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((6368, 6390), 'typing.cast', 'cast', (['Callable', '_model'], {}), '(Callable, _model)\n', (6372, 6390), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast\n'), ((6429, 6451), 'typing.cast', 'cast', (['Callable', '_model'], {}), '(Callable, _model)\n', (6433, 6451), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, List, Optional, Sequence, Tuple, Union, cast\n'), ((6642, 6669), 'langchain.schema.runnable.RunnableLambda', 'RunnableLambda', (['constructor'], {}), '(constructor)\n', (6656, 6669), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n')]
"""Schemas for the langchainplus API.""" from __future__ import annotations import logging import os from concurrent.futures import Future, ThreadPoolExecutor, wait from datetime import datetime from typing import Dict, List, Optional, Union, cast from uuid import UUID, uuid4 from pydantic import Field, PrivateAttr, root_validator, validator from langchainplus_sdk.client import LangChainPlusClient from langchainplus_sdk.schemas import RunBase, RunTypeEnum from langchainplus_sdk.utils import get_runtime_environment logger = logging.getLogger(__name__) def _make_thread_pool() -> ThreadPoolExecutor: """Ensure a thread pool exists in the current context.""" return ThreadPoolExecutor(max_workers=1) class RunTree(RunBase): """Run Schema with back-references for posting runs.""" name: str id: UUID = Field(default_factory=uuid4) start_time: datetime = Field(default_factory=datetime.utcnow) parent_run: Optional[RunTree] = Field(default=None, exclude=True) child_runs: List[RunTree] = Field( default_factory=list, exclude={"__all__": {"parent_run_id"}}, ) session_name: str = Field( default_factory=lambda: os.environ.get("LANGCHAIN_SESSION", "default") ) session_id: Optional[UUID] = Field(default=None) execution_order: int = 1 child_execution_order: int = Field(default=1, exclude=True) extra: Dict = Field(default_factory=dict) client: LangChainPlusClient = Field( default_factory=LangChainPlusClient, exclude=True ) executor: ThreadPoolExecutor = Field( default_factory=_make_thread_pool, exclude=True ) _futures: List[Future] = PrivateAttr(default_factory=list) class Config: arbitrary_types_allowed = True @validator("executor", pre=True) def validate_executor(cls, v: Optional[ThreadPoolExecutor]) -> ThreadPoolExecutor: """Ensure the executor is running.""" if v is None: return _make_thread_pool() if v._shutdown: raise ValueError("Executor has been shutdown.") return v @root_validator(pre=True) def infer_defaults(cls, values: dict) -> dict: """Assign name to the run.""" if "serialized" not in values: values["serialized"] = {"name": values["name"]} if "execution_order" not in values: values["execution_order"] = 1 if "child_execution_order" not in values: values["child_execution_order"] = values["execution_order"] if values.get("parent_run") is not None: values["parent_run_id"] = values["parent_run"].id extra = cast(dict, values.setdefault("extra", {})) runtime = cast(dict, extra.setdefault("runtime", {})) runtime.update(get_runtime_environment()) return values def end( self, *, outputs: Optional[Dict] = None, error: Optional[str] = None, end_time: Optional[datetime] = None, ) -> None: """Set the end time of the run and all child runs.""" self.end_time = end_time or datetime.utcnow() if outputs is not None: self.outputs = outputs if error is not None: self.error = error if self.parent_run: self.parent_run.child_execution_order = max( self.parent_run.child_execution_order, self.child_execution_order, ) def create_child( self, name: str, run_type: Union[str, RunTypeEnum], *, run_id: Optional[UUID] = None, serialized: Optional[Dict] = None, inputs: Optional[Dict] = None, outputs: Optional[Dict] = None, error: Optional[str] = None, reference_example_id: Optional[UUID] = None, start_time: Optional[datetime] = None, end_time: Optional[datetime] = None, extra: Optional[Dict] = None, ) -> RunTree: """Add a child run to the run tree.""" execution_order = self.child_execution_order + 1 serialized_ = serialized or {"name": name} run = RunTree( name=name, id=run_id or uuid4(), serialized=serialized_, inputs=inputs or {}, outputs=outputs or {}, error=error, run_type=run_type, reference_example_id=reference_example_id, start_time=start_time or datetime.utcnow(), end_time=end_time, execution_order=execution_order, child_execution_order=execution_order, extra=extra or {}, parent_run=self, session_name=self.session_name, client=self.client, executor=self.executor, ) self.child_runs.append(run) return run def post(self, exclude_child_runs: bool = True) -> Future: """Post the run tree to the API asynchronously.""" exclude = {"child_runs"} if exclude_child_runs else None kwargs = self.dict(exclude=exclude, exclude_none=True) self._futures.append( self.executor.submit( self.client.create_run, **kwargs, ) ) return self._futures[-1] def patch(self) -> Future: """Patch the run tree to the API in a background thread.""" self._futures.append( self.executor.submit( self.client.update_run, run_id=self.id, outputs=self.outputs.copy() if self.outputs else None, error=self.error, parent_run_id=self.parent_run_id, reference_example_id=self.reference_example_id, ) ) return self._futures[-1] def wait(self) -> None: """Wait for all _futures to complete.""" wait(self._futures) self._futures.clear()
[ "langchainplus_sdk.utils.get_runtime_environment" ]
[((533, 560), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (550, 560), False, 'import logging\n'), ((683, 716), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(1)'}), '(max_workers=1)\n', (701, 716), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((833, 861), 'pydantic.Field', 'Field', ([], {'default_factory': 'uuid4'}), '(default_factory=uuid4)\n', (838, 861), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((889, 927), 'pydantic.Field', 'Field', ([], {'default_factory': 'datetime.utcnow'}), '(default_factory=datetime.utcnow)\n', (894, 927), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((964, 997), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (969, 997), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1030, 1097), 'pydantic.Field', 'Field', ([], {'default_factory': 'list', 'exclude': "{'__all__': {'parent_run_id'}}"}), "(default_factory=list, exclude={'__all__': {'parent_run_id'}})\n", (1035, 1097), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1270, 1289), 'pydantic.Field', 'Field', ([], {'default': 'None'}), '(default=None)\n', (1275, 1289), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1352, 1382), 'pydantic.Field', 'Field', ([], {'default': '(1)', 'exclude': '(True)'}), '(default=1, exclude=True)\n', (1357, 1382), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1401, 1428), 'pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1406, 1428), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1463, 1519), 'pydantic.Field', 'Field', ([], {'default_factory': 'LangChainPlusClient', 'exclude': '(True)'}), '(default_factory=LangChainPlusClient, exclude=True)\n', (1468, 1519), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1569, 1623), 'pydantic.Field', 'Field', ([], {'default_factory': '_make_thread_pool', 'exclude': '(True)'}), '(default_factory=_make_thread_pool, exclude=True)\n', (1574, 1623), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1667, 1700), 'pydantic.PrivateAttr', 'PrivateAttr', ([], {'default_factory': 'list'}), '(default_factory=list)\n', (1678, 1700), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1765, 1796), 'pydantic.validator', 'validator', (['"""executor"""'], {'pre': '(True)'}), "('executor', pre=True)\n", (1774, 1796), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((2098, 2122), 'pydantic.root_validator', 'root_validator', ([], {'pre': '(True)'}), '(pre=True)\n', (2112, 2122), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((5856, 5875), 'concurrent.futures.wait', 'wait', (['self._futures'], {}), '(self._futures)\n', (5860, 5875), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((2774, 2799), 'langchainplus_sdk.utils.get_runtime_environment', 'get_runtime_environment', ([], {}), '()\n', (2797, 2799), False, 'from langchainplus_sdk.utils import get_runtime_environment\n'), ((3097, 3114), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (3112, 3114), False, 'from datetime import datetime\n'), ((1184, 1230), 'os.environ.get', 'os.environ.get', (['"""LANGCHAIN_SESSION"""', '"""default"""'], {}), "('LANGCHAIN_SESSION', 'default')\n", (1198, 1230), False, 'import os\n'), ((4176, 4183), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (4181, 4183), False, 'from uuid import UUID, uuid4\n'), ((4437, 4454), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (4452, 4454), False, 'from datetime import datetime\n')]
"""Schemas for the langchainplus API.""" from __future__ import annotations import logging import os from concurrent.futures import Future, ThreadPoolExecutor, wait from datetime import datetime from typing import Dict, List, Optional, Union, cast from uuid import UUID, uuid4 from pydantic import Field, PrivateAttr, root_validator, validator from langchainplus_sdk.client import LangChainPlusClient from langchainplus_sdk.schemas import RunBase, RunTypeEnum from langchainplus_sdk.utils import get_runtime_environment logger = logging.getLogger(__name__) def _make_thread_pool() -> ThreadPoolExecutor: """Ensure a thread pool exists in the current context.""" return ThreadPoolExecutor(max_workers=1) class RunTree(RunBase): """Run Schema with back-references for posting runs.""" name: str id: UUID = Field(default_factory=uuid4) start_time: datetime = Field(default_factory=datetime.utcnow) parent_run: Optional[RunTree] = Field(default=None, exclude=True) child_runs: List[RunTree] = Field( default_factory=list, exclude={"__all__": {"parent_run_id"}}, ) session_name: str = Field( default_factory=lambda: os.environ.get("LANGCHAIN_SESSION", "default") ) session_id: Optional[UUID] = Field(default=None) execution_order: int = 1 child_execution_order: int = Field(default=1, exclude=True) extra: Dict = Field(default_factory=dict) client: LangChainPlusClient = Field( default_factory=LangChainPlusClient, exclude=True ) executor: ThreadPoolExecutor = Field( default_factory=_make_thread_pool, exclude=True ) _futures: List[Future] = PrivateAttr(default_factory=list) class Config: arbitrary_types_allowed = True @validator("executor", pre=True) def validate_executor(cls, v: Optional[ThreadPoolExecutor]) -> ThreadPoolExecutor: """Ensure the executor is running.""" if v is None: return _make_thread_pool() if v._shutdown: raise ValueError("Executor has been shutdown.") return v @root_validator(pre=True) def infer_defaults(cls, values: dict) -> dict: """Assign name to the run.""" if "serialized" not in values: values["serialized"] = {"name": values["name"]} if "execution_order" not in values: values["execution_order"] = 1 if "child_execution_order" not in values: values["child_execution_order"] = values["execution_order"] if values.get("parent_run") is not None: values["parent_run_id"] = values["parent_run"].id extra = cast(dict, values.setdefault("extra", {})) runtime = cast(dict, extra.setdefault("runtime", {})) runtime.update(get_runtime_environment()) return values def end( self, *, outputs: Optional[Dict] = None, error: Optional[str] = None, end_time: Optional[datetime] = None, ) -> None: """Set the end time of the run and all child runs.""" self.end_time = end_time or datetime.utcnow() if outputs is not None: self.outputs = outputs if error is not None: self.error = error if self.parent_run: self.parent_run.child_execution_order = max( self.parent_run.child_execution_order, self.child_execution_order, ) def create_child( self, name: str, run_type: Union[str, RunTypeEnum], *, run_id: Optional[UUID] = None, serialized: Optional[Dict] = None, inputs: Optional[Dict] = None, outputs: Optional[Dict] = None, error: Optional[str] = None, reference_example_id: Optional[UUID] = None, start_time: Optional[datetime] = None, end_time: Optional[datetime] = None, extra: Optional[Dict] = None, ) -> RunTree: """Add a child run to the run tree.""" execution_order = self.child_execution_order + 1 serialized_ = serialized or {"name": name} run = RunTree( name=name, id=run_id or uuid4(), serialized=serialized_, inputs=inputs or {}, outputs=outputs or {}, error=error, run_type=run_type, reference_example_id=reference_example_id, start_time=start_time or datetime.utcnow(), end_time=end_time, execution_order=execution_order, child_execution_order=execution_order, extra=extra or {}, parent_run=self, session_name=self.session_name, client=self.client, executor=self.executor, ) self.child_runs.append(run) return run def post(self, exclude_child_runs: bool = True) -> Future: """Post the run tree to the API asynchronously.""" exclude = {"child_runs"} if exclude_child_runs else None kwargs = self.dict(exclude=exclude, exclude_none=True) self._futures.append( self.executor.submit( self.client.create_run, **kwargs, ) ) return self._futures[-1] def patch(self) -> Future: """Patch the run tree to the API in a background thread.""" self._futures.append( self.executor.submit( self.client.update_run, run_id=self.id, outputs=self.outputs.copy() if self.outputs else None, error=self.error, parent_run_id=self.parent_run_id, reference_example_id=self.reference_example_id, ) ) return self._futures[-1] def wait(self) -> None: """Wait for all _futures to complete.""" wait(self._futures) self._futures.clear()
[ "langchainplus_sdk.utils.get_runtime_environment" ]
[((533, 560), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (550, 560), False, 'import logging\n'), ((683, 716), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(1)'}), '(max_workers=1)\n', (701, 716), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((833, 861), 'pydantic.Field', 'Field', ([], {'default_factory': 'uuid4'}), '(default_factory=uuid4)\n', (838, 861), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((889, 927), 'pydantic.Field', 'Field', ([], {'default_factory': 'datetime.utcnow'}), '(default_factory=datetime.utcnow)\n', (894, 927), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((964, 997), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (969, 997), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1030, 1097), 'pydantic.Field', 'Field', ([], {'default_factory': 'list', 'exclude': "{'__all__': {'parent_run_id'}}"}), "(default_factory=list, exclude={'__all__': {'parent_run_id'}})\n", (1035, 1097), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1270, 1289), 'pydantic.Field', 'Field', ([], {'default': 'None'}), '(default=None)\n', (1275, 1289), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1352, 1382), 'pydantic.Field', 'Field', ([], {'default': '(1)', 'exclude': '(True)'}), '(default=1, exclude=True)\n', (1357, 1382), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1401, 1428), 'pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1406, 1428), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1463, 1519), 'pydantic.Field', 'Field', ([], {'default_factory': 'LangChainPlusClient', 'exclude': '(True)'}), '(default_factory=LangChainPlusClient, exclude=True)\n', (1468, 1519), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1569, 1623), 'pydantic.Field', 'Field', ([], {'default_factory': '_make_thread_pool', 'exclude': '(True)'}), '(default_factory=_make_thread_pool, exclude=True)\n', (1574, 1623), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1667, 1700), 'pydantic.PrivateAttr', 'PrivateAttr', ([], {'default_factory': 'list'}), '(default_factory=list)\n', (1678, 1700), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1765, 1796), 'pydantic.validator', 'validator', (['"""executor"""'], {'pre': '(True)'}), "('executor', pre=True)\n", (1774, 1796), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((2098, 2122), 'pydantic.root_validator', 'root_validator', ([], {'pre': '(True)'}), '(pre=True)\n', (2112, 2122), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((5856, 5875), 'concurrent.futures.wait', 'wait', (['self._futures'], {}), '(self._futures)\n', (5860, 5875), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((2774, 2799), 'langchainplus_sdk.utils.get_runtime_environment', 'get_runtime_environment', ([], {}), '()\n', (2797, 2799), False, 'from langchainplus_sdk.utils import get_runtime_environment\n'), ((3097, 3114), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (3112, 3114), False, 'from datetime import datetime\n'), ((1184, 1230), 'os.environ.get', 'os.environ.get', (['"""LANGCHAIN_SESSION"""', '"""default"""'], {}), "('LANGCHAIN_SESSION', 'default')\n", (1198, 1230), False, 'import os\n'), ((4176, 4183), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (4181, 4183), False, 'from uuid import UUID, uuid4\n'), ((4437, 4454), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (4452, 4454), False, 'from datetime import datetime\n')]
"""Schemas for the langchainplus API.""" from __future__ import annotations import logging import os from concurrent.futures import Future, ThreadPoolExecutor, wait from datetime import datetime from typing import Dict, List, Optional, Union, cast from uuid import UUID, uuid4 from pydantic import Field, PrivateAttr, root_validator, validator from langchainplus_sdk.client import LangChainPlusClient from langchainplus_sdk.schemas import RunBase, RunTypeEnum from langchainplus_sdk.utils import get_runtime_environment logger = logging.getLogger(__name__) def _make_thread_pool() -> ThreadPoolExecutor: """Ensure a thread pool exists in the current context.""" return ThreadPoolExecutor(max_workers=1) class RunTree(RunBase): """Run Schema with back-references for posting runs.""" name: str id: UUID = Field(default_factory=uuid4) start_time: datetime = Field(default_factory=datetime.utcnow) parent_run: Optional[RunTree] = Field(default=None, exclude=True) child_runs: List[RunTree] = Field( default_factory=list, exclude={"__all__": {"parent_run_id"}}, ) session_name: str = Field( default_factory=lambda: os.environ.get("LANGCHAIN_SESSION", "default") ) session_id: Optional[UUID] = Field(default=None) execution_order: int = 1 child_execution_order: int = Field(default=1, exclude=True) extra: Dict = Field(default_factory=dict) client: LangChainPlusClient = Field( default_factory=LangChainPlusClient, exclude=True ) executor: ThreadPoolExecutor = Field( default_factory=_make_thread_pool, exclude=True ) _futures: List[Future] = PrivateAttr(default_factory=list) class Config: arbitrary_types_allowed = True @validator("executor", pre=True) def validate_executor(cls, v: Optional[ThreadPoolExecutor]) -> ThreadPoolExecutor: """Ensure the executor is running.""" if v is None: return _make_thread_pool() if v._shutdown: raise ValueError("Executor has been shutdown.") return v @root_validator(pre=True) def infer_defaults(cls, values: dict) -> dict: """Assign name to the run.""" if "serialized" not in values: values["serialized"] = {"name": values["name"]} if "execution_order" not in values: values["execution_order"] = 1 if "child_execution_order" not in values: values["child_execution_order"] = values["execution_order"] if values.get("parent_run") is not None: values["parent_run_id"] = values["parent_run"].id extra = cast(dict, values.setdefault("extra", {})) runtime = cast(dict, extra.setdefault("runtime", {})) runtime.update(get_runtime_environment()) return values def end( self, *, outputs: Optional[Dict] = None, error: Optional[str] = None, end_time: Optional[datetime] = None, ) -> None: """Set the end time of the run and all child runs.""" self.end_time = end_time or datetime.utcnow() if outputs is not None: self.outputs = outputs if error is not None: self.error = error if self.parent_run: self.parent_run.child_execution_order = max( self.parent_run.child_execution_order, self.child_execution_order, ) def create_child( self, name: str, run_type: Union[str, RunTypeEnum], *, run_id: Optional[UUID] = None, serialized: Optional[Dict] = None, inputs: Optional[Dict] = None, outputs: Optional[Dict] = None, error: Optional[str] = None, reference_example_id: Optional[UUID] = None, start_time: Optional[datetime] = None, end_time: Optional[datetime] = None, extra: Optional[Dict] = None, ) -> RunTree: """Add a child run to the run tree.""" execution_order = self.child_execution_order + 1 serialized_ = serialized or {"name": name} run = RunTree( name=name, id=run_id or uuid4(), serialized=serialized_, inputs=inputs or {}, outputs=outputs or {}, error=error, run_type=run_type, reference_example_id=reference_example_id, start_time=start_time or datetime.utcnow(), end_time=end_time, execution_order=execution_order, child_execution_order=execution_order, extra=extra or {}, parent_run=self, session_name=self.session_name, client=self.client, executor=self.executor, ) self.child_runs.append(run) return run def post(self, exclude_child_runs: bool = True) -> Future: """Post the run tree to the API asynchronously.""" exclude = {"child_runs"} if exclude_child_runs else None kwargs = self.dict(exclude=exclude, exclude_none=True) self._futures.append( self.executor.submit( self.client.create_run, **kwargs, ) ) return self._futures[-1] def patch(self) -> Future: """Patch the run tree to the API in a background thread.""" self._futures.append( self.executor.submit( self.client.update_run, run_id=self.id, outputs=self.outputs.copy() if self.outputs else None, error=self.error, parent_run_id=self.parent_run_id, reference_example_id=self.reference_example_id, ) ) return self._futures[-1] def wait(self) -> None: """Wait for all _futures to complete.""" wait(self._futures) self._futures.clear()
[ "langchainplus_sdk.utils.get_runtime_environment" ]
[((533, 560), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (550, 560), False, 'import logging\n'), ((683, 716), 'concurrent.futures.ThreadPoolExecutor', 'ThreadPoolExecutor', ([], {'max_workers': '(1)'}), '(max_workers=1)\n', (701, 716), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((833, 861), 'pydantic.Field', 'Field', ([], {'default_factory': 'uuid4'}), '(default_factory=uuid4)\n', (838, 861), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((889, 927), 'pydantic.Field', 'Field', ([], {'default_factory': 'datetime.utcnow'}), '(default_factory=datetime.utcnow)\n', (894, 927), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((964, 997), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (969, 997), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1030, 1097), 'pydantic.Field', 'Field', ([], {'default_factory': 'list', 'exclude': "{'__all__': {'parent_run_id'}}"}), "(default_factory=list, exclude={'__all__': {'parent_run_id'}})\n", (1035, 1097), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1270, 1289), 'pydantic.Field', 'Field', ([], {'default': 'None'}), '(default=None)\n', (1275, 1289), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1352, 1382), 'pydantic.Field', 'Field', ([], {'default': '(1)', 'exclude': '(True)'}), '(default=1, exclude=True)\n', (1357, 1382), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1401, 1428), 'pydantic.Field', 'Field', ([], {'default_factory': 'dict'}), '(default_factory=dict)\n', (1406, 1428), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1463, 1519), 'pydantic.Field', 'Field', ([], {'default_factory': 'LangChainPlusClient', 'exclude': '(True)'}), '(default_factory=LangChainPlusClient, exclude=True)\n', (1468, 1519), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1569, 1623), 'pydantic.Field', 'Field', ([], {'default_factory': '_make_thread_pool', 'exclude': '(True)'}), '(default_factory=_make_thread_pool, exclude=True)\n', (1574, 1623), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1667, 1700), 'pydantic.PrivateAttr', 'PrivateAttr', ([], {'default_factory': 'list'}), '(default_factory=list)\n', (1678, 1700), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((1765, 1796), 'pydantic.validator', 'validator', (['"""executor"""'], {'pre': '(True)'}), "('executor', pre=True)\n", (1774, 1796), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((2098, 2122), 'pydantic.root_validator', 'root_validator', ([], {'pre': '(True)'}), '(pre=True)\n', (2112, 2122), False, 'from pydantic import Field, PrivateAttr, root_validator, validator\n'), ((5856, 5875), 'concurrent.futures.wait', 'wait', (['self._futures'], {}), '(self._futures)\n', (5860, 5875), False, 'from concurrent.futures import Future, ThreadPoolExecutor, wait\n'), ((2774, 2799), 'langchainplus_sdk.utils.get_runtime_environment', 'get_runtime_environment', ([], {}), '()\n', (2797, 2799), False, 'from langchainplus_sdk.utils import get_runtime_environment\n'), ((3097, 3114), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (3112, 3114), False, 'from datetime import datetime\n'), ((1184, 1230), 'os.environ.get', 'os.environ.get', (['"""LANGCHAIN_SESSION"""', '"""default"""'], {}), "('LANGCHAIN_SESSION', 'default')\n", (1198, 1230), False, 'import os\n'), ((4176, 4183), 'uuid.uuid4', 'uuid4', ([], {}), '()\n', (4181, 4183), False, 'from uuid import UUID, uuid4\n'), ((4437, 4454), 'datetime.datetime.utcnow', 'datetime.utcnow', ([], {}), '()\n', (4452, 4454), False, 'from datetime import datetime\n')]
import os import dotenv dotenv.load_dotenv() ### Load the credentials api_key = os.getenv("API_KEY", None) ibm_cloud_url = os.getenv("IBM_CLOUD_URL", None) project_id = os.getenv("PROJECT_ID", None) HUGGINGFACEHUB_API_TOKEN = os.getenv("HUGGINGFACEHUB_API_TOKEN", None) min_new_tokens=1 max_new_tokens=300 temperature=1 top_k=50 top_p=1 random_seed=42 repetition_penalty=1.2 from langchain.callbacks.base import BaseCallbackHandler class StreamHandler(BaseCallbackHandler): from uuid import UUID from langchain.schema.output import LLMResult from typing import Any def __init__(self, container, initial_text=""): self.container = container self.text = initial_text def on_llm_new_token(self, token: str, **kwargs) -> None: if token: self.text += token self.container.markdown(self.text) def on_llm_end(self, response: LLMResult, *, run_id: UUID, parent_run_id: UUID | None = None, **kwargs: Any) -> Any: return super().on_llm_end(response, run_id=run_id, parent_run_id=parent_run_id, **kwargs) ### IBM Watson def set_params_ibm(decoding_method): from ibm_watson_machine_learning.metanames import GenTextParamsMetaNames as GenParams if decoding_method == "sample": params = { GenParams.DECODING_METHOD: decoding_method, GenParams.MIN_NEW_TOKENS: min_new_tokens, GenParams.MAX_NEW_TOKENS: max_new_tokens, GenParams.TEMPERATURE: temperature, GenParams.TOP_K: top_k, GenParams.TOP_P: top_p, GenParams.RANDOM_SEED: random_seed, GenParams.REPETITION_PENALTY: repetition_penalty, GenParams.STOP_SEQUENCES: ["# [","\n\n","**"], } elif decoding_method == "greedy": params = { GenParams.DECODING_METHOD: decoding_method, GenParams.MIN_NEW_TOKENS: min_new_tokens, GenParams.MAX_NEW_TOKENS: max_new_tokens, GenParams.RANDOM_SEED: random_seed, GenParams.TEMPERATURE: temperature, GenParams.REPETITION_PENALTY: repetition_penalty, GenParams.STOP_SEQUENCES: ["#","[","\n\n","**"," * "], } else: params = None print("Decoding method not supported, please use 'sample' or 'greedy'.") return params ### IBM Research BAM def llm_param(decoding_method, model_id): from genai.credentials import Credentials from genai.model import Model from genai.extensions.langchain import LangChainInterface if bam_api_key is None or bam_api_url is None: llm = None print("Ensure you copied the .env file that you created earlier into the same directory as this notebook") else: creds = Credentials(bam_api_key, api_endpoint=bam_api_url) # decoding_method = "greedy" params = set_params_ibm(decoding_method) message_placeholder = None # You can define this if needed stream_handler = StreamHandler(message_placeholder) callback = False if callback is True: llm = LangChainInterface( model=model_id, credentials=creds, params=params, callbacks=[stream_handler] ) else: llm = LangChainInterface( model=model_id, credentials=creds, params=params ) llm = LangChainInterface(model=model_id, credentials=creds, params=params, project_id=project_id) return llm ### Hugging Face def hugfacelib(repo_id): from langchain.embeddings import HuggingFaceHubEmbeddings repo_id = "sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2" embedding = HuggingFaceHubEmbeddings( task="feature-extraction", repo_id = repo_id, huggingfacehub_api_token = HUGGINGFACEHUB_API_TOKEN ) return embedding
[ "langchain.embeddings.HuggingFaceHubEmbeddings" ]
[((24, 44), 'dotenv.load_dotenv', 'dotenv.load_dotenv', ([], {}), '()\n', (42, 44), False, 'import dotenv\n'), ((81, 107), 'os.getenv', 'os.getenv', (['"""API_KEY"""', 'None'], {}), "('API_KEY', None)\n", (90, 107), False, 'import os\n'), ((124, 156), 'os.getenv', 'os.getenv', (['"""IBM_CLOUD_URL"""', 'None'], {}), "('IBM_CLOUD_URL', None)\n", (133, 156), False, 'import os\n'), ((170, 199), 'os.getenv', 'os.getenv', (['"""PROJECT_ID"""', 'None'], {}), "('PROJECT_ID', None)\n", (179, 199), False, 'import os\n'), ((228, 271), 'os.getenv', 'os.getenv', (['"""HUGGINGFACEHUB_API_TOKEN"""', 'None'], {}), "('HUGGINGFACEHUB_API_TOKEN', None)\n", (237, 271), False, 'import os\n'), ((3772, 3895), 'langchain.embeddings.HuggingFaceHubEmbeddings', 'HuggingFaceHubEmbeddings', ([], {'task': '"""feature-extraction"""', 'repo_id': 'repo_id', 'huggingfacehub_api_token': 'HUGGINGFACEHUB_API_TOKEN'}), "(task='feature-extraction', repo_id=repo_id,\n huggingfacehub_api_token=HUGGINGFACEHUB_API_TOKEN)\n", (3796, 3895), False, 'from langchain.embeddings import HuggingFaceHubEmbeddings\n'), ((2763, 2813), 'genai.credentials.Credentials', 'Credentials', (['bam_api_key'], {'api_endpoint': 'bam_api_url'}), '(bam_api_key, api_endpoint=bam_api_url)\n', (2774, 2813), False, 'from genai.credentials import Credentials\n'), ((3465, 3560), 'genai.extensions.langchain.LangChainInterface', 'LangChainInterface', ([], {'model': 'model_id', 'credentials': 'creds', 'params': 'params', 'project_id': 'project_id'}), '(model=model_id, credentials=creds, params=params,\n project_id=project_id)\n', (3483, 3560), False, 'from genai.extensions.langchain import LangChainInterface\n'), ((3112, 3212), 'genai.extensions.langchain.LangChainInterface', 'LangChainInterface', ([], {'model': 'model_id', 'credentials': 'creds', 'params': 'params', 'callbacks': '[stream_handler]'}), '(model=model_id, credentials=creds, params=params,\n callbacks=[stream_handler])\n', (3130, 3212), False, 'from genai.extensions.langchain import LangChainInterface\n'), ((3319, 3387), 'genai.extensions.langchain.LangChainInterface', 'LangChainInterface', ([], {'model': 'model_id', 'credentials': 'creds', 'params': 'params'}), '(model=model_id, credentials=creds, params=params)\n', (3337, 3387), False, 'from genai.extensions.langchain import LangChainInterface\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.schema import ( AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string, ) def _get_verbosity() -> bool: return langchain.verbose def get_prompts( params: Dict[str, Any], prompts: List[str] ) -> Tuple[Dict[int, List], str, List[int], List[str]]: """Get prompts that are already cached.""" llm_string = str(sorted([(k, v) for k, v in params.items()])) missing_prompts = [] missing_prompt_idxs = [] existing_prompts = {} for i, prompt in enumerate(prompts): if langchain.llm_cache is not None: cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): existing_prompts[i] = cache_val else: missing_prompts.append(prompt) missing_prompt_idxs.append(i) return existing_prompts, llm_string, missing_prompt_idxs, missing_prompts def update_cache( existing_prompts: Dict[int, List], llm_string: str, missing_prompt_idxs: List[int], new_results: LLMResult, prompts: List[str], ) -> Optional[dict]: """Update the cache and get the LLM output.""" for i, result in enumerate(new_results.generations): existing_prompts[missing_prompt_idxs[i]] = result prompt = prompts[missing_prompt_idxs[i]] if langchain.llm_cache is not None: langchain.llm_cache.update(prompt, llm_string, result) llm_output = new_results.llm_output return llm_output class BaseLLM(BaseLanguageModel, ABC): """LLM wrapper should take in a prompt and return a string.""" cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @abstractmethod def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" @abstractmethod async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return self.generate(prompt_strings, stop=stop, callbacks=callbacks) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return await self.agenerate(prompt_strings, stop=stop, callbacks=callbacks) def generate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # If string is passed in directly no errors will be raised but outputs will # not make sense. if not isinstance(prompts, list): raise ValueError( "Argument 'prompts' is expected to be of type List[str], received" f" argument of type {type(prompts)}." ) params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( self._generate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(output) return output if len(missing_prompts) > 0: run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( self._generate(missing_prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) async def agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( await self._agenerate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else await self._agenerate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e, verbose=self.verbose) raise e await run_manager.on_llm_end(output, verbose=self.verbose) return output if len(missing_prompts) > 0: run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( await self._agenerate( missing_prompts, stop=stop, run_manager=run_manager ) if new_arg_supported else await self._agenerate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e) raise e await run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) def __call__( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" if not isinstance(prompt, str): raise ValueError( "Argument `prompt` is expected to be a string. Instead found " f"{type(prompt)}. If you want to run the LLM on multiple prompts, use " "`generate` instead." ) return ( self.generate([prompt], stop=stop, callbacks=callbacks) .generations[0][0] .text ) async def _call_async( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" result = await self.agenerate([prompt], stop=stop, callbacks=callbacks) return result.generations[0][0].text def predict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return self(text, stop=_stop) def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = self(text, stop=_stop) return AIMessage(content=content) async def apredict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(text, stop=_stop) async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = await self._call_async(text, stop=_stop) return AIMessage(content=content) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} def __str__(self) -> str: """Get a string representation of the object for printing.""" cls_name = f"\033[1m{self.__class__.__name__}\033[0m" return f"{cls_name}\nParams: {self._identifying_params}" @property @abstractmethod def _llm_type(self) -> str: """Return type of llm.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict def save(self, file_path: Union[Path, str]) -> None: """Save the LLM. Args: file_path: Path to file to save the LLM to. Example: .. code-block:: python llm.save(file_path="path/llm.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save prompt_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(prompt_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(prompt_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") class LLM(BaseLLM): """LLM class that expect subclasses to implement a simpler call method. The purpose of this class is to expose a simpler interface for working with LLMs, rather than expect the user to implement the full _generate method. """ @abstractmethod def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" async def _acall( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" raise NotImplementedError("Async generation not implemented for this LLM.") def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # TODO: add caching here. generations = [] new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") for prompt in prompts: text = ( self._call(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else self._call(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations) async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" generations = [] new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") for prompt in prompts: text = ( await self._acall(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else await self._acall(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((2302, 2339), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2307, 2339), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2413, 2446), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2418, 2446), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2501, 2534), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2506, 2534), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2683, 2699), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (2697, 2699), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((3135, 3178), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (3144, 3178), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((5507, 5573), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (5532, 5573), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7567, 7624), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7576, 7624), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((8170, 8241), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (8200, 8241), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((10390, 10447), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (10399, 10447), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11768, 11795), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (11785, 11795), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11947, 11973), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (11956, 11973), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12369, 12396), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (12386, 12396), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12566, 12592), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (12575, 12592), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((15773, 15807), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (15782, 15807), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16489, 16523), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (16498, 16523), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((1235, 1281), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (1261, 1281), False, 'import langchain\n'), ((2023, 2077), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result'], {}), '(prompt, llm_string, result)\n', (2049, 2077), False, 'import langchain\n'), ((2890, 2992), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (2903, 2992), False, 'import warnings\n'), ((13642, 13657), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (13646, 13657), False, 'from pathlib import Path\n'), ((13978, 14013), 'json.dump', 'json.dump', (['prompt_dict', 'f'], {'indent': '(4)'}), '(prompt_dict, f, indent=4)\n', (13987, 14013), False, 'import json\n'), ((5624, 5657), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (5641, 5657), False, 'import inspect\n'), ((8292, 8326), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (8309, 8326), False, 'import inspect\n'), ((14116, 14167), 'yaml.dump', 'yaml.dump', (['prompt_dict', 'f'], {'default_flow_style': '(False)'}), '(prompt_dict, f, default_flow_style=False)\n', (14125, 14167), False, 'import yaml\n'), ((15417, 15446), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (15434, 15446), False, 'import inspect\n'), ((15734, 15755), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (15744, 15755), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16118, 16148), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (16135, 16148), False, 'import inspect\n'), ((16450, 16471), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (16460, 16471), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.schema import ( AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string, ) def _get_verbosity() -> bool: return langchain.verbose def get_prompts( params: Dict[str, Any], prompts: List[str] ) -> Tuple[Dict[int, List], str, List[int], List[str]]: """Get prompts that are already cached.""" llm_string = str(sorted([(k, v) for k, v in params.items()])) missing_prompts = [] missing_prompt_idxs = [] existing_prompts = {} for i, prompt in enumerate(prompts): if langchain.llm_cache is not None: cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): existing_prompts[i] = cache_val else: missing_prompts.append(prompt) missing_prompt_idxs.append(i) return existing_prompts, llm_string, missing_prompt_idxs, missing_prompts def update_cache( existing_prompts: Dict[int, List], llm_string: str, missing_prompt_idxs: List[int], new_results: LLMResult, prompts: List[str], ) -> Optional[dict]: """Update the cache and get the LLM output.""" for i, result in enumerate(new_results.generations): existing_prompts[missing_prompt_idxs[i]] = result prompt = prompts[missing_prompt_idxs[i]] if langchain.llm_cache is not None: langchain.llm_cache.update(prompt, llm_string, result) llm_output = new_results.llm_output return llm_output class BaseLLM(BaseLanguageModel, ABC): """LLM wrapper should take in a prompt and return a string.""" cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @abstractmethod def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" @abstractmethod async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return self.generate(prompt_strings, stop=stop, callbacks=callbacks) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return await self.agenerate(prompt_strings, stop=stop, callbacks=callbacks) def generate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # If string is passed in directly no errors will be raised but outputs will # not make sense. if not isinstance(prompts, list): raise ValueError( "Argument 'prompts' is expected to be of type List[str], received" f" argument of type {type(prompts)}." ) params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( self._generate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(output) return output if len(missing_prompts) > 0: run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( self._generate(missing_prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) async def agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( await self._agenerate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else await self._agenerate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e, verbose=self.verbose) raise e await run_manager.on_llm_end(output, verbose=self.verbose) return output if len(missing_prompts) > 0: run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( await self._agenerate( missing_prompts, stop=stop, run_manager=run_manager ) if new_arg_supported else await self._agenerate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e) raise e await run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) def __call__( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" if not isinstance(prompt, str): raise ValueError( "Argument `prompt` is expected to be a string. Instead found " f"{type(prompt)}. If you want to run the LLM on multiple prompts, use " "`generate` instead." ) return ( self.generate([prompt], stop=stop, callbacks=callbacks) .generations[0][0] .text ) async def _call_async( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" result = await self.agenerate([prompt], stop=stop, callbacks=callbacks) return result.generations[0][0].text def predict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return self(text, stop=_stop) def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = self(text, stop=_stop) return AIMessage(content=content) async def apredict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(text, stop=_stop) async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = await self._call_async(text, stop=_stop) return AIMessage(content=content) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} def __str__(self) -> str: """Get a string representation of the object for printing.""" cls_name = f"\033[1m{self.__class__.__name__}\033[0m" return f"{cls_name}\nParams: {self._identifying_params}" @property @abstractmethod def _llm_type(self) -> str: """Return type of llm.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict def save(self, file_path: Union[Path, str]) -> None: """Save the LLM. Args: file_path: Path to file to save the LLM to. Example: .. code-block:: python llm.save(file_path="path/llm.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save prompt_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(prompt_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(prompt_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") class LLM(BaseLLM): """LLM class that expect subclasses to implement a simpler call method. The purpose of this class is to expose a simpler interface for working with LLMs, rather than expect the user to implement the full _generate method. """ @abstractmethod def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" async def _acall( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" raise NotImplementedError("Async generation not implemented for this LLM.") def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # TODO: add caching here. generations = [] new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") for prompt in prompts: text = ( self._call(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else self._call(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations) async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" generations = [] new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") for prompt in prompts: text = ( await self._acall(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else await self._acall(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((2302, 2339), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2307, 2339), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2413, 2446), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2418, 2446), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2501, 2534), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2506, 2534), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2683, 2699), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (2697, 2699), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((3135, 3178), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (3144, 3178), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((5507, 5573), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (5532, 5573), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7567, 7624), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7576, 7624), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((8170, 8241), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (8200, 8241), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((10390, 10447), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (10399, 10447), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11768, 11795), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (11785, 11795), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11947, 11973), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (11956, 11973), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12369, 12396), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (12386, 12396), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12566, 12592), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (12575, 12592), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((15773, 15807), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (15782, 15807), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16489, 16523), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (16498, 16523), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((1235, 1281), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (1261, 1281), False, 'import langchain\n'), ((2023, 2077), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result'], {}), '(prompt, llm_string, result)\n', (2049, 2077), False, 'import langchain\n'), ((2890, 2992), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (2903, 2992), False, 'import warnings\n'), ((13642, 13657), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (13646, 13657), False, 'from pathlib import Path\n'), ((13978, 14013), 'json.dump', 'json.dump', (['prompt_dict', 'f'], {'indent': '(4)'}), '(prompt_dict, f, indent=4)\n', (13987, 14013), False, 'import json\n'), ((5624, 5657), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (5641, 5657), False, 'import inspect\n'), ((8292, 8326), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (8309, 8326), False, 'import inspect\n'), ((14116, 14167), 'yaml.dump', 'yaml.dump', (['prompt_dict', 'f'], {'default_flow_style': '(False)'}), '(prompt_dict, f, default_flow_style=False)\n', (14125, 14167), False, 'import yaml\n'), ((15417, 15446), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (15434, 15446), False, 'import inspect\n'), ((15734, 15755), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (15744, 15755), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16118, 16148), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (16135, 16148), False, 'import inspect\n'), ((16450, 16471), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (16460, 16471), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.schema import ( AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string, ) def _get_verbosity() -> bool: return langchain.verbose def get_prompts( params: Dict[str, Any], prompts: List[str] ) -> Tuple[Dict[int, List], str, List[int], List[str]]: """Get prompts that are already cached.""" llm_string = str(sorted([(k, v) for k, v in params.items()])) missing_prompts = [] missing_prompt_idxs = [] existing_prompts = {} for i, prompt in enumerate(prompts): if langchain.llm_cache is not None: cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): existing_prompts[i] = cache_val else: missing_prompts.append(prompt) missing_prompt_idxs.append(i) return existing_prompts, llm_string, missing_prompt_idxs, missing_prompts def update_cache( existing_prompts: Dict[int, List], llm_string: str, missing_prompt_idxs: List[int], new_results: LLMResult, prompts: List[str], ) -> Optional[dict]: """Update the cache and get the LLM output.""" for i, result in enumerate(new_results.generations): existing_prompts[missing_prompt_idxs[i]] = result prompt = prompts[missing_prompt_idxs[i]] if langchain.llm_cache is not None: langchain.llm_cache.update(prompt, llm_string, result) llm_output = new_results.llm_output return llm_output class BaseLLM(BaseLanguageModel, ABC): """LLM wrapper should take in a prompt and return a string.""" cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @abstractmethod def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" @abstractmethod async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return self.generate(prompt_strings, stop=stop, callbacks=callbacks) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return await self.agenerate(prompt_strings, stop=stop, callbacks=callbacks) def generate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # If string is passed in directly no errors will be raised but outputs will # not make sense. if not isinstance(prompts, list): raise ValueError( "Argument 'prompts' is expected to be of type List[str], received" f" argument of type {type(prompts)}." ) params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( self._generate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(output) return output if len(missing_prompts) > 0: run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( self._generate(missing_prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) async def agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( await self._agenerate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else await self._agenerate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e, verbose=self.verbose) raise e await run_manager.on_llm_end(output, verbose=self.verbose) return output if len(missing_prompts) > 0: run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( await self._agenerate( missing_prompts, stop=stop, run_manager=run_manager ) if new_arg_supported else await self._agenerate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e) raise e await run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) def __call__( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" if not isinstance(prompt, str): raise ValueError( "Argument `prompt` is expected to be a string. Instead found " f"{type(prompt)}. If you want to run the LLM on multiple prompts, use " "`generate` instead." ) return ( self.generate([prompt], stop=stop, callbacks=callbacks) .generations[0][0] .text ) async def _call_async( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" result = await self.agenerate([prompt], stop=stop, callbacks=callbacks) return result.generations[0][0].text def predict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return self(text, stop=_stop) def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = self(text, stop=_stop) return AIMessage(content=content) async def apredict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(text, stop=_stop) async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = await self._call_async(text, stop=_stop) return AIMessage(content=content) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} def __str__(self) -> str: """Get a string representation of the object for printing.""" cls_name = f"\033[1m{self.__class__.__name__}\033[0m" return f"{cls_name}\nParams: {self._identifying_params}" @property @abstractmethod def _llm_type(self) -> str: """Return type of llm.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict def save(self, file_path: Union[Path, str]) -> None: """Save the LLM. Args: file_path: Path to file to save the LLM to. Example: .. code-block:: python llm.save(file_path="path/llm.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save prompt_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(prompt_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(prompt_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") class LLM(BaseLLM): """LLM class that expect subclasses to implement a simpler call method. The purpose of this class is to expose a simpler interface for working with LLMs, rather than expect the user to implement the full _generate method. """ @abstractmethod def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" async def _acall( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" raise NotImplementedError("Async generation not implemented for this LLM.") def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # TODO: add caching here. generations = [] new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") for prompt in prompts: text = ( self._call(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else self._call(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations) async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" generations = [] new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") for prompt in prompts: text = ( await self._acall(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else await self._acall(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((2302, 2339), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2307, 2339), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2413, 2446), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2418, 2446), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2501, 2534), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2506, 2534), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2683, 2699), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (2697, 2699), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((3135, 3178), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (3144, 3178), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((5507, 5573), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (5532, 5573), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7567, 7624), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7576, 7624), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((8170, 8241), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (8200, 8241), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((10390, 10447), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (10399, 10447), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11768, 11795), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (11785, 11795), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11947, 11973), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (11956, 11973), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12369, 12396), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (12386, 12396), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12566, 12592), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (12575, 12592), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((15773, 15807), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (15782, 15807), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16489, 16523), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (16498, 16523), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((1235, 1281), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (1261, 1281), False, 'import langchain\n'), ((2023, 2077), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result'], {}), '(prompt, llm_string, result)\n', (2049, 2077), False, 'import langchain\n'), ((2890, 2992), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (2903, 2992), False, 'import warnings\n'), ((13642, 13657), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (13646, 13657), False, 'from pathlib import Path\n'), ((13978, 14013), 'json.dump', 'json.dump', (['prompt_dict', 'f'], {'indent': '(4)'}), '(prompt_dict, f, indent=4)\n', (13987, 14013), False, 'import json\n'), ((5624, 5657), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (5641, 5657), False, 'import inspect\n'), ((8292, 8326), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (8309, 8326), False, 'import inspect\n'), ((14116, 14167), 'yaml.dump', 'yaml.dump', (['prompt_dict', 'f'], {'default_flow_style': '(False)'}), '(prompt_dict, f, default_flow_style=False)\n', (14125, 14167), False, 'import yaml\n'), ((15417, 15446), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (15434, 15446), False, 'import inspect\n'), ((15734, 15755), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (15744, 15755), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16118, 16148), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (16135, 16148), False, 'import inspect\n'), ((16450, 16471), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (16460, 16471), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n')]
"""Base interface for large language models to expose.""" import inspect import json import warnings from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Mapping, Optional, Sequence, Tuple, Union import yaml from pydantic import Extra, Field, root_validator, validator import langchain from langchain.base_language import BaseLanguageModel from langchain.callbacks.base import BaseCallbackManager from langchain.callbacks.manager import ( AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks, ) from langchain.schema import ( AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string, ) def _get_verbosity() -> bool: return langchain.verbose def get_prompts( params: Dict[str, Any], prompts: List[str] ) -> Tuple[Dict[int, List], str, List[int], List[str]]: """Get prompts that are already cached.""" llm_string = str(sorted([(k, v) for k, v in params.items()])) missing_prompts = [] missing_prompt_idxs = [] existing_prompts = {} for i, prompt in enumerate(prompts): if langchain.llm_cache is not None: cache_val = langchain.llm_cache.lookup(prompt, llm_string) if isinstance(cache_val, list): existing_prompts[i] = cache_val else: missing_prompts.append(prompt) missing_prompt_idxs.append(i) return existing_prompts, llm_string, missing_prompt_idxs, missing_prompts def update_cache( existing_prompts: Dict[int, List], llm_string: str, missing_prompt_idxs: List[int], new_results: LLMResult, prompts: List[str], ) -> Optional[dict]: """Update the cache and get the LLM output.""" for i, result in enumerate(new_results.generations): existing_prompts[missing_prompt_idxs[i]] = result prompt = prompts[missing_prompt_idxs[i]] if langchain.llm_cache is not None: langchain.llm_cache.update(prompt, llm_string, result) llm_output = new_results.llm_output return llm_output class BaseLLM(BaseLanguageModel, ABC): """LLM wrapper should take in a prompt and return a string.""" cache: Optional[bool] = None verbose: bool = Field(default_factory=_get_verbosity) """Whether to print out response text.""" callbacks: Callbacks = Field(default=None, exclude=True) callback_manager: Optional[BaseCallbackManager] = Field(default=None, exclude=True) class Config: """Configuration for this pydantic object.""" extra = Extra.forbid arbitrary_types_allowed = True @root_validator() def raise_deprecation(cls, values: Dict) -> Dict: """Raise deprecation warning if callback_manager is used.""" if values.get("callback_manager") is not None: warnings.warn( "callback_manager is deprecated. Please use callbacks instead.", DeprecationWarning, ) values["callbacks"] = values.pop("callback_manager", None) return values @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @abstractmethod def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" @abstractmethod async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompts.""" def generate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return self.generate(prompt_strings, stop=stop, callbacks=callbacks) async def agenerate_prompt( self, prompts: List[PromptValue], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: prompt_strings = [p.to_string() for p in prompts] return await self.agenerate(prompt_strings, stop=stop, callbacks=callbacks) def generate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # If string is passed in directly no errors will be raised but outputs will # not make sense. if not isinstance(prompts, list): raise ValueError( "Argument 'prompts' is expected to be of type List[str], received" f" argument of type {type(prompts)}." ) params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = CallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._generate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( self._generate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(output) return output if len(missing_prompts) > 0: run_manager = callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( self._generate(missing_prompts, stop=stop, run_manager=run_manager) if new_arg_supported else self._generate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: run_manager.on_llm_error(e) raise e run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) async def agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, callbacks: Callbacks = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" params = self.dict() params["stop"] = stop ( existing_prompts, llm_string, missing_prompt_idxs, missing_prompts, ) = get_prompts(params, prompts) disregard_cache = self.cache is not None and not self.cache callback_manager = AsyncCallbackManager.configure( callbacks, self.callbacks, self.verbose ) new_arg_supported = inspect.signature(self._agenerate).parameters.get( "run_manager" ) if langchain.llm_cache is None or disregard_cache: # This happens when langchain.cache is None, but self.cache is True if self.cache is not None and self.cache: raise ValueError( "Asked to cache, but no cache found at `langchain.cache`." ) run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, prompts, invocation_params=params ) try: output = ( await self._agenerate(prompts, stop=stop, run_manager=run_manager) if new_arg_supported else await self._agenerate(prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e, verbose=self.verbose) raise e await run_manager.on_llm_end(output, verbose=self.verbose) return output if len(missing_prompts) > 0: run_manager = await callback_manager.on_llm_start( {"name": self.__class__.__name__}, missing_prompts, invocation_params=params, ) try: new_results = ( await self._agenerate( missing_prompts, stop=stop, run_manager=run_manager ) if new_arg_supported else await self._agenerate(missing_prompts, stop=stop) ) except (KeyboardInterrupt, Exception) as e: await run_manager.on_llm_error(e) raise e await run_manager.on_llm_end(new_results) llm_output = update_cache( existing_prompts, llm_string, missing_prompt_idxs, new_results, prompts ) else: llm_output = {} generations = [existing_prompts[i] for i in range(len(prompts))] return LLMResult(generations=generations, llm_output=llm_output) def __call__( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" if not isinstance(prompt, str): raise ValueError( "Argument `prompt` is expected to be a string. Instead found " f"{type(prompt)}. If you want to run the LLM on multiple prompts, use " "`generate` instead." ) return ( self.generate([prompt], stop=stop, callbacks=callbacks) .generations[0][0] .text ) async def _call_async( self, prompt: str, stop: Optional[List[str]] = None, callbacks: Callbacks = None ) -> str: """Check Cache and run the LLM on the given prompt and input.""" result = await self.agenerate([prompt], stop=stop, callbacks=callbacks) return result.generations[0][0].text def predict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return self(text, stop=_stop) def predict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = self(text, stop=_stop) return AIMessage(content=content) async def apredict(self, text: str, *, stop: Optional[Sequence[str]] = None) -> str: if stop is None: _stop = None else: _stop = list(stop) return await self._call_async(text, stop=_stop) async def apredict_messages( self, messages: List[BaseMessage], *, stop: Optional[Sequence[str]] = None ) -> BaseMessage: text = get_buffer_string(messages) if stop is None: _stop = None else: _stop = list(stop) content = await self._call_async(text, stop=_stop) return AIMessage(content=content) @property def _identifying_params(self) -> Mapping[str, Any]: """Get the identifying parameters.""" return {} def __str__(self) -> str: """Get a string representation of the object for printing.""" cls_name = f"\033[1m{self.__class__.__name__}\033[0m" return f"{cls_name}\nParams: {self._identifying_params}" @property @abstractmethod def _llm_type(self) -> str: """Return type of llm.""" def dict(self, **kwargs: Any) -> Dict: """Return a dictionary of the LLM.""" starter_dict = dict(self._identifying_params) starter_dict["_type"] = self._llm_type return starter_dict def save(self, file_path: Union[Path, str]) -> None: """Save the LLM. Args: file_path: Path to file to save the LLM to. Example: .. code-block:: python llm.save(file_path="path/llm.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save prompt_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(prompt_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(prompt_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml") class LLM(BaseLLM): """LLM class that expect subclasses to implement a simpler call method. The purpose of this class is to expose a simpler interface for working with LLMs, rather than expect the user to implement the full _generate method. """ @abstractmethod def _call( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" async def _acall( self, prompt: str, stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> str: """Run the LLM on the given prompt and input.""" raise NotImplementedError("Async generation not implemented for this LLM.") def _generate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[CallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" # TODO: add caching here. generations = [] new_arg_supported = inspect.signature(self._call).parameters.get("run_manager") for prompt in prompts: text = ( self._call(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else self._call(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations) async def _agenerate( self, prompts: List[str], stop: Optional[List[str]] = None, run_manager: Optional[AsyncCallbackManagerForLLMRun] = None, ) -> LLMResult: """Run the LLM on the given prompt and input.""" generations = [] new_arg_supported = inspect.signature(self._acall).parameters.get("run_manager") for prompt in prompts: text = ( await self._acall(prompt, stop=stop, run_manager=run_manager) if new_arg_supported else await self._acall(prompt, stop=stop) ) generations.append([Generation(text=text)]) return LLMResult(generations=generations)
[ "langchain.callbacks.manager.AsyncCallbackManager.configure", "langchain.schema.Generation", "langchain.schema.get_buffer_string", "langchain.callbacks.manager.CallbackManager.configure", "langchain.schema.AIMessage", "langchain.llm_cache.lookup", "langchain.llm_cache.update", "langchain.schema.LLMResult" ]
[((2302, 2339), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (2307, 2339), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2413, 2446), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2418, 2446), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2501, 2534), 'pydantic.Field', 'Field', ([], {'default': 'None', 'exclude': '(True)'}), '(default=None, exclude=True)\n', (2506, 2534), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((2683, 2699), 'pydantic.root_validator', 'root_validator', ([], {}), '()\n', (2697, 2699), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((3135, 3178), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (3144, 3178), False, 'from pydantic import Extra, Field, root_validator, validator\n'), ((5507, 5573), 'langchain.callbacks.manager.CallbackManager.configure', 'CallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (5532, 5573), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((7567, 7624), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (7576, 7624), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((8170, 8241), 'langchain.callbacks.manager.AsyncCallbackManager.configure', 'AsyncCallbackManager.configure', (['callbacks', 'self.callbacks', 'self.verbose'], {}), '(callbacks, self.callbacks, self.verbose)\n', (8200, 8241), False, 'from langchain.callbacks.manager import AsyncCallbackManager, AsyncCallbackManagerForLLMRun, CallbackManager, CallbackManagerForLLMRun, Callbacks\n'), ((10390, 10447), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations', 'llm_output': 'llm_output'}), '(generations=generations, llm_output=llm_output)\n', (10399, 10447), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11768, 11795), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (11785, 11795), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((11947, 11973), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (11956, 11973), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12369, 12396), 'langchain.schema.get_buffer_string', 'get_buffer_string', (['messages'], {}), '(messages)\n', (12386, 12396), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((12566, 12592), 'langchain.schema.AIMessage', 'AIMessage', ([], {'content': 'content'}), '(content=content)\n', (12575, 12592), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((15773, 15807), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (15782, 15807), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16489, 16523), 'langchain.schema.LLMResult', 'LLMResult', ([], {'generations': 'generations'}), '(generations=generations)\n', (16498, 16523), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((1235, 1281), 'langchain.llm_cache.lookup', 'langchain.llm_cache.lookup', (['prompt', 'llm_string'], {}), '(prompt, llm_string)\n', (1261, 1281), False, 'import langchain\n'), ((2023, 2077), 'langchain.llm_cache.update', 'langchain.llm_cache.update', (['prompt', 'llm_string', 'result'], {}), '(prompt, llm_string, result)\n', (2049, 2077), False, 'import langchain\n'), ((2890, 2992), 'warnings.warn', 'warnings.warn', (['"""callback_manager is deprecated. Please use callbacks instead."""', 'DeprecationWarning'], {}), "('callback_manager is deprecated. Please use callbacks instead.',\n DeprecationWarning)\n", (2903, 2992), False, 'import warnings\n'), ((13642, 13657), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (13646, 13657), False, 'from pathlib import Path\n'), ((13978, 14013), 'json.dump', 'json.dump', (['prompt_dict', 'f'], {'indent': '(4)'}), '(prompt_dict, f, indent=4)\n', (13987, 14013), False, 'import json\n'), ((5624, 5657), 'inspect.signature', 'inspect.signature', (['self._generate'], {}), '(self._generate)\n', (5641, 5657), False, 'import inspect\n'), ((8292, 8326), 'inspect.signature', 'inspect.signature', (['self._agenerate'], {}), '(self._agenerate)\n', (8309, 8326), False, 'import inspect\n'), ((14116, 14167), 'yaml.dump', 'yaml.dump', (['prompt_dict', 'f'], {'default_flow_style': '(False)'}), '(prompt_dict, f, default_flow_style=False)\n', (14125, 14167), False, 'import yaml\n'), ((15417, 15446), 'inspect.signature', 'inspect.signature', (['self._call'], {}), '(self._call)\n', (15434, 15446), False, 'import inspect\n'), ((15734, 15755), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (15744, 15755), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n'), ((16118, 16148), 'inspect.signature', 'inspect.signature', (['self._acall'], {}), '(self._acall)\n', (16135, 16148), False, 'import inspect\n'), ((16450, 16471), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (16460, 16471), False, 'from langchain.schema import AIMessage, BaseMessage, Generation, LLMResult, PromptValue, get_buffer_string\n')]
import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Callable, Dict, List, Optional, Sequence import langchain from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat, ) from langchain.schema import AgentAction, AgentFinish, Generation, LLMResult LANGCHAIN_MODEL_NAME = "langchain-model" def import_comet_ml() -> Any: """Import comet_ml and raise an error if it is not installed.""" try: import comet_ml # noqa: F401 except ImportError: raise ImportError( "To use the comet_ml callback manager you need to have the " "`comet_ml` python package installed. Please install it with" " `pip install comet_ml`" ) return comet_ml def _get_experiment( workspace: Optional[str] = None, project_name: Optional[str] = None ) -> Any: comet_ml = import_comet_ml() experiment = comet_ml.Experiment( # type: ignore workspace=workspace, project_name=project_name, ) return experiment def _fetch_text_complexity_metrics(text: str) -> dict: textstat = import_textstat() text_complexity_metrics = { "flesch_reading_ease": textstat.flesch_reading_ease(text), "flesch_kincaid_grade": textstat.flesch_kincaid_grade(text), "smog_index": textstat.smog_index(text), "coleman_liau_index": textstat.coleman_liau_index(text), "automated_readability_index": textstat.automated_readability_index(text), "dale_chall_readability_score": textstat.dale_chall_readability_score(text), "difficult_words": textstat.difficult_words(text), "linsear_write_formula": textstat.linsear_write_formula(text), "gunning_fog": textstat.gunning_fog(text), "text_standard": textstat.text_standard(text), "fernandez_huerta": textstat.fernandez_huerta(text), "szigriszt_pazos": textstat.szigriszt_pazos(text), "gutierrez_polini": textstat.gutierrez_polini(text), "crawford": textstat.crawford(text), "gulpease_index": textstat.gulpease_index(text), "osman": textstat.osman(text), } return text_complexity_metrics def _summarize_metrics_for_generated_outputs(metrics: Sequence) -> dict: pd = import_pandas() metrics_df = pd.DataFrame(metrics) metrics_summary = metrics_df.describe() return metrics_summary.to_dict() class CometCallbackHandler(BaseMetadataCallbackHandler, BaseCallbackHandler): """Callback Handler that logs to Comet. Parameters: job_type (str): The type of comet_ml task such as "inference", "testing" or "qc" project_name (str): The comet_ml project name tags (list): Tags to add to the task task_name (str): Name of the comet_ml task visualize (bool): Whether to visualize the run. complexity_metrics (bool): Whether to log complexity metrics stream_logs (bool): Whether to stream callback actions to Comet This handler will utilize the associated callback method and formats the input of each callback function with metadata regarding the state of LLM run, and adds the response to the list of records for both the {method}_records and action. It then logs the response to Comet. """ def __init__( self, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, stream_logs: bool = True, ) -> None: """Initialize callback handler.""" self.comet_ml = import_comet_ml() super().__init__() self.task_type = task_type self.workspace = workspace self.project_name = project_name self.tags = tags self.visualizations = visualizations self.complexity_metrics = complexity_metrics self.custom_metrics = custom_metrics self.stream_logs = stream_logs self.temp_dir = tempfile.TemporaryDirectory() self.experiment = _get_experiment(workspace, project_name) self.experiment.log_other("Created from", "langchain") if tags: self.experiment.add_tags(tags) self.name = name if self.name: self.experiment.set_name(self.name) warning = ( "The comet_ml callback is currently in beta and is subject to change " "based on updates to `langchain`. Please report any issues to " "https://github.com/comet-ml/issue-tracking/issues with the tag " "`langchain`." ) self.comet_ml.LOGGER.warning(warning) self.callback_columns: list = [] self.action_records: list = [] self.complexity_metrics = complexity_metrics if self.visualizations: spacy = import_spacy() self.nlp = spacy.load("en_core_web_sm") else: self.nlp = None def _init_resp(self) -> Dict: return {k: None for k in self.callback_columns} def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: """Run when LLM starts.""" self.step += 1 self.llm_starts += 1 self.starts += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_start"}) metadata.update(flatten_dict(serialized)) metadata.update(self.get_custom_callback_meta()) for prompt in prompts: prompt_resp = deepcopy(metadata) prompt_resp["prompts"] = prompt self.on_llm_start_records.append(prompt_resp) self.action_records.append(prompt_resp) if self.stream_logs: self._log_stream(prompt, metadata, self.step) def on_llm_new_token(self, token: str, **kwargs: Any) -> None: """Run when LLM generates a new token.""" self.step += 1 self.llm_streams += 1 resp = self._init_resp() resp.update({"action": "on_llm_new_token", "token": token}) resp.update(self.get_custom_callback_meta()) self.action_records.append(resp) def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" self.step += 1 self.llm_ends += 1 self.ends += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_end"}) metadata.update(flatten_dict(response.llm_output or {})) metadata.update(self.get_custom_callback_meta()) output_complexity_metrics = [] output_custom_metrics = [] for prompt_idx, generations in enumerate(response.generations): for gen_idx, generation in enumerate(generations): text = generation.text generation_resp = deepcopy(metadata) generation_resp.update(flatten_dict(generation.dict())) complexity_metrics = self._get_complexity_metrics(text) if complexity_metrics: output_complexity_metrics.append(complexity_metrics) generation_resp.update(complexity_metrics) custom_metrics = self._get_custom_metrics( generation, prompt_idx, gen_idx ) if custom_metrics: output_custom_metrics.append(custom_metrics) generation_resp.update(custom_metrics) if self.stream_logs: self._log_stream(text, metadata, self.step) self.action_records.append(generation_resp) self.on_llm_end_records.append(generation_resp) self._log_text_metrics(output_complexity_metrics, step=self.step) self._log_text_metrics(output_custom_metrics, step=self.step) def on_llm_error(self, error: BaseException, **kwargs: Any) -> None: """Run when LLM errors.""" self.step += 1 self.errors += 1 def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], **kwargs: Any ) -> None: """Run when chain starts running.""" self.step += 1 self.chain_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_chain_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) for chain_input_key, chain_input_val in inputs.items(): if isinstance(chain_input_val, str): input_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_input_val, resp, self.step) input_resp.update({chain_input_key: chain_input_val}) self.action_records.append(input_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Input Value for {chain_input_key} will not be logged" ) def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" self.step += 1 self.chain_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_chain_end"}) resp.update(self.get_custom_callback_meta()) for chain_output_key, chain_output_val in outputs.items(): if isinstance(chain_output_val, str): output_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_output_val, resp, self.step) output_resp.update({chain_output_key: chain_output_val}) self.action_records.append(output_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Output Value for {chain_output_key} will not be logged" ) def on_chain_error(self, error: BaseException, **kwargs: Any) -> None: """Run when chain errors.""" self.step += 1 self.errors += 1 def on_tool_start( self, serialized: Dict[str, Any], input_str: str, **kwargs: Any ) -> None: """Run when tool starts running.""" self.step += 1 self.tool_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_tool_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(input_str, resp, self.step) resp.update({"input_str": input_str}) self.action_records.append(resp) def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" self.step += 1 self.tool_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_tool_end"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_tool_error(self, error: BaseException, **kwargs: Any) -> None: """Run when tool errors.""" self.step += 1 self.errors += 1 def on_text(self, text: str, **kwargs: Any) -> None: """ Run when agent is ending. """ self.step += 1 self.text_ctr += 1 resp = self._init_resp() resp.update({"action": "on_text"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(text, resp, self.step) resp.update({"text": text}) self.action_records.append(resp) def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> None: """Run when agent ends running.""" self.step += 1 self.agent_ends += 1 self.ends += 1 resp = self._init_resp() output = finish.return_values["output"] log = finish.log resp.update({"action": "on_agent_finish", "log": log}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run on agent action.""" self.step += 1 self.tool_starts += 1 self.starts += 1 tool = action.tool tool_input = str(action.tool_input) log = action.log resp = self._init_resp() resp.update({"action": "on_agent_action", "log": log, "tool": tool}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(tool_input, resp, self.step) resp.update({"tool_input": tool_input}) self.action_records.append(resp) def _get_complexity_metrics(self, text: str) -> dict: """Compute text complexity metrics using textstat. Parameters: text (str): The text to analyze. Returns: (dict): A dictionary containing the complexity metrics. """ resp = {} if self.complexity_metrics: text_complexity_metrics = _fetch_text_complexity_metrics(text) resp.update(text_complexity_metrics) return resp def _get_custom_metrics( self, generation: Generation, prompt_idx: int, gen_idx: int ) -> dict: """Compute Custom Metrics for an LLM Generated Output Args: generation (LLMResult): Output generation from an LLM prompt_idx (int): List index of the input prompt gen_idx (int): List index of the generated output Returns: dict: A dictionary containing the custom metrics. """ resp = {} if self.custom_metrics: custom_metrics = self.custom_metrics(generation, prompt_idx, gen_idx) resp.update(custom_metrics) return resp def flush_tracker( self, langchain_asset: Any = None, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = "comet-langchain-demo", tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, finish: bool = False, reset: bool = False, ) -> None: """Flush the tracker and setup the session. Everything after this will be a new table. Args: name: Name of the performed session so far so it is identifiable langchain_asset: The langchain asset to save. finish: Whether to finish the run. Returns: None """ self._log_session(langchain_asset) if langchain_asset: try: self._log_model(langchain_asset) except Exception: self.comet_ml.LOGGER.error( "Failed to export agent or LLM to Comet", exc_info=True, extra={"show_traceback": True}, ) if finish: self.experiment.end() if reset: self._reset( task_type, workspace, project_name, tags, name, visualizations, complexity_metrics, custom_metrics, ) def _log_stream(self, prompt: str, metadata: dict, step: int) -> None: self.experiment.log_text(prompt, metadata=metadata, step=step) def _log_model(self, langchain_asset: Any) -> None: model_parameters = self._get_llm_parameters(langchain_asset) self.experiment.log_parameters(model_parameters, prefix="model") langchain_asset_path = Path(self.temp_dir.name, "model.json") model_name = self.name if self.name else LANGCHAIN_MODEL_NAME try: if hasattr(langchain_asset, "save"): langchain_asset.save(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) except (ValueError, AttributeError, NotImplementedError) as e: if hasattr(langchain_asset, "save_agent"): langchain_asset.save_agent(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) else: self.comet_ml.LOGGER.error( f"{e}" " Could not save Langchain Asset " f"for {langchain_asset.__class__.__name__}" ) def _log_session(self, langchain_asset: Optional[Any] = None) -> None: try: llm_session_df = self._create_session_analysis_dataframe(langchain_asset) # Log the cleaned dataframe as a table self.experiment.log_table("langchain-llm-session.csv", llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: metadata = {"langchain_version": str(langchain.__version__)} # Log the langchain low-level records as a JSON file directly self.experiment.log_asset_data( self.action_records, "langchain-action_records.json", metadata=metadata ) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: self._log_visualizations(llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log visualizations to Comet", exc_info=True, extra={"show_traceback": True}, ) def _log_text_metrics(self, metrics: Sequence[dict], step: int) -> None: if not metrics: return metrics_summary = _summarize_metrics_for_generated_outputs(metrics) for key, value in metrics_summary.items(): self.experiment.log_metrics(value, prefix=key, step=step) def _log_visualizations(self, session_df: Any) -> None: if not (self.visualizations and self.nlp): return spacy = import_spacy() prompts = session_df["prompts"].tolist() outputs = session_df["text"].tolist() for idx, (prompt, output) in enumerate(zip(prompts, outputs)): doc = self.nlp(output) sentence_spans = list(doc.sents) for visualization in self.visualizations: try: html = spacy.displacy.render( sentence_spans, style=visualization, options={"compact": True}, jupyter=False, page=True, ) self.experiment.log_asset_data( html, name=f"langchain-viz-{visualization}-{idx}.html", metadata={"prompt": prompt}, step=idx, ) except Exception as e: self.comet_ml.LOGGER.warning( e, exc_info=True, extra={"show_traceback": True} ) return def _reset( self, task_type: Optional[str] = None, workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, ) -> None: _task_type = task_type if task_type else self.task_type _workspace = workspace if workspace else self.workspace _project_name = project_name if project_name else self.project_name _tags = tags if tags else self.tags _name = name if name else self.name _visualizations = visualizations if visualizations else self.visualizations _complexity_metrics = ( complexity_metrics if complexity_metrics else self.complexity_metrics ) _custom_metrics = custom_metrics if custom_metrics else self.custom_metrics self.__init__( # type: ignore task_type=_task_type, workspace=_workspace, project_name=_project_name, tags=_tags, name=_name, visualizations=_visualizations, complexity_metrics=_complexity_metrics, custom_metrics=_custom_metrics, ) self.reset_callback_meta() self.temp_dir = tempfile.TemporaryDirectory() def _create_session_analysis_dataframe(self, langchain_asset: Any = None) -> dict: pd = import_pandas() llm_parameters = self._get_llm_parameters(langchain_asset) num_generations_per_prompt = llm_parameters.get("n", 1) llm_start_records_df = pd.DataFrame(self.on_llm_start_records) # Repeat each input row based on the number of outputs generated per prompt llm_start_records_df = llm_start_records_df.loc[ llm_start_records_df.index.repeat(num_generations_per_prompt) ].reset_index(drop=True) llm_end_records_df = pd.DataFrame(self.on_llm_end_records) llm_session_df = pd.merge( llm_start_records_df, llm_end_records_df, left_index=True, right_index=True, suffixes=["_llm_start", "_llm_end"], ) return llm_session_df def _get_llm_parameters(self, langchain_asset: Any = None) -> dict: if not langchain_asset: return {} try: if hasattr(langchain_asset, "agent"): llm_parameters = langchain_asset.agent.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm_chain"): llm_parameters = langchain_asset.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm"): llm_parameters = langchain_asset.llm.dict() else: llm_parameters = langchain_asset.dict() except Exception: return {} return llm_parameters
[ "langchain.callbacks.utils.import_spacy", "langchain.callbacks.utils.import_pandas", "langchain.callbacks.utils.import_textstat", "langchain.callbacks.utils.flatten_dict" ]
[((1047, 1114), 'comet_ml.Experiment', 'comet_ml.Experiment', ([], {'workspace': 'workspace', 'project_name': 'project_name'}), '(workspace=workspace, project_name=project_name)\n', (1066, 1114), False, 'import comet_ml\n'), ((1249, 1266), 'langchain.callbacks.utils.import_textstat', 'import_textstat', ([], {}), '()\n', (1264, 1266), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((2400, 2415), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (2413, 2415), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((4318, 4347), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (4345, 4347), False, 'import tempfile\n'), ((16471, 16509), 'pathlib.Path', 'Path', (['self.temp_dir.name', '"""model.json"""'], {}), "(self.temp_dir.name, 'model.json')\n", (16475, 16509), False, 'from pathlib import Path\n'), ((19081, 19095), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (19093, 19095), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((21551, 21580), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (21578, 21580), False, 'import tempfile\n'), ((21682, 21697), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (21695, 21697), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5161, 5175), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (5173, 5175), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5701, 5725), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (5713, 5725), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5842, 5860), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (5850, 5860), False, 'from copy import deepcopy\n'), ((6776, 6815), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['(response.llm_output or {})'], {}), '(response.llm_output or {})\n', (6788, 6815), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((8664, 8688), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (8676, 8688), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((10795, 10819), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (10807, 10819), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((7159, 7177), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (7167, 7177), False, 'from copy import deepcopy\n'), ((8886, 8900), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (8894, 8900), False, 'from copy import deepcopy\n'), ((9828, 9842), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (9836, 9842), False, 'from copy import deepcopy\n')]
import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Callable, Dict, List, Optional, Sequence import langchain from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat, ) from langchain.schema import AgentAction, AgentFinish, Generation, LLMResult LANGCHAIN_MODEL_NAME = "langchain-model" def import_comet_ml() -> Any: """Import comet_ml and raise an error if it is not installed.""" try: import comet_ml # noqa: F401 except ImportError: raise ImportError( "To use the comet_ml callback manager you need to have the " "`comet_ml` python package installed. Please install it with" " `pip install comet_ml`" ) return comet_ml def _get_experiment( workspace: Optional[str] = None, project_name: Optional[str] = None ) -> Any: comet_ml = import_comet_ml() experiment = comet_ml.Experiment( # type: ignore workspace=workspace, project_name=project_name, ) return experiment def _fetch_text_complexity_metrics(text: str) -> dict: textstat = import_textstat() text_complexity_metrics = { "flesch_reading_ease": textstat.flesch_reading_ease(text), "flesch_kincaid_grade": textstat.flesch_kincaid_grade(text), "smog_index": textstat.smog_index(text), "coleman_liau_index": textstat.coleman_liau_index(text), "automated_readability_index": textstat.automated_readability_index(text), "dale_chall_readability_score": textstat.dale_chall_readability_score(text), "difficult_words": textstat.difficult_words(text), "linsear_write_formula": textstat.linsear_write_formula(text), "gunning_fog": textstat.gunning_fog(text), "text_standard": textstat.text_standard(text), "fernandez_huerta": textstat.fernandez_huerta(text), "szigriszt_pazos": textstat.szigriszt_pazos(text), "gutierrez_polini": textstat.gutierrez_polini(text), "crawford": textstat.crawford(text), "gulpease_index": textstat.gulpease_index(text), "osman": textstat.osman(text), } return text_complexity_metrics def _summarize_metrics_for_generated_outputs(metrics: Sequence) -> dict: pd = import_pandas() metrics_df = pd.DataFrame(metrics) metrics_summary = metrics_df.describe() return metrics_summary.to_dict() class CometCallbackHandler(BaseMetadataCallbackHandler, BaseCallbackHandler): """Callback Handler that logs to Comet. Parameters: job_type (str): The type of comet_ml task such as "inference", "testing" or "qc" project_name (str): The comet_ml project name tags (list): Tags to add to the task task_name (str): Name of the comet_ml task visualize (bool): Whether to visualize the run. complexity_metrics (bool): Whether to log complexity metrics stream_logs (bool): Whether to stream callback actions to Comet This handler will utilize the associated callback method and formats the input of each callback function with metadata regarding the state of LLM run, and adds the response to the list of records for both the {method}_records and action. It then logs the response to Comet. """ def __init__( self, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, stream_logs: bool = True, ) -> None: """Initialize callback handler.""" self.comet_ml = import_comet_ml() super().__init__() self.task_type = task_type self.workspace = workspace self.project_name = project_name self.tags = tags self.visualizations = visualizations self.complexity_metrics = complexity_metrics self.custom_metrics = custom_metrics self.stream_logs = stream_logs self.temp_dir = tempfile.TemporaryDirectory() self.experiment = _get_experiment(workspace, project_name) self.experiment.log_other("Created from", "langchain") if tags: self.experiment.add_tags(tags) self.name = name if self.name: self.experiment.set_name(self.name) warning = ( "The comet_ml callback is currently in beta and is subject to change " "based on updates to `langchain`. Please report any issues to " "https://github.com/comet-ml/issue-tracking/issues with the tag " "`langchain`." ) self.comet_ml.LOGGER.warning(warning) self.callback_columns: list = [] self.action_records: list = [] self.complexity_metrics = complexity_metrics if self.visualizations: spacy = import_spacy() self.nlp = spacy.load("en_core_web_sm") else: self.nlp = None def _init_resp(self) -> Dict: return {k: None for k in self.callback_columns} def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: """Run when LLM starts.""" self.step += 1 self.llm_starts += 1 self.starts += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_start"}) metadata.update(flatten_dict(serialized)) metadata.update(self.get_custom_callback_meta()) for prompt in prompts: prompt_resp = deepcopy(metadata) prompt_resp["prompts"] = prompt self.on_llm_start_records.append(prompt_resp) self.action_records.append(prompt_resp) if self.stream_logs: self._log_stream(prompt, metadata, self.step) def on_llm_new_token(self, token: str, **kwargs: Any) -> None: """Run when LLM generates a new token.""" self.step += 1 self.llm_streams += 1 resp = self._init_resp() resp.update({"action": "on_llm_new_token", "token": token}) resp.update(self.get_custom_callback_meta()) self.action_records.append(resp) def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" self.step += 1 self.llm_ends += 1 self.ends += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_end"}) metadata.update(flatten_dict(response.llm_output or {})) metadata.update(self.get_custom_callback_meta()) output_complexity_metrics = [] output_custom_metrics = [] for prompt_idx, generations in enumerate(response.generations): for gen_idx, generation in enumerate(generations): text = generation.text generation_resp = deepcopy(metadata) generation_resp.update(flatten_dict(generation.dict())) complexity_metrics = self._get_complexity_metrics(text) if complexity_metrics: output_complexity_metrics.append(complexity_metrics) generation_resp.update(complexity_metrics) custom_metrics = self._get_custom_metrics( generation, prompt_idx, gen_idx ) if custom_metrics: output_custom_metrics.append(custom_metrics) generation_resp.update(custom_metrics) if self.stream_logs: self._log_stream(text, metadata, self.step) self.action_records.append(generation_resp) self.on_llm_end_records.append(generation_resp) self._log_text_metrics(output_complexity_metrics, step=self.step) self._log_text_metrics(output_custom_metrics, step=self.step) def on_llm_error(self, error: BaseException, **kwargs: Any) -> None: """Run when LLM errors.""" self.step += 1 self.errors += 1 def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], **kwargs: Any ) -> None: """Run when chain starts running.""" self.step += 1 self.chain_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_chain_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) for chain_input_key, chain_input_val in inputs.items(): if isinstance(chain_input_val, str): input_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_input_val, resp, self.step) input_resp.update({chain_input_key: chain_input_val}) self.action_records.append(input_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Input Value for {chain_input_key} will not be logged" ) def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" self.step += 1 self.chain_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_chain_end"}) resp.update(self.get_custom_callback_meta()) for chain_output_key, chain_output_val in outputs.items(): if isinstance(chain_output_val, str): output_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_output_val, resp, self.step) output_resp.update({chain_output_key: chain_output_val}) self.action_records.append(output_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Output Value for {chain_output_key} will not be logged" ) def on_chain_error(self, error: BaseException, **kwargs: Any) -> None: """Run when chain errors.""" self.step += 1 self.errors += 1 def on_tool_start( self, serialized: Dict[str, Any], input_str: str, **kwargs: Any ) -> None: """Run when tool starts running.""" self.step += 1 self.tool_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_tool_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(input_str, resp, self.step) resp.update({"input_str": input_str}) self.action_records.append(resp) def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" self.step += 1 self.tool_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_tool_end"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_tool_error(self, error: BaseException, **kwargs: Any) -> None: """Run when tool errors.""" self.step += 1 self.errors += 1 def on_text(self, text: str, **kwargs: Any) -> None: """ Run when agent is ending. """ self.step += 1 self.text_ctr += 1 resp = self._init_resp() resp.update({"action": "on_text"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(text, resp, self.step) resp.update({"text": text}) self.action_records.append(resp) def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> None: """Run when agent ends running.""" self.step += 1 self.agent_ends += 1 self.ends += 1 resp = self._init_resp() output = finish.return_values["output"] log = finish.log resp.update({"action": "on_agent_finish", "log": log}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run on agent action.""" self.step += 1 self.tool_starts += 1 self.starts += 1 tool = action.tool tool_input = str(action.tool_input) log = action.log resp = self._init_resp() resp.update({"action": "on_agent_action", "log": log, "tool": tool}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(tool_input, resp, self.step) resp.update({"tool_input": tool_input}) self.action_records.append(resp) def _get_complexity_metrics(self, text: str) -> dict: """Compute text complexity metrics using textstat. Parameters: text (str): The text to analyze. Returns: (dict): A dictionary containing the complexity metrics. """ resp = {} if self.complexity_metrics: text_complexity_metrics = _fetch_text_complexity_metrics(text) resp.update(text_complexity_metrics) return resp def _get_custom_metrics( self, generation: Generation, prompt_idx: int, gen_idx: int ) -> dict: """Compute Custom Metrics for an LLM Generated Output Args: generation (LLMResult): Output generation from an LLM prompt_idx (int): List index of the input prompt gen_idx (int): List index of the generated output Returns: dict: A dictionary containing the custom metrics. """ resp = {} if self.custom_metrics: custom_metrics = self.custom_metrics(generation, prompt_idx, gen_idx) resp.update(custom_metrics) return resp def flush_tracker( self, langchain_asset: Any = None, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = "comet-langchain-demo", tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, finish: bool = False, reset: bool = False, ) -> None: """Flush the tracker and setup the session. Everything after this will be a new table. Args: name: Name of the performed session so far so it is identifiable langchain_asset: The langchain asset to save. finish: Whether to finish the run. Returns: None """ self._log_session(langchain_asset) if langchain_asset: try: self._log_model(langchain_asset) except Exception: self.comet_ml.LOGGER.error( "Failed to export agent or LLM to Comet", exc_info=True, extra={"show_traceback": True}, ) if finish: self.experiment.end() if reset: self._reset( task_type, workspace, project_name, tags, name, visualizations, complexity_metrics, custom_metrics, ) def _log_stream(self, prompt: str, metadata: dict, step: int) -> None: self.experiment.log_text(prompt, metadata=metadata, step=step) def _log_model(self, langchain_asset: Any) -> None: model_parameters = self._get_llm_parameters(langchain_asset) self.experiment.log_parameters(model_parameters, prefix="model") langchain_asset_path = Path(self.temp_dir.name, "model.json") model_name = self.name if self.name else LANGCHAIN_MODEL_NAME try: if hasattr(langchain_asset, "save"): langchain_asset.save(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) except (ValueError, AttributeError, NotImplementedError) as e: if hasattr(langchain_asset, "save_agent"): langchain_asset.save_agent(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) else: self.comet_ml.LOGGER.error( f"{e}" " Could not save Langchain Asset " f"for {langchain_asset.__class__.__name__}" ) def _log_session(self, langchain_asset: Optional[Any] = None) -> None: try: llm_session_df = self._create_session_analysis_dataframe(langchain_asset) # Log the cleaned dataframe as a table self.experiment.log_table("langchain-llm-session.csv", llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: metadata = {"langchain_version": str(langchain.__version__)} # Log the langchain low-level records as a JSON file directly self.experiment.log_asset_data( self.action_records, "langchain-action_records.json", metadata=metadata ) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: self._log_visualizations(llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log visualizations to Comet", exc_info=True, extra={"show_traceback": True}, ) def _log_text_metrics(self, metrics: Sequence[dict], step: int) -> None: if not metrics: return metrics_summary = _summarize_metrics_for_generated_outputs(metrics) for key, value in metrics_summary.items(): self.experiment.log_metrics(value, prefix=key, step=step) def _log_visualizations(self, session_df: Any) -> None: if not (self.visualizations and self.nlp): return spacy = import_spacy() prompts = session_df["prompts"].tolist() outputs = session_df["text"].tolist() for idx, (prompt, output) in enumerate(zip(prompts, outputs)): doc = self.nlp(output) sentence_spans = list(doc.sents) for visualization in self.visualizations: try: html = spacy.displacy.render( sentence_spans, style=visualization, options={"compact": True}, jupyter=False, page=True, ) self.experiment.log_asset_data( html, name=f"langchain-viz-{visualization}-{idx}.html", metadata={"prompt": prompt}, step=idx, ) except Exception as e: self.comet_ml.LOGGER.warning( e, exc_info=True, extra={"show_traceback": True} ) return def _reset( self, task_type: Optional[str] = None, workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, ) -> None: _task_type = task_type if task_type else self.task_type _workspace = workspace if workspace else self.workspace _project_name = project_name if project_name else self.project_name _tags = tags if tags else self.tags _name = name if name else self.name _visualizations = visualizations if visualizations else self.visualizations _complexity_metrics = ( complexity_metrics if complexity_metrics else self.complexity_metrics ) _custom_metrics = custom_metrics if custom_metrics else self.custom_metrics self.__init__( # type: ignore task_type=_task_type, workspace=_workspace, project_name=_project_name, tags=_tags, name=_name, visualizations=_visualizations, complexity_metrics=_complexity_metrics, custom_metrics=_custom_metrics, ) self.reset_callback_meta() self.temp_dir = tempfile.TemporaryDirectory() def _create_session_analysis_dataframe(self, langchain_asset: Any = None) -> dict: pd = import_pandas() llm_parameters = self._get_llm_parameters(langchain_asset) num_generations_per_prompt = llm_parameters.get("n", 1) llm_start_records_df = pd.DataFrame(self.on_llm_start_records) # Repeat each input row based on the number of outputs generated per prompt llm_start_records_df = llm_start_records_df.loc[ llm_start_records_df.index.repeat(num_generations_per_prompt) ].reset_index(drop=True) llm_end_records_df = pd.DataFrame(self.on_llm_end_records) llm_session_df = pd.merge( llm_start_records_df, llm_end_records_df, left_index=True, right_index=True, suffixes=["_llm_start", "_llm_end"], ) return llm_session_df def _get_llm_parameters(self, langchain_asset: Any = None) -> dict: if not langchain_asset: return {} try: if hasattr(langchain_asset, "agent"): llm_parameters = langchain_asset.agent.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm_chain"): llm_parameters = langchain_asset.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm"): llm_parameters = langchain_asset.llm.dict() else: llm_parameters = langchain_asset.dict() except Exception: return {} return llm_parameters
[ "langchain.callbacks.utils.import_spacy", "langchain.callbacks.utils.import_pandas", "langchain.callbacks.utils.import_textstat", "langchain.callbacks.utils.flatten_dict" ]
[((1047, 1114), 'comet_ml.Experiment', 'comet_ml.Experiment', ([], {'workspace': 'workspace', 'project_name': 'project_name'}), '(workspace=workspace, project_name=project_name)\n', (1066, 1114), False, 'import comet_ml\n'), ((1249, 1266), 'langchain.callbacks.utils.import_textstat', 'import_textstat', ([], {}), '()\n', (1264, 1266), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((2400, 2415), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (2413, 2415), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((4318, 4347), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (4345, 4347), False, 'import tempfile\n'), ((16471, 16509), 'pathlib.Path', 'Path', (['self.temp_dir.name', '"""model.json"""'], {}), "(self.temp_dir.name, 'model.json')\n", (16475, 16509), False, 'from pathlib import Path\n'), ((19081, 19095), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (19093, 19095), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((21551, 21580), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (21578, 21580), False, 'import tempfile\n'), ((21682, 21697), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (21695, 21697), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5161, 5175), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (5173, 5175), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5701, 5725), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (5713, 5725), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5842, 5860), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (5850, 5860), False, 'from copy import deepcopy\n'), ((6776, 6815), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['(response.llm_output or {})'], {}), '(response.llm_output or {})\n', (6788, 6815), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((8664, 8688), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (8676, 8688), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((10795, 10819), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (10807, 10819), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((7159, 7177), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (7167, 7177), False, 'from copy import deepcopy\n'), ((8886, 8900), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (8894, 8900), False, 'from copy import deepcopy\n'), ((9828, 9842), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (9836, 9842), False, 'from copy import deepcopy\n')]
import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Callable, Dict, List, Optional, Sequence import langchain from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat, ) from langchain.schema import AgentAction, AgentFinish, Generation, LLMResult LANGCHAIN_MODEL_NAME = "langchain-model" def import_comet_ml() -> Any: """Import comet_ml and raise an error if it is not installed.""" try: import comet_ml # noqa: F401 except ImportError: raise ImportError( "To use the comet_ml callback manager you need to have the " "`comet_ml` python package installed. Please install it with" " `pip install comet_ml`" ) return comet_ml def _get_experiment( workspace: Optional[str] = None, project_name: Optional[str] = None ) -> Any: comet_ml = import_comet_ml() experiment = comet_ml.Experiment( # type: ignore workspace=workspace, project_name=project_name, ) return experiment def _fetch_text_complexity_metrics(text: str) -> dict: textstat = import_textstat() text_complexity_metrics = { "flesch_reading_ease": textstat.flesch_reading_ease(text), "flesch_kincaid_grade": textstat.flesch_kincaid_grade(text), "smog_index": textstat.smog_index(text), "coleman_liau_index": textstat.coleman_liau_index(text), "automated_readability_index": textstat.automated_readability_index(text), "dale_chall_readability_score": textstat.dale_chall_readability_score(text), "difficult_words": textstat.difficult_words(text), "linsear_write_formula": textstat.linsear_write_formula(text), "gunning_fog": textstat.gunning_fog(text), "text_standard": textstat.text_standard(text), "fernandez_huerta": textstat.fernandez_huerta(text), "szigriszt_pazos": textstat.szigriszt_pazos(text), "gutierrez_polini": textstat.gutierrez_polini(text), "crawford": textstat.crawford(text), "gulpease_index": textstat.gulpease_index(text), "osman": textstat.osman(text), } return text_complexity_metrics def _summarize_metrics_for_generated_outputs(metrics: Sequence) -> dict: pd = import_pandas() metrics_df = pd.DataFrame(metrics) metrics_summary = metrics_df.describe() return metrics_summary.to_dict() class CometCallbackHandler(BaseMetadataCallbackHandler, BaseCallbackHandler): """Callback Handler that logs to Comet. Parameters: job_type (str): The type of comet_ml task such as "inference", "testing" or "qc" project_name (str): The comet_ml project name tags (list): Tags to add to the task task_name (str): Name of the comet_ml task visualize (bool): Whether to visualize the run. complexity_metrics (bool): Whether to log complexity metrics stream_logs (bool): Whether to stream callback actions to Comet This handler will utilize the associated callback method and formats the input of each callback function with metadata regarding the state of LLM run, and adds the response to the list of records for both the {method}_records and action. It then logs the response to Comet. """ def __init__( self, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, stream_logs: bool = True, ) -> None: """Initialize callback handler.""" self.comet_ml = import_comet_ml() super().__init__() self.task_type = task_type self.workspace = workspace self.project_name = project_name self.tags = tags self.visualizations = visualizations self.complexity_metrics = complexity_metrics self.custom_metrics = custom_metrics self.stream_logs = stream_logs self.temp_dir = tempfile.TemporaryDirectory() self.experiment = _get_experiment(workspace, project_name) self.experiment.log_other("Created from", "langchain") if tags: self.experiment.add_tags(tags) self.name = name if self.name: self.experiment.set_name(self.name) warning = ( "The comet_ml callback is currently in beta and is subject to change " "based on updates to `langchain`. Please report any issues to " "https://github.com/comet-ml/issue-tracking/issues with the tag " "`langchain`." ) self.comet_ml.LOGGER.warning(warning) self.callback_columns: list = [] self.action_records: list = [] self.complexity_metrics = complexity_metrics if self.visualizations: spacy = import_spacy() self.nlp = spacy.load("en_core_web_sm") else: self.nlp = None def _init_resp(self) -> Dict: return {k: None for k in self.callback_columns} def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: """Run when LLM starts.""" self.step += 1 self.llm_starts += 1 self.starts += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_start"}) metadata.update(flatten_dict(serialized)) metadata.update(self.get_custom_callback_meta()) for prompt in prompts: prompt_resp = deepcopy(metadata) prompt_resp["prompts"] = prompt self.on_llm_start_records.append(prompt_resp) self.action_records.append(prompt_resp) if self.stream_logs: self._log_stream(prompt, metadata, self.step) def on_llm_new_token(self, token: str, **kwargs: Any) -> None: """Run when LLM generates a new token.""" self.step += 1 self.llm_streams += 1 resp = self._init_resp() resp.update({"action": "on_llm_new_token", "token": token}) resp.update(self.get_custom_callback_meta()) self.action_records.append(resp) def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" self.step += 1 self.llm_ends += 1 self.ends += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_end"}) metadata.update(flatten_dict(response.llm_output or {})) metadata.update(self.get_custom_callback_meta()) output_complexity_metrics = [] output_custom_metrics = [] for prompt_idx, generations in enumerate(response.generations): for gen_idx, generation in enumerate(generations): text = generation.text generation_resp = deepcopy(metadata) generation_resp.update(flatten_dict(generation.dict())) complexity_metrics = self._get_complexity_metrics(text) if complexity_metrics: output_complexity_metrics.append(complexity_metrics) generation_resp.update(complexity_metrics) custom_metrics = self._get_custom_metrics( generation, prompt_idx, gen_idx ) if custom_metrics: output_custom_metrics.append(custom_metrics) generation_resp.update(custom_metrics) if self.stream_logs: self._log_stream(text, metadata, self.step) self.action_records.append(generation_resp) self.on_llm_end_records.append(generation_resp) self._log_text_metrics(output_complexity_metrics, step=self.step) self._log_text_metrics(output_custom_metrics, step=self.step) def on_llm_error(self, error: BaseException, **kwargs: Any) -> None: """Run when LLM errors.""" self.step += 1 self.errors += 1 def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], **kwargs: Any ) -> None: """Run when chain starts running.""" self.step += 1 self.chain_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_chain_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) for chain_input_key, chain_input_val in inputs.items(): if isinstance(chain_input_val, str): input_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_input_val, resp, self.step) input_resp.update({chain_input_key: chain_input_val}) self.action_records.append(input_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Input Value for {chain_input_key} will not be logged" ) def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" self.step += 1 self.chain_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_chain_end"}) resp.update(self.get_custom_callback_meta()) for chain_output_key, chain_output_val in outputs.items(): if isinstance(chain_output_val, str): output_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_output_val, resp, self.step) output_resp.update({chain_output_key: chain_output_val}) self.action_records.append(output_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Output Value for {chain_output_key} will not be logged" ) def on_chain_error(self, error: BaseException, **kwargs: Any) -> None: """Run when chain errors.""" self.step += 1 self.errors += 1 def on_tool_start( self, serialized: Dict[str, Any], input_str: str, **kwargs: Any ) -> None: """Run when tool starts running.""" self.step += 1 self.tool_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_tool_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(input_str, resp, self.step) resp.update({"input_str": input_str}) self.action_records.append(resp) def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" self.step += 1 self.tool_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_tool_end"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_tool_error(self, error: BaseException, **kwargs: Any) -> None: """Run when tool errors.""" self.step += 1 self.errors += 1 def on_text(self, text: str, **kwargs: Any) -> None: """ Run when agent is ending. """ self.step += 1 self.text_ctr += 1 resp = self._init_resp() resp.update({"action": "on_text"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(text, resp, self.step) resp.update({"text": text}) self.action_records.append(resp) def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> None: """Run when agent ends running.""" self.step += 1 self.agent_ends += 1 self.ends += 1 resp = self._init_resp() output = finish.return_values["output"] log = finish.log resp.update({"action": "on_agent_finish", "log": log}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run on agent action.""" self.step += 1 self.tool_starts += 1 self.starts += 1 tool = action.tool tool_input = str(action.tool_input) log = action.log resp = self._init_resp() resp.update({"action": "on_agent_action", "log": log, "tool": tool}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(tool_input, resp, self.step) resp.update({"tool_input": tool_input}) self.action_records.append(resp) def _get_complexity_metrics(self, text: str) -> dict: """Compute text complexity metrics using textstat. Parameters: text (str): The text to analyze. Returns: (dict): A dictionary containing the complexity metrics. """ resp = {} if self.complexity_metrics: text_complexity_metrics = _fetch_text_complexity_metrics(text) resp.update(text_complexity_metrics) return resp def _get_custom_metrics( self, generation: Generation, prompt_idx: int, gen_idx: int ) -> dict: """Compute Custom Metrics for an LLM Generated Output Args: generation (LLMResult): Output generation from an LLM prompt_idx (int): List index of the input prompt gen_idx (int): List index of the generated output Returns: dict: A dictionary containing the custom metrics. """ resp = {} if self.custom_metrics: custom_metrics = self.custom_metrics(generation, prompt_idx, gen_idx) resp.update(custom_metrics) return resp def flush_tracker( self, langchain_asset: Any = None, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = "comet-langchain-demo", tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, finish: bool = False, reset: bool = False, ) -> None: """Flush the tracker and setup the session. Everything after this will be a new table. Args: name: Name of the performed session so far so it is identifiable langchain_asset: The langchain asset to save. finish: Whether to finish the run. Returns: None """ self._log_session(langchain_asset) if langchain_asset: try: self._log_model(langchain_asset) except Exception: self.comet_ml.LOGGER.error( "Failed to export agent or LLM to Comet", exc_info=True, extra={"show_traceback": True}, ) if finish: self.experiment.end() if reset: self._reset( task_type, workspace, project_name, tags, name, visualizations, complexity_metrics, custom_metrics, ) def _log_stream(self, prompt: str, metadata: dict, step: int) -> None: self.experiment.log_text(prompt, metadata=metadata, step=step) def _log_model(self, langchain_asset: Any) -> None: model_parameters = self._get_llm_parameters(langchain_asset) self.experiment.log_parameters(model_parameters, prefix="model") langchain_asset_path = Path(self.temp_dir.name, "model.json") model_name = self.name if self.name else LANGCHAIN_MODEL_NAME try: if hasattr(langchain_asset, "save"): langchain_asset.save(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) except (ValueError, AttributeError, NotImplementedError) as e: if hasattr(langchain_asset, "save_agent"): langchain_asset.save_agent(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) else: self.comet_ml.LOGGER.error( f"{e}" " Could not save Langchain Asset " f"for {langchain_asset.__class__.__name__}" ) def _log_session(self, langchain_asset: Optional[Any] = None) -> None: try: llm_session_df = self._create_session_analysis_dataframe(langchain_asset) # Log the cleaned dataframe as a table self.experiment.log_table("langchain-llm-session.csv", llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: metadata = {"langchain_version": str(langchain.__version__)} # Log the langchain low-level records as a JSON file directly self.experiment.log_asset_data( self.action_records, "langchain-action_records.json", metadata=metadata ) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: self._log_visualizations(llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log visualizations to Comet", exc_info=True, extra={"show_traceback": True}, ) def _log_text_metrics(self, metrics: Sequence[dict], step: int) -> None: if not metrics: return metrics_summary = _summarize_metrics_for_generated_outputs(metrics) for key, value in metrics_summary.items(): self.experiment.log_metrics(value, prefix=key, step=step) def _log_visualizations(self, session_df: Any) -> None: if not (self.visualizations and self.nlp): return spacy = import_spacy() prompts = session_df["prompts"].tolist() outputs = session_df["text"].tolist() for idx, (prompt, output) in enumerate(zip(prompts, outputs)): doc = self.nlp(output) sentence_spans = list(doc.sents) for visualization in self.visualizations: try: html = spacy.displacy.render( sentence_spans, style=visualization, options={"compact": True}, jupyter=False, page=True, ) self.experiment.log_asset_data( html, name=f"langchain-viz-{visualization}-{idx}.html", metadata={"prompt": prompt}, step=idx, ) except Exception as e: self.comet_ml.LOGGER.warning( e, exc_info=True, extra={"show_traceback": True} ) return def _reset( self, task_type: Optional[str] = None, workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, ) -> None: _task_type = task_type if task_type else self.task_type _workspace = workspace if workspace else self.workspace _project_name = project_name if project_name else self.project_name _tags = tags if tags else self.tags _name = name if name else self.name _visualizations = visualizations if visualizations else self.visualizations _complexity_metrics = ( complexity_metrics if complexity_metrics else self.complexity_metrics ) _custom_metrics = custom_metrics if custom_metrics else self.custom_metrics self.__init__( # type: ignore task_type=_task_type, workspace=_workspace, project_name=_project_name, tags=_tags, name=_name, visualizations=_visualizations, complexity_metrics=_complexity_metrics, custom_metrics=_custom_metrics, ) self.reset_callback_meta() self.temp_dir = tempfile.TemporaryDirectory() def _create_session_analysis_dataframe(self, langchain_asset: Any = None) -> dict: pd = import_pandas() llm_parameters = self._get_llm_parameters(langchain_asset) num_generations_per_prompt = llm_parameters.get("n", 1) llm_start_records_df = pd.DataFrame(self.on_llm_start_records) # Repeat each input row based on the number of outputs generated per prompt llm_start_records_df = llm_start_records_df.loc[ llm_start_records_df.index.repeat(num_generations_per_prompt) ].reset_index(drop=True) llm_end_records_df = pd.DataFrame(self.on_llm_end_records) llm_session_df = pd.merge( llm_start_records_df, llm_end_records_df, left_index=True, right_index=True, suffixes=["_llm_start", "_llm_end"], ) return llm_session_df def _get_llm_parameters(self, langchain_asset: Any = None) -> dict: if not langchain_asset: return {} try: if hasattr(langchain_asset, "agent"): llm_parameters = langchain_asset.agent.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm_chain"): llm_parameters = langchain_asset.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm"): llm_parameters = langchain_asset.llm.dict() else: llm_parameters = langchain_asset.dict() except Exception: return {} return llm_parameters
[ "langchain.callbacks.utils.import_spacy", "langchain.callbacks.utils.import_pandas", "langchain.callbacks.utils.import_textstat", "langchain.callbacks.utils.flatten_dict" ]
[((1047, 1114), 'comet_ml.Experiment', 'comet_ml.Experiment', ([], {'workspace': 'workspace', 'project_name': 'project_name'}), '(workspace=workspace, project_name=project_name)\n', (1066, 1114), False, 'import comet_ml\n'), ((1249, 1266), 'langchain.callbacks.utils.import_textstat', 'import_textstat', ([], {}), '()\n', (1264, 1266), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((2400, 2415), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (2413, 2415), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((4318, 4347), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (4345, 4347), False, 'import tempfile\n'), ((16471, 16509), 'pathlib.Path', 'Path', (['self.temp_dir.name', '"""model.json"""'], {}), "(self.temp_dir.name, 'model.json')\n", (16475, 16509), False, 'from pathlib import Path\n'), ((19081, 19095), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (19093, 19095), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((21551, 21580), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (21578, 21580), False, 'import tempfile\n'), ((21682, 21697), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (21695, 21697), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5161, 5175), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (5173, 5175), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5701, 5725), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (5713, 5725), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5842, 5860), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (5850, 5860), False, 'from copy import deepcopy\n'), ((6776, 6815), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['(response.llm_output or {})'], {}), '(response.llm_output or {})\n', (6788, 6815), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((8664, 8688), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (8676, 8688), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((10795, 10819), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (10807, 10819), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((7159, 7177), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (7167, 7177), False, 'from copy import deepcopy\n'), ((8886, 8900), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (8894, 8900), False, 'from copy import deepcopy\n'), ((9828, 9842), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (9836, 9842), False, 'from copy import deepcopy\n')]
import tempfile from copy import deepcopy from pathlib import Path from typing import Any, Callable, Dict, List, Optional, Sequence import langchain from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.utils import ( BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat, ) from langchain.schema import AgentAction, AgentFinish, Generation, LLMResult LANGCHAIN_MODEL_NAME = "langchain-model" def import_comet_ml() -> Any: """Import comet_ml and raise an error if it is not installed.""" try: import comet_ml # noqa: F401 except ImportError: raise ImportError( "To use the comet_ml callback manager you need to have the " "`comet_ml` python package installed. Please install it with" " `pip install comet_ml`" ) return comet_ml def _get_experiment( workspace: Optional[str] = None, project_name: Optional[str] = None ) -> Any: comet_ml = import_comet_ml() experiment = comet_ml.Experiment( # type: ignore workspace=workspace, project_name=project_name, ) return experiment def _fetch_text_complexity_metrics(text: str) -> dict: textstat = import_textstat() text_complexity_metrics = { "flesch_reading_ease": textstat.flesch_reading_ease(text), "flesch_kincaid_grade": textstat.flesch_kincaid_grade(text), "smog_index": textstat.smog_index(text), "coleman_liau_index": textstat.coleman_liau_index(text), "automated_readability_index": textstat.automated_readability_index(text), "dale_chall_readability_score": textstat.dale_chall_readability_score(text), "difficult_words": textstat.difficult_words(text), "linsear_write_formula": textstat.linsear_write_formula(text), "gunning_fog": textstat.gunning_fog(text), "text_standard": textstat.text_standard(text), "fernandez_huerta": textstat.fernandez_huerta(text), "szigriszt_pazos": textstat.szigriszt_pazos(text), "gutierrez_polini": textstat.gutierrez_polini(text), "crawford": textstat.crawford(text), "gulpease_index": textstat.gulpease_index(text), "osman": textstat.osman(text), } return text_complexity_metrics def _summarize_metrics_for_generated_outputs(metrics: Sequence) -> dict: pd = import_pandas() metrics_df = pd.DataFrame(metrics) metrics_summary = metrics_df.describe() return metrics_summary.to_dict() class CometCallbackHandler(BaseMetadataCallbackHandler, BaseCallbackHandler): """Callback Handler that logs to Comet. Parameters: job_type (str): The type of comet_ml task such as "inference", "testing" or "qc" project_name (str): The comet_ml project name tags (list): Tags to add to the task task_name (str): Name of the comet_ml task visualize (bool): Whether to visualize the run. complexity_metrics (bool): Whether to log complexity metrics stream_logs (bool): Whether to stream callback actions to Comet This handler will utilize the associated callback method and formats the input of each callback function with metadata regarding the state of LLM run, and adds the response to the list of records for both the {method}_records and action. It then logs the response to Comet. """ def __init__( self, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, stream_logs: bool = True, ) -> None: """Initialize callback handler.""" self.comet_ml = import_comet_ml() super().__init__() self.task_type = task_type self.workspace = workspace self.project_name = project_name self.tags = tags self.visualizations = visualizations self.complexity_metrics = complexity_metrics self.custom_metrics = custom_metrics self.stream_logs = stream_logs self.temp_dir = tempfile.TemporaryDirectory() self.experiment = _get_experiment(workspace, project_name) self.experiment.log_other("Created from", "langchain") if tags: self.experiment.add_tags(tags) self.name = name if self.name: self.experiment.set_name(self.name) warning = ( "The comet_ml callback is currently in beta and is subject to change " "based on updates to `langchain`. Please report any issues to " "https://github.com/comet-ml/issue-tracking/issues with the tag " "`langchain`." ) self.comet_ml.LOGGER.warning(warning) self.callback_columns: list = [] self.action_records: list = [] self.complexity_metrics = complexity_metrics if self.visualizations: spacy = import_spacy() self.nlp = spacy.load("en_core_web_sm") else: self.nlp = None def _init_resp(self) -> Dict: return {k: None for k in self.callback_columns} def on_llm_start( self, serialized: Dict[str, Any], prompts: List[str], **kwargs: Any ) -> None: """Run when LLM starts.""" self.step += 1 self.llm_starts += 1 self.starts += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_start"}) metadata.update(flatten_dict(serialized)) metadata.update(self.get_custom_callback_meta()) for prompt in prompts: prompt_resp = deepcopy(metadata) prompt_resp["prompts"] = prompt self.on_llm_start_records.append(prompt_resp) self.action_records.append(prompt_resp) if self.stream_logs: self._log_stream(prompt, metadata, self.step) def on_llm_new_token(self, token: str, **kwargs: Any) -> None: """Run when LLM generates a new token.""" self.step += 1 self.llm_streams += 1 resp = self._init_resp() resp.update({"action": "on_llm_new_token", "token": token}) resp.update(self.get_custom_callback_meta()) self.action_records.append(resp) def on_llm_end(self, response: LLMResult, **kwargs: Any) -> None: """Run when LLM ends running.""" self.step += 1 self.llm_ends += 1 self.ends += 1 metadata = self._init_resp() metadata.update({"action": "on_llm_end"}) metadata.update(flatten_dict(response.llm_output or {})) metadata.update(self.get_custom_callback_meta()) output_complexity_metrics = [] output_custom_metrics = [] for prompt_idx, generations in enumerate(response.generations): for gen_idx, generation in enumerate(generations): text = generation.text generation_resp = deepcopy(metadata) generation_resp.update(flatten_dict(generation.dict())) complexity_metrics = self._get_complexity_metrics(text) if complexity_metrics: output_complexity_metrics.append(complexity_metrics) generation_resp.update(complexity_metrics) custom_metrics = self._get_custom_metrics( generation, prompt_idx, gen_idx ) if custom_metrics: output_custom_metrics.append(custom_metrics) generation_resp.update(custom_metrics) if self.stream_logs: self._log_stream(text, metadata, self.step) self.action_records.append(generation_resp) self.on_llm_end_records.append(generation_resp) self._log_text_metrics(output_complexity_metrics, step=self.step) self._log_text_metrics(output_custom_metrics, step=self.step) def on_llm_error(self, error: BaseException, **kwargs: Any) -> None: """Run when LLM errors.""" self.step += 1 self.errors += 1 def on_chain_start( self, serialized: Dict[str, Any], inputs: Dict[str, Any], **kwargs: Any ) -> None: """Run when chain starts running.""" self.step += 1 self.chain_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_chain_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) for chain_input_key, chain_input_val in inputs.items(): if isinstance(chain_input_val, str): input_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_input_val, resp, self.step) input_resp.update({chain_input_key: chain_input_val}) self.action_records.append(input_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Input Value for {chain_input_key} will not be logged" ) def on_chain_end(self, outputs: Dict[str, Any], **kwargs: Any) -> None: """Run when chain ends running.""" self.step += 1 self.chain_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_chain_end"}) resp.update(self.get_custom_callback_meta()) for chain_output_key, chain_output_val in outputs.items(): if isinstance(chain_output_val, str): output_resp = deepcopy(resp) if self.stream_logs: self._log_stream(chain_output_val, resp, self.step) output_resp.update({chain_output_key: chain_output_val}) self.action_records.append(output_resp) else: self.comet_ml.LOGGER.warning( f"Unexpected data format provided! " f"Output Value for {chain_output_key} will not be logged" ) def on_chain_error(self, error: BaseException, **kwargs: Any) -> None: """Run when chain errors.""" self.step += 1 self.errors += 1 def on_tool_start( self, serialized: Dict[str, Any], input_str: str, **kwargs: Any ) -> None: """Run when tool starts running.""" self.step += 1 self.tool_starts += 1 self.starts += 1 resp = self._init_resp() resp.update({"action": "on_tool_start"}) resp.update(flatten_dict(serialized)) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(input_str, resp, self.step) resp.update({"input_str": input_str}) self.action_records.append(resp) def on_tool_end(self, output: str, **kwargs: Any) -> None: """Run when tool ends running.""" self.step += 1 self.tool_ends += 1 self.ends += 1 resp = self._init_resp() resp.update({"action": "on_tool_end"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_tool_error(self, error: BaseException, **kwargs: Any) -> None: """Run when tool errors.""" self.step += 1 self.errors += 1 def on_text(self, text: str, **kwargs: Any) -> None: """ Run when agent is ending. """ self.step += 1 self.text_ctr += 1 resp = self._init_resp() resp.update({"action": "on_text"}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(text, resp, self.step) resp.update({"text": text}) self.action_records.append(resp) def on_agent_finish(self, finish: AgentFinish, **kwargs: Any) -> None: """Run when agent ends running.""" self.step += 1 self.agent_ends += 1 self.ends += 1 resp = self._init_resp() output = finish.return_values["output"] log = finish.log resp.update({"action": "on_agent_finish", "log": log}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(output, resp, self.step) resp.update({"output": output}) self.action_records.append(resp) def on_agent_action(self, action: AgentAction, **kwargs: Any) -> Any: """Run on agent action.""" self.step += 1 self.tool_starts += 1 self.starts += 1 tool = action.tool tool_input = str(action.tool_input) log = action.log resp = self._init_resp() resp.update({"action": "on_agent_action", "log": log, "tool": tool}) resp.update(self.get_custom_callback_meta()) if self.stream_logs: self._log_stream(tool_input, resp, self.step) resp.update({"tool_input": tool_input}) self.action_records.append(resp) def _get_complexity_metrics(self, text: str) -> dict: """Compute text complexity metrics using textstat. Parameters: text (str): The text to analyze. Returns: (dict): A dictionary containing the complexity metrics. """ resp = {} if self.complexity_metrics: text_complexity_metrics = _fetch_text_complexity_metrics(text) resp.update(text_complexity_metrics) return resp def _get_custom_metrics( self, generation: Generation, prompt_idx: int, gen_idx: int ) -> dict: """Compute Custom Metrics for an LLM Generated Output Args: generation (LLMResult): Output generation from an LLM prompt_idx (int): List index of the input prompt gen_idx (int): List index of the generated output Returns: dict: A dictionary containing the custom metrics. """ resp = {} if self.custom_metrics: custom_metrics = self.custom_metrics(generation, prompt_idx, gen_idx) resp.update(custom_metrics) return resp def flush_tracker( self, langchain_asset: Any = None, task_type: Optional[str] = "inference", workspace: Optional[str] = None, project_name: Optional[str] = "comet-langchain-demo", tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, finish: bool = False, reset: bool = False, ) -> None: """Flush the tracker and setup the session. Everything after this will be a new table. Args: name: Name of the performed session so far so it is identifiable langchain_asset: The langchain asset to save. finish: Whether to finish the run. Returns: None """ self._log_session(langchain_asset) if langchain_asset: try: self._log_model(langchain_asset) except Exception: self.comet_ml.LOGGER.error( "Failed to export agent or LLM to Comet", exc_info=True, extra={"show_traceback": True}, ) if finish: self.experiment.end() if reset: self._reset( task_type, workspace, project_name, tags, name, visualizations, complexity_metrics, custom_metrics, ) def _log_stream(self, prompt: str, metadata: dict, step: int) -> None: self.experiment.log_text(prompt, metadata=metadata, step=step) def _log_model(self, langchain_asset: Any) -> None: model_parameters = self._get_llm_parameters(langchain_asset) self.experiment.log_parameters(model_parameters, prefix="model") langchain_asset_path = Path(self.temp_dir.name, "model.json") model_name = self.name if self.name else LANGCHAIN_MODEL_NAME try: if hasattr(langchain_asset, "save"): langchain_asset.save(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) except (ValueError, AttributeError, NotImplementedError) as e: if hasattr(langchain_asset, "save_agent"): langchain_asset.save_agent(langchain_asset_path) self.experiment.log_model(model_name, str(langchain_asset_path)) else: self.comet_ml.LOGGER.error( f"{e}" " Could not save Langchain Asset " f"for {langchain_asset.__class__.__name__}" ) def _log_session(self, langchain_asset: Optional[Any] = None) -> None: try: llm_session_df = self._create_session_analysis_dataframe(langchain_asset) # Log the cleaned dataframe as a table self.experiment.log_table("langchain-llm-session.csv", llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: metadata = {"langchain_version": str(langchain.__version__)} # Log the langchain low-level records as a JSON file directly self.experiment.log_asset_data( self.action_records, "langchain-action_records.json", metadata=metadata ) except Exception: self.comet_ml.LOGGER.warning( "Failed to log session data to Comet", exc_info=True, extra={"show_traceback": True}, ) try: self._log_visualizations(llm_session_df) except Exception: self.comet_ml.LOGGER.warning( "Failed to log visualizations to Comet", exc_info=True, extra={"show_traceback": True}, ) def _log_text_metrics(self, metrics: Sequence[dict], step: int) -> None: if not metrics: return metrics_summary = _summarize_metrics_for_generated_outputs(metrics) for key, value in metrics_summary.items(): self.experiment.log_metrics(value, prefix=key, step=step) def _log_visualizations(self, session_df: Any) -> None: if not (self.visualizations and self.nlp): return spacy = import_spacy() prompts = session_df["prompts"].tolist() outputs = session_df["text"].tolist() for idx, (prompt, output) in enumerate(zip(prompts, outputs)): doc = self.nlp(output) sentence_spans = list(doc.sents) for visualization in self.visualizations: try: html = spacy.displacy.render( sentence_spans, style=visualization, options={"compact": True}, jupyter=False, page=True, ) self.experiment.log_asset_data( html, name=f"langchain-viz-{visualization}-{idx}.html", metadata={"prompt": prompt}, step=idx, ) except Exception as e: self.comet_ml.LOGGER.warning( e, exc_info=True, extra={"show_traceback": True} ) return def _reset( self, task_type: Optional[str] = None, workspace: Optional[str] = None, project_name: Optional[str] = None, tags: Optional[Sequence] = None, name: Optional[str] = None, visualizations: Optional[List[str]] = None, complexity_metrics: bool = False, custom_metrics: Optional[Callable] = None, ) -> None: _task_type = task_type if task_type else self.task_type _workspace = workspace if workspace else self.workspace _project_name = project_name if project_name else self.project_name _tags = tags if tags else self.tags _name = name if name else self.name _visualizations = visualizations if visualizations else self.visualizations _complexity_metrics = ( complexity_metrics if complexity_metrics else self.complexity_metrics ) _custom_metrics = custom_metrics if custom_metrics else self.custom_metrics self.__init__( # type: ignore task_type=_task_type, workspace=_workspace, project_name=_project_name, tags=_tags, name=_name, visualizations=_visualizations, complexity_metrics=_complexity_metrics, custom_metrics=_custom_metrics, ) self.reset_callback_meta() self.temp_dir = tempfile.TemporaryDirectory() def _create_session_analysis_dataframe(self, langchain_asset: Any = None) -> dict: pd = import_pandas() llm_parameters = self._get_llm_parameters(langchain_asset) num_generations_per_prompt = llm_parameters.get("n", 1) llm_start_records_df = pd.DataFrame(self.on_llm_start_records) # Repeat each input row based on the number of outputs generated per prompt llm_start_records_df = llm_start_records_df.loc[ llm_start_records_df.index.repeat(num_generations_per_prompt) ].reset_index(drop=True) llm_end_records_df = pd.DataFrame(self.on_llm_end_records) llm_session_df = pd.merge( llm_start_records_df, llm_end_records_df, left_index=True, right_index=True, suffixes=["_llm_start", "_llm_end"], ) return llm_session_df def _get_llm_parameters(self, langchain_asset: Any = None) -> dict: if not langchain_asset: return {} try: if hasattr(langchain_asset, "agent"): llm_parameters = langchain_asset.agent.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm_chain"): llm_parameters = langchain_asset.llm_chain.llm.dict() elif hasattr(langchain_asset, "llm"): llm_parameters = langchain_asset.llm.dict() else: llm_parameters = langchain_asset.dict() except Exception: return {} return llm_parameters
[ "langchain.callbacks.utils.import_spacy", "langchain.callbacks.utils.import_pandas", "langchain.callbacks.utils.import_textstat", "langchain.callbacks.utils.flatten_dict" ]
[((1047, 1114), 'comet_ml.Experiment', 'comet_ml.Experiment', ([], {'workspace': 'workspace', 'project_name': 'project_name'}), '(workspace=workspace, project_name=project_name)\n', (1066, 1114), False, 'import comet_ml\n'), ((1249, 1266), 'langchain.callbacks.utils.import_textstat', 'import_textstat', ([], {}), '()\n', (1264, 1266), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((2400, 2415), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (2413, 2415), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((4318, 4347), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (4345, 4347), False, 'import tempfile\n'), ((16471, 16509), 'pathlib.Path', 'Path', (['self.temp_dir.name', '"""model.json"""'], {}), "(self.temp_dir.name, 'model.json')\n", (16475, 16509), False, 'from pathlib import Path\n'), ((19081, 19095), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (19093, 19095), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((21551, 21580), 'tempfile.TemporaryDirectory', 'tempfile.TemporaryDirectory', ([], {}), '()\n', (21578, 21580), False, 'import tempfile\n'), ((21682, 21697), 'langchain.callbacks.utils.import_pandas', 'import_pandas', ([], {}), '()\n', (21695, 21697), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5161, 5175), 'langchain.callbacks.utils.import_spacy', 'import_spacy', ([], {}), '()\n', (5173, 5175), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5701, 5725), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (5713, 5725), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((5842, 5860), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (5850, 5860), False, 'from copy import deepcopy\n'), ((6776, 6815), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['(response.llm_output or {})'], {}), '(response.llm_output or {})\n', (6788, 6815), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((8664, 8688), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (8676, 8688), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((10795, 10819), 'langchain.callbacks.utils.flatten_dict', 'flatten_dict', (['serialized'], {}), '(serialized)\n', (10807, 10819), False, 'from langchain.callbacks.utils import BaseMetadataCallbackHandler, flatten_dict, import_pandas, import_spacy, import_textstat\n'), ((7159, 7177), 'copy.deepcopy', 'deepcopy', (['metadata'], {}), '(metadata)\n', (7167, 7177), False, 'from copy import deepcopy\n'), ((8886, 8900), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (8894, 8900), False, 'from copy import deepcopy\n'), ((9828, 9842), 'copy.deepcopy', 'deepcopy', (['resp'], {}), '(resp)\n', (9836, 9842), False, 'from copy import deepcopy\n')]
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import asyncio import functools import itertools import logging import uuid from enum import Enum from typing import ( Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, ) from urllib.parse import urlparse, urlunparse from langsmith import Client, RunEvaluator from langsmith.schemas import Dataset, DataType, Example from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.manager import Callbacks from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.chains.base import Chain from langchain.chat_models.openai import ChatOpenAI from langchain.evaluation.loading import load_evaluator from langchain.evaluation.schema import EvaluatorType, StringEvaluator from langchain.schema import ChatResult, LLMResult from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import BaseMessage, messages_from_dict from langchain.smith.evaluation.config import EvalConfig, RunEvalConfig from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain logger = logging.getLogger(__name__) MODEL_OR_CHAIN_FACTORY = Union[Callable[[], Chain], BaseLanguageModel] class InputFormatError(Exception): """Raised when the input format is invalid.""" ## Shared Utilities def _get_eval_project_url(api_url: str, project_id: str) -> str: """Get the project url from the api url.""" parsed = urlparse(api_url) hostname = parsed.hostname or "" if "api." in hostname: hostname = hostname.replace("api.", "", 1) if "localhost" in hostname: # Remove the port hostname = "localhost" url = urlunparse(parsed._replace(netloc=hostname)) return f"{url}/projects/p/{project_id}?eval=true" def _wrap_in_chain_factory( llm_or_chain_factory: Union[Chain, MODEL_OR_CHAIN_FACTORY], dataset_name: str = "<my_dataset>", ) -> MODEL_OR_CHAIN_FACTORY: """Forgive the user if they pass in a chain without memory instead of a chain factory. It's a common mistake. Raise a more helpful error message as well.""" if isinstance(llm_or_chain_factory, Chain): chain = llm_or_chain_factory chain_class = chain.__class__.__name__ if llm_or_chain_factory.memory is not None: memory_class = chain.memory.__class__.__name__ raise ValueError( "Cannot directly evaluate a chain with stateful memory." " To evaluate this chain, pass in a chain constructor" " that initializes fresh memory each time it is called." " This will safegaurd against information" " leakage between dataset examples." "\nFor example:\n\n" "def chain_constructor():\n" f" new_memory = {memory_class}(...)\n" f" return {chain_class}" "(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) logger.warning( "Directly passing in a chain is not recommended as chains may have state." " This can lead to unexpected behavior as the " "same chain instance could be used across multiple datasets. Instead," " please pass a chain constructor that creates a new " "chain with fresh memory each time it is called. This will safeguard" " against information leakage between dataset examples. " "\nFor example:\n\n" "def chain_constructor():\n" f" return {chain_class}(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) return lambda: chain elif isinstance(llm_or_chain_factory, BaseLanguageModel): return llm_or_chain_factory elif callable(llm_or_chain_factory): _model = llm_or_chain_factory() if isinstance(_model, BaseLanguageModel): return _model return llm_or_chain_factory return llm_or_chain_factory def _first_example(examples: Iterator[Example]) -> Tuple[Example, Iterator[Example]]: """Get the first example while chaining it back and preserving the iterator.""" try: example: Example = next(examples) except StopIteration: raise ValueError("No examples provided.") return example, itertools.chain([example], examples) def _get_prompt(inputs: Dict[str, Any]) -> str: """Get prompt from inputs. Args: inputs: The input dictionary. Returns: A string prompt. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") prompts = [] if "prompt" in inputs: if not isinstance(inputs["prompt"], str): raise InputFormatError( "Expected string for 'prompt', got" f" {type(inputs['prompt']).__name__}" ) prompts = [inputs["prompt"]] elif "prompts" in inputs: if not isinstance(inputs["prompts"], list) or not all( isinstance(i, str) for i in inputs["prompts"] ): raise InputFormatError( "Expected list of strings for 'prompts'," f" got {type(inputs['prompts']).__name__}" ) prompts = inputs["prompts"] elif len(inputs) == 1: prompt_ = next(iter(inputs.values())) if isinstance(prompt_, str): prompts = [prompt_] elif isinstance(prompt_, list) and all(isinstance(i, str) for i in prompt_): prompts = prompt_ else: raise InputFormatError(f"LLM Run expects string prompt input. Got {inputs}") else: raise InputFormatError( f"LLM Run expects 'prompt' or 'prompts' in inputs. Got {inputs}" ) if len(prompts) == 1: return prompts[0] else: raise InputFormatError( f"LLM Run expects single prompt input. Got {len(prompts)} prompts." ) def _get_messages(inputs: Dict[str, Any]) -> List[BaseMessage]: """Get Chat Messages from inputs. Args: inputs: The input dictionary. Returns: A list of chat messages. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") if "messages" in inputs: single_input = inputs["messages"] elif len(inputs) == 1: single_input = next(iter(inputs.values())) else: raise InputFormatError( f"Chat Run expects 'messages' in inputs when example has multiple" f" input keys. Got {inputs}" ) if isinstance(single_input, list) and all( isinstance(i, dict) for i in single_input ): raw_messages = [single_input] elif isinstance(single_input, list) and all( isinstance(i, list) for i in single_input ): raw_messages = single_input else: raise InputFormatError( f"Chat Run expects List[dict] or List[List[dict]] values for" f" 'messages' key input. Got {inputs}" ) if len(raw_messages) == 1: return messages_from_dict(raw_messages[0]) else: raise InputFormatError( f"Chat Run expects single List[dict] or List[List[dict]] 'messages'" f" input. Got {len(raw_messages)} messages from inputs {inputs}" ) def _get_project_name( project_name: Optional[str], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, ) -> str: """ Get the project name. Args: project_name: The project name if manually specified. llm_or_chain_factory: The Chain or language model constructor. Returns: The project name. """ if project_name is not None: return project_name if isinstance(llm_or_chain_factory, BaseLanguageModel): model_name = llm_or_chain_factory.__class__.__name__ else: model_name = llm_or_chain_factory().__class__.__name__ hex = uuid.uuid4().hex return f"{hex}-{model_name}" ## Shared Validation Utilities def _validate_example_inputs_for_language_model( first_example: Example, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: if input_mapper: prompt_input = input_mapper(first_example.inputs) if not isinstance(prompt_input, str) and not ( isinstance(prompt_input, list) and all(isinstance(msg, BaseMessage) for msg in prompt_input) ): raise InputFormatError( "When using an input_mapper to prepare dataset example inputs" " for an LLM or chat model, the output must a single string or" " a list of chat messages." f"\nGot: {prompt_input} of type {type(prompt_input)}." ) else: try: _get_prompt(first_example.inputs) except InputFormatError: try: _get_messages(first_example.inputs) except InputFormatError: raise InputFormatError( "Example inputs do not match language model input format. " "Expected a dictionary with messages or a single prompt." f" Got: {first_example.inputs}" " Please update your dataset OR provide an input_mapper" " to convert the example.inputs to a compatible format" " for the llm or chat model you wish to evaluate." ) def _validate_example_inputs_for_chain( first_example: Example, chain: Chain, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: """Validate that the example inputs match the chain input keys.""" if input_mapper: first_inputs = input_mapper(first_example.inputs) if not isinstance(first_inputs, dict): raise InputFormatError( "When using an input_mapper to prepare dataset example" " inputs for a chain, the mapped value must be a dictionary." f"\nGot: {first_inputs} of type {type(first_inputs)}." ) if not set(first_inputs.keys()) == set(chain.input_keys): raise InputFormatError( "When using an input_mapper to prepare dataset example inputs" " for a chain mapped value must have keys that match the chain's" " expected input keys." f"\nExpected: {chain.input_keys}. Got: {first_inputs.keys()}" ) else: first_inputs = first_example.inputs if len(first_inputs) == 1 and len(chain.input_keys) == 1: # We can pass this through the run method. # Refrain from calling to validate. pass elif not set(first_inputs.keys()) == set(chain.input_keys): raise InputFormatError( "Example inputs do not match chain input keys." " Please provide an input_mapper to convert the example.inputs" " to a compatible format for the chain you wish to evaluate." f"Expected: {chain.input_keys}. " f"Got: {first_inputs.keys()}" ) def _validate_example_inputs( examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, input_mapper: Optional[Callable[[Dict], Any]], ) -> Iterator[Example]: """Validate that the example inputs are valid for the model.""" first_example, examples = _first_example(examples) if isinstance(llm_or_chain_factory, BaseLanguageModel): _validate_example_inputs_for_language_model(first_example, input_mapper) else: chain = llm_or_chain_factory() _validate_example_inputs_for_chain(first_example, chain, input_mapper) return examples ## Shared Evaluator Setup Utilities def _setup_evaluation( llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, examples: Iterator[Example], evaluation: Optional[RunEvalConfig], data_type: DataType, ) -> Tuple[Optional[List[RunEvaluator]], Iterator[Example]]: """Configure the evaluators to run on the results of the chain.""" if evaluation: first_example, examples = _first_example(examples) if isinstance(llm_or_chain_factory, BaseLanguageModel): run_inputs, run_outputs = None, None run_type = "llm" else: run_type = "chain" if data_type in (DataType.chat, DataType.llm): val = data_type.value if isinstance(data_type, Enum) else data_type raise ValueError( "Cannot evaluate a chain on dataset with " f"data_type={val}. " "Please specify a dataset with the default 'kv' data type." ) chain = llm_or_chain_factory() run_inputs = chain.input_keys run_outputs = chain.output_keys run_evaluators = _load_run_evaluators( evaluation, run_type, data_type, list(first_example.outputs) if first_example.outputs else None, run_inputs, run_outputs, ) else: # TODO: Create a default helpfulness evaluator run_evaluators = None return run_evaluators, examples def _determine_input_key( config: RunEvalConfig, run_inputs: Optional[List[str]], run_type: str, ) -> Optional[str]: if config.input_key: input_key = config.input_key if run_inputs and input_key not in run_inputs: raise ValueError(f"Input key {input_key} not in run inputs {run_inputs}") elif run_type == "llm": input_key = None elif run_inputs and len(run_inputs) == 1: input_key = run_inputs[0] else: raise ValueError( f"Must specify input key for model with multiple inputs: {run_inputs}" ) return input_key def _determine_prediction_key( config: RunEvalConfig, run_outputs: Optional[List[str]], run_type: str, ) -> Optional[str]: if config.prediction_key: prediction_key = config.prediction_key if run_outputs and prediction_key not in run_outputs: raise ValueError( f"Prediction key {prediction_key} not in run outputs {run_outputs}" ) elif run_type == "llm": prediction_key = None elif run_outputs and len(run_outputs) == 1: prediction_key = run_outputs[0] else: raise ValueError( f"Must specify prediction key for model" f" with multiple outputs: {run_outputs}" ) return prediction_key def _determine_reference_key( config: RunEvalConfig, example_outputs: Optional[List[str]], ) -> Optional[str]: if config.reference_key: reference_key = config.reference_key if example_outputs and reference_key not in example_outputs: raise ValueError( f"Reference key {reference_key} not in Dataset" f" example outputs: {example_outputs}" ) elif example_outputs and len(example_outputs) == 1: reference_key = list(example_outputs)[0] else: reference_key = None return reference_key def _construct_run_evaluator( eval_config: Union[EvaluatorType, EvalConfig], eval_llm: BaseLanguageModel, run_type: str, data_type: DataType, example_outputs: Optional[List[str]], reference_key: Optional[str], input_key: Optional[str], prediction_key: Optional[str], ) -> RunEvaluator: if isinstance(eval_config, EvaluatorType): evaluator_ = load_evaluator(eval_config, llm=eval_llm) eval_type_tag = eval_config.value else: kwargs = {"llm": eval_llm, **eval_config.get_kwargs()} evaluator_ = load_evaluator(eval_config.evaluator_type, **kwargs) eval_type_tag = eval_config.evaluator_type.value if isinstance(evaluator_, StringEvaluator): if evaluator_.requires_reference and reference_key is None: raise ValueError( f"Must specify reference_key in RunEvalConfig to use" f" evaluator of type {eval_type_tag} with" f" dataset with multiple output keys: {example_outputs}." ) run_evaluator = StringRunEvaluatorChain.from_run_and_data_type( evaluator_, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, tags=[eval_type_tag], ) else: raise NotImplementedError( f"Run evaluator for {eval_type_tag} is not implemented" ) return run_evaluator def _load_run_evaluators( config: RunEvalConfig, run_type: str, data_type: DataType, example_outputs: Optional[List[str]], run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], ) -> List[RunEvaluator]: """ Load run evaluators from a configuration. Args: config: Configuration for the run evaluators. Returns: A list of run evaluators. """ eval_llm = config.eval_llm or ChatOpenAI(model="gpt-4", temperature=0.0) run_evaluators = [] input_key = _determine_input_key(config, run_inputs, run_type) prediction_key = _determine_prediction_key(config, run_outputs, run_type) reference_key = _determine_reference_key(config, example_outputs) for eval_config in config.evaluators: run_evaluator = _construct_run_evaluator( eval_config, eval_llm, run_type, data_type, example_outputs, reference_key, input_key, prediction_key, ) run_evaluators.append(run_evaluator) custom_evaluators = config.custom_evaluators or [] for custom_evaluator in custom_evaluators: if isinstance(custom_evaluator, RunEvaluator): run_evaluators.append(custom_evaluator) elif isinstance(custom_evaluator, StringEvaluator): run_evaluators.append( StringRunEvaluatorChain.from_run_and_data_type( custom_evaluator, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, ) ) else: raise ValueError( f"Unsupported custom evaluator: {custom_evaluator}." f" Expected RunEvaluator or StringEvaluator." ) return run_evaluators ### Async Helpers async def _arun_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], *, tags: Optional[List[str]] = None, callbacks: Callbacks = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """Asynchronously run the language model. Args: llm: The language model to run. inputs: The input dictionary. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map inputs to the expected format. Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): return await llm.apredict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): return await llm.apredict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format" f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: prompt = _get_prompt(inputs) llm_output: Union[str, BaseMessage] = await llm.apredict( prompt, callbacks=callbacks, tags=tags ) except InputFormatError: messages = _get_messages(inputs) llm_output = await llm.apredict_messages( messages, callbacks=callbacks, tags=tags ) return llm_output async def _arun_chain( chain: Chain, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str]: """Run a chain asynchronously on inputs.""" if input_mapper is not None: inputs_ = input_mapper(inputs) output: Union[dict, str] = await chain.acall( inputs_, callbacks=callbacks, tags=tags ) else: inputs_ = next(iter(inputs.values())) if len(inputs) == 1 else inputs output = await chain.acall(inputs_, callbacks=callbacks, tags=tags) return output async def _arun_llm_or_chain( example: Example, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, n_repetitions: int, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: """Asynchronously run the Chain or language model. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. n_repetitions: The number of times to run the model on each example. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map the input to the expected format. Returns: A list of outputs. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None outputs = [] chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) for _ in range(n_repetitions): try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = await _arun_llm( llm_or_chain_factory, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = await _arun_chain( chain, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) outputs.append(output) except Exception as e: logger.warning( f"{chain_or_llm} failed for example {example.id}. Error: {e}" ) outputs.append({"Error": str(e)}) if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return outputs async def _gather_with_concurrency( n: int, initializer: Callable[[], Coroutine[Any, Any, Any]], *async_funcs: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ], ) -> List[Any]: """Run coroutines with a concurrency limit. Args: n: The maximum number of concurrent tasks. initializer: A coroutine that initializes shared resources for the tasks. async_funcs: The async_funcs to be run concurrently. Returns: A list of results from the coroutines. """ semaphore = asyncio.Semaphore(n) job_state = {"num_processed": 0} callback_queue: asyncio.Queue[Sequence[BaseCallbackHandler]] = asyncio.Queue() for _ in range(n): callback_queue.put_nowait(await initializer()) async def run_coroutine_with_semaphore( async_func: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ] ) -> Any: async with semaphore: callbacks = await callback_queue.get() try: result = await async_func(callbacks, job_state) finally: callback_queue.put_nowait(callbacks) return result results = await asyncio.gather( *(run_coroutine_with_semaphore(function) for function in async_funcs) ) while callback_queue: try: callbacks = callback_queue.get_nowait() except asyncio.QueueEmpty: break for callback in callbacks: if isinstance(callback, (LangChainTracer, EvaluatorCallbackHandler)): callback.wait_for_futures() return results async def _callbacks_initializer( project_name: Optional[str], client: Client, run_evaluators: Sequence[RunEvaluator], evaluation_handler_collector: List[EvaluatorCallbackHandler], ) -> List[BaseTracer]: """ Initialize a tracer to share across tasks. Args: project_name: The project name for the tracer. client: The client to use for the tracer. run_evaluators: The evaluators to run. evaluation_handler_collector: A list to collect the evaluators. Used to wait for the evaluators to finish. Returns: The callbacks for this thread. """ callbacks: List[BaseTracer] = [] if project_name: callbacks.append( LangChainTracer( project_name=project_name, client=client, use_threading=False ) ) evaluator_project_name = f"{project_name}-evaluators" if project_name else None if run_evaluators: callback = EvaluatorCallbackHandler( client=client, evaluators=run_evaluators, # We already have concurrency, don't want to overload the machine max_workers=1, project_name=evaluator_project_name, ) callbacks.append(callback) evaluation_handler_collector.append(callback) return callbacks async def _arun_on_examples( client: Client, examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, data_type: DataType = DataType.kv, ) -> Dict[str, Any]: """ Asynchronously run the chain on examples and store traces to the specified project name. Args: client: LangSmith client to use to log feedback and runs. examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating concurrency_level: The number of async tasks to run concurrently. num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Project name to use when tracing runs. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. data_type: The dataset's data type. This is used to determine determine how to deserialize the reference data and model compatibility. Returns: A dictionary mapping example ids to the model outputs. """ llm_or_chain_factory = _wrap_in_chain_factory(llm_or_chain_factory) project_name = _get_project_name(project_name, llm_or_chain_factory) run_evaluators, examples = _setup_evaluation( llm_or_chain_factory, examples, evaluation, data_type ) examples = _validate_example_inputs(examples, llm_or_chain_factory, input_mapper) results: Dict[str, List[Any]] = {} async def process_example( example: Example, callbacks: List[BaseCallbackHandler], job_state: dict ) -> None: """Process a single example.""" result = await _arun_llm_or_chain( example, llm_or_chain_factory, num_repetitions, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) results[str(example.id)] = result job_state["num_processed"] += 1 if verbose: print( f"Processed examples: {job_state['num_processed']}", end="\r", flush=True, ) evaluation_handlers: List[EvaluatorCallbackHandler] = [] await _gather_with_concurrency( concurrency_level, functools.partial( _callbacks_initializer, project_name=project_name, client=client, evaluation_handler_collector=evaluation_handlers, run_evaluators=run_evaluators or [], ), *(functools.partial(process_example, e) for e in examples), ) for handler in evaluation_handlers: handler.wait_for_futures() return results ## Sync Utilities def _run_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """ Run the language model on the example. Args: llm: The language model to run. inputs: The input dictionary. callbacks: The callbacks to use during the run. tags: Optional tags to add to the run. input_mapper: function to map to the inputs dictionary from an Example Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): llm_output: Union[str, BaseMessage] = llm.predict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): llm_output = llm.predict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format: " f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: llm_prompts = _get_prompt(inputs) llm_output = llm.predict(llm_prompts, callbacks=callbacks, tags=tags) except InputFormatError: llm_messages = _get_messages(inputs) llm_output = llm.predict_messages(llm_messages, callbacks=callbacks) return llm_output def _run_chain( chain: Chain, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[Dict, str]: """Run a chain on inputs.""" if input_mapper is not None: inputs_ = input_mapper(inputs) output: Union[dict, str] = chain(inputs_, callbacks=callbacks, tags=tags) else: inputs_ = next(iter(inputs.values())) if len(inputs) == 1 else inputs output = chain(inputs_, callbacks=callbacks, tags=tags) return output def _run_llm_or_chain( example: Example, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, n_repetitions: int, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: """ Run the Chain or language model synchronously. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. n_repetitions: The number of times to run the model on each example. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. Returns: Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: The outputs of the model or chain. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None outputs = [] chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) for _ in range(n_repetitions): try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = _run_llm( llm_or_chain_factory, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = _run_chain( chain, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) outputs.append(output) except Exception as e: logger.warning( f"{chain_or_llm} failed for example {example.id}. Error: {e}" ) outputs.append({"Error": str(e)}) if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return outputs def _run_on_examples( client: Client, examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, data_type: DataType = DataType.kv, ) -> Dict[str, Any]: """ Run the Chain or language model on examples and store traces to the specified project name. Args: client: LangSmith client to use to log feedback and runs. examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. data_type: The dataset's data type. This is used to determine determine how to deserialize the reference data and model compatibility. Returns: A dictionary mapping example ids to the model outputs. """ results: Dict[str, Any] = {} llm_or_chain_factory = _wrap_in_chain_factory(llm_or_chain_factory) project_name = _get_project_name(project_name, llm_or_chain_factory) tracer = LangChainTracer( project_name=project_name, client=client, use_threading=False ) evaluator_project_name = f"{project_name}-evaluators" run_evaluators, examples = _setup_evaluation( llm_or_chain_factory, examples, evaluation, data_type ) examples = _validate_example_inputs(examples, llm_or_chain_factory, input_mapper) evalution_handler = EvaluatorCallbackHandler( evaluators=run_evaluators or [], client=client, project_name=evaluator_project_name, ) callbacks: List[BaseCallbackHandler] = [tracer, evalution_handler] for i, example in enumerate(examples): result = _run_llm_or_chain( example, llm_or_chain_factory, num_repetitions, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) if verbose: print(f"{i+1} processed", flush=True, end="\r") results[str(example.id)] = result tracer.wait_for_futures() evalution_handler.wait_for_futures() return results ## Public API def _prepare_eval_run( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, project_name: Optional[str], ) -> Tuple[MODEL_OR_CHAIN_FACTORY, str, Dataset, Iterator[Example]]: llm_or_chain_factory = _wrap_in_chain_factory(llm_or_chain_factory, dataset_name) project_name = _get_project_name(project_name, llm_or_chain_factory) try: project = client.create_project(project_name) except ValueError as e: if "already exists " not in str(e): raise e raise ValueError( f"Project {project_name} already exists. Please use a different name." ) project_url = _get_eval_project_url(client.api_url, project.id) print( f"View the evaluation results for project '{project_name}' at:\n{project_url}" ) dataset = client.read_dataset(dataset_name=dataset_name) examples = client.list_examples(dataset_id=str(dataset.id)) return llm_or_chain_factory, project_name, dataset, examples async def arun_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Dict[str, Any]: """ Asynchronously run the Chain or language model on a dataset and store traces to the specified project name. Args: client: LangSmith client to use to read the dataset, and to log feedback and run traces. dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating concurrency_level: The number of async tasks to run concurrently. num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. For the synchronous version, see :func:`run_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import RunEvalConfig, arun_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", RunEvalConfig.Criteria("helpfulness"), RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() await arun_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) await arun_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 llm_or_chain_factory, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name ) results = await _arun_on_examples( client, examples, llm_or_chain_factory, concurrency_level=concurrency_level, num_repetitions=num_repetitions, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) return { "project_name": project_name, "results": results, } def run_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, num_repetitions: int = 1, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Dict[str, Any]: """ Run the Chain or language model on a dataset and store traces to the specified project name. Args: client: LangSmith client to use to access the dataset and to log feedback and run traces. dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Configuration for evaluators to run on the results of the chain num_repetitions: Number of times to run the model on each example. This is useful when testing success rates or generating confidence intervals. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. For the (usually faster) async version of this function, see :func:`arun_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import RunEvalConfig, run_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", RunEvalConfig.Criteria("helpfulness"), RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 llm_or_chain_factory, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name ) results = _run_on_examples( client, examples, llm_or_chain_factory, num_repetitions=num_repetitions, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) return { "project_name": project_name, "results": results, }
[ "langchain.schema.messages.messages_from_dict", "langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.chat_models.openai.ChatOpenAI", "langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler", "langchain.evaluation.loading.load_evaluator" ]
[((1370, 1397), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1387, 1397), False, 'import logging\n'), ((1708, 1725), 'urllib.parse.urlparse', 'urlparse', (['api_url'], {}), '(api_url)\n', (1716, 1725), False, 'from urllib.parse import urlparse, urlunparse\n'), ((24648, 24668), 'asyncio.Semaphore', 'asyncio.Semaphore', (['n'], {}), '(n)\n', (24665, 24668), False, 'import asyncio\n'), ((24774, 24789), 'asyncio.Queue', 'asyncio.Queue', ([], {}), '()\n', (24787, 24789), False, 'import asyncio\n'), ((37766, 37844), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)'}), '(project_name=project_name, client=client, use_threading=False)\n', (37781, 37844), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((38145, 38258), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'evaluators': '(run_evaluators or [])', 'client': 'client', 'project_name': 'evaluator_project_name'}), '(evaluators=run_evaluators or [], client=client,\n project_name=evaluator_project_name)\n', (38169, 38258), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler\n'), ((4660, 4696), 'itertools.chain', 'itertools.chain', (['[example]', 'examples'], {}), '([example], examples)\n', (4675, 4696), False, 'import itertools\n'), ((7539, 7574), 'langchain.schema.messages.messages_from_dict', 'messages_from_dict', (['raw_messages[0]'], {}), '(raw_messages[0])\n', (7557, 7574), False, 'from langchain.schema.messages import BaseMessage, messages_from_dict\n'), ((8394, 8406), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (8404, 8406), False, 'import uuid\n'), ((16022, 16063), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config'], {'llm': 'eval_llm'}), '(eval_config, llm=eval_llm)\n', (16036, 16063), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((16200, 16252), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config.evaluator_type'], {}), '(eval_config.evaluator_type, **kwargs)\n', (16214, 16252), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((16698, 16888), 'langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type', 'StringRunEvaluatorChain.from_run_and_data_type', (['evaluator_', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key', 'tags': '[eval_type_tag]'}), '(evaluator_, run_type,\n data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key, tags=[eval_type_tag])\n', (16744, 16888), False, 'from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain\n'), ((17574, 17616), 'langchain.chat_models.openai.ChatOpenAI', 'ChatOpenAI', ([], {'model': '"""gpt-4"""', 'temperature': '(0.0)'}), "(model='gpt-4', temperature=0.0)\n", (17584, 17616), False, 'from langchain.chat_models.openai import ChatOpenAI\n'), ((26719, 26841), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'client': 'client', 'evaluators': 'run_evaluators', 'max_workers': '(1)', 'project_name': 'evaluator_project_name'}), '(client=client, evaluators=run_evaluators,\n max_workers=1, project_name=evaluator_project_name)\n', (26743, 26841), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler\n'), ((26474, 26552), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)'}), '(project_name=project_name, client=client, use_threading=False)\n', (26489, 26552), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((30344, 30523), 'functools.partial', 'functools.partial', (['_callbacks_initializer'], {'project_name': 'project_name', 'client': 'client', 'evaluation_handler_collector': 'evaluation_handlers', 'run_evaluators': '(run_evaluators or [])'}), '(_callbacks_initializer, project_name=project_name, client\n =client, evaluation_handler_collector=evaluation_handlers,\n run_evaluators=run_evaluators or [])\n', (30361, 30523), False, 'import functools\n'), ((18522, 18696), 'langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type', 'StringRunEvaluatorChain.from_run_and_data_type', (['custom_evaluator', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key'}), '(custom_evaluator, run_type,\n data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key)\n', (18568, 18696), False, 'from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain\n'), ((30597, 30634), 'functools.partial', 'functools.partial', (['process_example', 'e'], {}), '(process_example, e)\n', (30614, 30634), False, 'import functools\n')]
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session from langchain.utils import get_from_env try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento RETURN_VAL_TYPE = Sequence[Generation] def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) class BaseCache(ABC): """Base interface for cache.""" @abstractmethod def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" @abstractmethod def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" @abstractmethod def clear(self, **kwargs: Any) -> None: """Clear cache that can take additional keyword arguments.""" class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" # TODO - implement a TTL policy in Redis def __init__(self, redis_: Any): """Initialize by passing in Redis instance.""" try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) self.redis.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, ) except ValueError: redis = RedisVectorstore( embedding_function=self.embedding.embed_query, index_name=index_name, redis_url=self.redis_url, ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations = [] # Read from a Hash results = llm_cache.similarity_search_limit_score( query=prompt, k=1, score_threshold=self.score_threshold, ) if results: for document in results: for text in document.metadata["return_val"]: generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) # Write to vectorstore metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": [generation.text for generation in return_val], } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((1483, 1510), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1500, 1510), False, 'import logging\n'), ((3955, 3973), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3971, 3973), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((4130, 4162), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4136, 4162), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4173, 4205), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4179, 4205), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4216, 4249), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (4222, 4249), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4265, 4279), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (4271, 4279), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2453, 2481), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2463, 2481), False, 'import json\n'), ((6683, 6726), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (6696, 6726), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((15254, 15261), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (15259, 15261), False, 'from gptcache import Cache\n'), ((16587, 16619), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (16590, 16619), False, 'from gptcache.adapter.api import get\n'), ((17506, 17552), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (17509, 17552), False, 'from gptcache.adapter.api import put\n'), ((21401, 21443), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (21431, 21443), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21467, 21523), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (21478, 21523), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((2498, 2527), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2508, 2527), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5029, 5049), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (5036, 5049), False, 'from sqlalchemy.orm import Session\n'), ((6136, 6156), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6143, 6156), False, 'from sqlalchemy.orm import Session\n'), ((6338, 6358), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6345, 6358), False, 'from sqlalchemy.orm import Session\n'), ((10691, 10807), 'langchain.vectorstores.redis.Redis.from_existing_index', 'RedisVectorstore.from_existing_index', ([], {'embedding': 'self.embedding', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding=self.embedding, index_name=\n index_name, redis_url=self.redis_url)\n', (10727, 10807), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((15328, 15370), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (15345, 15370), False, 'import inspect\n'), ((17775, 17805), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (17779, 17805), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast\n'), ((18705, 18725), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (18714, 18725), False, 'from datetime import timedelta\n'), ((21268, 21294), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (21292, 21294), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21330, 21378), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (21342, 21378), False, 'from langchain.utils import get_from_env\n'), ((8403, 8522), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (8416, 8522), False, 'import warnings\n'), ((10913, 11030), 'langchain.vectorstores.redis.Redis', 'RedisVectorstore', ([], {'embedding_function': 'self.embedding.embed_query', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding_function=self.embedding.embed_query, index_name=\n index_name, redis_url=self.redis_url)\n', (10929, 11030), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((12815, 12943), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (12828, 12943), False, 'import warnings\n'), ((16673, 16702), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (16683, 16702), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6046, 6056), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (6051, 6056), False, 'from langchain.load.dump import dumps\n'), ((7862, 7883), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (7872, 7883), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((15714, 15752), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (15730, 15752), False, 'from gptcache.manager.factory import get_data_manager\n'), ((16726, 16741), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (16736, 16741), False, 'import json\n'), ((5184, 5197), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (5189, 5197), False, 'from langchain.load.load import loads\n'), ((12266, 12287), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (12276, 12287), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5754, 5777), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (5764, 5777), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((4799, 4833), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (4805, 4833), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session from langchain.utils import get_from_env try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento RETURN_VAL_TYPE = Sequence[Generation] def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) class BaseCache(ABC): """Base interface for cache.""" @abstractmethod def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" @abstractmethod def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" @abstractmethod def clear(self, **kwargs: Any) -> None: """Clear cache that can take additional keyword arguments.""" class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" # TODO - implement a TTL policy in Redis def __init__(self, redis_: Any): """Initialize by passing in Redis instance.""" try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) self.redis.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, ) except ValueError: redis = RedisVectorstore( embedding_function=self.embedding.embed_query, index_name=index_name, redis_url=self.redis_url, ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations = [] # Read from a Hash results = llm_cache.similarity_search_limit_score( query=prompt, k=1, score_threshold=self.score_threshold, ) if results: for document in results: for text in document.metadata["return_val"]: generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) # Write to vectorstore metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": [generation.text for generation in return_val], } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((1483, 1510), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1500, 1510), False, 'import logging\n'), ((3955, 3973), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3971, 3973), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((4130, 4162), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4136, 4162), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4173, 4205), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4179, 4205), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4216, 4249), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (4222, 4249), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4265, 4279), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (4271, 4279), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2453, 2481), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2463, 2481), False, 'import json\n'), ((6683, 6726), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (6696, 6726), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((15254, 15261), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (15259, 15261), False, 'from gptcache import Cache\n'), ((16587, 16619), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (16590, 16619), False, 'from gptcache.adapter.api import get\n'), ((17506, 17552), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (17509, 17552), False, 'from gptcache.adapter.api import put\n'), ((21401, 21443), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (21431, 21443), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21467, 21523), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (21478, 21523), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((2498, 2527), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2508, 2527), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5029, 5049), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (5036, 5049), False, 'from sqlalchemy.orm import Session\n'), ((6136, 6156), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6143, 6156), False, 'from sqlalchemy.orm import Session\n'), ((6338, 6358), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6345, 6358), False, 'from sqlalchemy.orm import Session\n'), ((10691, 10807), 'langchain.vectorstores.redis.Redis.from_existing_index', 'RedisVectorstore.from_existing_index', ([], {'embedding': 'self.embedding', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding=self.embedding, index_name=\n index_name, redis_url=self.redis_url)\n', (10727, 10807), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((15328, 15370), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (15345, 15370), False, 'import inspect\n'), ((17775, 17805), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (17779, 17805), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast\n'), ((18705, 18725), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (18714, 18725), False, 'from datetime import timedelta\n'), ((21268, 21294), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (21292, 21294), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21330, 21378), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (21342, 21378), False, 'from langchain.utils import get_from_env\n'), ((8403, 8522), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (8416, 8522), False, 'import warnings\n'), ((10913, 11030), 'langchain.vectorstores.redis.Redis', 'RedisVectorstore', ([], {'embedding_function': 'self.embedding.embed_query', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding_function=self.embedding.embed_query, index_name=\n index_name, redis_url=self.redis_url)\n', (10929, 11030), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((12815, 12943), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (12828, 12943), False, 'import warnings\n'), ((16673, 16702), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (16683, 16702), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6046, 6056), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (6051, 6056), False, 'from langchain.load.dump import dumps\n'), ((7862, 7883), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (7872, 7883), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((15714, 15752), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (15730, 15752), False, 'from gptcache.manager.factory import get_data_manager\n'), ((16726, 16741), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (16736, 16741), False, 'import json\n'), ((5184, 5197), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (5189, 5197), False, 'from langchain.load.load import loads\n'), ((12266, 12287), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (12276, 12287), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5754, 5777), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (5764, 5777), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((4799, 4833), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (4805, 4833), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session from langchain.utils import get_from_env try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento RETURN_VAL_TYPE = Sequence[Generation] def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) class BaseCache(ABC): """Base interface for cache.""" @abstractmethod def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" @abstractmethod def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" @abstractmethod def clear(self, **kwargs: Any) -> None: """Clear cache that can take additional keyword arguments.""" class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" # TODO - implement a TTL policy in Redis def __init__(self, redis_: Any): """Initialize by passing in Redis instance.""" try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) self.redis.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, ) except ValueError: redis = RedisVectorstore( embedding_function=self.embedding.embed_query, index_name=index_name, redis_url=self.redis_url, ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations = [] # Read from a Hash results = llm_cache.similarity_search_limit_score( query=prompt, k=1, score_threshold=self.score_threshold, ) if results: for document in results: for text in document.metadata["return_val"]: generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) # Write to vectorstore metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": [generation.text for generation in return_val], } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((1483, 1510), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1500, 1510), False, 'import logging\n'), ((3955, 3973), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3971, 3973), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((4130, 4162), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4136, 4162), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4173, 4205), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4179, 4205), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4216, 4249), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (4222, 4249), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4265, 4279), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (4271, 4279), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2453, 2481), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2463, 2481), False, 'import json\n'), ((6683, 6726), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (6696, 6726), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((15254, 15261), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (15259, 15261), False, 'from gptcache import Cache\n'), ((16587, 16619), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (16590, 16619), False, 'from gptcache.adapter.api import get\n'), ((17506, 17552), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (17509, 17552), False, 'from gptcache.adapter.api import put\n'), ((21401, 21443), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (21431, 21443), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21467, 21523), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (21478, 21523), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((2498, 2527), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2508, 2527), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5029, 5049), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (5036, 5049), False, 'from sqlalchemy.orm import Session\n'), ((6136, 6156), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6143, 6156), False, 'from sqlalchemy.orm import Session\n'), ((6338, 6358), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6345, 6358), False, 'from sqlalchemy.orm import Session\n'), ((10691, 10807), 'langchain.vectorstores.redis.Redis.from_existing_index', 'RedisVectorstore.from_existing_index', ([], {'embedding': 'self.embedding', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding=self.embedding, index_name=\n index_name, redis_url=self.redis_url)\n', (10727, 10807), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((15328, 15370), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (15345, 15370), False, 'import inspect\n'), ((17775, 17805), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (17779, 17805), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast\n'), ((18705, 18725), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (18714, 18725), False, 'from datetime import timedelta\n'), ((21268, 21294), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (21292, 21294), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21330, 21378), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (21342, 21378), False, 'from langchain.utils import get_from_env\n'), ((8403, 8522), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (8416, 8522), False, 'import warnings\n'), ((10913, 11030), 'langchain.vectorstores.redis.Redis', 'RedisVectorstore', ([], {'embedding_function': 'self.embedding.embed_query', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding_function=self.embedding.embed_query, index_name=\n index_name, redis_url=self.redis_url)\n', (10929, 11030), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((12815, 12943), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (12828, 12943), False, 'import warnings\n'), ((16673, 16702), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (16683, 16702), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6046, 6056), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (6051, 6056), False, 'from langchain.load.dump import dumps\n'), ((7862, 7883), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (7872, 7883), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((15714, 15752), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (15730, 15752), False, 'from gptcache.manager.factory import get_data_manager\n'), ((16726, 16741), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (16736, 16741), False, 'import json\n'), ((5184, 5197), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (5189, 5197), False, 'from langchain.load.load import loads\n'), ((12266, 12287), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (12276, 12287), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5754, 5777), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (5764, 5777), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((4799, 4833), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (4805, 4833), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
""" .. warning:: Beta Feature! **Cache** provides an optional caching layer for LLMs. Cache is useful for two reasons: - It can save you money by reducing the number of API calls you make to the LLM provider if you're often requesting the same completion multiple times. - It can speed up your application by reducing the number of API calls you make to the LLM provider. Cache directly competes with Memory. See documentation for Pros and Cons. **Class hierarchy:** .. code-block:: BaseCache --> <name>Cache # Examples: InMemoryCache, RedisCache, GPTCache """ from __future__ import annotations import hashlib import inspect import json import logging import warnings from abc import ABC, abstractmethod from datetime import timedelta from typing import ( TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast, ) from sqlalchemy import Column, Integer, String, create_engine, select from sqlalchemy.engine.base import Engine from sqlalchemy.orm import Session from langchain.utils import get_from_env try: from sqlalchemy.orm import declarative_base except ImportError: from sqlalchemy.ext.declarative import declarative_base from langchain.embeddings.base import Embeddings from langchain.load.dump import dumps from langchain.load.load import loads from langchain.schema import ChatGeneration, Generation from langchain.vectorstores.redis import Redis as RedisVectorstore logger = logging.getLogger(__file__) if TYPE_CHECKING: import momento RETURN_VAL_TYPE = Sequence[Generation] def _hash(_input: str) -> str: """Use a deterministic hashing approach.""" return hashlib.md5(_input.encode()).hexdigest() def _dump_generations_to_json(generations: RETURN_VAL_TYPE) -> str: """Dump generations to json. Args: generations (RETURN_VAL_TYPE): A list of language model generations. Returns: str: Json representing a list of generations. """ return json.dumps([generation.dict() for generation in generations]) def _load_generations_from_json(generations_json: str) -> RETURN_VAL_TYPE: """Load generations from json. Args: generations_json (str): A string of json representing a list of generations. Raises: ValueError: Could not decode json string to list of generations. Returns: RETURN_VAL_TYPE: A list of generations. """ try: results = json.loads(generations_json) return [Generation(**generation_dict) for generation_dict in results] except json.JSONDecodeError: raise ValueError( f"Could not decode json to list of generations: {generations_json}" ) class BaseCache(ABC): """Base interface for cache.""" @abstractmethod def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" @abstractmethod def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" @abstractmethod def clear(self, **kwargs: Any) -> None: """Clear cache that can take additional keyword arguments.""" class InMemoryCache(BaseCache): """Cache that stores things in memory.""" def __init__(self) -> None: """Initialize with empty cache.""" self._cache: Dict[Tuple[str, str], RETURN_VAL_TYPE] = {} def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" return self._cache.get((prompt, llm_string), None) def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" self._cache[(prompt, llm_string)] = return_val def clear(self, **kwargs: Any) -> None: """Clear cache.""" self._cache = {} Base = declarative_base() class FullLLMCache(Base): # type: ignore """SQLite table for full LLM Cache (all generations).""" __tablename__ = "full_llm_cache" prompt = Column(String, primary_key=True) llm = Column(String, primary_key=True) idx = Column(Integer, primary_key=True) response = Column(String) class SQLAlchemyCache(BaseCache): """Cache that uses SQAlchemy as a backend.""" def __init__(self, engine: Engine, cache_schema: Type[FullLLMCache] = FullLLMCache): """Initialize by creating all tables.""" self.engine = engine self.cache_schema = cache_schema self.cache_schema.metadata.create_all(self.engine) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" stmt = ( select(self.cache_schema.response) .where(self.cache_schema.prompt == prompt) # type: ignore .where(self.cache_schema.llm == llm_string) .order_by(self.cache_schema.idx) ) with Session(self.engine) as session: rows = session.execute(stmt).fetchall() if rows: try: return [loads(row[0]) for row in rows] except Exception: logger.warning( "Retrieving a cache value that could not be deserialized " "properly. This is likely due to the cache being in an " "older format. Please recreate your cache to avoid this " "error." ) # In a previous life we stored the raw text directly # in the table, so assume it's in that format. return [Generation(text=row[0]) for row in rows] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update based on prompt and llm_string.""" items = [ self.cache_schema(prompt=prompt, llm=llm_string, response=dumps(gen), idx=i) for i, gen in enumerate(return_val) ] with Session(self.engine) as session, session.begin(): for item in items: session.merge(item) def clear(self, **kwargs: Any) -> None: """Clear cache.""" with Session(self.engine) as session: session.query(self.cache_schema).delete() session.commit() class SQLiteCache(SQLAlchemyCache): """Cache that uses SQLite as a backend.""" def __init__(self, database_path: str = ".langchain.db"): """Initialize by creating the engine and all tables.""" engine = create_engine(f"sqlite:///{database_path}") super().__init__(engine) class RedisCache(BaseCache): """Cache that uses Redis as a backend.""" # TODO - implement a TTL policy in Redis def __init__(self, redis_: Any): """Initialize by passing in Redis instance.""" try: from redis import Redis except ImportError: raise ValueError( "Could not import redis python package. " "Please install it with `pip install redis`." ) if not isinstance(redis_, Redis): raise ValueError("Please pass in Redis object.") self.redis = redis_ def _key(self, prompt: str, llm_string: str) -> str: """Compute key from prompt and llm_string""" return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" generations = [] # Read from a Redis HASH results = self.redis.hgetall(self._key(prompt, llm_string)) if results: for _, text in results.items(): generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisCache does not" " support caching ChatModel outputs." ) return # Write to a Redis HASH key = self._key(prompt, llm_string) self.redis.hset( key, mapping={ str(idx): generation.text for idx, generation in enumerate(return_val) }, ) def clear(self, **kwargs: Any) -> None: """Clear cache. If `asynchronous` is True, flush asynchronously.""" asynchronous = kwargs.get("asynchronous", False) self.redis.flushdb(asynchronous=asynchronous, **kwargs) class RedisSemanticCache(BaseCache): """Cache that uses Redis as a vector-store backend.""" # TODO - implement a TTL policy in Redis def __init__( self, redis_url: str, embedding: Embeddings, score_threshold: float = 0.2 ): """Initialize by passing in the `init` GPTCache func Args: redis_url (str): URL to connect to Redis. embedding (Embedding): Embedding provider for semantic encoding and search. score_threshold (float, 0.2): Example: .. code-block:: python import langchain from langchain.cache import RedisSemanticCache from langchain.embeddings import OpenAIEmbeddings langchain.llm_cache = RedisSemanticCache( redis_url="redis://localhost:6379", embedding=OpenAIEmbeddings() ) """ self._cache_dict: Dict[str, RedisVectorstore] = {} self.redis_url = redis_url self.embedding = embedding self.score_threshold = score_threshold def _index_name(self, llm_string: str) -> str: hashed_index = _hash(llm_string) return f"cache:{hashed_index}" def _get_llm_cache(self, llm_string: str) -> RedisVectorstore: index_name = self._index_name(llm_string) # return vectorstore client for the specific llm string if index_name in self._cache_dict: return self._cache_dict[index_name] # create new vectorstore client for the specific llm string try: self._cache_dict[index_name] = RedisVectorstore.from_existing_index( embedding=self.embedding, index_name=index_name, redis_url=self.redis_url, ) except ValueError: redis = RedisVectorstore( embedding_function=self.embedding.embed_query, index_name=index_name, redis_url=self.redis_url, ) _embedding = self.embedding.embed_query(text="test") redis._create_index(dim=len(_embedding)) self._cache_dict[index_name] = redis return self._cache_dict[index_name] def clear(self, **kwargs: Any) -> None: """Clear semantic cache for a given llm_string.""" index_name = self._index_name(kwargs["llm_string"]) if index_name in self._cache_dict: self._cache_dict[index_name].drop_index( index_name=index_name, delete_documents=True, redis_url=self.redis_url ) del self._cache_dict[index_name] def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up based on prompt and llm_string.""" llm_cache = self._get_llm_cache(llm_string) generations = [] # Read from a Hash results = llm_cache.similarity_search_limit_score( query=prompt, k=1, score_threshold=self.score_threshold, ) if results: for document in results: for text in document.metadata["return_val"]: generations.append(Generation(text=text)) return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache based on prompt and llm_string.""" for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "RedisSemanticCache only supports caching of " f"normal LLM generations, got {type(gen)}" ) if isinstance(gen, ChatGeneration): warnings.warn( "NOTE: Generation has not been cached. RedisSentimentCache does not" " support caching ChatModel outputs." ) return llm_cache = self._get_llm_cache(llm_string) # Write to vectorstore metadata = { "llm_string": llm_string, "prompt": prompt, "return_val": [generation.text for generation in return_val], } llm_cache.add_texts(texts=[prompt], metadatas=[metadata]) class GPTCache(BaseCache): """Cache that uses GPTCache as a backend.""" def __init__( self, init_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = None, ): """Initialize by passing in init function (default: `None`). Args: init_func (Optional[Callable[[Any], None]]): init `GPTCache` function (default: `None`) Example: .. code-block:: python # Initialize GPTCache with a custom init function import gptcache from gptcache.processor.pre import get_prompt from gptcache.manager.factory import get_data_manager # Avoid multiple caches using the same file, causing different llm model caches to affect each other def init_gptcache(cache_obj: gptcache.Cache, llm str): cache_obj.init( pre_embedding_func=get_prompt, data_manager=manager_factory( manager="map", data_dir=f"map_cache_{llm}" ), ) langchain.llm_cache = GPTCache(init_gptcache) """ try: import gptcache # noqa: F401 except ImportError: raise ImportError( "Could not import gptcache python package. " "Please install it with `pip install gptcache`." ) self.init_gptcache_func: Union[ Callable[[Any, str], None], Callable[[Any], None], None ] = init_func self.gptcache_dict: Dict[str, Any] = {} def _new_gptcache(self, llm_string: str) -> Any: """New gptcache object""" from gptcache import Cache from gptcache.manager.factory import get_data_manager from gptcache.processor.pre import get_prompt _gptcache = Cache() if self.init_gptcache_func is not None: sig = inspect.signature(self.init_gptcache_func) if len(sig.parameters) == 2: self.init_gptcache_func(_gptcache, llm_string) # type: ignore[call-arg] else: self.init_gptcache_func(_gptcache) # type: ignore[call-arg] else: _gptcache.init( pre_embedding_func=get_prompt, data_manager=get_data_manager(data_path=llm_string), ) self.gptcache_dict[llm_string] = _gptcache return _gptcache def _get_gptcache(self, llm_string: str) -> Any: """Get a cache object. When the corresponding llm model cache does not exist, it will be created.""" _gptcache = self.gptcache_dict.get(llm_string, None) if not _gptcache: _gptcache = self._new_gptcache(llm_string) return _gptcache def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Look up the cache data. First, retrieve the corresponding cache object using the `llm_string` parameter, and then retrieve the data from the cache based on the `prompt`. """ from gptcache.adapter.api import get _gptcache = self._get_gptcache(llm_string) res = get(prompt, cache_obj=_gptcache) if res: return [ Generation(**generation_dict) for generation_dict in json.loads(res) ] return None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Update cache. First, retrieve the corresponding cache object using the `llm_string` parameter, and then store the `prompt` and `return_val` in the cache object. """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "GPTCache only supports caching of normal LLM generations, " f"got {type(gen)}" ) from gptcache.adapter.api import put _gptcache = self._get_gptcache(llm_string) handled_data = json.dumps([generation.dict() for generation in return_val]) put(prompt, handled_data, cache_obj=_gptcache) return None def clear(self, **kwargs: Any) -> None: """Clear cache.""" from gptcache import Cache for gptcache_instance in self.gptcache_dict.values(): gptcache_instance = cast(Cache, gptcache_instance) gptcache_instance.flush() self.gptcache_dict.clear() def _ensure_cache_exists(cache_client: momento.CacheClient, cache_name: str) -> None: """Create cache if it doesn't exist. Raises: SdkException: Momento service or network error Exception: Unexpected response """ from momento.responses import CreateCache create_cache_response = cache_client.create_cache(cache_name) if isinstance(create_cache_response, CreateCache.Success) or isinstance( create_cache_response, CreateCache.CacheAlreadyExists ): return None elif isinstance(create_cache_response, CreateCache.Error): raise create_cache_response.inner_exception else: raise Exception(f"Unexpected response cache creation: {create_cache_response}") def _validate_ttl(ttl: Optional[timedelta]) -> None: if ttl is not None and ttl <= timedelta(seconds=0): raise ValueError(f"ttl must be positive but was {ttl}.") class MomentoCache(BaseCache): """Cache that uses Momento as a backend. See https://gomomento.com/""" def __init__( self, cache_client: momento.CacheClient, cache_name: str, *, ttl: Optional[timedelta] = None, ensure_cache_exists: bool = True, ): """Instantiate a prompt cache using Momento as a backend. Note: to instantiate the cache client passed to MomentoCache, you must have a Momento account. See https://gomomento.com/. Args: cache_client (CacheClient): The Momento cache client. cache_name (str): The name of the cache to use to store the data. ttl (Optional[timedelta], optional): The time to live for the cache items. Defaults to None, ie use the client default TTL. ensure_cache_exists (bool, optional): Create the cache if it doesn't exist. Defaults to True. Raises: ImportError: Momento python package is not installed. TypeError: cache_client is not of type momento.CacheClientObject ValueError: ttl is non-null and non-negative """ try: from momento import CacheClient except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if not isinstance(cache_client, CacheClient): raise TypeError("cache_client must be a momento.CacheClient object.") _validate_ttl(ttl) if ensure_cache_exists: _ensure_cache_exists(cache_client, cache_name) self.cache_client = cache_client self.cache_name = cache_name self.ttl = ttl @classmethod def from_client_params( cls, cache_name: str, ttl: timedelta, *, configuration: Optional[momento.config.Configuration] = None, auth_token: Optional[str] = None, **kwargs: Any, ) -> MomentoCache: """Construct cache from CacheClient parameters.""" try: from momento import CacheClient, Configurations, CredentialProvider except ImportError: raise ImportError( "Could not import momento python package. " "Please install it with `pip install momento`." ) if configuration is None: configuration = Configurations.Laptop.v1() auth_token = auth_token or get_from_env("auth_token", "MOMENTO_AUTH_TOKEN") credentials = CredentialProvider.from_string(auth_token) cache_client = CacheClient(configuration, credentials, default_ttl=ttl) return cls(cache_client, cache_name, ttl=ttl, **kwargs) def __key(self, prompt: str, llm_string: str) -> str: """Compute cache key from prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Returns: str: The cache key. """ return _hash(prompt + llm_string) def lookup(self, prompt: str, llm_string: str) -> Optional[RETURN_VAL_TYPE]: """Lookup llm generations in cache by prompt and associated model and settings. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model version and settings. Raises: SdkException: Momento service or network error Returns: Optional[RETURN_VAL_TYPE]: A list of language model generations. """ from momento.responses import CacheGet generations: RETURN_VAL_TYPE = [] get_response = self.cache_client.get( self.cache_name, self.__key(prompt, llm_string) ) if isinstance(get_response, CacheGet.Hit): value = get_response.value_string generations = _load_generations_from_json(value) elif isinstance(get_response, CacheGet.Miss): pass elif isinstance(get_response, CacheGet.Error): raise get_response.inner_exception return generations if generations else None def update(self, prompt: str, llm_string: str, return_val: RETURN_VAL_TYPE) -> None: """Store llm generations in cache. Args: prompt (str): The prompt run through the language model. llm_string (str): The language model string. return_val (RETURN_VAL_TYPE): A list of language model generations. Raises: SdkException: Momento service or network error Exception: Unexpected response """ for gen in return_val: if not isinstance(gen, Generation): raise ValueError( "Momento only supports caching of normal LLM generations, " f"got {type(gen)}" ) key = self.__key(prompt, llm_string) value = _dump_generations_to_json(return_val) set_response = self.cache_client.set(self.cache_name, key, value, self.ttl) from momento.responses import CacheSet if isinstance(set_response, CacheSet.Success): pass elif isinstance(set_response, CacheSet.Error): raise set_response.inner_exception else: raise Exception(f"Unexpected response: {set_response}") def clear(self, **kwargs: Any) -> None: """Clear the cache. Raises: SdkException: Momento service or network error """ from momento.responses import CacheFlush flush_response = self.cache_client.flush_cache(self.cache_name) if isinstance(flush_response, CacheFlush.Success): pass elif isinstance(flush_response, CacheFlush.Error): raise flush_response.inner_exception
[ "langchain.utils.get_from_env", "langchain.schema.Generation", "langchain.load.dump.dumps", "langchain.vectorstores.redis.Redis.from_existing_index", "langchain.vectorstores.redis.Redis", "langchain.load.load.loads" ]
[((1483, 1510), 'logging.getLogger', 'logging.getLogger', (['__file__'], {}), '(__file__)\n', (1500, 1510), False, 'import logging\n'), ((3955, 3973), 'sqlalchemy.ext.declarative.declarative_base', 'declarative_base', ([], {}), '()\n', (3971, 3973), False, 'from sqlalchemy.ext.declarative import declarative_base\n'), ((4130, 4162), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4136, 4162), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4173, 4205), 'sqlalchemy.Column', 'Column', (['String'], {'primary_key': '(True)'}), '(String, primary_key=True)\n', (4179, 4205), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4216, 4249), 'sqlalchemy.Column', 'Column', (['Integer'], {'primary_key': '(True)'}), '(Integer, primary_key=True)\n', (4222, 4249), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((4265, 4279), 'sqlalchemy.Column', 'Column', (['String'], {}), '(String)\n', (4271, 4279), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((2453, 2481), 'json.loads', 'json.loads', (['generations_json'], {}), '(generations_json)\n', (2463, 2481), False, 'import json\n'), ((6683, 6726), 'sqlalchemy.create_engine', 'create_engine', (['f"""sqlite:///{database_path}"""'], {}), "(f'sqlite:///{database_path}')\n", (6696, 6726), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n'), ((15254, 15261), 'gptcache.Cache', 'Cache', ([], {}), '()\n', (15259, 15261), False, 'from gptcache import Cache\n'), ((16587, 16619), 'gptcache.adapter.api.get', 'get', (['prompt'], {'cache_obj': '_gptcache'}), '(prompt, cache_obj=_gptcache)\n', (16590, 16619), False, 'from gptcache.adapter.api import get\n'), ((17506, 17552), 'gptcache.adapter.api.put', 'put', (['prompt', 'handled_data'], {'cache_obj': '_gptcache'}), '(prompt, handled_data, cache_obj=_gptcache)\n', (17509, 17552), False, 'from gptcache.adapter.api import put\n'), ((21401, 21443), 'momento.CredentialProvider.from_string', 'CredentialProvider.from_string', (['auth_token'], {}), '(auth_token)\n', (21431, 21443), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21467, 21523), 'momento.CacheClient', 'CacheClient', (['configuration', 'credentials'], {'default_ttl': 'ttl'}), '(configuration, credentials, default_ttl=ttl)\n', (21478, 21523), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((2498, 2527), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (2508, 2527), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5029, 5049), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (5036, 5049), False, 'from sqlalchemy.orm import Session\n'), ((6136, 6156), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6143, 6156), False, 'from sqlalchemy.orm import Session\n'), ((6338, 6358), 'sqlalchemy.orm.Session', 'Session', (['self.engine'], {}), '(self.engine)\n', (6345, 6358), False, 'from sqlalchemy.orm import Session\n'), ((10691, 10807), 'langchain.vectorstores.redis.Redis.from_existing_index', 'RedisVectorstore.from_existing_index', ([], {'embedding': 'self.embedding', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding=self.embedding, index_name=\n index_name, redis_url=self.redis_url)\n', (10727, 10807), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((15328, 15370), 'inspect.signature', 'inspect.signature', (['self.init_gptcache_func'], {}), '(self.init_gptcache_func)\n', (15345, 15370), False, 'import inspect\n'), ((17775, 17805), 'typing.cast', 'cast', (['Cache', 'gptcache_instance'], {}), '(Cache, gptcache_instance)\n', (17779, 17805), False, 'from typing import TYPE_CHECKING, Any, Callable, Dict, Optional, Sequence, Tuple, Type, Union, cast\n'), ((18705, 18725), 'datetime.timedelta', 'timedelta', ([], {'seconds': '(0)'}), '(seconds=0)\n', (18714, 18725), False, 'from datetime import timedelta\n'), ((21268, 21294), 'momento.Configurations.Laptop.v1', 'Configurations.Laptop.v1', ([], {}), '()\n', (21292, 21294), False, 'from momento import CacheClient, Configurations, CredentialProvider\n'), ((21330, 21378), 'langchain.utils.get_from_env', 'get_from_env', (['"""auth_token"""', '"""MOMENTO_AUTH_TOKEN"""'], {}), "('auth_token', 'MOMENTO_AUTH_TOKEN')\n", (21342, 21378), False, 'from langchain.utils import get_from_env\n'), ((8403, 8522), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisCache does not support caching ChatModel outputs.'\n )\n", (8416, 8522), False, 'import warnings\n'), ((10913, 11030), 'langchain.vectorstores.redis.Redis', 'RedisVectorstore', ([], {'embedding_function': 'self.embedding.embed_query', 'index_name': 'index_name', 'redis_url': 'self.redis_url'}), '(embedding_function=self.embedding.embed_query, index_name=\n index_name, redis_url=self.redis_url)\n', (10929, 11030), True, 'from langchain.vectorstores.redis import Redis as RedisVectorstore\n'), ((12815, 12943), 'warnings.warn', 'warnings.warn', (['"""NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs."""'], {}), "(\n 'NOTE: Generation has not been cached. RedisSentimentCache does not support caching ChatModel outputs.'\n )\n", (12828, 12943), False, 'import warnings\n'), ((16673, 16702), 'langchain.schema.Generation', 'Generation', ([], {}), '(**generation_dict)\n', (16683, 16702), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((6046, 6056), 'langchain.load.dump.dumps', 'dumps', (['gen'], {}), '(gen)\n', (6051, 6056), False, 'from langchain.load.dump import dumps\n'), ((7862, 7883), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (7872, 7883), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((15714, 15752), 'gptcache.manager.factory.get_data_manager', 'get_data_manager', ([], {'data_path': 'llm_string'}), '(data_path=llm_string)\n', (15730, 15752), False, 'from gptcache.manager.factory import get_data_manager\n'), ((16726, 16741), 'json.loads', 'json.loads', (['res'], {}), '(res)\n', (16736, 16741), False, 'import json\n'), ((5184, 5197), 'langchain.load.load.loads', 'loads', (['row[0]'], {}), '(row[0])\n', (5189, 5197), False, 'from langchain.load.load import loads\n'), ((12266, 12287), 'langchain.schema.Generation', 'Generation', ([], {'text': 'text'}), '(text=text)\n', (12276, 12287), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((5754, 5777), 'langchain.schema.Generation', 'Generation', ([], {'text': 'row[0]'}), '(text=row[0])\n', (5764, 5777), False, 'from langchain.schema import ChatGeneration, Generation\n'), ((4799, 4833), 'sqlalchemy.select', 'select', (['self.cache_schema.response'], {}), '(self.cache_schema.response)\n', (4805, 4833), False, 'from sqlalchemy import Column, Integer, String, create_engine, select\n')]
"""Utilities for running language models or Chains over datasets.""" from __future__ import annotations import asyncio import functools import inspect import itertools import logging import uuid import warnings from enum import Enum from typing import ( TYPE_CHECKING, Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, cast, ) from urllib.parse import urlparse, urlunparse from langsmith import Client, RunEvaluator from langsmith.schemas import Dataset, DataType, Example from langchain.callbacks.base import BaseCallbackHandler from langchain.callbacks.manager import Callbacks from langchain.callbacks.tracers.base import BaseTracer from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler from langchain.callbacks.tracers.langchain import LangChainTracer from langchain.chains.base import Chain from langchain.evaluation.loading import load_evaluator from langchain.evaluation.schema import EvaluatorType, StringEvaluator from langchain.schema import ChatResult, LLMResult from langchain.schema.language_model import BaseLanguageModel from langchain.schema.messages import BaseMessage, messages_from_dict from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda from langchain.smith.evaluation.config import EvalConfig, RunEvalConfig from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain if TYPE_CHECKING: import pandas as pd logger = logging.getLogger(__name__) MODEL_OR_CHAIN_FACTORY = Union[ Callable[[], Union[Chain, Runnable]], BaseLanguageModel, Callable[[dict], Any], Runnable, Chain, ] MCF = Union[Callable[[], Union[Chain, Runnable]], BaseLanguageModel] class InputFormatError(Exception): """Raised when the input format is invalid.""" ## Shared Utilities class TestResult(dict): """A dictionary of the results of a single test run.""" def to_dataframe(self) -> pd.DataFrame: """Convert the results to a dataframe.""" try: import pandas as pd except ImportError as e: raise ImportError( "Pandas is required to convert the results to a dataframe." " to install pandas, run `pip install pandas`." ) from e indices = [] records = [] for example_id, result in self["results"].items(): feedback = result["feedback"] records.append( {**{f.key: f.score for f in feedback}, "output": result["output"]} ) indices.append(example_id) return pd.DataFrame(records, index=indices) def _get_eval_project_url(api_url: str, project_id: str) -> str: """Get the project url from the api url.""" parsed = urlparse(api_url) hostname = parsed.hostname or "" if "api." in hostname: hostname = hostname.replace("api.", "", 1) if "localhost" in hostname: # Remove the port hostname = "localhost" url = urlunparse(parsed._replace(netloc=hostname)) return f"{url}/projects/p/{project_id}?eval=true" def _wrap_in_chain_factory( llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, dataset_name: str = "<my_dataset>", ) -> MCF: """Forgive the user if they pass in a chain without memory instead of a chain factory. It's a common mistake. Raise a more helpful error message as well.""" if isinstance(llm_or_chain_factory, Chain): chain = llm_or_chain_factory chain_class = chain.__class__.__name__ if llm_or_chain_factory.memory is not None: memory_class = chain.memory.__class__.__name__ raise ValueError( "Cannot directly evaluate a chain with stateful memory." " To evaluate this chain, pass in a chain constructor" " that initializes fresh memory each time it is called." " This will safegaurd against information" " leakage between dataset examples." "\nFor example:\n\n" "def chain_constructor():\n" f" new_memory = {memory_class}(...)\n" f" return {chain_class}" "(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) logger.warning( "Directly passing in a chain is not recommended as chains may have state." " This can lead to unexpected behavior as the " "same chain instance could be used across multiple datasets. Instead," " please pass a chain constructor that creates a new " "chain with fresh memory each time it is called. This will safeguard" " against information leakage between dataset examples. " "\nFor example:\n\n" "def chain_constructor():\n" f" return {chain_class}(memory=new_memory, ...)\n\n" f'run_on_dataset("{dataset_name}", chain_constructor, ...)' ) return lambda: chain elif isinstance(llm_or_chain_factory, BaseLanguageModel): return llm_or_chain_factory elif isinstance(llm_or_chain_factory, Runnable): # Memory may exist here, but it's not elegant to check all those cases. lcf = llm_or_chain_factory return lambda: lcf elif callable(llm_or_chain_factory): try: _model = llm_or_chain_factory() # type: ignore[call-arg] except TypeError: # It's an arbitrary function, wrap it in a RunnableLambda user_func = cast(Callable, llm_or_chain_factory) sig = inspect.signature(user_func) logger.info(f"Wrapping function {sig} as RunnableLambda.") wrapped = RunnableLambda(user_func) return lambda: wrapped constructor = cast(Callable, llm_or_chain_factory) if isinstance(_model, BaseLanguageModel): # It's not uncommon to do an LLM constructor instead of raw LLM, # so we'll unpack it for the user. return _model elif not isinstance(_model, Runnable): # This is unlikely to happen - a constructor for a model function return lambda: RunnableLambda(constructor) else: # Typical correct case return constructor # noqa return llm_or_chain_factory def _first_example(examples: Iterator[Example]) -> Tuple[Example, Iterator[Example]]: """Get the first example while chaining it back and preserving the iterator.""" try: example: Example = next(examples) except StopIteration: raise ValueError("No examples provided.") return example, itertools.chain([example], examples) def _get_prompt(inputs: Dict[str, Any]) -> str: """Get prompt from inputs. Args: inputs: The input dictionary. Returns: A string prompt. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") prompts = [] if "prompt" in inputs: if not isinstance(inputs["prompt"], str): raise InputFormatError( "Expected string for 'prompt', got" f" {type(inputs['prompt']).__name__}" ) prompts = [inputs["prompt"]] elif "prompts" in inputs: if not isinstance(inputs["prompts"], list) or not all( isinstance(i, str) for i in inputs["prompts"] ): raise InputFormatError( "Expected list of strings for 'prompts'," f" got {type(inputs['prompts']).__name__}" ) prompts = inputs["prompts"] elif len(inputs) == 1: prompt_ = next(iter(inputs.values())) if isinstance(prompt_, str): prompts = [prompt_] elif isinstance(prompt_, list) and all(isinstance(i, str) for i in prompt_): prompts = prompt_ else: raise InputFormatError(f"LLM Run expects string prompt input. Got {inputs}") else: raise InputFormatError( f"LLM Run expects 'prompt' or 'prompts' in inputs. Got {inputs}" ) if len(prompts) == 1: return prompts[0] else: raise InputFormatError( f"LLM Run expects single prompt input. Got {len(prompts)} prompts." ) def _get_messages(inputs: Dict[str, Any]) -> List[BaseMessage]: """Get Chat Messages from inputs. Args: inputs: The input dictionary. Returns: A list of chat messages. Raises: InputFormatError: If the input format is invalid. """ if not inputs: raise InputFormatError("Inputs should not be empty.") if "messages" in inputs: single_input = inputs["messages"] elif len(inputs) == 1: single_input = next(iter(inputs.values())) else: raise InputFormatError( f"Chat Run expects 'messages' in inputs when example has multiple" f" input keys. Got {inputs}" ) if isinstance(single_input, list) and all( isinstance(i, dict) for i in single_input ): raw_messages = [single_input] elif isinstance(single_input, list) and all( isinstance(i, list) for i in single_input ): raw_messages = single_input else: raise InputFormatError( f"Chat Run expects List[dict] or List[List[dict]] values for" f" 'messages' key input. Got {inputs}" ) if len(raw_messages) == 1: return messages_from_dict(raw_messages[0]) else: raise InputFormatError( f"Chat Run expects single List[dict] or List[List[dict]] 'messages'" f" input. Got {len(raw_messages)} messages from inputs {inputs}" ) def _get_project_name( project_name: Optional[str], llm_or_chain_factory: MCF, ) -> str: """ Get the project name. Args: project_name: The project name if manually specified. llm_or_chain_factory: The Chain or language model constructor. Returns: The project name. """ if project_name is not None: return project_name if isinstance(llm_or_chain_factory, BaseLanguageModel): model_name = llm_or_chain_factory.__class__.__name__ else: model_name = llm_or_chain_factory().__class__.__name__ hex = uuid.uuid4().hex return f"{hex}-{model_name}" ## Shared Validation Utilities def _validate_example_inputs_for_language_model( first_example: Example, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: if input_mapper: prompt_input = input_mapper(first_example.inputs) if not isinstance(prompt_input, str) and not ( isinstance(prompt_input, list) and all(isinstance(msg, BaseMessage) for msg in prompt_input) ): raise InputFormatError( "When using an input_mapper to prepare dataset example inputs" " for an LLM or chat model, the output must a single string or" " a list of chat messages." f"\nGot: {prompt_input} of type {type(prompt_input)}." ) else: try: _get_prompt(first_example.inputs) except InputFormatError: try: _get_messages(first_example.inputs) except InputFormatError: raise InputFormatError( "Example inputs do not match language model input format. " "Expected a dictionary with messages or a single prompt." f" Got: {first_example.inputs}" " Please update your dataset OR provide an input_mapper" " to convert the example.inputs to a compatible format" " for the llm or chat model you wish to evaluate." ) def _validate_example_inputs_for_chain( first_example: Example, chain: Chain, input_mapper: Optional[Callable[[Dict], Any]], ) -> None: """Validate that the example inputs match the chain input keys.""" if input_mapper: first_inputs = input_mapper(first_example.inputs) missing_keys = set(chain.input_keys).difference(first_inputs) if not isinstance(first_inputs, dict): raise InputFormatError( "When using an input_mapper to prepare dataset example" " inputs for a chain, the mapped value must be a dictionary." f"\nGot: {first_inputs} of type {type(first_inputs)}." ) if missing_keys: raise InputFormatError( "Missing keys after loading example using input_mapper." f"\nExpected: {chain.input_keys}. Got: {first_inputs.keys()}" ) else: first_inputs = first_example.inputs missing_keys = set(chain.input_keys).difference(first_inputs) if len(first_inputs) == 1 and len(chain.input_keys) == 1: # We can pass this through the run method. # Refrain from calling to validate. pass elif missing_keys: raise InputFormatError( "Example inputs missing expected chain input keys." " Please provide an input_mapper to convert the example.inputs" " to a compatible format for the chain you wish to evaluate." f"Expected: {chain.input_keys}. " f"Got: {first_inputs.keys()}" ) def _validate_example_inputs( examples: Iterator[Example], llm_or_chain_factory: MCF, input_mapper: Optional[Callable[[Dict], Any]], ) -> Iterator[Example]: """Validate that the example inputs are valid for the model.""" first_example, examples = _first_example(examples) if isinstance(llm_or_chain_factory, BaseLanguageModel): _validate_example_inputs_for_language_model(first_example, input_mapper) else: chain = llm_or_chain_factory() if isinstance(chain, Chain): # Otherwise it's a runnable _validate_example_inputs_for_chain(first_example, chain, input_mapper) elif isinstance(chain, Runnable): logger.debug(f"Skipping input validation for {chain}") return examples ## Shared Evaluator Setup Utilities def _setup_evaluation( llm_or_chain_factory: MCF, examples: Iterator[Example], evaluation: Optional[RunEvalConfig], data_type: DataType, ) -> Tuple[Optional[List[RunEvaluator]], Iterator[Example]]: """Configure the evaluators to run on the results of the chain.""" if evaluation: first_example, examples = _first_example(examples) if isinstance(llm_or_chain_factory, BaseLanguageModel): run_inputs, run_outputs = None, None run_type = "llm" else: run_type = "chain" if data_type in (DataType.chat, DataType.llm): val = data_type.value if isinstance(data_type, Enum) else data_type raise ValueError( "Cannot evaluate a chain on dataset with " f"data_type={val}. " "Please specify a dataset with the default 'kv' data type." ) chain = llm_or_chain_factory() run_inputs = chain.input_keys if isinstance(chain, Chain) else None run_outputs = chain.output_keys if isinstance(chain, Chain) else None run_evaluators = _load_run_evaluators( evaluation, run_type, data_type, list(first_example.outputs) if first_example.outputs else None, run_inputs, run_outputs, ) else: # TODO: Create a default helpfulness evaluator run_evaluators = None return run_evaluators, examples def _determine_input_key( config: RunEvalConfig, run_inputs: Optional[List[str]], ) -> Optional[str]: input_key = None if config.input_key: input_key = config.input_key if run_inputs and input_key not in run_inputs: raise ValueError(f"Input key {input_key} not in run inputs {run_inputs}") elif run_inputs and len(run_inputs) == 1: input_key = run_inputs[0] elif run_inputs is not None and len(run_inputs) > 1: raise ValueError( f"Must specify input key for model with multiple inputs: {run_inputs}" ) return input_key def _determine_prediction_key( config: RunEvalConfig, run_outputs: Optional[List[str]], ) -> Optional[str]: prediction_key = None if config.prediction_key: prediction_key = config.prediction_key if run_outputs and prediction_key not in run_outputs: raise ValueError( f"Prediction key {prediction_key} not in run outputs {run_outputs}" ) elif run_outputs and len(run_outputs) == 1: prediction_key = run_outputs[0] elif run_outputs is not None and len(run_outputs) > 1: raise ValueError( f"Must specify prediction key for model" f" with multiple outputs: {run_outputs}" ) return prediction_key def _determine_reference_key( config: RunEvalConfig, example_outputs: Optional[List[str]], ) -> Optional[str]: if config.reference_key: reference_key = config.reference_key if example_outputs and reference_key not in example_outputs: raise ValueError( f"Reference key {reference_key} not in Dataset" f" example outputs: {example_outputs}" ) elif example_outputs and len(example_outputs) == 1: reference_key = list(example_outputs)[0] else: reference_key = None return reference_key def _construct_run_evaluator( eval_config: Union[EvaluatorType, str, EvalConfig], eval_llm: Optional[BaseLanguageModel], run_type: str, data_type: DataType, example_outputs: Optional[List[str]], reference_key: Optional[str], input_key: Optional[str], prediction_key: Optional[str], ) -> RunEvaluator: if isinstance(eval_config, (EvaluatorType, str)): if not isinstance(eval_config, EvaluatorType): eval_config = EvaluatorType(eval_config) evaluator_ = load_evaluator(eval_config, llm=eval_llm) eval_type_tag = eval_config.value else: kwargs = {"llm": eval_llm, **eval_config.get_kwargs()} evaluator_ = load_evaluator(eval_config.evaluator_type, **kwargs) eval_type_tag = eval_config.evaluator_type.value if isinstance(evaluator_, StringEvaluator): if evaluator_.requires_reference and reference_key is None: raise ValueError( f"Must specify reference_key in RunEvalConfig to use" f" evaluator of type {eval_type_tag} with" f" dataset with multiple output keys: {example_outputs}." ) run_evaluator = StringRunEvaluatorChain.from_run_and_data_type( evaluator_, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, tags=[eval_type_tag], ) else: raise NotImplementedError( f"Run evaluator for {eval_type_tag} is not implemented" ) return run_evaluator def _get_keys( config: RunEvalConfig, run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], example_outputs: Optional[List[str]], ) -> Tuple[Optional[str], Optional[str], Optional[str]]: input_key = _determine_input_key(config, run_inputs) prediction_key = _determine_prediction_key(config, run_outputs) reference_key = _determine_reference_key(config, example_outputs) return input_key, prediction_key, reference_key def _load_run_evaluators( config: RunEvalConfig, run_type: str, data_type: DataType, example_outputs: Optional[List[str]], run_inputs: Optional[List[str]], run_outputs: Optional[List[str]], ) -> List[RunEvaluator]: """ Load run evaluators from a configuration. Args: config: Configuration for the run evaluators. Returns: A list of run evaluators. """ run_evaluators = [] input_key, prediction_key, reference_key = None, None, None if ( config.evaluators or any([isinstance(e, EvaluatorType) for e in config.evaluators]) or ( config.custom_evaluators and any([isinstance(e, StringEvaluator) for e in config.custom_evaluators]) ) ): input_key, prediction_key, reference_key = _get_keys( config, run_inputs, run_outputs, example_outputs ) for eval_config in config.evaluators: run_evaluator = _construct_run_evaluator( eval_config, config.eval_llm, run_type, data_type, example_outputs, reference_key, input_key, prediction_key, ) run_evaluators.append(run_evaluator) custom_evaluators = config.custom_evaluators or [] for custom_evaluator in custom_evaluators: if isinstance(custom_evaluator, RunEvaluator): run_evaluators.append(custom_evaluator) elif isinstance(custom_evaluator, StringEvaluator): run_evaluators.append( StringRunEvaluatorChain.from_run_and_data_type( custom_evaluator, run_type, data_type, input_key=input_key, prediction_key=prediction_key, reference_key=reference_key, ) ) else: raise ValueError( f"Unsupported custom evaluator: {custom_evaluator}." f" Expected RunEvaluator or StringEvaluator." ) return run_evaluators ### Async Helpers async def _arun_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], *, tags: Optional[List[str]] = None, callbacks: Callbacks = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """Asynchronously run the language model. Args: llm: The language model to run. inputs: The input dictionary. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map inputs to the expected format. Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): return await llm.apredict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): return await llm.apredict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format" f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: prompt = _get_prompt(inputs) llm_output: Union[str, BaseMessage] = await llm.apredict( prompt, callbacks=callbacks, tags=tags ) except InputFormatError: messages = _get_messages(inputs) llm_output = await llm.apredict_messages( messages, callbacks=callbacks, tags=tags ) return llm_output async def _arun_chain( chain: Union[Chain, Runnable], inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str]: """Run a chain asynchronously on inputs.""" inputs_ = inputs if input_mapper is None else input_mapper(inputs) if isinstance(chain, Chain): if isinstance(inputs_, dict) and len(inputs_) == 1: val = next(iter(inputs_.values())) output = await chain.acall(val, callbacks=callbacks, tags=tags) else: output = await chain.acall(inputs_, callbacks=callbacks, tags=tags) else: runnable_config = RunnableConfig(tags=tags or [], callbacks=callbacks) output = await chain.ainvoke(inputs_, config=runnable_config) return output async def _arun_llm_or_chain( example: Example, llm_or_chain_factory: MCF, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str, LLMResult, ChatResult]: """Asynchronously run the Chain or language model. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. input_mapper: Optional function to map the input to the expected format. Returns: A list of outputs. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) result = None try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = await _arun_llm( llm_or_chain_factory, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = await _arun_chain( chain, example.inputs, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) result = output except Exception as e: logger.warning(f"{chain_or_llm} failed for example {example.id}. Error: {e}") result = {"Error": str(e)} if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return result async def _gather_with_concurrency( n: int, initializer: Callable[[], Coroutine[Any, Any, Any]], *async_funcs: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ], ) -> List[Any]: """Run coroutines with a concurrency limit. Args: n: The maximum number of concurrent tasks. initializer: A coroutine that initializes shared resources for the tasks. async_funcs: The async_funcs to be run concurrently. Returns: A list of results from the coroutines. """ semaphore = asyncio.Semaphore(n) job_state = {"num_processed": 0} callback_queue: asyncio.Queue[Sequence[BaseCallbackHandler]] = asyncio.Queue() for _ in range(n): callback_queue.put_nowait(await initializer()) async def run_coroutine_with_semaphore( async_func: Callable[ [Sequence[BaseCallbackHandler], Dict], Coroutine[Any, Any, Any] ] ) -> Any: async with semaphore: callbacks = await callback_queue.get() try: result = await async_func(callbacks, job_state) finally: callback_queue.put_nowait(callbacks) return result results = await asyncio.gather( *(run_coroutine_with_semaphore(function) for function in async_funcs) ) while callback_queue: try: callbacks = callback_queue.get_nowait() except asyncio.QueueEmpty: break for callback in callbacks: if isinstance(callback, (LangChainTracer, EvaluatorCallbackHandler)): callback.wait_for_futures() return results async def _callbacks_initializer( project_name: Optional[str], client: Client, run_evaluators: Sequence[RunEvaluator], evaluation_handler_collector: List[EvaluatorCallbackHandler], ) -> List[BaseTracer]: """ Initialize a tracer to share across tasks. Args: project_name: The project name for the tracer. client: The client to use for the tracer. run_evaluators: The evaluators to run. evaluation_handler_collector: A list to collect the evaluators. Used to wait for the evaluators to finish. Returns: The callbacks for this thread. """ callbacks: List[BaseTracer] = [] if project_name: callbacks.append( LangChainTracer( project_name=project_name, client=client, use_threading=False ) ) if run_evaluators: callback = EvaluatorCallbackHandler( client=client, evaluators=run_evaluators, # We already have concurrency, don't want to overload the machine max_workers=1, ) callbacks.append(callback) evaluation_handler_collector.append(callback) return callbacks async def _arun_on_examples( client: Client, examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, data_type: DataType = DataType.kv, ) -> Dict[str, Any]: """ Asynchronously run the chain on examples and store traces to the specified project name. Args: client: LangSmith client to use to log feedback and runs. examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating concurrency_level: The number of async tasks to run concurrently. project_name: Project name to use when tracing runs. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. data_type: The dataset's data type. This is used to determine determine how to deserialize the reference data and model compatibility. Returns: A dictionary mapping example ids to the model outputs. """ wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory) project_name = _get_project_name(project_name, wrapped_model) run_evaluators, examples = _setup_evaluation( wrapped_model, examples, evaluation, data_type ) examples = _validate_example_inputs(examples, wrapped_model, input_mapper) results: Dict[str, dict] = {} async def process_example( example: Example, callbacks: List[BaseCallbackHandler], job_state: dict ) -> None: """Process a single example.""" result = await _arun_llm_or_chain( example, wrapped_model, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) results[str(example.id)] = {"output": result} job_state["num_processed"] += 1 if verbose: print( f"Processed examples: {job_state['num_processed']}", end="\r", flush=True, ) evaluation_handlers: List[EvaluatorCallbackHandler] = [] await _gather_with_concurrency( concurrency_level, functools.partial( _callbacks_initializer, project_name=project_name, client=client, evaluation_handler_collector=evaluation_handlers, run_evaluators=run_evaluators or [], ), *(functools.partial(process_example, e) for e in examples), ) all_feedback = {} for handler in evaluation_handlers: handler.wait_for_futures() all_feedback.update(handler.logged_feedback) # join the results and feedback on the example id for example_id, output_dict in results.items(): feedback = all_feedback.get(example_id, []) output_dict["feedback"] = feedback return results ## Sync Utilities def _run_llm( llm: BaseLanguageModel, inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[str, BaseMessage]: """ Run the language model on the example. Args: llm: The language model to run. inputs: The input dictionary. callbacks: The callbacks to use during the run. tags: Optional tags to add to the run. input_mapper: function to map to the inputs dictionary from an Example Returns: The LLMResult or ChatResult. Raises: ValueError: If the LLM type is unsupported. InputFormatError: If the input format is invalid. """ if input_mapper is not None: prompt_or_messages = input_mapper(inputs) if isinstance(prompt_or_messages, str): llm_output: Union[str, BaseMessage] = llm.predict( prompt_or_messages, callbacks=callbacks, tags=tags ) elif isinstance(prompt_or_messages, list) and all( isinstance(msg, BaseMessage) for msg in prompt_or_messages ): llm_output = llm.predict_messages( prompt_or_messages, callbacks=callbacks, tags=tags ) else: raise InputFormatError( "Input mapper returned invalid format: " f" {prompt_or_messages}" "\nExpected a single string or list of chat messages." ) else: try: llm_prompts = _get_prompt(inputs) llm_output = llm.predict(llm_prompts, callbacks=callbacks, tags=tags) except InputFormatError: llm_messages = _get_messages(inputs) llm_output = llm.predict_messages(llm_messages, callbacks=callbacks) return llm_output def _run_chain( chain: Union[Chain, Runnable], inputs: Dict[str, Any], callbacks: Callbacks, *, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[Dict, str]: """Run a chain on inputs.""" inputs_ = inputs if input_mapper is None else input_mapper(inputs) if isinstance(chain, Chain): if isinstance(inputs_, dict) and len(inputs_) == 1: val = next(iter(inputs_.values())) output = chain(val, callbacks=callbacks, tags=tags) else: output = chain(inputs_, callbacks=callbacks, tags=tags) else: runnable_config = RunnableConfig(tags=tags or [], callbacks=callbacks) output = chain.invoke(inputs_, config=runnable_config) return output def _run_llm_or_chain( example: Example, llm_or_chain_factory: MCF, *, tags: Optional[List[str]] = None, callbacks: Optional[List[BaseCallbackHandler]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, ) -> Union[dict, str, LLMResult, ChatResult]: """ Run the Chain or language model synchronously. Args: example: The example to run. llm_or_chain_factory: The Chain or language model constructor to run. tags: Optional tags to add to the run. callbacks: Optional callbacks to use during the run. Returns: Union[List[dict], List[str], List[LLMResult], List[ChatResult]]: The outputs of the model or chain. """ if callbacks: previous_example_ids = [ getattr(tracer, "example_id", None) for tracer in callbacks ] for tracer in callbacks: if hasattr(tracer, "example_id"): tracer.example_id = example.id else: previous_example_ids = None chain_or_llm = ( "LLM" if isinstance(llm_or_chain_factory, BaseLanguageModel) else "Chain" ) result = None try: if isinstance(llm_or_chain_factory, BaseLanguageModel): output: Any = _run_llm( llm_or_chain_factory, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) else: chain = llm_or_chain_factory() output = _run_chain( chain, example.inputs, callbacks, tags=tags, input_mapper=input_mapper, ) result = output except Exception as e: logger.warning( f"{chain_or_llm} failed for example {example.id} with inputs:" f" {example.inputs}.\nError: {e}", ) result = {"Error": str(e)} if callbacks and previous_example_ids: for example_id, tracer in zip(previous_example_ids, callbacks): if hasattr(tracer, "example_id"): tracer.example_id = example_id return result def _run_on_examples( client: Client, examples: Iterator[Example], llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, data_type: DataType = DataType.kv, ) -> Dict[str, Any]: """ Run the Chain or language model on examples and store traces to the specified project name. Args: client: LangSmith client to use to log feedback and runs. examples: Examples to run the model or chain over. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. data_type: The dataset's data type. This is used to determine determine how to deserialize the reference data and model compatibility. Returns: A dictionary mapping example ids to the model outputs. """ results: Dict[str, dict] = {} wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory) project_name = _get_project_name(project_name, wrapped_model) tracer = LangChainTracer( project_name=project_name, client=client, use_threading=False ) run_evaluators, examples = _setup_evaluation( wrapped_model, examples, evaluation, data_type ) examples = _validate_example_inputs(examples, wrapped_model, input_mapper) evaluation_handler = EvaluatorCallbackHandler( evaluators=run_evaluators or [], client=client, ) callbacks: List[BaseCallbackHandler] = [tracer, evaluation_handler] for i, example in enumerate(examples): result = _run_llm_or_chain( example, wrapped_model, tags=tags, callbacks=callbacks, input_mapper=input_mapper, ) if verbose: print(f"{i+1} processed", flush=True, end="\r") results[str(example.id)] = {"output": result} tracer.wait_for_futures() evaluation_handler.wait_for_futures() all_feedback = evaluation_handler.logged_feedback # join the results and feedback on the example id for example_id, output_dict in results.items(): feedback = all_feedback.get(example_id, []) output_dict["feedback"] = feedback return results ## Public API def _prepare_eval_run( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, project_name: Optional[str], ) -> Tuple[MCF, str, Dataset, Iterator[Example]]: wrapped_model = _wrap_in_chain_factory(llm_or_chain_factory, dataset_name) project_name = _get_project_name(project_name, wrapped_model) try: project = client.create_project(project_name) except ValueError as e: if "already exists " not in str(e): raise e raise ValueError( f"Project {project_name} already exists. Please use a different name." ) project_url = _get_eval_project_url(client.api_url, project.id) print( f"View the evaluation results for project '{project_name}' at:\n{project_url}" ) dataset = client.read_dataset(dataset_name=dataset_name) examples = client.list_examples(dataset_id=str(dataset.id)) return wrapped_model, project_name, dataset, examples async def arun_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, **kwargs: Any, ) -> Dict[str, Any]: """ Asynchronously run the Chain or language model on a dataset and store traces to the specified project name. Args: client: LangSmith client to use to read the dataset, and to log feedback and run traces. dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Optional evaluation configuration to use when evaluating concurrency_level: The number of async tasks to run concurrently. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. For the synchronous version, see :func:`run_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import RunEvalConfig, arun_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", RunEvalConfig.Criteria("helpfulness"), RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() await arun_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) await arun_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 if kwargs: warnings.warn( "The following arguments are deprecated and will " "be removed in a future release: " f"{kwargs.keys()}.", DeprecationWarning, ) wrapped_model, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name ) results = await _arun_on_examples( client, examples, wrapped_model, concurrency_level=concurrency_level, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) return TestResult( project_name=project_name, results=results, ) def _handle_coroutine(coro: Coroutine) -> Any: """ Handles a coroutine from a sync context. Args: coro (asyncio.coroutine): The coroutine to be handled. Returns: any: The result of the executed coroutine. """ # Check if there's a running event loop try: loop = asyncio.get_event_loop() except RuntimeError: # No event loop return asyncio.run(coro) if loop.is_running(): return loop.run_until_complete(coro) else: return asyncio.run(coro) def run_on_dataset( client: Client, dataset_name: str, llm_or_chain_factory: MODEL_OR_CHAIN_FACTORY, *, evaluation: Optional[RunEvalConfig] = None, concurrency_level: int = 5, project_name: Optional[str] = None, verbose: bool = False, tags: Optional[List[str]] = None, input_mapper: Optional[Callable[[Dict], Any]] = None, **kwargs: Any, ) -> Dict[str, Any]: """ Run the Chain or language model on a dataset and store traces to the specified project name. Args: client: LangSmith client to use to access the dataset and to log feedback and run traces. dataset_name: Name of the dataset to run the chain on. llm_or_chain_factory: Language model or Chain constructor to run over the dataset. The Chain constructor is used to permit independent calls on each example without carrying over state. evaluation: Configuration for evaluators to run on the results of the chain concurrency_level: The number of async tasks to run concurrently. project_name: Name of the project to store the traces in. Defaults to {dataset_name}-{chain class name}-{datetime}. verbose: Whether to print progress. tags: Tags to add to each run in the project. input_mapper: A function to map to the inputs dictionary from an Example to the format expected by the model to be evaluated. This is useful if your model needs to deserialize more complex schema or if your dataset has inputs with keys that differ from what is expected by your chain or agent. Returns: A dictionary containing the run's project name and the resulting model outputs. For the (usually faster) async version of this function, see :func:`arun_on_dataset`. Examples -------- .. code-block:: python from langsmith import Client from langchain.chat_models import ChatOpenAI from langchain.chains import LLMChain from langchain.smith import RunEvalConfig, run_on_dataset # Chains may have memory. Passing in a constructor function lets the # evaluation framework avoid cross-contamination between runs. def construct_chain(): llm = ChatOpenAI(temperature=0) chain = LLMChain.from_string( llm, "What's the answer to {your_input_key}" ) return chain # Load off-the-shelf evaluators via config or the EvaluatorType (string or enum) evaluation_config = RunEvalConfig( evaluators=[ "qa", # "Correctness" against a reference answer "embedding_distance", RunEvalConfig.Criteria("helpfulness"), RunEvalConfig.Criteria({ "fifth-grader-score": "Do you have to be smarter than a fifth grader to answer this question?" }), ] ) client = Client() run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) You can also create custom evaluators by subclassing the :class:`StringEvaluator <langchain.evaluation.schema.StringEvaluator>` or LangSmith's `RunEvaluator` classes. .. code-block:: python from typing import Optional from langchain.evaluation import StringEvaluator class MyStringEvaluator(StringEvaluator): @property def requires_input(self) -> bool: return False @property def requires_reference(self) -> bool: return True @property def evaluation_name(self) -> str: return "exact_match" def _evaluate_strings(self, prediction, reference=None, input=None, **kwargs) -> dict: return {"score": prediction == reference} evaluation_config = RunEvalConfig( custom_evaluators = [MyStringEvaluator()], ) run_on_dataset( client, "<my_dataset_name>", construct_chain, evaluation=evaluation_config, ) """ # noqa: E501 if kwargs: warnings.warn( "The following arguments are deprecated and " "will be removed in a future release: " f"{kwargs.keys()}.", DeprecationWarning, ) wrapped_model, project_name, dataset, examples = _prepare_eval_run( client, dataset_name, llm_or_chain_factory, project_name ) if concurrency_level in (0, 1): results = _run_on_examples( client, examples, wrapped_model, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) else: # TODO: Use runnables and the batch method coro = _arun_on_examples( client, examples, wrapped_model, concurrency_level=concurrency_level, project_name=project_name, verbose=verbose, tags=tags, evaluation=evaluation, input_mapper=input_mapper, data_type=dataset.data_type, ) results = _handle_coroutine(coro) return TestResult( project_name=project_name, results=results, )
[ "langchain.schema.messages.messages_from_dict", "langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type", "langchain.evaluation.schema.EvaluatorType", "langchain.callbacks.tracers.langchain.LangChainTracer", "langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler", "langchain.schema.runnable.RunnableLambda", "langchain.schema.runnable.RunnableConfig", "langchain.evaluation.loading.load_evaluator" ]
[((1500, 1527), 'logging.getLogger', 'logging.getLogger', (['__name__'], {}), '(__name__)\n', (1517, 1527), False, 'import logging\n'), ((2799, 2816), 'urllib.parse.urlparse', 'urlparse', (['api_url'], {}), '(api_url)\n', (2807, 2816), False, 'from urllib.parse import urlparse, urlunparse\n'), ((27519, 27539), 'asyncio.Semaphore', 'asyncio.Semaphore', (['n'], {}), '(n)\n', (27536, 27539), False, 'import asyncio\n'), ((27645, 27660), 'asyncio.Queue', 'asyncio.Queue', ([], {}), '()\n', (27658, 27660), False, 'import asyncio\n'), ((40275, 40353), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)'}), '(project_name=project_name, client=client, use_threading=False)\n', (40290, 40353), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((40583, 40655), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'evaluators': '(run_evaluators or [])', 'client': 'client'}), '(evaluators=run_evaluators or [], client=client)\n', (40607, 40655), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler\n'), ((2634, 2670), 'pandas.DataFrame', 'pd.DataFrame', (['records'], {'index': 'indices'}), '(records, index=indices)\n', (2646, 2670), True, 'import pandas as pd\n'), ((6729, 6765), 'itertools.chain', 'itertools.chain', (['[example]', 'examples'], {}), '([example], examples)\n', (6744, 6765), False, 'import itertools\n'), ((9608, 9643), 'langchain.schema.messages.messages_from_dict', 'messages_from_dict', (['raw_messages[0]'], {}), '(raw_messages[0])\n', (9626, 9643), False, 'from langchain.schema.messages import BaseMessage, messages_from_dict\n'), ((10444, 10456), 'uuid.uuid4', 'uuid.uuid4', ([], {}), '()\n', (10454, 10456), False, 'import uuid\n'), ((18358, 18399), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config'], {'llm': 'eval_llm'}), '(eval_config, llm=eval_llm)\n', (18372, 18399), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((18536, 18588), 'langchain.evaluation.loading.load_evaluator', 'load_evaluator', (['eval_config.evaluator_type'], {}), '(eval_config.evaluator_type, **kwargs)\n', (18550, 18588), False, 'from langchain.evaluation.loading import load_evaluator\n'), ((19034, 19224), 'langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type', 'StringRunEvaluatorChain.from_run_and_data_type', (['evaluator_', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key', 'tags': '[eval_type_tag]'}), '(evaluator_, run_type,\n data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key, tags=[eval_type_tag])\n', (19080, 19224), False, 'from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain\n'), ((24688, 24740), 'langchain.schema.runnable.RunnableConfig', 'RunnableConfig', ([], {'tags': '(tags or [])', 'callbacks': 'callbacks'}), '(tags=tags or [], callbacks=callbacks)\n', (24702, 24740), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((29506, 29591), 'langchain.callbacks.tracers.evaluation.EvaluatorCallbackHandler', 'EvaluatorCallbackHandler', ([], {'client': 'client', 'evaluators': 'run_evaluators', 'max_workers': '(1)'}), '(client=client, evaluators=run_evaluators,\n max_workers=1)\n', (29530, 29591), False, 'from langchain.callbacks.tracers.evaluation import EvaluatorCallbackHandler\n'), ((36030, 36082), 'langchain.schema.runnable.RunnableConfig', 'RunnableConfig', ([], {'tags': '(tags or [])', 'callbacks': 'callbacks'}), '(tags=tags or [], callbacks=callbacks)\n', (36044, 36082), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((47863, 47887), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (47885, 47887), False, 'import asyncio\n'), ((48059, 48076), 'asyncio.run', 'asyncio.run', (['coro'], {}), '(coro)\n', (48070, 48076), False, 'import asyncio\n'), ((18310, 18336), 'langchain.evaluation.schema.EvaluatorType', 'EvaluatorType', (['eval_config'], {}), '(eval_config)\n', (18323, 18336), False, 'from langchain.evaluation.schema import EvaluatorType, StringEvaluator\n'), ((29345, 29423), 'langchain.callbacks.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'project_name': 'project_name', 'client': 'client', 'use_threading': '(False)'}), '(project_name=project_name, client=client, use_threading=False)\n', (29360, 29423), False, 'from langchain.callbacks.tracers.langchain import LangChainTracer\n'), ((32818, 32997), 'functools.partial', 'functools.partial', (['_callbacks_initializer'], {'project_name': 'project_name', 'client': 'client', 'evaluation_handler_collector': 'evaluation_handlers', 'run_evaluators': '(run_evaluators or [])'}), '(_callbacks_initializer, project_name=project_name, client\n =client, evaluation_handler_collector=evaluation_handlers,\n run_evaluators=run_evaluators or [])\n', (32835, 32997), False, 'import functools\n'), ((47945, 47962), 'asyncio.run', 'asyncio.run', (['coro'], {}), '(coro)\n', (47956, 47962), False, 'import asyncio\n'), ((5873, 5909), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (5877, 5909), False, 'from typing import TYPE_CHECKING, Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, cast\n'), ((21499, 21673), 'langchain.smith.evaluation.string_run_evaluator.StringRunEvaluatorChain.from_run_and_data_type', 'StringRunEvaluatorChain.from_run_and_data_type', (['custom_evaluator', 'run_type', 'data_type'], {'input_key': 'input_key', 'prediction_key': 'prediction_key', 'reference_key': 'reference_key'}), '(custom_evaluator, run_type,\n data_type, input_key=input_key, prediction_key=prediction_key,\n reference_key=reference_key)\n', (21545, 21673), False, 'from langchain.smith.evaluation.string_run_evaluator import StringRunEvaluatorChain\n'), ((33071, 33108), 'functools.partial', 'functools.partial', (['process_example', 'e'], {}), '(process_example, e)\n', (33088, 33108), False, 'import functools\n'), ((5613, 5649), 'typing.cast', 'cast', (['Callable', 'llm_or_chain_factory'], {}), '(Callable, llm_or_chain_factory)\n', (5617, 5649), False, 'from typing import TYPE_CHECKING, Any, Callable, Coroutine, Dict, Iterator, List, Optional, Sequence, Tuple, Union, cast\n'), ((5668, 5696), 'inspect.signature', 'inspect.signature', (['user_func'], {}), '(user_func)\n', (5685, 5696), False, 'import inspect\n'), ((5790, 5815), 'langchain.schema.runnable.RunnableLambda', 'RunnableLambda', (['user_func'], {}), '(user_func)\n', (5804, 5815), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n'), ((6262, 6289), 'langchain.schema.runnable.RunnableLambda', 'RunnableLambda', (['constructor'], {}), '(constructor)\n', (6276, 6289), False, 'from langchain.schema.runnable import Runnable, RunnableConfig, RunnableLambda\n')]
"""Base interface that all chains should implement.""" import json from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import BaseModel, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import BaseMemory def _get_verbosity() -> bool: return langchain.verbose class Chain(BaseModel, ABC): """Base interface that all chains should implement.""" memory: Optional[BaseMemory] = None callback_manager: BaseCallbackManager = Field( default_factory=get_callback_manager, exclude=True ) verbose: bool = Field( default_factory=_get_verbosity ) # Whether to print the response text class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Input keys this chain expects.""" @property @abstractmethod def output_keys(self) -> List[str]: """Output keys this chain expects.""" def _validate_inputs(self, inputs: Dict[str, str]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, str]) -> None: if set(outputs) != set(self.output_keys): raise ValueError( f"Did not get output keys that were expected. " f"Got: {set(outputs)}. Expected: {set(self.output_keys)}." ) @abstractmethod def _call(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" async def _acall(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = self._call(inputs) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) if self.callback_manager.is_async: await self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) else: self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = await self._acall(inputs) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_chain_error(e, verbose=self.verbose) else: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_chain_end(outputs, verbose=self.verbose) else: self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prep outputs.""" self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prep inputs.""" if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs def apply(self, input_list: List[Dict[str, Any]]) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs) for inputs in input_list] def run(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0])[self.output_keys[0]] if kwargs and not args: return self(kwargs)[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return (await self.acall(args[0]))[self.output_keys[0]] if kwargs and not args: return (await self.acall(kwargs))[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain.""" if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict() _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml")
[ "langchain.callbacks.get_callback_manager" ]
[((646, 703), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager', 'exclude': '(True)'}), '(default_factory=get_callback_manager, exclude=True)\n', (651, 703), False, 'from pydantic import BaseModel, Field, validator\n'), ((738, 775), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (743, 775), False, 'from pydantic import BaseModel, Field, validator\n'), ((1075, 1127), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (1084, 1127), False, 'from pydantic import BaseModel, Field, validator\n'), ((1457, 1500), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (1466, 1500), False, 'from pydantic import BaseModel, Field, validator\n'), ((1428, 1450), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1448, 1450), False, 'from langchain.callbacks import get_callback_manager\n'), ((10078, 10093), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (10082, 10093), False, 'from pathlib import Path\n'), ((10413, 10447), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (10422, 10447), False, 'import json\n'), ((10550, 10600), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (10559, 10600), False, 'import yaml\n')]
"""Base interface that all chains should implement.""" import json from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import BaseModel, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import BaseMemory def _get_verbosity() -> bool: return langchain.verbose class Chain(BaseModel, ABC): """Base interface that all chains should implement.""" memory: Optional[BaseMemory] = None callback_manager: BaseCallbackManager = Field( default_factory=get_callback_manager, exclude=True ) verbose: bool = Field( default_factory=_get_verbosity ) # Whether to print the response text class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Input keys this chain expects.""" @property @abstractmethod def output_keys(self) -> List[str]: """Output keys this chain expects.""" def _validate_inputs(self, inputs: Dict[str, str]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, str]) -> None: if set(outputs) != set(self.output_keys): raise ValueError( f"Did not get output keys that were expected. " f"Got: {set(outputs)}. Expected: {set(self.output_keys)}." ) @abstractmethod def _call(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" async def _acall(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = self._call(inputs) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) if self.callback_manager.is_async: await self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) else: self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = await self._acall(inputs) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_chain_error(e, verbose=self.verbose) else: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_chain_end(outputs, verbose=self.verbose) else: self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prep outputs.""" self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prep inputs.""" if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs def apply(self, input_list: List[Dict[str, Any]]) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs) for inputs in input_list] def run(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0])[self.output_keys[0]] if kwargs and not args: return self(kwargs)[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return (await self.acall(args[0]))[self.output_keys[0]] if kwargs and not args: return (await self.acall(kwargs))[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain.""" if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict() _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml")
[ "langchain.callbacks.get_callback_manager" ]
[((646, 703), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager', 'exclude': '(True)'}), '(default_factory=get_callback_manager, exclude=True)\n', (651, 703), False, 'from pydantic import BaseModel, Field, validator\n'), ((738, 775), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (743, 775), False, 'from pydantic import BaseModel, Field, validator\n'), ((1075, 1127), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (1084, 1127), False, 'from pydantic import BaseModel, Field, validator\n'), ((1457, 1500), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (1466, 1500), False, 'from pydantic import BaseModel, Field, validator\n'), ((1428, 1450), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1448, 1450), False, 'from langchain.callbacks import get_callback_manager\n'), ((10078, 10093), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (10082, 10093), False, 'from pathlib import Path\n'), ((10413, 10447), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (10422, 10447), False, 'import json\n'), ((10550, 10600), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (10559, 10600), False, 'import yaml\n')]
"""Base interface that all chains should implement.""" import json from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import BaseModel, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import BaseMemory def _get_verbosity() -> bool: return langchain.verbose class Chain(BaseModel, ABC): """Base interface that all chains should implement.""" memory: Optional[BaseMemory] = None callback_manager: BaseCallbackManager = Field( default_factory=get_callback_manager, exclude=True ) verbose: bool = Field( default_factory=_get_verbosity ) # Whether to print the response text class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Input keys this chain expects.""" @property @abstractmethod def output_keys(self) -> List[str]: """Output keys this chain expects.""" def _validate_inputs(self, inputs: Dict[str, str]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, str]) -> None: if set(outputs) != set(self.output_keys): raise ValueError( f"Did not get output keys that were expected. " f"Got: {set(outputs)}. Expected: {set(self.output_keys)}." ) @abstractmethod def _call(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" async def _acall(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = self._call(inputs) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) if self.callback_manager.is_async: await self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) else: self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = await self._acall(inputs) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_chain_error(e, verbose=self.verbose) else: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_chain_end(outputs, verbose=self.verbose) else: self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prep outputs.""" self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prep inputs.""" if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs def apply(self, input_list: List[Dict[str, Any]]) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs) for inputs in input_list] def run(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0])[self.output_keys[0]] if kwargs and not args: return self(kwargs)[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return (await self.acall(args[0]))[self.output_keys[0]] if kwargs and not args: return (await self.acall(kwargs))[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain.""" if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict() _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml")
[ "langchain.callbacks.get_callback_manager" ]
[((646, 703), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager', 'exclude': '(True)'}), '(default_factory=get_callback_manager, exclude=True)\n', (651, 703), False, 'from pydantic import BaseModel, Field, validator\n'), ((738, 775), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (743, 775), False, 'from pydantic import BaseModel, Field, validator\n'), ((1075, 1127), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (1084, 1127), False, 'from pydantic import BaseModel, Field, validator\n'), ((1457, 1500), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (1466, 1500), False, 'from pydantic import BaseModel, Field, validator\n'), ((1428, 1450), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1448, 1450), False, 'from langchain.callbacks import get_callback_manager\n'), ((10078, 10093), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (10082, 10093), False, 'from pathlib import Path\n'), ((10413, 10447), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (10422, 10447), False, 'import json\n'), ((10550, 10600), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (10559, 10600), False, 'import yaml\n')]
"""Base interface that all chains should implement.""" import json from abc import ABC, abstractmethod from pathlib import Path from typing import Any, Dict, List, Optional, Union import yaml from pydantic import BaseModel, Field, validator import langchain from langchain.callbacks import get_callback_manager from langchain.callbacks.base import BaseCallbackManager from langchain.schema import BaseMemory def _get_verbosity() -> bool: return langchain.verbose class Chain(BaseModel, ABC): """Base interface that all chains should implement.""" memory: Optional[BaseMemory] = None callback_manager: BaseCallbackManager = Field( default_factory=get_callback_manager, exclude=True ) verbose: bool = Field( default_factory=_get_verbosity ) # Whether to print the response text class Config: """Configuration for this pydantic object.""" arbitrary_types_allowed = True @property def _chain_type(self) -> str: raise NotImplementedError("Saving not supported for this chain type.") @validator("callback_manager", pre=True, always=True) def set_callback_manager( cls, callback_manager: Optional[BaseCallbackManager] ) -> BaseCallbackManager: """If callback manager is None, set it. This allows users to pass in None as callback manager, which is a nice UX. """ return callback_manager or get_callback_manager() @validator("verbose", pre=True, always=True) def set_verbose(cls, verbose: Optional[bool]) -> bool: """If verbose is None, set it. This allows users to pass in None as verbose to access the global setting. """ if verbose is None: return _get_verbosity() else: return verbose @property @abstractmethod def input_keys(self) -> List[str]: """Input keys this chain expects.""" @property @abstractmethod def output_keys(self) -> List[str]: """Output keys this chain expects.""" def _validate_inputs(self, inputs: Dict[str, str]) -> None: """Check that all inputs are present.""" missing_keys = set(self.input_keys).difference(inputs) if missing_keys: raise ValueError(f"Missing some input keys: {missing_keys}") def _validate_outputs(self, outputs: Dict[str, str]) -> None: if set(outputs) != set(self.output_keys): raise ValueError( f"Did not get output keys that were expected. " f"Got: {set(outputs)}. Expected: {set(self.output_keys)}." ) @abstractmethod def _call(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" async def _acall(self, inputs: Dict[str, str]) -> Dict[str, str]: """Run the logic of this chain and return the output.""" raise NotImplementedError("Async call not supported for this chain type.") def __call__( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = self._call(inputs) except (KeyboardInterrupt, Exception) as e: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) async def acall( self, inputs: Union[Dict[str, Any], Any], return_only_outputs: bool = False ) -> Dict[str, Any]: """Run the logic of this chain and add to output if desired. Args: inputs: Dictionary of inputs, or single input if chain expects only one param. return_only_outputs: boolean for whether to return only outputs in the response. If True, only new keys generated by this chain will be returned. If False, both input keys and new keys generated by this chain will be returned. Defaults to False. """ inputs = self.prep_inputs(inputs) if self.callback_manager.is_async: await self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) else: self.callback_manager.on_chain_start( {"name": self.__class__.__name__}, inputs, verbose=self.verbose, ) try: outputs = await self._acall(inputs) except (KeyboardInterrupt, Exception) as e: if self.callback_manager.is_async: await self.callback_manager.on_chain_error(e, verbose=self.verbose) else: self.callback_manager.on_chain_error(e, verbose=self.verbose) raise e if self.callback_manager.is_async: await self.callback_manager.on_chain_end(outputs, verbose=self.verbose) else: self.callback_manager.on_chain_end(outputs, verbose=self.verbose) return self.prep_outputs(inputs, outputs, return_only_outputs) def prep_outputs( self, inputs: Dict[str, str], outputs: Dict[str, str], return_only_outputs: bool = False, ) -> Dict[str, str]: """Validate and prep outputs.""" self._validate_outputs(outputs) if self.memory is not None: self.memory.save_context(inputs, outputs) if return_only_outputs: return outputs else: return {**inputs, **outputs} def prep_inputs(self, inputs: Union[Dict[str, Any], Any]) -> Dict[str, str]: """Validate and prep inputs.""" if not isinstance(inputs, dict): _input_keys = set(self.input_keys) if self.memory is not None: # If there are multiple input keys, but some get set by memory so that # only one is not set, we can still figure out which key it is. _input_keys = _input_keys.difference(self.memory.memory_variables) if len(_input_keys) != 1: raise ValueError( f"A single string input was passed in, but this chain expects " f"multiple inputs ({_input_keys}). When a chain expects " f"multiple inputs, please call it by passing in a dictionary, " "eg `chain({'foo': 1, 'bar': 2})`" ) inputs = {list(_input_keys)[0]: inputs} if self.memory is not None: external_context = self.memory.load_memory_variables(inputs) inputs = dict(inputs, **external_context) self._validate_inputs(inputs) return inputs def apply(self, input_list: List[Dict[str, Any]]) -> List[Dict[str, str]]: """Call the chain on all inputs in the list.""" return [self(inputs) for inputs in input_list] def run(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return self(args[0])[self.output_keys[0]] if kwargs and not args: return self(kwargs)[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) async def arun(self, *args: Any, **kwargs: Any) -> str: """Run the chain as text in, text out or multiple variables, text out.""" if len(self.output_keys) != 1: raise ValueError( f"`run` not supported when there is not exactly " f"one output key. Got {self.output_keys}." ) if args and not kwargs: if len(args) != 1: raise ValueError("`run` supports only one positional argument.") return (await self.acall(args[0]))[self.output_keys[0]] if kwargs and not args: return (await self.acall(kwargs))[self.output_keys[0]] raise ValueError( f"`run` supported with either positional arguments or keyword arguments" f" but not both. Got args: {args} and kwargs: {kwargs}." ) def dict(self, **kwargs: Any) -> Dict: """Return dictionary representation of chain.""" if self.memory is not None: raise ValueError("Saving of memory is not yet supported.") _dict = super().dict() _dict["_type"] = self._chain_type return _dict def save(self, file_path: Union[Path, str]) -> None: """Save the chain. Args: file_path: Path to file to save the chain to. Example: .. code-block:: python chain.save(file_path="path/chain.yaml") """ # Convert file to Path object. if isinstance(file_path, str): save_path = Path(file_path) else: save_path = file_path directory_path = save_path.parent directory_path.mkdir(parents=True, exist_ok=True) # Fetch dictionary to save chain_dict = self.dict() if save_path.suffix == ".json": with open(file_path, "w") as f: json.dump(chain_dict, f, indent=4) elif save_path.suffix == ".yaml": with open(file_path, "w") as f: yaml.dump(chain_dict, f, default_flow_style=False) else: raise ValueError(f"{save_path} must be json or yaml")
[ "langchain.callbacks.get_callback_manager" ]
[((646, 703), 'pydantic.Field', 'Field', ([], {'default_factory': 'get_callback_manager', 'exclude': '(True)'}), '(default_factory=get_callback_manager, exclude=True)\n', (651, 703), False, 'from pydantic import BaseModel, Field, validator\n'), ((738, 775), 'pydantic.Field', 'Field', ([], {'default_factory': '_get_verbosity'}), '(default_factory=_get_verbosity)\n', (743, 775), False, 'from pydantic import BaseModel, Field, validator\n'), ((1075, 1127), 'pydantic.validator', 'validator', (['"""callback_manager"""'], {'pre': '(True)', 'always': '(True)'}), "('callback_manager', pre=True, always=True)\n", (1084, 1127), False, 'from pydantic import BaseModel, Field, validator\n'), ((1457, 1500), 'pydantic.validator', 'validator', (['"""verbose"""'], {'pre': '(True)', 'always': '(True)'}), "('verbose', pre=True, always=True)\n", (1466, 1500), False, 'from pydantic import BaseModel, Field, validator\n'), ((1428, 1450), 'langchain.callbacks.get_callback_manager', 'get_callback_manager', ([], {}), '()\n', (1448, 1450), False, 'from langchain.callbacks import get_callback_manager\n'), ((10078, 10093), 'pathlib.Path', 'Path', (['file_path'], {}), '(file_path)\n', (10082, 10093), False, 'from pathlib import Path\n'), ((10413, 10447), 'json.dump', 'json.dump', (['chain_dict', 'f'], {'indent': '(4)'}), '(chain_dict, f, indent=4)\n', (10422, 10447), False, 'import json\n'), ((10550, 10600), 'yaml.dump', 'yaml.dump', (['chain_dict', 'f'], {'default_flow_style': '(False)'}), '(chain_dict, f, default_flow_style=False)\n', (10559, 10600), False, 'import yaml\n')]
from typing import Dict, List, Optional from langchain.agents.load_tools import ( _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS, ) from langflow.custom import customs from langflow.interface.base import LangChainTypeCreator from langflow.interface.tools.constants import ( ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS, ) from langflow.interface.tools.util import get_tool_params from langflow.services.utils import get_settings_manager from langflow.template.field.base import TemplateField from langflow.template.template.base import Template from langflow.utils import util from langflow.utils.util import build_template_from_class TOOL_INPUTS = { "str": TemplateField( field_type="str", required=True, is_list=False, show=True, placeholder="", value="", ), "llm": TemplateField( field_type="BaseLanguageModel", required=True, is_list=False, show=True ), "func": TemplateField( field_type="function", required=True, is_list=False, show=True, multiline=True, ), "code": TemplateField( field_type="str", required=True, is_list=False, show=True, value="", multiline=True, ), "path": TemplateField( field_type="file", required=True, is_list=False, show=True, value="", suffixes=[".json", ".yaml", ".yml"], file_types=["json", "yaml", "yml"], ), } class ToolCreator(LangChainTypeCreator): type_name: str = "tools" tools_dict: Optional[Dict] = None @property def type_to_loader_dict(self) -> Dict: settings_manager = get_settings_manager() if self.tools_dict is None: all_tools = {} for tool, tool_fcn in ALL_TOOLS_NAMES.items(): tool_params = get_tool_params(tool_fcn) tool_name = tool_params.get("name") or tool if ( tool_name in settings_manager.settings.TOOLS or settings_manager.settings.DEV ): if tool_name == "JsonSpec": tool_params["path"] = tool_params.pop("dict_") # type: ignore all_tools[tool_name] = { "type": tool, "params": tool_params, "fcn": tool_fcn, } self.tools_dict = all_tools return self.tools_dict def get_signature(self, name: str) -> Optional[Dict]: """Get the signature of a tool.""" base_classes = ["Tool", "BaseTool"] fields = [] params = [] tool_params = {} # Raise error if name is not in tools if name not in self.type_to_loader_dict.keys(): raise ValueError("Tool not found") tool_type: str = self.type_to_loader_dict[name]["type"] # type: ignore # if tool_type in _BASE_TOOLS.keys(): # params = [] if tool_type in _LLM_TOOLS.keys(): params = ["llm"] elif tool_type in _EXTRA_LLM_TOOLS.keys(): extra_keys = _EXTRA_LLM_TOOLS[tool_type][1] params = ["llm"] + extra_keys elif tool_type in _EXTRA_OPTIONAL_TOOLS.keys(): extra_keys = _EXTRA_OPTIONAL_TOOLS[tool_type][1] params = extra_keys # elif tool_type == "Tool": # params = ["name", "description", "func"] elif tool_type in CUSTOM_TOOLS: # Get custom tool params params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes = ["function"] if node := customs.get_custom_nodes("tools").get(tool_type): return node elif tool_type in FILE_TOOLS: params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes += [name] elif tool_type in OTHER_TOOLS: tool_dict = build_template_from_class(tool_type, OTHER_TOOLS) fields = tool_dict["template"] # Pop unnecessary fields and add name fields.pop("_type") # type: ignore fields.pop("return_direct") # type: ignore fields.pop("verbose") # type: ignore tool_params = { "name": fields.pop("name")["value"], # type: ignore "description": fields.pop("description")["value"], # type: ignore } fields = [ TemplateField(name=name, field_type=field["type"], **field) for name, field in fields.items() # type: ignore ] base_classes += tool_dict["base_classes"] # Copy the field and add the name for param in params: field = TOOL_INPUTS.get(param, TOOL_INPUTS["str"]).copy() field.name = param field.advanced = False if param == "aiosession": field.show = False field.required = False fields.append(field) template = Template(fields=fields, type_name=tool_type) tool_params = {**tool_params, **self.type_to_loader_dict[name]["params"]} return { "template": util.format_dict(template.to_dict()), **tool_params, "base_classes": base_classes, } def to_list(self) -> List[str]: """List all load tools""" return list(self.type_to_loader_dict.keys()) tool_creator = ToolCreator()
[ "langchain.agents.load_tools._LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys" ]
[((709, 811), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'placeholder': '""""""', 'value': '""""""'}), "(field_type='str', required=True, is_list=False, show=True,\n placeholder='', value='')\n", (722, 811), False, 'from langflow.template.field.base import TemplateField\n'), ((875, 965), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""BaseLanguageModel"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)'}), "(field_type='BaseLanguageModel', required=True, is_list=False,\n show=True)\n", (888, 965), False, 'from langflow.template.field.base import TemplateField\n'), ((989, 1087), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""function"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'multiline': '(True)'}), "(field_type='function', required=True, is_list=False, show=\n True, multiline=True)\n", (1002, 1087), False, 'from langflow.template.field.base import TemplateField\n'), ((1143, 1245), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'multiline': '(True)'}), "(field_type='str', required=True, is_list=False, show=True,\n value='', multiline=True)\n", (1156, 1245), False, 'from langflow.template.field.base import TemplateField\n'), ((1310, 1474), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""file"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'suffixes': "['.json', '.yaml', '.yml']", 'file_types': "['json', 'yaml', 'yml']"}), "(field_type='file', required=True, is_list=False, show=True,\n value='', suffixes=['.json', '.yaml', '.yml'], file_types=['json',\n 'yaml', 'yml'])\n", (1323, 1474), False, 'from langflow.template.field.base import TemplateField\n'), ((1728, 1750), 'langflow.services.utils.get_settings_manager', 'get_settings_manager', ([], {}), '()\n', (1748, 1750), False, 'from langflow.services.utils import get_settings_manager\n'), ((5139, 5183), 'langflow.template.template.base.Template', 'Template', ([], {'fields': 'fields', 'type_name': 'tool_type'}), '(fields=fields, type_name=tool_type)\n', (5147, 5183), False, 'from langflow.template.template.base import Template\n'), ((1849, 1872), 'langflow.interface.tools.constants.ALL_TOOLS_NAMES.items', 'ALL_TOOLS_NAMES.items', ([], {}), '()\n', (1870, 1872), False, 'from langflow.interface.tools.constants import ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS\n'), ((3091, 3108), 'langchain.agents.load_tools._LLM_TOOLS.keys', '_LLM_TOOLS.keys', ([], {}), '()\n', (3106, 3108), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((1904, 1929), 'langflow.interface.tools.util.get_tool_params', 'get_tool_params', (['tool_fcn'], {}), '(tool_fcn)\n', (1919, 1929), False, 'from langflow.interface.tools.util import get_tool_params\n'), ((3165, 3188), 'langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys', '_EXTRA_LLM_TOOLS.keys', ([], {}), '()\n', (3186, 3188), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((3314, 3342), 'langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys', '_EXTRA_OPTIONAL_TOOLS.keys', ([], {}), '()\n', (3340, 3342), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((4038, 4087), 'langflow.utils.util.build_template_from_class', 'build_template_from_class', (['tool_type', 'OTHER_TOOLS'], {}), '(tool_type, OTHER_TOOLS)\n', (4063, 4087), False, 'from langflow.utils.util import build_template_from_class\n'), ((3746, 3779), 'langflow.custom.customs.get_custom_nodes', 'customs.get_custom_nodes', (['"""tools"""'], {}), "('tools')\n", (3770, 3779), False, 'from langflow.custom import customs\n'), ((4571, 4630), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'name': 'name', 'field_type': "field['type']"}), "(name=name, field_type=field['type'], **field)\n", (4584, 4630), False, 'from langflow.template.field.base import TemplateField\n')]
from typing import Dict, List, Optional from langchain.agents.load_tools import ( _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS, ) from langflow.custom import customs from langflow.interface.base import LangChainTypeCreator from langflow.interface.tools.constants import ( ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS, ) from langflow.interface.tools.util import get_tool_params from langflow.services.utils import get_settings_manager from langflow.template.field.base import TemplateField from langflow.template.template.base import Template from langflow.utils import util from langflow.utils.util import build_template_from_class TOOL_INPUTS = { "str": TemplateField( field_type="str", required=True, is_list=False, show=True, placeholder="", value="", ), "llm": TemplateField( field_type="BaseLanguageModel", required=True, is_list=False, show=True ), "func": TemplateField( field_type="function", required=True, is_list=False, show=True, multiline=True, ), "code": TemplateField( field_type="str", required=True, is_list=False, show=True, value="", multiline=True, ), "path": TemplateField( field_type="file", required=True, is_list=False, show=True, value="", suffixes=[".json", ".yaml", ".yml"], file_types=["json", "yaml", "yml"], ), } class ToolCreator(LangChainTypeCreator): type_name: str = "tools" tools_dict: Optional[Dict] = None @property def type_to_loader_dict(self) -> Dict: settings_manager = get_settings_manager() if self.tools_dict is None: all_tools = {} for tool, tool_fcn in ALL_TOOLS_NAMES.items(): tool_params = get_tool_params(tool_fcn) tool_name = tool_params.get("name") or tool if ( tool_name in settings_manager.settings.TOOLS or settings_manager.settings.DEV ): if tool_name == "JsonSpec": tool_params["path"] = tool_params.pop("dict_") # type: ignore all_tools[tool_name] = { "type": tool, "params": tool_params, "fcn": tool_fcn, } self.tools_dict = all_tools return self.tools_dict def get_signature(self, name: str) -> Optional[Dict]: """Get the signature of a tool.""" base_classes = ["Tool", "BaseTool"] fields = [] params = [] tool_params = {} # Raise error if name is not in tools if name not in self.type_to_loader_dict.keys(): raise ValueError("Tool not found") tool_type: str = self.type_to_loader_dict[name]["type"] # type: ignore # if tool_type in _BASE_TOOLS.keys(): # params = [] if tool_type in _LLM_TOOLS.keys(): params = ["llm"] elif tool_type in _EXTRA_LLM_TOOLS.keys(): extra_keys = _EXTRA_LLM_TOOLS[tool_type][1] params = ["llm"] + extra_keys elif tool_type in _EXTRA_OPTIONAL_TOOLS.keys(): extra_keys = _EXTRA_OPTIONAL_TOOLS[tool_type][1] params = extra_keys # elif tool_type == "Tool": # params = ["name", "description", "func"] elif tool_type in CUSTOM_TOOLS: # Get custom tool params params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes = ["function"] if node := customs.get_custom_nodes("tools").get(tool_type): return node elif tool_type in FILE_TOOLS: params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes += [name] elif tool_type in OTHER_TOOLS: tool_dict = build_template_from_class(tool_type, OTHER_TOOLS) fields = tool_dict["template"] # Pop unnecessary fields and add name fields.pop("_type") # type: ignore fields.pop("return_direct") # type: ignore fields.pop("verbose") # type: ignore tool_params = { "name": fields.pop("name")["value"], # type: ignore "description": fields.pop("description")["value"], # type: ignore } fields = [ TemplateField(name=name, field_type=field["type"], **field) for name, field in fields.items() # type: ignore ] base_classes += tool_dict["base_classes"] # Copy the field and add the name for param in params: field = TOOL_INPUTS.get(param, TOOL_INPUTS["str"]).copy() field.name = param field.advanced = False if param == "aiosession": field.show = False field.required = False fields.append(field) template = Template(fields=fields, type_name=tool_type) tool_params = {**tool_params, **self.type_to_loader_dict[name]["params"]} return { "template": util.format_dict(template.to_dict()), **tool_params, "base_classes": base_classes, } def to_list(self) -> List[str]: """List all load tools""" return list(self.type_to_loader_dict.keys()) tool_creator = ToolCreator()
[ "langchain.agents.load_tools._LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys" ]
[((709, 811), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'placeholder': '""""""', 'value': '""""""'}), "(field_type='str', required=True, is_list=False, show=True,\n placeholder='', value='')\n", (722, 811), False, 'from langflow.template.field.base import TemplateField\n'), ((875, 965), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""BaseLanguageModel"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)'}), "(field_type='BaseLanguageModel', required=True, is_list=False,\n show=True)\n", (888, 965), False, 'from langflow.template.field.base import TemplateField\n'), ((989, 1087), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""function"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'multiline': '(True)'}), "(field_type='function', required=True, is_list=False, show=\n True, multiline=True)\n", (1002, 1087), False, 'from langflow.template.field.base import TemplateField\n'), ((1143, 1245), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'multiline': '(True)'}), "(field_type='str', required=True, is_list=False, show=True,\n value='', multiline=True)\n", (1156, 1245), False, 'from langflow.template.field.base import TemplateField\n'), ((1310, 1474), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""file"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'suffixes': "['.json', '.yaml', '.yml']", 'file_types': "['json', 'yaml', 'yml']"}), "(field_type='file', required=True, is_list=False, show=True,\n value='', suffixes=['.json', '.yaml', '.yml'], file_types=['json',\n 'yaml', 'yml'])\n", (1323, 1474), False, 'from langflow.template.field.base import TemplateField\n'), ((1728, 1750), 'langflow.services.utils.get_settings_manager', 'get_settings_manager', ([], {}), '()\n', (1748, 1750), False, 'from langflow.services.utils import get_settings_manager\n'), ((5139, 5183), 'langflow.template.template.base.Template', 'Template', ([], {'fields': 'fields', 'type_name': 'tool_type'}), '(fields=fields, type_name=tool_type)\n', (5147, 5183), False, 'from langflow.template.template.base import Template\n'), ((1849, 1872), 'langflow.interface.tools.constants.ALL_TOOLS_NAMES.items', 'ALL_TOOLS_NAMES.items', ([], {}), '()\n', (1870, 1872), False, 'from langflow.interface.tools.constants import ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS\n'), ((3091, 3108), 'langchain.agents.load_tools._LLM_TOOLS.keys', '_LLM_TOOLS.keys', ([], {}), '()\n', (3106, 3108), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((1904, 1929), 'langflow.interface.tools.util.get_tool_params', 'get_tool_params', (['tool_fcn'], {}), '(tool_fcn)\n', (1919, 1929), False, 'from langflow.interface.tools.util import get_tool_params\n'), ((3165, 3188), 'langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys', '_EXTRA_LLM_TOOLS.keys', ([], {}), '()\n', (3186, 3188), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((3314, 3342), 'langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys', '_EXTRA_OPTIONAL_TOOLS.keys', ([], {}), '()\n', (3340, 3342), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((4038, 4087), 'langflow.utils.util.build_template_from_class', 'build_template_from_class', (['tool_type', 'OTHER_TOOLS'], {}), '(tool_type, OTHER_TOOLS)\n', (4063, 4087), False, 'from langflow.utils.util import build_template_from_class\n'), ((3746, 3779), 'langflow.custom.customs.get_custom_nodes', 'customs.get_custom_nodes', (['"""tools"""'], {}), "('tools')\n", (3770, 3779), False, 'from langflow.custom import customs\n'), ((4571, 4630), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'name': 'name', 'field_type': "field['type']"}), "(name=name, field_type=field['type'], **field)\n", (4584, 4630), False, 'from langflow.template.field.base import TemplateField\n')]
from typing import Dict, List, Optional from langchain.agents.load_tools import ( _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS, ) from langflow.custom import customs from langflow.interface.base import LangChainTypeCreator from langflow.interface.tools.constants import ( ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS, ) from langflow.interface.tools.util import get_tool_params from langflow.services.utils import get_settings_manager from langflow.template.field.base import TemplateField from langflow.template.template.base import Template from langflow.utils import util from langflow.utils.util import build_template_from_class TOOL_INPUTS = { "str": TemplateField( field_type="str", required=True, is_list=False, show=True, placeholder="", value="", ), "llm": TemplateField( field_type="BaseLanguageModel", required=True, is_list=False, show=True ), "func": TemplateField( field_type="function", required=True, is_list=False, show=True, multiline=True, ), "code": TemplateField( field_type="str", required=True, is_list=False, show=True, value="", multiline=True, ), "path": TemplateField( field_type="file", required=True, is_list=False, show=True, value="", suffixes=[".json", ".yaml", ".yml"], file_types=["json", "yaml", "yml"], ), } class ToolCreator(LangChainTypeCreator): type_name: str = "tools" tools_dict: Optional[Dict] = None @property def type_to_loader_dict(self) -> Dict: settings_manager = get_settings_manager() if self.tools_dict is None: all_tools = {} for tool, tool_fcn in ALL_TOOLS_NAMES.items(): tool_params = get_tool_params(tool_fcn) tool_name = tool_params.get("name") or tool if ( tool_name in settings_manager.settings.TOOLS or settings_manager.settings.DEV ): if tool_name == "JsonSpec": tool_params["path"] = tool_params.pop("dict_") # type: ignore all_tools[tool_name] = { "type": tool, "params": tool_params, "fcn": tool_fcn, } self.tools_dict = all_tools return self.tools_dict def get_signature(self, name: str) -> Optional[Dict]: """Get the signature of a tool.""" base_classes = ["Tool", "BaseTool"] fields = [] params = [] tool_params = {} # Raise error if name is not in tools if name not in self.type_to_loader_dict.keys(): raise ValueError("Tool not found") tool_type: str = self.type_to_loader_dict[name]["type"] # type: ignore # if tool_type in _BASE_TOOLS.keys(): # params = [] if tool_type in _LLM_TOOLS.keys(): params = ["llm"] elif tool_type in _EXTRA_LLM_TOOLS.keys(): extra_keys = _EXTRA_LLM_TOOLS[tool_type][1] params = ["llm"] + extra_keys elif tool_type in _EXTRA_OPTIONAL_TOOLS.keys(): extra_keys = _EXTRA_OPTIONAL_TOOLS[tool_type][1] params = extra_keys # elif tool_type == "Tool": # params = ["name", "description", "func"] elif tool_type in CUSTOM_TOOLS: # Get custom tool params params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes = ["function"] if node := customs.get_custom_nodes("tools").get(tool_type): return node elif tool_type in FILE_TOOLS: params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes += [name] elif tool_type in OTHER_TOOLS: tool_dict = build_template_from_class(tool_type, OTHER_TOOLS) fields = tool_dict["template"] # Pop unnecessary fields and add name fields.pop("_type") # type: ignore fields.pop("return_direct") # type: ignore fields.pop("verbose") # type: ignore tool_params = { "name": fields.pop("name")["value"], # type: ignore "description": fields.pop("description")["value"], # type: ignore } fields = [ TemplateField(name=name, field_type=field["type"], **field) for name, field in fields.items() # type: ignore ] base_classes += tool_dict["base_classes"] # Copy the field and add the name for param in params: field = TOOL_INPUTS.get(param, TOOL_INPUTS["str"]).copy() field.name = param field.advanced = False if param == "aiosession": field.show = False field.required = False fields.append(field) template = Template(fields=fields, type_name=tool_type) tool_params = {**tool_params, **self.type_to_loader_dict[name]["params"]} return { "template": util.format_dict(template.to_dict()), **tool_params, "base_classes": base_classes, } def to_list(self) -> List[str]: """List all load tools""" return list(self.type_to_loader_dict.keys()) tool_creator = ToolCreator()
[ "langchain.agents.load_tools._LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys" ]
[((709, 811), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'placeholder': '""""""', 'value': '""""""'}), "(field_type='str', required=True, is_list=False, show=True,\n placeholder='', value='')\n", (722, 811), False, 'from langflow.template.field.base import TemplateField\n'), ((875, 965), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""BaseLanguageModel"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)'}), "(field_type='BaseLanguageModel', required=True, is_list=False,\n show=True)\n", (888, 965), False, 'from langflow.template.field.base import TemplateField\n'), ((989, 1087), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""function"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'multiline': '(True)'}), "(field_type='function', required=True, is_list=False, show=\n True, multiline=True)\n", (1002, 1087), False, 'from langflow.template.field.base import TemplateField\n'), ((1143, 1245), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'multiline': '(True)'}), "(field_type='str', required=True, is_list=False, show=True,\n value='', multiline=True)\n", (1156, 1245), False, 'from langflow.template.field.base import TemplateField\n'), ((1310, 1474), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""file"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'suffixes': "['.json', '.yaml', '.yml']", 'file_types': "['json', 'yaml', 'yml']"}), "(field_type='file', required=True, is_list=False, show=True,\n value='', suffixes=['.json', '.yaml', '.yml'], file_types=['json',\n 'yaml', 'yml'])\n", (1323, 1474), False, 'from langflow.template.field.base import TemplateField\n'), ((1728, 1750), 'langflow.services.utils.get_settings_manager', 'get_settings_manager', ([], {}), '()\n', (1748, 1750), False, 'from langflow.services.utils import get_settings_manager\n'), ((5139, 5183), 'langflow.template.template.base.Template', 'Template', ([], {'fields': 'fields', 'type_name': 'tool_type'}), '(fields=fields, type_name=tool_type)\n', (5147, 5183), False, 'from langflow.template.template.base import Template\n'), ((1849, 1872), 'langflow.interface.tools.constants.ALL_TOOLS_NAMES.items', 'ALL_TOOLS_NAMES.items', ([], {}), '()\n', (1870, 1872), False, 'from langflow.interface.tools.constants import ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS\n'), ((3091, 3108), 'langchain.agents.load_tools._LLM_TOOLS.keys', '_LLM_TOOLS.keys', ([], {}), '()\n', (3106, 3108), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((1904, 1929), 'langflow.interface.tools.util.get_tool_params', 'get_tool_params', (['tool_fcn'], {}), '(tool_fcn)\n', (1919, 1929), False, 'from langflow.interface.tools.util import get_tool_params\n'), ((3165, 3188), 'langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys', '_EXTRA_LLM_TOOLS.keys', ([], {}), '()\n', (3186, 3188), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((3314, 3342), 'langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys', '_EXTRA_OPTIONAL_TOOLS.keys', ([], {}), '()\n', (3340, 3342), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((4038, 4087), 'langflow.utils.util.build_template_from_class', 'build_template_from_class', (['tool_type', 'OTHER_TOOLS'], {}), '(tool_type, OTHER_TOOLS)\n', (4063, 4087), False, 'from langflow.utils.util import build_template_from_class\n'), ((3746, 3779), 'langflow.custom.customs.get_custom_nodes', 'customs.get_custom_nodes', (['"""tools"""'], {}), "('tools')\n", (3770, 3779), False, 'from langflow.custom import customs\n'), ((4571, 4630), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'name': 'name', 'field_type': "field['type']"}), "(name=name, field_type=field['type'], **field)\n", (4584, 4630), False, 'from langflow.template.field.base import TemplateField\n')]
from typing import Dict, List, Optional from langchain.agents.load_tools import ( _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS, ) from langflow.custom import customs from langflow.interface.base import LangChainTypeCreator from langflow.interface.tools.constants import ( ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS, ) from langflow.interface.tools.util import get_tool_params from langflow.services.utils import get_settings_manager from langflow.template.field.base import TemplateField from langflow.template.template.base import Template from langflow.utils import util from langflow.utils.util import build_template_from_class TOOL_INPUTS = { "str": TemplateField( field_type="str", required=True, is_list=False, show=True, placeholder="", value="", ), "llm": TemplateField( field_type="BaseLanguageModel", required=True, is_list=False, show=True ), "func": TemplateField( field_type="function", required=True, is_list=False, show=True, multiline=True, ), "code": TemplateField( field_type="str", required=True, is_list=False, show=True, value="", multiline=True, ), "path": TemplateField( field_type="file", required=True, is_list=False, show=True, value="", suffixes=[".json", ".yaml", ".yml"], file_types=["json", "yaml", "yml"], ), } class ToolCreator(LangChainTypeCreator): type_name: str = "tools" tools_dict: Optional[Dict] = None @property def type_to_loader_dict(self) -> Dict: settings_manager = get_settings_manager() if self.tools_dict is None: all_tools = {} for tool, tool_fcn in ALL_TOOLS_NAMES.items(): tool_params = get_tool_params(tool_fcn) tool_name = tool_params.get("name") or tool if ( tool_name in settings_manager.settings.TOOLS or settings_manager.settings.DEV ): if tool_name == "JsonSpec": tool_params["path"] = tool_params.pop("dict_") # type: ignore all_tools[tool_name] = { "type": tool, "params": tool_params, "fcn": tool_fcn, } self.tools_dict = all_tools return self.tools_dict def get_signature(self, name: str) -> Optional[Dict]: """Get the signature of a tool.""" base_classes = ["Tool", "BaseTool"] fields = [] params = [] tool_params = {} # Raise error if name is not in tools if name not in self.type_to_loader_dict.keys(): raise ValueError("Tool not found") tool_type: str = self.type_to_loader_dict[name]["type"] # type: ignore # if tool_type in _BASE_TOOLS.keys(): # params = [] if tool_type in _LLM_TOOLS.keys(): params = ["llm"] elif tool_type in _EXTRA_LLM_TOOLS.keys(): extra_keys = _EXTRA_LLM_TOOLS[tool_type][1] params = ["llm"] + extra_keys elif tool_type in _EXTRA_OPTIONAL_TOOLS.keys(): extra_keys = _EXTRA_OPTIONAL_TOOLS[tool_type][1] params = extra_keys # elif tool_type == "Tool": # params = ["name", "description", "func"] elif tool_type in CUSTOM_TOOLS: # Get custom tool params params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes = ["function"] if node := customs.get_custom_nodes("tools").get(tool_type): return node elif tool_type in FILE_TOOLS: params = self.type_to_loader_dict[name]["params"] # type: ignore base_classes += [name] elif tool_type in OTHER_TOOLS: tool_dict = build_template_from_class(tool_type, OTHER_TOOLS) fields = tool_dict["template"] # Pop unnecessary fields and add name fields.pop("_type") # type: ignore fields.pop("return_direct") # type: ignore fields.pop("verbose") # type: ignore tool_params = { "name": fields.pop("name")["value"], # type: ignore "description": fields.pop("description")["value"], # type: ignore } fields = [ TemplateField(name=name, field_type=field["type"], **field) for name, field in fields.items() # type: ignore ] base_classes += tool_dict["base_classes"] # Copy the field and add the name for param in params: field = TOOL_INPUTS.get(param, TOOL_INPUTS["str"]).copy() field.name = param field.advanced = False if param == "aiosession": field.show = False field.required = False fields.append(field) template = Template(fields=fields, type_name=tool_type) tool_params = {**tool_params, **self.type_to_loader_dict[name]["params"]} return { "template": util.format_dict(template.to_dict()), **tool_params, "base_classes": base_classes, } def to_list(self) -> List[str]: """List all load tools""" return list(self.type_to_loader_dict.keys()) tool_creator = ToolCreator()
[ "langchain.agents.load_tools._LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys", "langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys" ]
[((709, 811), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'placeholder': '""""""', 'value': '""""""'}), "(field_type='str', required=True, is_list=False, show=True,\n placeholder='', value='')\n", (722, 811), False, 'from langflow.template.field.base import TemplateField\n'), ((875, 965), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""BaseLanguageModel"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)'}), "(field_type='BaseLanguageModel', required=True, is_list=False,\n show=True)\n", (888, 965), False, 'from langflow.template.field.base import TemplateField\n'), ((989, 1087), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""function"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'multiline': '(True)'}), "(field_type='function', required=True, is_list=False, show=\n True, multiline=True)\n", (1002, 1087), False, 'from langflow.template.field.base import TemplateField\n'), ((1143, 1245), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""str"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'multiline': '(True)'}), "(field_type='str', required=True, is_list=False, show=True,\n value='', multiline=True)\n", (1156, 1245), False, 'from langflow.template.field.base import TemplateField\n'), ((1310, 1474), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'field_type': '"""file"""', 'required': '(True)', 'is_list': '(False)', 'show': '(True)', 'value': '""""""', 'suffixes': "['.json', '.yaml', '.yml']", 'file_types': "['json', 'yaml', 'yml']"}), "(field_type='file', required=True, is_list=False, show=True,\n value='', suffixes=['.json', '.yaml', '.yml'], file_types=['json',\n 'yaml', 'yml'])\n", (1323, 1474), False, 'from langflow.template.field.base import TemplateField\n'), ((1728, 1750), 'langflow.services.utils.get_settings_manager', 'get_settings_manager', ([], {}), '()\n', (1748, 1750), False, 'from langflow.services.utils import get_settings_manager\n'), ((5139, 5183), 'langflow.template.template.base.Template', 'Template', ([], {'fields': 'fields', 'type_name': 'tool_type'}), '(fields=fields, type_name=tool_type)\n', (5147, 5183), False, 'from langflow.template.template.base import Template\n'), ((1849, 1872), 'langflow.interface.tools.constants.ALL_TOOLS_NAMES.items', 'ALL_TOOLS_NAMES.items', ([], {}), '()\n', (1870, 1872), False, 'from langflow.interface.tools.constants import ALL_TOOLS_NAMES, CUSTOM_TOOLS, FILE_TOOLS, OTHER_TOOLS\n'), ((3091, 3108), 'langchain.agents.load_tools._LLM_TOOLS.keys', '_LLM_TOOLS.keys', ([], {}), '()\n', (3106, 3108), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((1904, 1929), 'langflow.interface.tools.util.get_tool_params', 'get_tool_params', (['tool_fcn'], {}), '(tool_fcn)\n', (1919, 1929), False, 'from langflow.interface.tools.util import get_tool_params\n'), ((3165, 3188), 'langchain.agents.load_tools._EXTRA_LLM_TOOLS.keys', '_EXTRA_LLM_TOOLS.keys', ([], {}), '()\n', (3186, 3188), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((3314, 3342), 'langchain.agents.load_tools._EXTRA_OPTIONAL_TOOLS.keys', '_EXTRA_OPTIONAL_TOOLS.keys', ([], {}), '()\n', (3340, 3342), False, 'from langchain.agents.load_tools import _EXTRA_LLM_TOOLS, _EXTRA_OPTIONAL_TOOLS, _LLM_TOOLS\n'), ((4038, 4087), 'langflow.utils.util.build_template_from_class', 'build_template_from_class', (['tool_type', 'OTHER_TOOLS'], {}), '(tool_type, OTHER_TOOLS)\n', (4063, 4087), False, 'from langflow.utils.util import build_template_from_class\n'), ((3746, 3779), 'langflow.custom.customs.get_custom_nodes', 'customs.get_custom_nodes', (['"""tools"""'], {}), "('tools')\n", (3770, 3779), False, 'from langflow.custom import customs\n'), ((4571, 4630), 'langflow.template.field.base.TemplateField', 'TemplateField', ([], {'name': 'name', 'field_type': "field['type']"}), "(name=name, field_type=field['type'], **field)\n", (4584, 4630), False, 'from langflow.template.field.base import TemplateField\n')]
import langchain_visualizer # isort:skip # noqa: F401 import asyncio import vcr_langchain as vcr from fvalues import FValue from langchain import PromptTemplate from langchain.llms import OpenAI # ========================== Start of langchain example code ========================== # https://langchain.readthedocs.io/en/latest/modules/prompts/getting_started.html # An example prompt with one input variable one_input_prompt = PromptTemplate( input_variables=["adjective"], template="Tell me a {adjective} joke." ) prompt = one_input_prompt.format(adjective="funny") # ================================== Execute example ================================== def test_prompt(): assert prompt.parts == ( "Tell me a ", FValue(source="adjective", value="funny", formatted="funny"), " joke.", ) @vcr.use_cassette() async def one_input_prompt_demo(): agent = OpenAI(model_name="text-ada-001", temperature=0) return agent(prompt) def test_llm_usage_succeeds(): """ Check that it works like a regular prompt. Also, record playback for easy visualization. """ result = asyncio.get_event_loop().run_until_complete(one_input_prompt_demo()) assert result.strip().startswith("Why did the chicken cross the road?") if __name__ == "__main__": from langchain_visualizer import visualize visualize(one_input_prompt_demo)
[ "langchain_visualizer.visualize", "langchain.llms.OpenAI", "langchain.PromptTemplate" ]
[((434, 524), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['adjective']", 'template': '"""Tell me a {adjective} joke."""'}), "(input_variables=['adjective'], template=\n 'Tell me a {adjective} joke.')\n", (448, 524), False, 'from langchain import PromptTemplate\n'), ((837, 855), 'vcr_langchain.use_cassette', 'vcr.use_cassette', ([], {}), '()\n', (853, 855), True, 'import vcr_langchain as vcr\n'), ((903, 951), 'langchain.llms.OpenAI', 'OpenAI', ([], {'model_name': '"""text-ada-001"""', 'temperature': '(0)'}), "(model_name='text-ada-001', temperature=0)\n", (909, 951), False, 'from langchain.llms import OpenAI\n'), ((1362, 1394), 'langchain_visualizer.visualize', 'visualize', (['one_input_prompt_demo'], {}), '(one_input_prompt_demo)\n', (1371, 1394), False, 'from langchain_visualizer import visualize\n'), ((748, 808), 'fvalues.FValue', 'FValue', ([], {'source': '"""adjective"""', 'value': '"""funny"""', 'formatted': '"""funny"""'}), "(source='adjective', value='funny', formatted='funny')\n", (754, 808), False, 'from fvalues import FValue\n'), ((1136, 1160), 'asyncio.get_event_loop', 'asyncio.get_event_loop', ([], {}), '()\n', (1158, 1160), False, 'import asyncio\n')]
import threading import time import unittest import unittest.mock from typing import Any, Dict from uuid import UUID import pytest from langsmith import Client from langchain_core.outputs import LLMResult from langchain_core.tracers.langchain import LangChainTracer from langchain_core.tracers.schemas import Run def test_example_id_assignment_threadsafe() -> None: """Test that example assigned at callback start/end is honored.""" example_ids = {} def mock_create_run(**kwargs: Any) -> Any: example_ids[kwargs.get("id")] = kwargs.get("reference_example_id") return unittest.mock.MagicMock() client = unittest.mock.MagicMock(spec=Client) client.create_run = mock_create_run tracer = LangChainTracer(client=client) old_persist_run_single = tracer._persist_run_single def new_persist_run_single(run: Run) -> None: time.sleep(0.01) old_persist_run_single(run) with unittest.mock.patch.object( tracer, "_persist_run_single", new=new_persist_run_single ): run_id_1 = UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a") run_id_2 = UUID("f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1") example_id_1 = UUID("57e42c57-8c79-4d9f-8765-bf6cd3a98055") tracer.example_id = example_id_1 tracer.on_llm_start({"name": "example_1"}, ["foo"], run_id=run_id_1) tracer.on_llm_end(LLMResult(generations=[], llm_output={}), run_id=run_id_1) example_id_2 = UUID("4f31216e-7c26-4027-a5fd-0bbf9ace17dc") tracer.example_id = example_id_2 tracer.on_llm_start({"name": "example_2"}, ["foo"], run_id=run_id_2) tracer.on_llm_end(LLMResult(generations=[], llm_output={}), run_id=run_id_2) tracer.example_id = None expected_example_ids = { run_id_1: example_id_1, run_id_2: example_id_2, } tracer.wait_for_futures() assert example_ids == expected_example_ids def test_log_lock() -> None: """Test that example assigned at callback start/end is honored.""" client = unittest.mock.MagicMock(spec=Client) tracer = LangChainTracer(client=client) with unittest.mock.patch.object(tracer, "_persist_run_single", new=lambda _: _): run_id_1 = UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a") lock = threading.Lock() tracer.on_chain_start({"name": "example_1"}, {"input": lock}, run_id=run_id_1) tracer.on_chain_end({}, run_id=run_id_1) tracer.wait_for_futures() class LangChainProjectNameTest(unittest.TestCase): """ Test that the project name is set correctly for runs. """ class SetProperTracerProjectTestCase: def __init__( self, test_name: str, envvars: Dict[str, str], expected_project_name: str ): self.test_name = test_name self.envvars = envvars self.expected_project_name = expected_project_name def test_correct_get_tracer_project(self) -> None: cases = [ self.SetProperTracerProjectTestCase( test_name="default to 'default' when no project provided", envvars={}, expected_project_name="default", ), self.SetProperTracerProjectTestCase( test_name="use session_name for legacy tracers", envvars={"LANGCHAIN_SESSION": "old_timey_session"}, expected_project_name="old_timey_session", ), self.SetProperTracerProjectTestCase( test_name="use LANGCHAIN_PROJECT over SESSION_NAME", envvars={ "LANGCHAIN_SESSION": "old_timey_session", "LANGCHAIN_PROJECT": "modern_session", }, expected_project_name="modern_session", ), ] for case in cases: with self.subTest(msg=case.test_name): with pytest.MonkeyPatch.context() as mp: for k, v in case.envvars.items(): mp.setenv(k, v) client = unittest.mock.MagicMock(spec=Client) tracer = LangChainTracer(client=client) projects = [] def mock_create_run(**kwargs: Any) -> Any: projects.append(kwargs.get("project_name")) return unittest.mock.MagicMock() client.create_run = mock_create_run tracer.on_llm_start( {"name": "example_1"}, ["foo"], run_id=UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a"), ) tracer.wait_for_futures() assert ( len(projects) == 1 and projects[0] == case.expected_project_name )
[ "langchain_core.tracers.langchain.LangChainTracer", "langchain_core.outputs.LLMResult" ]
[((640, 676), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (663, 676), False, 'import unittest\n'), ((730, 760), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (745, 760), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((2059, 2095), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (2082, 2095), False, 'import unittest\n'), ((2109, 2139), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (2124, 2139), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((600, 625), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {}), '()\n', (623, 625), False, 'import unittest\n'), ((876, 892), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (886, 892), False, 'import time\n'), ((939, 1029), 'unittest.mock.patch.object', 'unittest.mock.patch.object', (['tracer', '"""_persist_run_single"""'], {'new': 'new_persist_run_single'}), "(tracer, '_persist_run_single', new=\n new_persist_run_single)\n", (965, 1029), False, 'import unittest\n'), ((1059, 1103), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (1063, 1103), False, 'from uuid import UUID\n'), ((1123, 1167), 'uuid.UUID', 'UUID', (['"""f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1"""'], {}), "('f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1')\n", (1127, 1167), False, 'from uuid import UUID\n'), ((1191, 1235), 'uuid.UUID', 'UUID', (['"""57e42c57-8c79-4d9f-8765-bf6cd3a98055"""'], {}), "('57e42c57-8c79-4d9f-8765-bf6cd3a98055')\n", (1195, 1235), False, 'from uuid import UUID\n'), ((1462, 1506), 'uuid.UUID', 'UUID', (['"""4f31216e-7c26-4027-a5fd-0bbf9ace17dc"""'], {}), "('4f31216e-7c26-4027-a5fd-0bbf9ace17dc')\n", (1466, 1506), False, 'from uuid import UUID\n'), ((2150, 2224), 'unittest.mock.patch.object', 'unittest.mock.patch.object', (['tracer', '"""_persist_run_single"""'], {'new': '(lambda _: _)'}), "(tracer, '_persist_run_single', new=lambda _: _)\n", (2176, 2224), False, 'import unittest\n'), ((2245, 2289), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (2249, 2289), False, 'from uuid import UUID\n'), ((2305, 2321), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2319, 2321), False, 'import threading\n'), ((1380, 1420), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': '[]', 'llm_output': '{}'}), '(generations=[], llm_output={})\n', (1389, 1420), False, 'from langchain_core.outputs import LLMResult\n'), ((1651, 1691), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': '[]', 'llm_output': '{}'}), '(generations=[], llm_output={})\n', (1660, 1691), False, 'from langchain_core.outputs import LLMResult\n'), ((3929, 3957), 'pytest.MonkeyPatch.context', 'pytest.MonkeyPatch.context', ([], {}), '()\n', (3955, 3957), False, 'import pytest\n'), ((4089, 4125), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (4112, 4125), False, 'import unittest\n'), ((4155, 4185), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (4170, 4185), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((4383, 4408), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {}), '()\n', (4406, 4408), False, 'import unittest\n'), ((4619, 4663), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (4623, 4663), False, 'from uuid import UUID\n')]
import threading import time import unittest import unittest.mock from typing import Any, Dict from uuid import UUID import pytest from langsmith import Client from langchain_core.outputs import LLMResult from langchain_core.tracers.langchain import LangChainTracer from langchain_core.tracers.schemas import Run def test_example_id_assignment_threadsafe() -> None: """Test that example assigned at callback start/end is honored.""" example_ids = {} def mock_create_run(**kwargs: Any) -> Any: example_ids[kwargs.get("id")] = kwargs.get("reference_example_id") return unittest.mock.MagicMock() client = unittest.mock.MagicMock(spec=Client) client.create_run = mock_create_run tracer = LangChainTracer(client=client) old_persist_run_single = tracer._persist_run_single def new_persist_run_single(run: Run) -> None: time.sleep(0.01) old_persist_run_single(run) with unittest.mock.patch.object( tracer, "_persist_run_single", new=new_persist_run_single ): run_id_1 = UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a") run_id_2 = UUID("f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1") example_id_1 = UUID("57e42c57-8c79-4d9f-8765-bf6cd3a98055") tracer.example_id = example_id_1 tracer.on_llm_start({"name": "example_1"}, ["foo"], run_id=run_id_1) tracer.on_llm_end(LLMResult(generations=[], llm_output={}), run_id=run_id_1) example_id_2 = UUID("4f31216e-7c26-4027-a5fd-0bbf9ace17dc") tracer.example_id = example_id_2 tracer.on_llm_start({"name": "example_2"}, ["foo"], run_id=run_id_2) tracer.on_llm_end(LLMResult(generations=[], llm_output={}), run_id=run_id_2) tracer.example_id = None expected_example_ids = { run_id_1: example_id_1, run_id_2: example_id_2, } tracer.wait_for_futures() assert example_ids == expected_example_ids def test_log_lock() -> None: """Test that example assigned at callback start/end is honored.""" client = unittest.mock.MagicMock(spec=Client) tracer = LangChainTracer(client=client) with unittest.mock.patch.object(tracer, "_persist_run_single", new=lambda _: _): run_id_1 = UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a") lock = threading.Lock() tracer.on_chain_start({"name": "example_1"}, {"input": lock}, run_id=run_id_1) tracer.on_chain_end({}, run_id=run_id_1) tracer.wait_for_futures() class LangChainProjectNameTest(unittest.TestCase): """ Test that the project name is set correctly for runs. """ class SetProperTracerProjectTestCase: def __init__( self, test_name: str, envvars: Dict[str, str], expected_project_name: str ): self.test_name = test_name self.envvars = envvars self.expected_project_name = expected_project_name def test_correct_get_tracer_project(self) -> None: cases = [ self.SetProperTracerProjectTestCase( test_name="default to 'default' when no project provided", envvars={}, expected_project_name="default", ), self.SetProperTracerProjectTestCase( test_name="use session_name for legacy tracers", envvars={"LANGCHAIN_SESSION": "old_timey_session"}, expected_project_name="old_timey_session", ), self.SetProperTracerProjectTestCase( test_name="use LANGCHAIN_PROJECT over SESSION_NAME", envvars={ "LANGCHAIN_SESSION": "old_timey_session", "LANGCHAIN_PROJECT": "modern_session", }, expected_project_name="modern_session", ), ] for case in cases: with self.subTest(msg=case.test_name): with pytest.MonkeyPatch.context() as mp: for k, v in case.envvars.items(): mp.setenv(k, v) client = unittest.mock.MagicMock(spec=Client) tracer = LangChainTracer(client=client) projects = [] def mock_create_run(**kwargs: Any) -> Any: projects.append(kwargs.get("project_name")) return unittest.mock.MagicMock() client.create_run = mock_create_run tracer.on_llm_start( {"name": "example_1"}, ["foo"], run_id=UUID("9d878ab3-e5ca-4218-aef6-44cbdc90160a"), ) tracer.wait_for_futures() assert ( len(projects) == 1 and projects[0] == case.expected_project_name )
[ "langchain_core.tracers.langchain.LangChainTracer", "langchain_core.outputs.LLMResult" ]
[((640, 676), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (663, 676), False, 'import unittest\n'), ((730, 760), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (745, 760), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((2059, 2095), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (2082, 2095), False, 'import unittest\n'), ((2109, 2139), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (2124, 2139), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((600, 625), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {}), '()\n', (623, 625), False, 'import unittest\n'), ((876, 892), 'time.sleep', 'time.sleep', (['(0.01)'], {}), '(0.01)\n', (886, 892), False, 'import time\n'), ((939, 1029), 'unittest.mock.patch.object', 'unittest.mock.patch.object', (['tracer', '"""_persist_run_single"""'], {'new': 'new_persist_run_single'}), "(tracer, '_persist_run_single', new=\n new_persist_run_single)\n", (965, 1029), False, 'import unittest\n'), ((1059, 1103), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (1063, 1103), False, 'from uuid import UUID\n'), ((1123, 1167), 'uuid.UUID', 'UUID', (['"""f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1"""'], {}), "('f1f9fa53-8b2f-4742-bdbc-38215f7bd1e1')\n", (1127, 1167), False, 'from uuid import UUID\n'), ((1191, 1235), 'uuid.UUID', 'UUID', (['"""57e42c57-8c79-4d9f-8765-bf6cd3a98055"""'], {}), "('57e42c57-8c79-4d9f-8765-bf6cd3a98055')\n", (1195, 1235), False, 'from uuid import UUID\n'), ((1462, 1506), 'uuid.UUID', 'UUID', (['"""4f31216e-7c26-4027-a5fd-0bbf9ace17dc"""'], {}), "('4f31216e-7c26-4027-a5fd-0bbf9ace17dc')\n", (1466, 1506), False, 'from uuid import UUID\n'), ((2150, 2224), 'unittest.mock.patch.object', 'unittest.mock.patch.object', (['tracer', '"""_persist_run_single"""'], {'new': '(lambda _: _)'}), "(tracer, '_persist_run_single', new=lambda _: _)\n", (2176, 2224), False, 'import unittest\n'), ((2245, 2289), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (2249, 2289), False, 'from uuid import UUID\n'), ((2305, 2321), 'threading.Lock', 'threading.Lock', ([], {}), '()\n', (2319, 2321), False, 'import threading\n'), ((1380, 1420), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': '[]', 'llm_output': '{}'}), '(generations=[], llm_output={})\n', (1389, 1420), False, 'from langchain_core.outputs import LLMResult\n'), ((1651, 1691), 'langchain_core.outputs.LLMResult', 'LLMResult', ([], {'generations': '[]', 'llm_output': '{}'}), '(generations=[], llm_output={})\n', (1660, 1691), False, 'from langchain_core.outputs import LLMResult\n'), ((3929, 3957), 'pytest.MonkeyPatch.context', 'pytest.MonkeyPatch.context', ([], {}), '()\n', (3955, 3957), False, 'import pytest\n'), ((4089, 4125), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {'spec': 'Client'}), '(spec=Client)\n', (4112, 4125), False, 'import unittest\n'), ((4155, 4185), 'langchain_core.tracers.langchain.LangChainTracer', 'LangChainTracer', ([], {'client': 'client'}), '(client=client)\n', (4170, 4185), False, 'from langchain_core.tracers.langchain import LangChainTracer\n'), ((4383, 4408), 'unittest.mock.MagicMock', 'unittest.mock.MagicMock', ([], {}), '()\n', (4406, 4408), False, 'import unittest\n'), ((4619, 4663), 'uuid.UUID', 'UUID', (['"""9d878ab3-e5ca-4218-aef6-44cbdc90160a"""'], {}), "('9d878ab3-e5ca-4218-aef6-44cbdc90160a')\n", (4623, 4663), False, 'from uuid import UUID\n')]
# coding=utf-8 import json import hashlib from datetime import datetime import os import time import openai import flet as ft import re import shutil from flet import ( ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons, ) from prompt_engineering import * import codecs # 赋值固定的api_key # 测试用 openai.api_key = None openai.api_key = read_APIKEY() ''' 每一行对话的类 在调用的同时,会将数据存入聊天记录文件 ''' class chat_row(ft.UserControl): def __init__(self, role, content): super(chat_row, self).__init__() self.role = role self.content = content self.hash = save_now_chat( chat_json_path=chat_json_path, role=self.role, content=self.content) def build(self): self.role_dropdown = ft.Dropdown( value=self.role, width=150, options=[ ft.dropdown.Option("system"), ft.dropdown.Option("user"), ft.dropdown.Option("assistant"), ], on_change=self.role_change ) self.content_textfield = ft.TextField( label="消息内容", value=self.content, filled=True, expand=True, multiline=True, on_change=self.content_change ) self.generalize_btn = ft.ElevatedButton( '概括本消息', on_click=self.sum_change ) self.sum_field = ft.TextField ( visible=False, label="概括内容", value='概括中', filled=True, expand=True, multiline=True, # on_change=self.content_change ) self.one_chat_row = ft.Row ( [ self.role_dropdown, self.content_textfield, self.generalize_btn, self.sum_field, ] ) return self.one_chat_row def sum_change(self,e): self.generalize_btn.disabled = True self.update() self.one_chat_row.controls.append(self.sum_field) self.update() self.summary = chatGPT_sum(self.content) renew_now_chat ( chat_json_path=chat_json_path, hash_val=self.hash, summary=self.summary ) self.sum_field.visible = True self.sum_field.value = self.summary self.generalize_btn.disabled = False self.update () return self.one_chat_row def role_change(self, e): self.role = self.role_dropdown.value renew_now_chat( chat_json_path=chat_json_path, hash_val=self.hash, role=self.role) def content_change(self, e): self.content = self.content_textfield.value renew_now_chat( chat_json_path=chat_json_path, hash_val=self.hash, content=self.content) ''' 调用chatGPT获取关键词和概括的函数 ''' # 文本切块成 def split_text(text, divide_length): return [text[i:i+divide_length] for i in range(0, len(text), divide_length)] # 获取概括文本 def chatGPT_sum_old(content): divided_content = split_text(content,divide_length=1000) chain_summary = [] for single_content in divided_content: composition_analysis_message = [{"role" : "user", "content" : f'''为以下文本创建概括: {content} 概括内容:'''}] chatGPT_raw_response = openai.ChatCompletion.create ( model="gpt-3.5-turbo", messages=composition_analysis_message ) summary = decode_chr ( chatGPT_raw_response.choices[0].message['content'].strip () ) chain_summary.append(summary) print(summary) chain_summary_long = '\n'.join ( chain_summary ) return chain_summary_long def chatGPT_sum(content,chain_type='map_reduce'): try: # 试图用langchain但无法打包 import langchain from langchain.llms import OpenAI from langchain.chains.summarize import load_summarize_chain from langchain.chains import AnalyzeDocumentChain from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.document_loaders import UnstructuredFileLoader from langchain.docstore.document import Document from langchain.prompts import PromptTemplate ''' 准备步骤 ''' # 创建文本分块器,每块长度为1000 text_splitter = RecursiveCharacterTextSplitter ( chunk_size=1000, chunk_overlap=0 ) # 创建大语言模型,将温度拉到最低以提高精确性 llm = OpenAI ( temperature=0, openai_api_key=openai.api_key ) ''' 加载需要概括的内容 ''' # 加载外部文件 # loader = UnstructuredFileLoader ( 'game_log_conversation.txt' ) # docs = loader.load () # split_docs = text_splitter.split_documents(docs) # 加载函数输入的字符串 split_content = text_splitter.split_text ( content ) split_docs = [Document ( page_content=t ) for t in split_content] ''' 总结文本 ''' # 创建prompt模板 prompt_template = """为以下文本创建概括: {text} 概括内容:""" PROMPT = PromptTemplate ( template=prompt_template, input_variables=["text"] ) # 创建总结链,模式为map_reduce # 第一种 # 有模板的总结链 summary_chain = load_summarize_chain ( llm, chain_type="map_reduce", return_intermediate_steps=True, map_prompt=PROMPT, combine_prompt=PROMPT ,verbose=True) # 带参数带模板总结 chain_summary = summary_chain ( {"input_documents" : split_docs}, # return_only_outputs=True, ) # 第二种 # 无模板的总结链 # summary_chain = load_summarize_chain ( llm, chain_type=chain_type, verbose=True ) # 直接总结 # chain_summary = summary_chain.run ( split_docs ) chain_summary_long = '\n'.join(chain_summary['intermediate_steps']) return chain_summary_long except Exception as error: print(error) return chatGPT_sum_old(content) # 获取关键词(还是用jieba吧) def chatGPT_getkeyword(content): chatGPT_raw_response = openai.Completion.create ( model="text-ada-001", prompt=f"你要总结这一文本的关键词,并以python列表的形式返回数个关键词字符串:{content}。", temperature=0 ) keyword = decode_chr ( chatGPT_raw_response.choices[0]['text'].strip () ) print(keyword) return keyword # 获取关键词(还是用jieba吧) def chatGPT_getsummary(content): chatGPT_raw_response = openai.Completion.create ( model="text-ada-001", prompt=f"你要在10字以内概括这段文本:{content}。", temperature=0 ) keyword = decode_chr ( chatGPT_raw_response.choices[0]['text'].strip () ) print(keyword) return keyword ''' 读写chatlog的相关函数 ''' # 创建chat数据记录的方法,将聊天数据结构基于 JSON 文件进行存储 def create_chat_json(save_path='./chatlog'): # 如果保存目录不存在,则创建目录 if not os.path.exists(save_path): os.makedirs(save_path) # 获取当前时间并格式化为文件名格式 now = datetime.now().strftime('%Y-%m-%d_%H-%M-%S') chat_json_name = f"chat_{now}.json" # 如果文件不存在,则创建文件 chat_json_path = os.path.join(save_path, chat_json_name) if not os.path.exists(chat_json_path): with open(chat_json_path, 'w') as f: json.dump([], f, default=str) return chat_json_path # 储存现有的每一句聊天信息 def save_now_chat(chat_json_path: str, role: str, content: str) -> str: ''' 将每一段聊天信息储存到一个 JSON 文件中。以哈希值作为唯一标识符 :param chat_json_path: 聊天文件的路径 :param role: 发言者 :param content: 发言内容 :return hash_val: 哈希值 ''' now = datetime.now().timestamp() str_to_hash = str(now) + role + content hash_obj = hashlib.blake2b(str_to_hash.encode('utf-8'), digest_size=8) hash_val = hash_obj.hexdigest() # 读取之前的内容 try: with open(chat_json_path, 'r') as f: chats = json.load(f) except FileNotFoundError: chats = [] # 添加新的聊天信息 message = { 'role': role, 'content': content, 'keyword': [], 'summary': '' } chats.append({ 'chat_seq': len(chats) + 1, 'hash': hash_val, 'created_time': now, 'altered_time': None, 'message': message }) # 将新的聊天信息写入文件 with open(chat_json_path, 'w') as f: json.dump(chats, f, default=str,indent=4) return hash_val # 根据聊天信息的哈希值,更新现有历史聊天列表 def renew_now_chat(chat_json_path: str, hash_val: str, role: str = None, content: str = None , summary=None) -> None: ''' 根据聊天信息的哈希值,更新现有历史聊天列表 :param chat_json_path: 聊天文件的路径 :param hash_val: 聊天信息的哈希值 :param role: 更新后的发言者(可选) :param content: 更新后的发言内容(可选) :return: None ''' # 读取chat_renew_data.json文件 with open(chat_json_path, 'r') as f: data = json.load(f) # 找出哈希值相同的键值对,并更新它们的role和content for chat_item in data: if hash_val == chat_item['hash']: if role: chat_item['message']['role'] = role if content: chat_item['message']['content'] = content if summary : chat_item['message']['summary'] = summary chat_item["altered_time"] = datetime.now().timestamp() break # 将更新后的数据写回文件 with open(chat_json_path, 'w') as f: json.dump(data, f, default=str,indent=4) # 根据聊天信息的哈希值,获取单个键值对内的role和content def get_one_role_and_content( chat_json_path: str, hash_value: str) -> (str, str): ''' 根据聊天信息的哈希值,获取单个键值对内的role和content :param chat_json_path: 聊天文件的路径 :param hash_value: 聊天信息的哈希值 :return: (发言者, 发言内容) ''' with open(chat_json_path) as f: data = json.load(f) for chat_item in data: for message in chat_item['message']: if hash_value == hashlib.blake2b( str(message["created_time"] + message["role"] + message["content"]).encode('utf-8'), digest_size=8).hexdigest(): return message["role"], message["content"] return None # 读取特定文件内的所有role和content def get_combined_data(chat_json_path: str) -> list[dict[str, str]]: ''' 获取特定文件内的所有role和content :param chat_json_path: JSON 聊天文件的路径 :return: 包含所有发言者和发言内容(若有概括则返回概括)的列表 ''' with open(chat_json_path) as f: data = json.load(f) result = [] for chat_item in data: if chat_item['message']['summary'] != '': result.append({ "role": chat_item['message']["role"], "content": chat_item['message']["summary"] }) else: result.append({ "role": chat_item['message']["role"], "content": chat_item['message']["content"] }) return result # 概括chatlog def summarize_chatlog(chatlog_json_path) : with open ( chatlog_json_path, 'r' ) as f : chatlog = json.load ( f ) for message in chatlog : if 'summary' not in message or not message['summary'] : content = message['message']['content'] # assuming content is always the second item in the message list if len ( content ) > 100 : summary = chatGPT_sum ( content ) message['summary'] = summary with open ( chatlog_json_path, 'w' ) as f : json.dump ( chatlog, f ,indent=4) # 概括chatlog def summarize_chatlog(chatlog_json_path) : with open ( chatlog_json_path, 'r' ) as f : chatlog = json.load ( f ) for message in chatlog : if 'summary' not in message or not message['summary'] : content = message['message']['content'] # assuming content is always the second item in the message list if len ( content ) > 100 : summary = chatGPT_sum ( content ) message['summary'] = summary with open ( chatlog_json_path, 'w' ) as f : json.dump ( chatlog, f ,indent=4) # 获取chatlog关键词 def get_chatlog_keyword(chatlog_json_path) : with open ( chatlog_json_path, 'r' ) as f : chatlog = json.load ( f ) for message in chatlog : if 'keyword' not in message or not message['keyword'] : content = message['message']['content'] # assuming content is always the second item in the message list keywords = chatGPT_getkeyword ( content ) message['keyword'] = keywords with open ( chatlog_json_path, 'w' ) as f : json.dump ( chatlog, f,indent=4 ) # 创建chat数据记录 chat_json_path = create_chat_json() ''' 加工message方法,先对向chatGPT发送的请求进行处理 ''' def process_message() : pass def cut_message(message) : ''' 剪切接收到的message,如果超过4000token长度就从最早的消息开始剪切,剪切到小于4000token为止 :param message: :return: ''' total_length = 0 # Iterate over contents in the list for message_dict in message : # message_dict['content'] = message_dict['content'] # 计算content长度 if message_dict['content'].isalpha () : length = len ( message_dict['content'].split () ) else : length = len ( message_dict['content'] ) total_length += length while total_length > 4000: if not message:raise Exception('最后一条消息长度大于4000字符了,请编辑消息或重置对话') removed_content = message.pop(0) removed_length = len ( removed_content['content'] ) if removed_content['content'].isalpha () : removed_length = len ( removed_content['content'].split () ) total_length -= removed_length return message ''' 读写API-KEY的函数 ''' # 定义 write_APIKEY 函数 def write_APIKEY(APIKEY=None): # 以追加模式打开或创建 APIKEY.txt 文件 with open("APIKEY.txt", "a") as f: # 写入字符串并换行 if APIKEY: f.write(APIKEY.strip() + "\n") # 定义 read_APIKEY 函数 def read_APIKEY(): # 以读取模式打开 APIKEY.txt 文件 with open("APIKEY.txt", "r") as f: # 读取所有行并存入列表 lines = f.readlines() # 如果列表不为空,返回最后一行的值,去掉换行符 if lines: return lines[-1].strip() ''' 读写settings.txt的函数 ''' # 读取settings.txt文件的设置项值,并返回一个字典 def read_settings(): """ 读取settings.txt文件,如果该文件不存在,则创建一个空白的文件 然后读取所有行,每行格式为:设置项名称 = 设置项的具体值 将每行拆分为键值对,添加到字典中,并返回该字典 :return: 包含settings.txt内的所有行的键值对形式的字典 """ settings_dict = {} try: with open('settings.txt', 'r', encoding='utf-8') as file: for line in file: line = line.strip() if line: key, value = line.split('=', maxsplit=1) settings_dict[key.strip()] = value.strip() except FileNotFoundError: with open('settings.txt', 'w', encoding='utf-8') as file: pass # 如果文件不存在,则创建一个空白的txt文件,不需要做任何操作 return settings_dict # 将字典中的多个键值对写入/修改settings.txt文件的设置项值 def write_settings(settings): """ 将多个键值对写入/更新settings.txt文件 如果文件不存在则创建一个空的文件 :param settings: 包含键值对的字典 """ with open('settings.txt', 'r+') as f: lines = f.readlines() f.seek(0) for key, value in settings.items(): for i, line in enumerate(lines): if key in line: lines[i] = key + ' = ' + value + '\n' break else: f.write(key + ' = ' + value + '\n') f.writelines(lines) ''' 读写字体文件的函数(适用于windows) ''' def replace_font_file(path): old_path = os.path.join(".", "asset", "font.ttf") try: os.remove(old_path) except OSError: pass try: shutil.copy(path, 'assets/font.ttf') except FileNotFoundError: with open('assets/font.ttf','a'): pass shutil.copy(path, 'assets/font.ttf') print("替换成功!") ''' 其他函数 ''' # 字符转码 def decode_chr(s): s = s.replace('\\\\','\\') pattern = re.compile(r'(\\u[0-9a-fA-F]{4}|\n)') result = '' pos = 0 while True: match = pattern.search(s, pos) if match is None: break result += s[pos:match.start()] if match.group() == '\n': result += '\n' else: result += chr(int(match.group()[2:], 16)) pos = match.end() result += s[pos:] return result def convert_content_to_unicode(txt_path): # 读取txt文件,转换为Python列表 with open(txt_path, 'r', encoding='utf-8') as f: txt_data = f.read() data = json.loads(txt_data) # 转换content内容为Unicode编码 for item in data: content = item["message"]["content"] item["message"]["content"] = content.encode('unicode_escape').decode() # item["message"]["content"] = item["message"]["content"].replace('\\u',r'\u') # 将Python列表转换为json格式字符串 json_data = json.dumps(data, ensure_ascii=False) # 保存json文件到cache文件夹 if not os.path.exists("./cache"): os.makedirs("./cache") # 创建cache文件夹 json_path = f"./cache/{os.path.basename(txt_path)[:-4]}.json" with open(json_path, 'w', encoding='utf-8') as f: f.write(json_data) return json_path # markdown检测 def markdown_check(gpt_msg): pass ''' 程序主窗体 ''' def ft_interface(page: ft.Page): # 设置字体与主题 page.title = 'BillyGPT' page.fonts = {'A75方正像素12': './assets/font.ttf'} page.theme = ft.Theme(font_family='A75方正像素12') page.dark_theme = page.theme # 设置主页面聊天区域的滚动列表 chat_area = ft.ListView( expand=True, spacing=10, auto_scroll=True, padding=20) ''' 添加选择上传文件、保存文件、打开文件夹按钮 ''' # 导入聊天记录 def import_chatlog(e: FilePickerResultEvent): try: clear_page() selected_file = ( ", ".join(map(lambda f: f.path, e.files) ) if e.files else "Cancelled!" ) if selected_file[-4:] == 'json': full_chatlog = get_combined_data(selected_file) print(full_chatlog) for chat_row_content in full_chatlog: role = chat_row_content['role'] content = chat_row_content['content'] chat_area.controls.append(chat_row(role, content)) page.update() elif selected_file[-4:] == '.txt': json_path = convert_content_to_unicode(selected_file) full_chatlog = get_combined_data(json_path) for chat_row_content in full_chatlog: role = decode_chr(chat_row_content['role']) content = decode_chr(chat_row_content['content']) chat_area.controls.append(chat_row(role, content)) page.update() except Exception as e: chat_area.controls.append( Text(f'出现如下报错\n{e}\n请检查导入的聊天记录是否正确,或联系开发者微信B1lli_official')) page.update() import_chatlog_dialog = FilePicker(on_result=import_chatlog) # 导出聊天记录 def save_file_result(e: FilePickerResultEvent): save_file_path.value = e.path if e.path else "Cancelled!" if save_file_path.value != "Cancelled!": shutil.copy(chat_json_path, save_file_path.value) path = save_file_path.value with open ( path, 'r', encoding='utf-8' ) as f : content = f.read () processed_content = decode_chr ( content ) with open ( path, 'w', encoding='utf-8' ) as f : f.write ( processed_content ) save_file_dialog = FilePicker(on_result=save_file_result) save_file_path = Text() # 打开文件夹(还没做) def get_directory_result(e: FilePickerResultEvent): directory_path.value = e.path if e.path else "Cancelled!" directory_path.update() get_directory_dialog = FilePicker(on_result=get_directory_result) directory_path = Text() # 隐藏所有 page.overlay.extend( [import_chatlog_dialog, save_file_dialog, get_directory_dialog]) ''' 添加改变字体按钮 并调用改变字体方法 ''' def change_font_clicked(e: FilePickerResultEvent): selected_file = ( ", ".join ( map ( lambda f : f.path, e.files ) ) if e.files else "Cancelled!" ) replace_font_file(selected_file) ''' 添加设置对话和按钮 ''' def save_settings(e): settings_dlg.open = False write_APIKEY(apikey_field.value) openai.api_key = read_APIKEY() page.update() def cancel_settings(e): settings_dlg.open = False page.update() change_font_dialog = FilePicker(on_result=change_font_clicked) page.overlay.extend([change_font_dialog]) apikey_field = ft.TextField(hint_text='在此输入apikey',) my_wechat = ft.Text('如有任何bug请联系我:B1lli_official',size=15) github_page_btn = ft.ElevatedButton( '打开本项目的GitHub页面', tooltip='如果你给这个项目点了star,你就是忠实用户了,请打开本页面后进入项目群!', on_click=lambda _:page.launch_url('https://github.com/createrX12/BillyGPT') ) change_font_btn = ft.ElevatedButton( '选择新的字体文件', on_click=lambda _:change_font_dialog.pick_files( allowed_extensions=['ttf'], allow_multiple=False, dialog_title='选择字体文件导入', ) ) settings_dlg = ft.AlertDialog( title=ft.Text("Settings"), content=ft.Column( [apikey_field,my_wechat,github_page_btn,change_font_btn], height=200, ), actions=[ ft.TextButton("保存", on_click=save_settings), ft.TextButton("取消", on_click=cancel_settings) ], actions_alignment=ft.MainAxisAlignment.END, shape=ft.RoundedRectangleBorder(radius=10) ) def open_dlg_modal(e): page.dialog = settings_dlg settings_dlg.open = True page.update() settings_btn = ft.IconButton( icon=ft.icons.SETTINGS_OUTLINED, icon_color="#9ecaff", bgcolor='#202429', icon_size=20, tooltip="Settings", on_click=open_dlg_modal, ) ''' 添加启动应用程序获取apikey窗口 ''' def save_settings_open(e): write_APIKEY(apikey_field_open.value) openai.api_key = read_APIKEY() open_setting_apikey_dlg.open = False page.update() get_apikey_btn = ft.ElevatedButton( '从openAI获取apikey', on_click=lambda _:page.launch_url('https://platform.openai.com/account/api-keys') ) write_APIKEY() openai.api_key = read_APIKEY() if not openai.api_key: apikey_field_open = ft.TextField(label="输入你的apikey") open_setting_apikey_dlg = ft.AlertDialog( open=True, modal=True, title=ft.Text("欢迎使用BillyGPT"), content=ft.Column( [apikey_field_open,get_apikey_btn], tight=True), actions=[ ft.ElevatedButton( text="开始使用", on_click=save_settings_open)], actions_alignment="end", ) page.dialog = open_setting_apikey_dlg openai.api_key = read_APIKEY() ''' 添加聊天行 ''' def add_msg(e): chatPO_btn.disabled = True chat_area.controls.append(chat_row('user', chat_text.value)) chat_text.value = "" page.update() chat_area.controls.append( chat_row( 'assistant', chatGPT())) page.update() def add_msg_composition(e): chatPO_btn.disabled = True chat_area.controls.append(chat_row('user', chat_text.value)) chat_area.controls.append( chat_row( 'assistant', chatGPT_PO( chat_text.value))) page.update() ''' 添加ctrl+enter发送消息方法 ''' def on_keyboard(e: ft.KeyboardEvent) : if e.ctrl: if e.key == 'Enter': add_msg(e) page.on_keyboard_event = on_keyboard ''' 添加模板选择功能 可以读取模板文件,实时搜索,选择时可以切换身份 ''' class DropdownSearchBar(ft.UserControl): def __init__(self): super(DropdownSearchBar, self).__init__() self.item_number = ft.Text(size=9,italic=True) def dropdown_search(self): _object_ = ft.Container( width=450, height=50, bgcolor="white10", border_radius=6, padding=ft.padding.only(top=15), clip_behavior=ft.ClipBehavior.HARD_EDGE, animate=ft.animation.Animation(400,"decelerate"), content=ft.Column( horizontal_alignment=ft.CrossAxisAlignment.CENTER, alignment=ft.MainAxisAlignment.START, controls=[ Row( spacing=10, vertical_alignment=ft.CrossAxisAlignment.CENTER, controls=[ ft.Icon( ft.icons.SEARCH_ROUNDED, size=15, opacity=0.90, ), ft.TextField( border_color='transparent', height=20, text_size=12, content_padding=2, cursor_color='black', cursor_width=1, hint_text='搜索模板', on_change=None ), self.item_number, ], ), ft.Column( scroll='auto', expand=True ) ] ) ) return _object_ def build(self): return ft.Container() page.add(DropdownSearchBar()) # def template_change(self, e): # template_name = template_dropbox.value # renew_now_chat( # chat_json_path=chat_json_path, # hash_val=self.hash, # role=self.role) # # # template_dropbox = ft.Dropdown( # value=role, # width=150, # options=[ # ft.dropdown.Option("system"), # ft.dropdown.Option("user"), # ft.dropdown.Option("assistant"), # ], # on_change=template_change # ) ''' 设置布局 添加控件 ''' # 模板选择下拉框 page.add( Row( [ ElevatedButton ( "清空聊天日志", icon=icons.CLEANING_SERVICES, on_click=lambda _ : clear_page(), ), ElevatedButton( "导入聊天日志", icon=icons.FILE_DOWNLOAD_OUTLINED, on_click=lambda _: import_chatlog_dialog.pick_files( allowed_extensions=['json','txt'], allow_multiple=False, dialog_title='选择聊天记录文件导入', initial_directory='./chatlog' ), ), ElevatedButton( "导出聊天日志", icon=icons.FILE_UPLOAD_OUTLINED, on_click=lambda _: save_file_dialog.save_file( file_name=f"chat_{datetime.now().strftime('%Y-%m-%d_%H-%M-%S')}.txt", # file_type=ft.FilePickerFileType.CUSTOM ), disabled=page.web, ), settings_btn ], alignment=ft.MainAxisAlignment.END, ), ) page.add(chat_area) chat_text = ft.TextField( hint_text="想和chatGPT说些什么?", filled=True, expand=True, multiline=True) chat_btn = ft.ElevatedButton("对话", on_click=add_msg, tooltip='随便聊聊') chatPO_btn = ft.ElevatedButton( "思维链优化对话", on_click=add_msg_composition, tooltip='认真提问\n仅在对话开始时可用' ) view = ft.Column( controls=[ ft.Row( controls=[ chat_text, chat_btn, chatPO_btn ] ), ], ) page.horizontal_alignment = ft.CrossAxisAlignment.CENTER page.add(view) ''' 添加清空页面方法 ''' def clear_page(): global chat_json_path chat_area.controls.clear () page.update() chat_json_path = create_chat_json() chatPO_btn.disabled = False ''' 聊天方法,向api发送请求 ''' def chatGPT(message=None): try: # global role_template if not message: message = get_combined_data(chat_json_path) # message = get_combined_data ( chat_json_path ) chatGPT_raw_response = openai.ChatCompletion.create( model="gpt-3.5-turbo", messages=message ) # if role_template:pass chatGPT_response = decode_chr( chatGPT_raw_response.choices[0].message['content']) return chatGPT_response.strip() except openai.error.AuthenticationError as error: chat_area.controls.append( ft.Text(f'出现如下报错\n{str(error)}\n请在设置中更新可用的apikey')) page.update() except openai.error.InvalidRequestError as error: chat_area.controls.append( ft.Text(f'出现如下报错\n{str(error)}\n聊天上下文过长,请对长文本调用概括模块概括后重试')) page.update() # summarize_chatlog(chat_json_path) # chat_area.controls.append( # ft.Text(f'概括完毕,已发送概括后消息')) # page.update() # message = cut_message ( get_combined_data ( chat_json_path ) ) # chatGPT_raw_response = openai.ChatCompletion.create( # model="gpt-3.5-turbo", # messages=message # ) # chatGPT_response = decode_chr( # chatGPT_raw_response.choices[0].message['content']) # return chatGPT_response.strip() except Exception as error: print(openai.api_key) chat_area.controls.append( ft.Text(f'出现如下报错\n{str ( error )}\n请联系开发者微信B1lli_official')) page.update() def chatGPT_PO(initial_prompt=None): ''' PO:Prompt Optimization 该方法是提示优化方法,工作原理如下: 它会尝试判定用户输入的每一条提示,并进行类别判断。 如果判断是问题或者要求,它会分析该类型提示一般需要哪些信息 对于缺少细节信息的提示,它会向用户提问具体细节 :param initial_prompt:初始prompt :return chatGPT_response.strip():基于提示工程思维链优化后的chatGPT方法 ''' try: initial_prompt = chat_text.value chat_text.value = '' page.update() chat_area.controls.append( ft.Text(f'正在分析提示词组成结构,请耐心等待', color='#1cc9a0')) page.update() composition_analysis_message = prompt_composition_analysis( initial_prompt) chat_area.controls.append( ft.Text(f'提示词组成结构分析完毕,正在根据组成结构逐步生成详细结果,耗时较长,请耐心等待', color='#1cc9a0')) page.update() chatGPT_raw_response = composition_stepped_reply( composition_analysis_message) chatGPT_response = decode_chr( chatGPT_raw_response.choices[0].message['content']) return chatGPT_response.strip() except openai.error.AuthenticationError as error: chat_area.controls.append( ft.Text(f'出现如下报错\n{str(error)}\n请在设置中更新可用的apikey')) page.update() except Exception as error: print(openai.api_key) chat_area.controls.append( ft.Text(f'出现如下报错\n{str ( error )}\n请联系开发者微信B1lli_official')) page.update() ''' 版本信息 ''' ver_text = ft.Text('BillyGPT V5.3.1 By B1lli', size=10) page.add(ver_text) if __name__ == '__main__': # 在客户端运行 ft.app(target=ft_interface, assets_dir='assets') # 在内网运行 # ft.app ( port=8550, target=ft_interface, view=ft.WEB_BROWSER ,assets_dir='assets') content = '''接下来你看到的是一个文字冒险游戏的剧情,你是剧情的女主角,名为钰鹤,下面是以对话形式展现出的剧情,你需要做出合理的判断,并说明原因:???:马路上的那个人怎么没撑伞 被雨淋湿的青年: ???:(他的样子好像有点奇怪··· ???:睫··车开过来了那个人却没有发现.....!? ???:再这样下去,他会被车撞到!? ???:危险………! ???:(我为什么在这里 ???:只记得自已跳入车道,之后的事完全没印象开眼睛就发现自己躺在医院病床上了 ???:不知道那个人事故之后怎么样了 穿西装的男子:小姐·你从刚才就在发呆你听到我说的话了吗? ???:啊…………不好意思 ???:对了,我想起来了。有人说发现我和某个人在一起而那个人身上带看违法药品 ???:(本以为要继续接受盘问结果莫名其妙地就被带到了这座岛上 穿西装的男子:埃我知道,遇到这种事头脑难免会陷入一片混乱 穿西装的男子:正因为情况特殊你更应该振作起来才行你说得出自己的全名吗? 鹤:嗯,我是钰鹤 穿西装的男子:真是太好了。看来你还算比较冷静至少说得出自已的名字。 穿西装的男子:那么,你应该也能够理解我刚才的说明吧钰小姐。 鹤:嗯……你是 穿西装的男子:真是的· 穿西装的男子:真是的我的名字是『今部』。看来你真的睡糊涂了。 鹤:对了,他叫今部·· 鹤:在被带来这里之前接受警方盘问正当事情始终理不出头绪时是他帮我脱离了困境 鹤:(但我完全不认识他不知道他是个怎么样的人 今部:没办法,只好挑重点再重新说明一次,这次请你务必专心听。 今部:首先第一个重点,正如刚刚所说的尔被怀疑涉及参与了某起事件现在被列为重点参考证人。 今部:由于事发现场,与你在一起的男子有交易非法药物的嫌疑。 今部:和他出现在同一个地点的人,也就是你,自然也被怀疑参与了违法买卖 鹤:你说我有嫌疑可是我真的什么都不知道也不认识那个男的 今部:可惜有自击者指出,曾经好几次看到男一女在事发地点附近的夜店出没进行非法药物交易 鹤:可是,我真的没有 今部:我了解。但现在的难点是没有人能证明你是无辜的。 鹤:那我该怎么做才好现在那个被环疑的男人在哪里?他看到我的脸应该就会知道抓错人了。 今部:你说的没错,请他真接和你见面确认,也许是最快的方法。 今部:只可惜他大概很难为你提供不在场证明 鹤:为什么…………·? 今部:你认得这名男子吧? 鹤:这个人的脸··我见过… 今部:这名涉嫌交易非法药物的男子可能因为车祸的冲击丧失了部分过去的记忆。 鹤:什么?丧失记忆 今部:是的。我想应该只是暂时性的失忆。但是他完全想不起来那天在夜店里和他在一起的女人是谁。 鹤:(怎么会难道真的没有人可以证明我的清白了? 今部:我帮你作证吧 今部:你没听清楚吗?我说我要帮你作证,证明你不认识那名男子,也和非法药物的犯罪无关。 鹤:你要帮我作证这样就能证明我的清白了吗? 今部:只要钰小姐愿意配合协助我。 今部:我的提议绝不会陷你干不利 鹤:(要我协助他可是如果他提出什么难题或是奇怪的条件,我该怎么办 今部:你似平还没完全了解你的处境 今部:事情发展到这地步,你没有选择的余地。请回我「好」或「YES」,二择其一。你会接受我的提议吧? 今部:我想请你帮的忙,就是在那名失忆男子的面前假装自己是『他的恋人』 鹤:要我扮演他的恋人·!? 今部:其实发生交通事故之前他和某位女性在一起,而那位女性很可能是他的女友。 今部:那名女性最后被目击到的地点就是那间夜店,随后不久她便失踪行踪成谜 今部:为了找出那位失踪的女性必须找回这名男子因车祸失去的记忆。 今部:所以才希望你假扮成他的恋人从他口中打听任何可能的线索找出他女友的所在地, 今部:在你协助的这段期间,我也会在别处进行搜索。只要找到和他出现在夜店的那名女性,就能顺利证明你的清白了。 鹤:可是我要怎么在素未谋面的陌生人面前假扮他的恋人,这未免太强人所难了· 今部:这不是担心做不做得到的时候你非做不可。 今部:麻烦你了。 鹤:为了洗清我的嫌疑也别无选择了 鹤:好我知道了 今部:谢谢你愿意协助据说这名男子周围的人都叫他也『葵』。 鹤:名字是『葵』吗 今部:啊,对了。差点忘了告诉你你从今天开始必须暂时住在这座小岛上。 鹤:什么? 今部:住宿的相关细节等下会由设施员工为你说明 鹤:等、等一下 看守员:不好意思我过来带钰小姐到她的房间。 今部:那就之后再见了钰小姐。 鹤:我到底会被带去挪里话说回来,在抵达小岛前好像坐了很久的船 鹤:该不会在找到他的女友之前我都回不了家? 鹤:..·嗯? 鹤:从这里能跳望到海面 鹤:这里真的是一座孤岛 看守员:今天的海面很平静呢 鹤:风景很美。大海和天空一望无际 看守员:毕竟这座设施坐落于岛上视野最佳的位置。 看守员:以前政府在这座岛上盖了监由国家管理 看守员:但战后因财政困难,没有经费维持而关闭后来由民间企业『西海普制药」将整座岛买下,加以利用。 看守员:我个现在所在的这栋建筑设施就是『西海普医疗中心』。 鹤:『西海普制药』? 看守员:那是一家专门研发、制造医疗用品及药品的制药厂。 看守员:这里就是『西海普制药』所管理的设施之一,用于收容病患以进行外地疗养 看守员:··表面上是这样。不过我想你也已经察觉到,实际用途不仅如此。 看守员:这里的收容人都有自己的房间只要在安全管理上不引发问题都能在允许的范围内从事自由活动。 看守员:也个的房间受到24小时全天候的监视一旦房间内发生任何意外,设施的勺人员都能马上应对。 鹤:·这么说的话,我的行动也会受到监视吗? 看守员:不会。我收到的指示是要带你去员工用宿舍的房间。那里不在监视的范围内,你大可放心。 看守员:一直站在这里说话也不是办法,我价们继续往前走吧 看守员:这里就是钰小姐的房间可供你自由使用 看守员:另外,还要给你这个 鹤:智能手机? 看守员:是的。在岛上的期间无法使用私人手机,请改用这台『SABOT』。 看守员:请收下。可以试着开启一下电源。 看守员:启动之后,手机内的引导功能就会为你介绍各功能的使用方法 看守员:这只手机也能当作你在岛上的身份证请务必随身携带 看守员:注意千万不要弄丢使用上也请多加小心 看守员:此外,在岛上只要出示这支手机证明身份,就能免费购物。不需要另外支付现金。 看守员:今部先生已经事先为钰小姐预付了一笔费用当作在岛上的生活费你不用担心金钱的问题。 看守员:最后说一件事关于明天会面请你明早10点到我们刚才经过的收容所入口处集合。 鹤:好的我明白了 看守员:那么,请好好休息 鹤:·是。 鹤:天呐没想到我被卷入了这么棘手的事件 鹤:今后,到底会怎样? 鹤:(太阳升起来了差不多该前往看守员所说的地点了 今部:早安,钰小姐。 今部:你好像很紧张。 鹤:嗯…是啊。 今部:那也是难免的。昨天离开后,我也重新思考了一遍 今部:你是普通人,不是职业演员:老实说我也不认为你能很自然地扮成他的恋人。 今部:今天就先以他的友人身份见面视状况往后再做调整吧 今部:请你在了解他的为人以后再看准合适的时机:告诉他你是他的女友。 看守员:早安。今部先生,钰小姐。让两位久等了。 今部:阿,已经到会面时间了吗那就拜托你好好表现了,钰小姐。 鹤:太、太强人所难了· 看守员:那我们走吧。 看守员:这间就是会面室。 看守员:进去以后,你会看见里面有块玻璃隔板将房间隔成会面者和收容人用的两间小房间。 看守员:会面过程中不会有职员陪同司,我们会在房间外待命,有任何问题或需要帮忙时,请随时叫我个 看守员:也许和收容人独处会让你感到不安但会面室的玻璃采用不易被打破的强化玻璃亚克力板制成, 看守员:即使搬起椅子用力砸玻璃房间内的隔板也绝对不会碎。这点你大可放心。 鹤:那个 看守员:是,请说。 鹤:请问·我该怎么面对收容人? 看守员:该怎么面对?如果你是他的朋友,就以平常心对待即可。 鹤:(我就是无法拿捏所谓的「平常心」 看守员:请问怎么了?还有什么事让你感到不安吗? 鹤:啊,没有没事。 看守员:等会面时间结束了,我会敲门提醒你就算事情没说完,也请尽快结束会面离开房间。 鹤:就是这人人吗· 葵:干嘛?一直町着别人的脸看 葵:我听说有人认识的人想找我谈谈 葵: 葵:不过很抱,我对你完全没印象。 葵:你叫什么名字? 葵:钰鹤 葵:嗯名字也没有印象。 葵:我们真的认识? 鹤:与其说认识,不如说 葵:啊,还是算了。反正我都不记得了问了也没有意义。 葵:如果我们以前真的见过面那很抱款,我现在都想不起来了。 葵:那么,你找我有什么事? 葵:特地跑到这种地方,应该不会只是为了闲聊吧? 葵:应该是有什么事想问我? 鹤:(就算这么说,我也不知道该问什么好. 鹤:(既然要装作认识的人要是表现得太客气的话反而会很不自然吧· 葵:怎么了? 鹤:啊,嗯那 葵:。。。 鹤:再这样下去会被怀疑的得随便问点问题才行··· 鹤:(总之,先问问看关于那场意外他还记得什么吧 你面临着2个选择,第一个:我听人说 ,你正涉及某件危险的事。第二个:你还记得意外发生前 ,你跟谁在一起吗 ?''' # a = chatGPT_sum(content) # print(a) ''' 你是一个文字冒险游戏的女主角,名为钰鹤,钰是姓氏,鹤是名字。下面是游戏剧情,你需要根据剧情做出合理的判断,并说明原因: 钰鹤被警方怀疑参与了某起事件,今部帮助她脱离困境,但他们需要找到一个可以证明钰鹤无辜的方法,最终他们决定让钰鹤和被环疑的男子见面,以确认钰鹤的无辜。 鹤被要求去小岛上假扮一名失忆男子的恋人,以帮助他找回记忆,以洗清自己的嫌疑。小岛上有一座监狱,拥有最佳的视野,海面平静,风景美丽。 鹤被卷入一个棘手的事件,被带到一个由民间企业『西海普制药」管理的设施,收容病患以进行外地疗养,但实际用途不仅如此。鹤被给予一台『SABOT』智能手机,可以用来证明身份,免费购物,以及收到一笔费用当作在岛上的生活费。最后,鹤被要求在第二天10点到 鹤面对着葵,不知道该如何面对,看守员提醒他以平常心对待,葵表示自己对鹤完全没有印象, 你要向葵提问,你面临着2个提问选择: 1.我听人说,你正涉及某件危险的事。 2.你还记得意外发生前,你跟谁在一起吗? 你选择哪个,为什么?'''
[ "langchain.chains.summarize.load_summarize_chain", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.llms.OpenAI", "langchain.docstore.document.Document", "langchain.prompts.PromptTemplate" ]
[((6345, 6470), 'openai.Completion.create', 'openai.Completion.create', ([], {'model': '"""text-ada-001"""', 'prompt': 'f"""你要总结这一文本的关键词,并以python列表的形式返回数个关键词字符串:{content}。"""', 'temperature': '(0)'}), "(model='text-ada-001', prompt=\n f'你要总结这一文本的关键词,并以python列表的形式返回数个关键词字符串:{content}。', temperature=0)\n", (6369, 6470), False, 'import openai\n'), ((6706, 6809), 'openai.Completion.create', 'openai.Completion.create', ([], {'model': '"""text-ada-001"""', 'prompt': 'f"""你要在10字以内概括这段文本:{content}。"""', 'temperature': '(0)'}), "(model='text-ada-001', prompt=\n f'你要在10字以内概括这段文本:{content}。', temperature=0)\n", (6730, 6809), False, 'import openai\n'), ((7340, 7379), 'os.path.join', 'os.path.join', (['save_path', 'chat_json_name'], {}), '(save_path, chat_json_name)\n', (7352, 7379), False, 'import os\n'), ((15982, 16020), 'os.path.join', 'os.path.join', (['"""."""', '"""asset"""', '"""font.ttf"""'], {}), "('.', 'asset', 'font.ttf')\n", (15994, 16020), False, 'import os\n'), ((16410, 16449), 're.compile', 're.compile', (['"""(\\\\\\\\u[0-9a-fA-F]{4}|\\\\n)"""'], {}), "('(\\\\\\\\u[0-9a-fA-F]{4}|\\\\n)')\n", (16420, 16449), False, 'import re\n'), ((17333, 17369), 'json.dumps', 'json.dumps', (['data'], {'ensure_ascii': '(False)'}), '(data, ensure_ascii=False)\n', (17343, 17369), False, 'import json\n'), ((17889, 17922), 'flet.Theme', 'ft.Theme', ([], {'font_family': '"""A75方正像素12"""'}), "(font_family='A75方正像素12')\n", (17897, 17922), True, 'import flet as ft\n'), ((17998, 18064), 'flet.ListView', 'ft.ListView', ([], {'expand': '(True)', 'spacing': '(10)', 'auto_scroll': '(True)', 'padding': '(20)'}), '(expand=True, spacing=10, auto_scroll=True, padding=20)\n', (18009, 18064), True, 'import flet as ft\n'), ((19524, 19560), 'flet.FilePicker', 'FilePicker', ([], {'on_result': 'import_chatlog'}), '(on_result=import_chatlog)\n', (19534, 19560), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((20147, 20185), 'flet.FilePicker', 'FilePicker', ([], {'on_result': 'save_file_result'}), '(on_result=save_file_result)\n', (20157, 20185), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((20208, 20214), 'flet.Text', 'Text', ([], {}), '()\n', (20212, 20214), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((20422, 20464), 'flet.FilePicker', 'FilePicker', ([], {'on_result': 'get_directory_result'}), '(on_result=get_directory_result)\n', (20432, 20464), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((20487, 20493), 'flet.Text', 'Text', ([], {}), '()\n', (20491, 20493), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((21232, 21273), 'flet.FilePicker', 'FilePicker', ([], {'on_result': 'change_font_clicked'}), '(on_result=change_font_clicked)\n', (21242, 21273), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((21343, 21379), 'flet.TextField', 'ft.TextField', ([], {'hint_text': '"""在此输入apikey"""'}), "(hint_text='在此输入apikey')\n", (21355, 21379), True, 'import flet as ft\n'), ((21398, 21444), 'flet.Text', 'ft.Text', (['"""如有任何bug请联系我:B1lli_official"""'], {'size': '(15)'}), "('如有任何bug请联系我:B1lli_official', size=15)\n", (21405, 21444), True, 'import flet as ft\n'), ((22531, 22686), 'flet.IconButton', 'ft.IconButton', ([], {'icon': 'ft.icons.SETTINGS_OUTLINED', 'icon_color': '"""#9ecaff"""', 'bgcolor': '"""#202429"""', 'icon_size': '(20)', 'tooltip': '"""Settings"""', 'on_click': 'open_dlg_modal'}), "(icon=ft.icons.SETTINGS_OUTLINED, icon_color='#9ecaff',\n bgcolor='#202429', icon_size=20, tooltip='Settings', on_click=\n open_dlg_modal)\n", (22544, 22686), True, 'import flet as ft\n'), ((28829, 28915), 'flet.TextField', 'ft.TextField', ([], {'hint_text': '"""想和chatGPT说些什么?"""', 'filled': '(True)', 'expand': '(True)', 'multiline': '(True)'}), "(hint_text='想和chatGPT说些什么?', filled=True, expand=True,\n multiline=True)\n", (28841, 28915), True, 'import flet as ft\n'), ((28965, 29022), 'flet.ElevatedButton', 'ft.ElevatedButton', (['"""对话"""'], {'on_click': 'add_msg', 'tooltip': '"""随便聊聊"""'}), "('对话', on_click=add_msg, tooltip='随便聊聊')\n", (28982, 29022), True, 'import flet as ft\n'), ((29041, 29131), 'flet.ElevatedButton', 'ft.ElevatedButton', (['"""思维链优化对话"""'], {'on_click': 'add_msg_composition', 'tooltip': '"""认真提问\n仅在对话开始时可用"""'}), "('思维链优化对话', on_click=add_msg_composition, tooltip=\n '认真提问\\n仅在对话开始时可用')\n", (29058, 29131), True, 'import flet as ft\n'), ((33098, 33143), 'flet.Text', 'ft.Text', (['"""BillyGPT V5.3.1 By B1lli"""'], {'size': '(10)'}), "('BillyGPT V5.3.1 By B1lli', size=10)\n", (33105, 33143), True, 'import flet as ft\n'), ((33219, 33267), 'flet.app', 'ft.app', ([], {'target': 'ft_interface', 'assets_dir': '"""assets"""'}), "(target=ft_interface, assets_dir='assets')\n", (33225, 33267), True, 'import flet as ft\n'), ((1177, 1300), 'flet.TextField', 'ft.TextField', ([], {'label': '"""消息内容"""', 'value': 'self.content', 'filled': '(True)', 'expand': '(True)', 'multiline': '(True)', 'on_change': 'self.content_change'}), "(label='消息内容', value=self.content, filled=True, expand=True,\n multiline=True, on_change=self.content_change)\n", (1189, 1300), True, 'import flet as ft\n'), ((1419, 1471), 'flet.ElevatedButton', 'ft.ElevatedButton', (['"""概括本消息"""'], {'on_click': 'self.sum_change'}), "('概括本消息', on_click=self.sum_change)\n", (1436, 1471), True, 'import flet as ft\n'), ((1523, 1624), 'flet.TextField', 'ft.TextField', ([], {'visible': '(False)', 'label': '"""概括内容"""', 'value': '"""概括中"""', 'filled': '(True)', 'expand': '(True)', 'multiline': '(True)'}), "(visible=False, label='概括内容', value='概括中', filled=True, expand=\n True, multiline=True)\n", (1535, 1624), True, 'import flet as ft\n'), ((1787, 1880), 'flet.Row', 'ft.Row', (['[self.role_dropdown, self.content_textfield, self.generalize_btn, self.\n sum_field]'], {}), '([self.role_dropdown, self.content_textfield, self.generalize_btn,\n self.sum_field])\n', (1793, 1880), True, 'import flet as ft\n'), ((3578, 3673), 'openai.ChatCompletion.create', 'openai.ChatCompletion.create', ([], {'model': '"""gpt-3.5-turbo"""', 'messages': 'composition_analysis_message'}), "(model='gpt-3.5-turbo', messages=\n composition_analysis_message)\n", (3606, 3673), False, 'import openai\n'), ((4597, 4661), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': '(1000)', 'chunk_overlap': '(0)'}), '(chunk_size=1000, chunk_overlap=0)\n', (4627, 4661), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter\n'), ((4713, 4765), 'langchain.llms.OpenAI', 'OpenAI', ([], {'temperature': '(0)', 'openai_api_key': 'openai.api_key'}), '(temperature=0, openai_api_key=openai.api_key)\n', (4719, 4765), False, 'from langchain.llms import OpenAI\n'), ((5344, 5410), 'langchain.prompts.PromptTemplate', 'PromptTemplate', ([], {'template': 'prompt_template', 'input_variables': "['text']"}), "(template=prompt_template, input_variables=['text'])\n", (5358, 5410), False, 'from langchain.prompts import PromptTemplate\n'), ((5504, 5651), 'langchain.chains.summarize.load_summarize_chain', 'load_summarize_chain', (['llm'], {'chain_type': '"""map_reduce"""', 'return_intermediate_steps': '(True)', 'map_prompt': 'PROMPT', 'combine_prompt': 'PROMPT', 'verbose': '(True)'}), "(llm, chain_type='map_reduce',\n return_intermediate_steps=True, map_prompt=PROMPT, combine_prompt=\n PROMPT, verbose=True)\n", (5524, 5651), False, 'from langchain.chains.summarize import load_summarize_chain\n'), ((7113, 7138), 'os.path.exists', 'os.path.exists', (['save_path'], {}), '(save_path)\n', (7127, 7138), False, 'import os\n'), ((7149, 7171), 'os.makedirs', 'os.makedirs', (['save_path'], {}), '(save_path)\n', (7160, 7171), False, 'import os\n'), ((7392, 7422), 'os.path.exists', 'os.path.exists', (['chat_json_path'], {}), '(chat_json_path)\n', (7406, 7422), False, 'import os\n'), ((8550, 8592), 'json.dump', 'json.dump', (['chats', 'f'], {'default': 'str', 'indent': '(4)'}), '(chats, f, default=str, indent=4)\n', (8559, 8592), False, 'import json\n'), ((9068, 9080), 'json.load', 'json.load', (['f'], {}), '(f)\n', (9077, 9080), False, 'import json\n'), ((9595, 9636), 'json.dump', 'json.dump', (['data', 'f'], {'default': 'str', 'indent': '(4)'}), '(data, f, default=str, indent=4)\n', (9604, 9636), False, 'import json\n'), ((9973, 9985), 'json.load', 'json.load', (['f'], {}), '(f)\n', (9982, 9985), False, 'import json\n'), ((10709, 10721), 'json.load', 'json.load', (['f'], {}), '(f)\n', (10718, 10721), False, 'import json\n'), ((11353, 11365), 'json.load', 'json.load', (['f'], {}), '(f)\n', (11362, 11365), False, 'import json\n'), ((11782, 11813), 'json.dump', 'json.dump', (['chatlog', 'f'], {'indent': '(4)'}), '(chatlog, f, indent=4)\n', (11791, 11813), False, 'import json\n'), ((11945, 11957), 'json.load', 'json.load', (['f'], {}), '(f)\n', (11954, 11957), False, 'import json\n'), ((12374, 12405), 'json.dump', 'json.dump', (['chatlog', 'f'], {'indent': '(4)'}), '(chatlog, f, indent=4)\n', (12383, 12405), False, 'import json\n'), ((12544, 12556), 'json.load', 'json.load', (['f'], {}), '(f)\n', (12553, 12556), False, 'import json\n'), ((12934, 12965), 'json.dump', 'json.dump', (['chatlog', 'f'], {'indent': '(4)'}), '(chatlog, f, indent=4)\n', (12943, 12965), False, 'import json\n'), ((16040, 16059), 'os.remove', 'os.remove', (['old_path'], {}), '(old_path)\n', (16049, 16059), False, 'import os\n'), ((16116, 16152), 'shutil.copy', 'shutil.copy', (['path', '"""assets/font.ttf"""'], {}), "(path, 'assets/font.ttf')\n", (16127, 16152), False, 'import shutil\n'), ((16996, 17016), 'json.loads', 'json.loads', (['txt_data'], {}), '(txt_data)\n', (17006, 17016), False, 'import json\n'), ((17409, 17434), 'os.path.exists', 'os.path.exists', (['"""./cache"""'], {}), "('./cache')\n", (17423, 17434), False, 'import os\n'), ((17445, 17467), 'os.makedirs', 'os.makedirs', (['"""./cache"""'], {}), "('./cache')\n", (17456, 17467), False, 'import os\n'), ((23254, 23286), 'flet.TextField', 'ft.TextField', ([], {'label': '"""输入你的apikey"""'}), "(label='输入你的apikey')\n", (23266, 23286), True, 'import flet as ft\n'), ((5111, 5135), 'langchain.docstore.document.Document', 'Document', ([], {'page_content': 't'}), '(page_content=t)\n', (5119, 5135), False, 'from langchain.docstore.document import Document\n'), ((7209, 7223), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (7221, 7223), False, 'from datetime import datetime\n'), ((7483, 7512), 'json.dump', 'json.dump', (['[]', 'f'], {'default': 'str'}), '([], f, default=str)\n', (7492, 7512), False, 'import json\n'), ((7813, 7827), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (7825, 7827), False, 'from datetime import datetime\n'), ((8092, 8104), 'json.load', 'json.load', (['f'], {}), '(f)\n', (8101, 8104), False, 'import json\n'), ((16254, 16290), 'shutil.copy', 'shutil.copy', (['path', '"""assets/font.ttf"""'], {}), "(path, 'assets/font.ttf')\n", (16265, 16290), False, 'import shutil\n'), ((19760, 19809), 'shutil.copy', 'shutil.copy', (['chat_json_path', 'save_file_path.value'], {}), '(chat_json_path, save_file_path.value)\n', (19771, 19809), False, 'import shutil\n'), ((21967, 21986), 'flet.Text', 'ft.Text', (['"""Settings"""'], {}), "('Settings')\n", (21974, 21986), True, 'import flet as ft\n'), ((22005, 22091), 'flet.Column', 'ft.Column', (['[apikey_field, my_wechat, github_page_btn, change_font_btn]'], {'height': '(200)'}), '([apikey_field, my_wechat, github_page_btn, change_font_btn],\n height=200)\n', (22014, 22091), True, 'import flet as ft\n'), ((22342, 22378), 'flet.RoundedRectangleBorder', 'ft.RoundedRectangleBorder', ([], {'radius': '(10)'}), '(radius=10)\n', (22367, 22378), True, 'import flet as ft\n'), ((24929, 24957), 'flet.Text', 'ft.Text', ([], {'size': '(9)', 'italic': '(True)'}), '(size=9, italic=True)\n', (24936, 24957), True, 'import flet as ft\n'), ((26874, 26888), 'flet.Container', 'ft.Container', ([], {}), '()\n', (26886, 26888), True, 'import flet as ft\n'), ((30010, 30079), 'openai.ChatCompletion.create', 'openai.ChatCompletion.create', ([], {'model': '"""gpt-3.5-turbo"""', 'messages': 'message'}), "(model='gpt-3.5-turbo', messages=message)\n", (30038, 30079), False, 'import openai\n'), ((17511, 17537), 'os.path.basename', 'os.path.basename', (['txt_path'], {}), '(txt_path)\n', (17527, 17537), False, 'import os\n'), ((22158, 22201), 'flet.TextButton', 'ft.TextButton', (['"""保存"""'], {'on_click': 'save_settings'}), "('保存', on_click=save_settings)\n", (22171, 22201), True, 'import flet as ft\n'), ((22216, 22261), 'flet.TextButton', 'ft.TextButton', (['"""取消"""'], {'on_click': 'cancel_settings'}), "('取消', on_click=cancel_settings)\n", (22229, 22261), True, 'import flet as ft\n'), ((23406, 23429), 'flet.Text', 'ft.Text', (['"""欢迎使用BillyGPT"""'], {}), "('欢迎使用BillyGPT')\n", (23413, 23429), True, 'import flet as ft\n'), ((23452, 23510), 'flet.Column', 'ft.Column', (['[apikey_field_open, get_apikey_btn]'], {'tight': '(True)'}), '([apikey_field_open, get_apikey_btn], tight=True)\n', (23461, 23510), True, 'import flet as ft\n'), ((29217, 29267), 'flet.Row', 'ft.Row', ([], {'controls': '[chat_text, chat_btn, chatPO_btn]'}), '(controls=[chat_text, chat_btn, chatPO_btn])\n', (29223, 29267), True, 'import flet as ft\n'), ((32038, 32084), 'flet.Text', 'ft.Text', (['f"""正在分析提示词组成结构,请耐心等待"""'], {'color': '"""#1cc9a0"""'}), "(f'正在分析提示词组成结构,请耐心等待', color='#1cc9a0')\n", (32045, 32084), True, 'import flet as ft\n'), ((32276, 32344), 'flet.Text', 'ft.Text', (['f"""提示词组成结构分析完毕,正在根据组成结构逐步生成详细结果,耗时较长,请耐心等待"""'], {'color': '"""#1cc9a0"""'}), "(f'提示词组成结构分析完毕,正在根据组成结构逐步生成详细结果,耗时较长,请耐心等待', color='#1cc9a0')\n", (32283, 32344), True, 'import flet as ft\n'), ((949, 977), 'flet.dropdown.Option', 'ft.dropdown.Option', (['"""system"""'], {}), "('system')\n", (967, 977), True, 'import flet as ft\n'), ((996, 1022), 'flet.dropdown.Option', 'ft.dropdown.Option', (['"""user"""'], {}), "('user')\n", (1014, 1022), True, 'import flet as ft\n'), ((1041, 1072), 'flet.dropdown.Option', 'ft.dropdown.Option', (['"""assistant"""'], {}), "('assistant')\n", (1059, 1072), True, 'import flet as ft\n'), ((9477, 9491), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (9489, 9491), False, 'from datetime import datetime\n'), ((19405, 19466), 'flet.Text', 'Text', (['f"""出现如下报错\n{e}\n请检查导入的聊天记录是否正确,或联系开发者微信B1lli_official"""'], {}), '(f"""出现如下报错\n{e}\n请检查导入的聊天记录是否正确,或联系开发者微信B1lli_official""")\n', (19409, 19466), False, 'from flet import ElevatedButton, FilePicker, FilePickerResultEvent, Page, Row, Text, icons\n'), ((23586, 23645), 'flet.ElevatedButton', 'ft.ElevatedButton', ([], {'text': '"""开始使用"""', 'on_click': 'save_settings_open'}), "(text='开始使用', on_click=save_settings_open)\n", (23603, 23645), True, 'import flet as ft\n'), ((25184, 25207), 'flet.padding.only', 'ft.padding.only', ([], {'top': '(15)'}), '(top=15)\n', (25199, 25207), True, 'import flet as ft\n'), ((25292, 25333), 'flet.animation.Animation', 'ft.animation.Animation', (['(400)', '"""decelerate"""'], {}), "(400, 'decelerate')\n", (25314, 25333), True, 'import flet as ft\n'), ((26617, 26654), 'flet.Column', 'ft.Column', ([], {'scroll': '"""auto"""', 'expand': '(True)'}), "(scroll='auto', expand=True)\n", (26626, 26654), True, 'import flet as ft\n'), ((25755, 25809), 'flet.Icon', 'ft.Icon', (['ft.icons.SEARCH_ROUNDED'], {'size': '(15)', 'opacity': '(0.9)'}), '(ft.icons.SEARCH_ROUNDED, size=15, opacity=0.9)\n', (25762, 25809), True, 'import flet as ft\n'), ((25992, 26157), 'flet.TextField', 'ft.TextField', ([], {'border_color': '"""transparent"""', 'height': '(20)', 'text_size': '(12)', 'content_padding': '(2)', 'cursor_color': '"""black"""', 'cursor_width': '(1)', 'hint_text': '"""搜索模板"""', 'on_change': 'None'}), "(border_color='transparent', height=20, text_size=12,\n content_padding=2, cursor_color='black', cursor_width=1, hint_text=\n '搜索模板', on_change=None)\n", (26004, 26157), True, 'import flet as ft\n'), ((28471, 28485), 'datetime.datetime.now', 'datetime.now', ([], {}), '()\n', (28483, 28485), False, 'from datetime import datetime\n')]
#################################################################################### # Copyright 2022 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # https://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. #################################################################################### # Author: Adam Paternostro # Summary: Create vector embeddings using Vertex AI # To setup your environemtn # python3 -m venv .venv # source .venv/bin/activate # pip install --only-binary :all: greenlet # pip install langchain pip install langchain==0.0.307 # pip install google-cloud-aiplatform # run it: python sample-prompt-vector-embeddings.py # deactivate import langchain from langchain.llms import VertexAI from langchain.embeddings import VertexAIEmbeddings # https://github.com/GoogleCloudPlatform/vertex-ai-samples/blob/main/notebooks/official/matching_engine/sdk_matching_engine_create_stack_overflow_embeddings_vertex.ipynb embeddings = VertexAIEmbeddings() text = "This text will be encoded for semantic search." text_embedding = embeddings.embed_query(text) print(text_embedding)
[ "langchain.embeddings.VertexAIEmbeddings" ]
[((1403, 1423), 'langchain.embeddings.VertexAIEmbeddings', 'VertexAIEmbeddings', ([], {}), '()\n', (1421, 1423), False, 'from langchain.embeddings import VertexAIEmbeddings\n')]
import os import langchain.text_splitter from langchain import PromptTemplate, LLMChain from langchain.callbacks.manager import CallbackManager from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler from langchain.llms import LlamaCpp try: from extensions.telegram_bot.source.generators.abstract_generator import AbstractGenerator except ImportError: from source.generators.abstract_generator import AbstractGenerator # Place where path to LLM file stored llm = None # Callbacks support token-wise streaming class Generator(AbstractGenerator): model_change_allowed = False # if model changing allowed without stopping. preset_change_allowed = True # if preset_file changing allowed. def __init__(self, model_path, n_ctx=2048, seed=0, n_gpu_layers=0): callback_manager = CallbackManager([StreamingStdOutCallbackHandler()]) self.llm = LlamaCpp( model_path=model_path, n_ctx=n_ctx, callback_manager=callback_manager, verbose=True, ) def generate_answer( self, prompt, generation_params, eos_token, stopping_strings, default_answer, turn_template="", **kwargs ): if "max_tokens" in generation_params: llm.max_tokens = generation_params["max_tokens"] if "temperature" in generation_params: llm.temperature = generation_params["temperature"] if "top_p" in generation_params: llm.top_p = generation_params["top_p"] if "top_k" in generation_params: llm.top_k = generation_params["top_k"] prompt_template = PromptTemplate(template="{prompt}", input_variables=["prompt"]) llm.stop = stopping_strings llm_chain = LLMChain(prompt=prompt_template, llm=self.llm) answer = llm_chain.run(prompt) return answer def tokens_count(self, text: str): splitter = langchain.text_splitter.TokenTextSplitter() length = len(splitter.split_text(text)) return length def get_model_list( self, ): bins = [] for i in os.listdir("models"): if i.endswith(".bin"): bins.append(i) return bins def load_model(self, model_file: str): pass
[ "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.llms.LlamaCpp", "langchain.LLMChain", "langchain.PromptTemplate" ]
[((905, 1003), 'langchain.llms.LlamaCpp', 'LlamaCpp', ([], {'model_path': 'model_path', 'n_ctx': 'n_ctx', 'callback_manager': 'callback_manager', 'verbose': '(True)'}), '(model_path=model_path, n_ctx=n_ctx, callback_manager=\n callback_manager, verbose=True)\n', (913, 1003), False, 'from langchain.llms import LlamaCpp\n'), ((1631, 1694), 'langchain.PromptTemplate', 'PromptTemplate', ([], {'template': '"""{prompt}"""', 'input_variables': "['prompt']"}), "(template='{prompt}', input_variables=['prompt'])\n", (1645, 1694), False, 'from langchain import PromptTemplate, LLMChain\n'), ((1751, 1797), 'langchain.LLMChain', 'LLMChain', ([], {'prompt': 'prompt_template', 'llm': 'self.llm'}), '(prompt=prompt_template, llm=self.llm)\n', (1759, 1797), False, 'from langchain import PromptTemplate, LLMChain\n'), ((2113, 2133), 'os.listdir', 'os.listdir', (['"""models"""'], {}), "('models')\n", (2123, 2133), False, 'import os\n'), ((851, 883), 'langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler', 'StreamingStdOutCallbackHandler', ([], {}), '()\n', (881, 883), False, 'from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n')]
import os import openai from dotenv import load_dotenv from langchain.chat_models import AzureChatOpenAI from langchain.chat_models import ChatOpenAI from langchain.embeddings import OpenAIEmbeddings from langchain.callbacks.base import BaseCallbackHandler from langchain.vectorstores import FAISS from langchain.chains import ConversationalRetrievalChain from langchain.chains.conversational_retrieval.base import BaseConversationalRetrievalChain from langchain.prompts import PromptTemplate from langchain.document_loaders import (UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader) import langchain.text_splitter as text_splitter from langchain.text_splitter import (RecursiveCharacterTextSplitter, CharacterTextSplitter) from typing import List import streamlit import glob REQUEST_TIMEOUT_DEFAULT = 10 TEMPERATURE_DEFAULT = 0.0 CHAT_MODEL_NAME_DEFAULT = "gpt-3.5-turbo" OPENAI_EMBEDDING_DEPLOYMENT_NAME_DEFAULT = "text-embedding-ada-002" CHUNK_SIZE_DEFAULT = 1000 CHUNK_OVERLAP_DEFAULT = 0 class StreamHandler(BaseCallbackHandler): def __init__(self, container, initial_text=""): self.container = container self.text = initial_text def on_llm_new_token(self, token: str, **kwargs) -> None: self.text += token self.container.markdown(self.text) class DocChatbot: llm: ChatOpenAI condense_question_llm: ChatOpenAI embeddings: OpenAIEmbeddings vector_db: FAISS chatchain: BaseConversationalRetrievalChain # configuration for API calls request_timeout: int temperature: float chat_model_name : str api_key : str def init_llm_openai(self, streaming: bool, condense_question_container = None, answer_container = None) -> None: # init for LLM using openai.com api self.llm = ChatOpenAI( temperature=self.temperature, openai_api_key=self.api_key, request_timeout=self.request_timeout, model=self.chat_model_name, # Model name is needed for openai.com only streaming=streaming, callbacks=[StreamHandler(answer_container)] if streaming else [] ) # type: ignore if streaming: self.condense_question_llm = ChatOpenAI( temperature=self.temperature, openai_api_key=self.api_key, request_timeout=self.request_timeout, streaming=True, model=self.chat_model_name, callbacks=[StreamHandler(condense_question_container, "🤔...")] ) # type: ignore else: self.condense_question_llm = self.llm def init_llm_azure(self, streaming: bool, condense_question_container = None, answer_container = None) -> None: # init for LLM using Azure OpenAI Service assert(os.getenv("OPENAI_GPT_DEPLOYMENT_NAME") is not None) assert(os.getenv("OPENAI_API_BASE") is not None) assert(os.getenv("OPENAI_EMBEDDING_DEPLOYMENT_NAME") is not None) assert(len(self.api_key) == 32) self.llm = AzureChatOpenAI( deployment_name=os.getenv("OPENAI_GPT_DEPLOYMENT_NAME"), temperature=self.temperature, openai_api_version="2023-05-15", openai_api_type="azure", openai_api_base=os.getenv("OPENAI_API_BASE"), openai_api_key=self.api_key, request_timeout=self.request_timeout, streaming=streaming, callbacks=[StreamHandler(answer_container)] if streaming else [] ) # type: ignore if streaming: self.condense_question_llm = AzureChatOpenAI( deployment_name=os.getenv("OPENAI_GPT_DEPLOYMENT_NAME"), temperature=self.temperature, openai_api_version="2023-05-15", openai_api_type="azure", openai_api_base=os.getenv("OPENAI_API_BASE"), openai_api_key=os.getenv("OPENAI_API_KEY"), request_timeout=self.request_timeout, model=self.chat_model_name, streaming=True, callbacks=[StreamHandler(condense_question_container, "🤔...")] ) # type: ignore else: self.condense_question_llm = self.llm def __init__(self) -> None: #init for LLM and Embeddings, without support for streaming #load environment variables load_dotenv() assert(os.getenv("OPENAI_API_KEY") is not None) self.api_key = str(os.getenv("OPENAI_API_KEY")) self.request_timeout = REQUEST_TIMEOUT_DEFAULT if os.getenv("REQUEST_TIMEOUT") is None else int(os.getenv("REQUEST_TIMEOUT")) self.temperature = TEMPERATURE_DEFAULT if os.getenv("TEMPERATURE") is None else float(os.getenv("TEMPERATURE")) self.chat_model_name = CHAT_MODEL_NAME_DEFAULT if os.getenv("CHAT_MODEL_NAME") is None else str(os.getenv("CHAT_MODEL_NAME")) #check if user is using API from openai.com or Azure OpenAI Service by inspecting the api key if self.api_key.startswith("sk-"): # user is using API from openai.com assert(len(self.api_key) == 51) self.init_llm_openai(False) else: # user is using Azure OpenAI Service self.init_llm_azure(False) embedding_deployment = OPENAI_EMBEDDING_DEPLOYMENT_NAME_DEFAULT if os.getenv("OPENAI_EMBEDDING_DEPLOYMENT_NAME") is None else str(os.getenv("OPENAI_EMBEDDING_DEPLOYMENT_NAME")) self.embeddings = OpenAIEmbeddings( deployment=embedding_deployment, chunk_size=1 ) # type: ignore def init_streaming(self, condense_question_container, answer_container) -> None: #init for LLM and Embeddings, with support for streaming if self.api_key.startswith("sk-"): # user is using API from openai.com self.init_llm_openai(True, condense_question_container, answer_container) else: # user is using Azure OpenAI Service self.init_llm_azure(True, condense_question_container, answer_container) # def init_streaming(self, condense_question_container, answer_container) -> None: # #init for LLM and Embeddings, with support for streaming # api_key = str(os.getenv("OPENAI_API_KEY")) # temperature=float(os.getenv("TEMPERATURE")) # request_timeout=int(os.getenv("REQUEST_TIMEOUT")) # model_name=str(os.getenv("CHAT_MODEL_NAME")) # if api_key.startswith("sk-"): # # user is using API from openai.com # self.llm = ChatOpenAI( # temperature=temperature, # openai_api_key=api_key, # request_timeout=request_timeout, # streaming=True, # model=model_name, # callbacks=[StreamHandler(answer_container)] # ) # type: ignore # self.condense_question_llm = ChatOpenAI( # temperature=temperature, # openai_api_key=api_key, # request_timeout=request_timeout, # streaming=True, # model=model_name, # callbacks=[StreamHandler(condense_question_container, "🤔...")] # ) # type: ignore # else: # # user is using Azure OpenAI Service # self.llm = AzureChatOpenAI( # deployment_name=os.getenv("OPENAI_GPT_DEPLOYMENT_NAME"), # temperature=temperature, # openai_api_version="2023-05-15", # openai_api_type="azure", # openai_api_base=os.getenv("OPENAI_API_BASE"), # openai_api_key=os.getenv("OPENAI_API_KEY"), # request_timeout=request_timeout, # model=model_name, # streaming=True, # callbacks=[StreamHandler(answer_container)] # ) # type: ignore # self.condense_question_llm = AzureChatOpenAI( # deployment_name=os.getenv("OPENAI_GPT_DEPLOYMENT_NAME"), # temperature=temperature, # openai_api_version="2023-05-15", # openai_api_type="azure", # openai_api_base=os.getenv("OPENAI_API_BASE"), # openai_api_key=os.getenv("OPENAI_API_KEY"), # request_timeout=request_timeout, # model=model_name, # streaming=True, # callbacks=[StreamHandler(condense_question_container, "🤔...")] # ) # type: ignore def init_chatchain(self, chain_type : str = "stuff") -> None: # init for ConversationalRetrievalChain CONDENSE_QUESTION_PROMPT = PromptTemplate.from_template("""Given the following conversation and a follow up input, rephrase the standalone question. The standanlone question to be generated should be in the same language with the input. For example, if the input is in Chinese, the follow up question or the standalone question below should be in Chinese too. Chat History: {chat_history} Follow Up Input: {question} Standalone Question:""" ) # stuff chain_type seems working better than others self.chatchain = ConversationalRetrievalChain.from_llm(llm=self.llm, retriever=self.vector_db.as_retriever(), condense_question_prompt=CONDENSE_QUESTION_PROMPT, condense_question_llm=self.condense_question_llm, chain_type=chain_type, return_source_documents=True, verbose=False) # combine_docs_chain_kwargs=dict(return_map_steps=False)) # get answer from query, return answer and source documents def get_answer_with_source(self, query, chat_history): result = self.chatchain({ "question": query, "chat_history": chat_history }, return_only_outputs=True) return result['answer'], result['source_documents'] # get answer from query. # This function is for streamlit app and the chat history is in a format aligned with openai api def get_answer(self, query, chat_history): ''' Here's the format for chat history: [{"role": "assistant", "content": "How can I help you?"}, {"role": "user", "content": "What is your name?"}] The input for the Chain is in a format like this: [("How can I help you?", "What is your name?")] That is, it's a list of question and answer pairs. So need to transform the chat history to the format for the Chain ''' chat_history_for_chain = [] for i in range(0, len(chat_history), 2): chat_history_for_chain.append(( chat_history[i]['content'], chat_history[i+1]['content'] if chat_history[i+1] is not None else "" )) result = self.chatchain({ "question": query, "chat_history": chat_history_for_chain }, return_only_outputs=True) return result['answer'], result['source_documents'] # load vector db from local def load_vector_db_from_local(self, path: str, index_name: str): self.vector_db = FAISS.load_local(path, self.embeddings, index_name) print(f"Loaded vector db from local: {path}/{index_name}") # save vector db to local def save_vector_db_to_local(self, path: str, index_name: str): FAISS.save_local(self.vector_db, path, index_name) print("Vector db saved to local") # split documents, generate embeddings and ingest to vector db def init_vector_db_from_documents(self, file_list: List[str]): chunk_size = CHUNK_SIZE_DEFAULT if os.getenv("CHUNK_SIZE") is None else int(os.getenv("CHUNK_SIZE")) chunk_overlap = CHUNK_OVERLAP_DEFAULT if os.getenv("CHUNK_OVERLAP") is None else int(os.getenv("CHUNK_OVERLAP")) text_splitter = RecursiveCharacterTextSplitter(chunk_size=chunk_size, chunk_overlap=chunk_overlap) docs = [] for file in file_list: print(f"Loading file: {file}") ext_name = os.path.splitext(file)[-1] # print(ext_name) if ext_name == ".pptx": loader = UnstructuredPowerPointLoader(file) elif ext_name == ".docx": loader = UnstructuredWordDocumentLoader(file) elif ext_name == ".pdf": loader = PyPDFLoader(file) elif ext_name == ".csv": loader = CSVLoader(file_path=file) elif ext_name == ".xml": loader = MWDumpLoader(file_path=file, encoding="utf8") else: # process .txt, .html loader = UnstructuredFileLoader(file) doc = loader.load_and_split(text_splitter) docs.extend(doc) print("Processed document: " + file) print("Generating embeddings and ingesting to vector db.") self.vector_db = FAISS.from_documents(docs, self.embeddings) print("Vector db initialized.") # Get indexes available def get_available_indexes(self, path: str): return [os.path.splitext(os.path.basename(file))[0] for file in glob.glob(f"{path}/*.faiss")]
[ "langchain.document_loaders.UnstructuredWordDocumentLoader", "langchain.document_loaders.UnstructuredFileLoader", "langchain.vectorstores.FAISS.load_local", "langchain.text_splitter.RecursiveCharacterTextSplitter", "langchain.document_loaders.UnstructuredPowerPointLoader", "langchain.vectorstores.FAISS.save_local", "langchain.vectorstores.FAISS.from_documents", "langchain.document_loaders.CSVLoader", "langchain.document_loaders.MWDumpLoader", "langchain.document_loaders.PyPDFLoader", "langchain.prompts.PromptTemplate.from_template", "langchain.embeddings.OpenAIEmbeddings" ]
[((4507, 4520), 'dotenv.load_dotenv', 'load_dotenv', ([], {}), '()\n', (4518, 4520), False, 'from dotenv import load_dotenv\n'), ((5613, 5676), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'deployment': 'embedding_deployment', 'chunk_size': '(1)'}), '(deployment=embedding_deployment, chunk_size=1)\n', (5629, 5676), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((8836, 9340), 'langchain.prompts.PromptTemplate.from_template', 'PromptTemplate.from_template', (['"""Given the following conversation and a follow up input, rephrase the standalone question. \n The standanlone question to be generated should be in the same language with the input. \n For example, if the input is in Chinese, the follow up question or the standalone question below should be in Chinese too.\n Chat History:\n {chat_history}\n\n Follow Up Input:\n {question}\n\n Standalone Question:"""'], {}), '(\n """Given the following conversation and a follow up input, rephrase the standalone question. \n The standanlone question to be generated should be in the same language with the input. \n For example, if the input is in Chinese, the follow up question or the standalone question below should be in Chinese too.\n Chat History:\n {chat_history}\n\n Follow Up Input:\n {question}\n\n Standalone Question:"""\n )\n', (8864, 9340), False, 'from langchain.prompts import PromptTemplate\n'), ((11740, 11791), 'langchain.vectorstores.FAISS.load_local', 'FAISS.load_local', (['path', 'self.embeddings', 'index_name'], {}), '(path, self.embeddings, index_name)\n', (11756, 11791), False, 'from langchain.vectorstores import FAISS\n'), ((11965, 12015), 'langchain.vectorstores.FAISS.save_local', 'FAISS.save_local', (['self.vector_db', 'path', 'index_name'], {}), '(self.vector_db, path, index_name)\n', (11981, 12015), False, 'from langchain.vectorstores import FAISS\n'), ((12448, 12535), 'langchain.text_splitter.RecursiveCharacterTextSplitter', 'RecursiveCharacterTextSplitter', ([], {'chunk_size': 'chunk_size', 'chunk_overlap': 'chunk_overlap'}), '(chunk_size=chunk_size, chunk_overlap=\n chunk_overlap)\n', (12478, 12535), False, 'from langchain.text_splitter import RecursiveCharacterTextSplitter, CharacterTextSplitter\n'), ((13530, 13573), 'langchain.vectorstores.FAISS.from_documents', 'FAISS.from_documents', (['docs', 'self.embeddings'], {}), '(docs, self.embeddings)\n', (13550, 13573), False, 'from langchain.vectorstores import FAISS\n'), ((2909, 2948), 'os.getenv', 'os.getenv', (['"""OPENAI_GPT_DEPLOYMENT_NAME"""'], {}), "('OPENAI_GPT_DEPLOYMENT_NAME')\n", (2918, 2948), False, 'import os\n'), ((2977, 3005), 'os.getenv', 'os.getenv', (['"""OPENAI_API_BASE"""'], {}), "('OPENAI_API_BASE')\n", (2986, 3005), False, 'import os\n'), ((3034, 3079), 'os.getenv', 'os.getenv', (['"""OPENAI_EMBEDDING_DEPLOYMENT_NAME"""'], {}), "('OPENAI_EMBEDDING_DEPLOYMENT_NAME')\n", (3043, 3079), False, 'import os\n'), ((4536, 4563), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (4545, 4563), False, 'import os\n'), ((4604, 4631), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (4613, 4631), False, 'import os\n'), ((3198, 3237), 'os.getenv', 'os.getenv', (['"""OPENAI_GPT_DEPLOYMENT_NAME"""'], {}), "('OPENAI_GPT_DEPLOYMENT_NAME')\n", (3207, 3237), False, 'import os\n'), ((3391, 3419), 'os.getenv', 'os.getenv', (['"""OPENAI_API_BASE"""'], {}), "('OPENAI_API_BASE')\n", (3400, 3419), False, 'import os\n'), ((4691, 4719), 'os.getenv', 'os.getenv', (['"""REQUEST_TIMEOUT"""'], {}), "('REQUEST_TIMEOUT')\n", (4700, 4719), False, 'import os\n'), ((4737, 4765), 'os.getenv', 'os.getenv', (['"""REQUEST_TIMEOUT"""'], {}), "('REQUEST_TIMEOUT')\n", (4746, 4765), False, 'import os\n'), ((4817, 4841), 'os.getenv', 'os.getenv', (['"""TEMPERATURE"""'], {}), "('TEMPERATURE')\n", (4826, 4841), False, 'import os\n'), ((4861, 4885), 'os.getenv', 'os.getenv', (['"""TEMPERATURE"""'], {}), "('TEMPERATURE')\n", (4870, 4885), False, 'import os\n'), ((4945, 4973), 'os.getenv', 'os.getenv', (['"""CHAT_MODEL_NAME"""'], {}), "('CHAT_MODEL_NAME')\n", (4954, 4973), False, 'import os\n'), ((4991, 5019), 'os.getenv', 'os.getenv', (['"""CHAT_MODEL_NAME"""'], {}), "('CHAT_MODEL_NAME')\n", (5000, 5019), False, 'import os\n'), ((5477, 5522), 'os.getenv', 'os.getenv', (['"""OPENAI_EMBEDDING_DEPLOYMENT_NAME"""'], {}), "('OPENAI_EMBEDDING_DEPLOYMENT_NAME')\n", (5486, 5522), False, 'import os\n'), ((5540, 5585), 'os.getenv', 'os.getenv', (['"""OPENAI_EMBEDDING_DEPLOYMENT_NAME"""'], {}), "('OPENAI_EMBEDDING_DEPLOYMENT_NAME')\n", (5549, 5585), False, 'import os\n'), ((12237, 12260), 'os.getenv', 'os.getenv', (['"""CHUNK_SIZE"""'], {}), "('CHUNK_SIZE')\n", (12246, 12260), False, 'import os\n'), ((12278, 12301), 'os.getenv', 'os.getenv', (['"""CHUNK_SIZE"""'], {}), "('CHUNK_SIZE')\n", (12287, 12301), False, 'import os\n'), ((12352, 12378), 'os.getenv', 'os.getenv', (['"""CHUNK_OVERLAP"""'], {}), "('CHUNK_OVERLAP')\n", (12361, 12378), False, 'import os\n'), ((12396, 12422), 'os.getenv', 'os.getenv', (['"""CHUNK_OVERLAP"""'], {}), "('CHUNK_OVERLAP')\n", (12405, 12422), False, 'import os\n'), ((12647, 12669), 'os.path.splitext', 'os.path.splitext', (['file'], {}), '(file)\n', (12663, 12669), False, 'import os\n'), ((12766, 12800), 'langchain.document_loaders.UnstructuredPowerPointLoader', 'UnstructuredPowerPointLoader', (['file'], {}), '(file)\n', (12794, 12800), False, 'from langchain.document_loaders import UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader\n'), ((13763, 13791), 'glob.glob', 'glob.glob', (['f"""{path}/*.faiss"""'], {}), "(f'{path}/*.faiss')\n", (13772, 13791), False, 'import glob\n'), ((3760, 3799), 'os.getenv', 'os.getenv', (['"""OPENAI_GPT_DEPLOYMENT_NAME"""'], {}), "('OPENAI_GPT_DEPLOYMENT_NAME')\n", (3769, 3799), False, 'import os\n'), ((3969, 3997), 'os.getenv', 'os.getenv', (['"""OPENAI_API_BASE"""'], {}), "('OPENAI_API_BASE')\n", (3978, 3997), False, 'import os\n'), ((4030, 4057), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (4039, 4057), False, 'import os\n'), ((12864, 12900), 'langchain.document_loaders.UnstructuredWordDocumentLoader', 'UnstructuredWordDocumentLoader', (['file'], {}), '(file)\n', (12894, 12900), False, 'from langchain.document_loaders import UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader\n'), ((13724, 13746), 'os.path.basename', 'os.path.basename', (['file'], {}), '(file)\n', (13740, 13746), False, 'import os\n'), ((12963, 12980), 'langchain.document_loaders.PyPDFLoader', 'PyPDFLoader', (['file'], {}), '(file)\n', (12974, 12980), False, 'from langchain.document_loaders import UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader\n'), ((13043, 13068), 'langchain.document_loaders.CSVLoader', 'CSVLoader', ([], {'file_path': 'file'}), '(file_path=file)\n', (13052, 13068), False, 'from langchain.document_loaders import UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader\n'), ((13131, 13176), 'langchain.document_loaders.MWDumpLoader', 'MWDumpLoader', ([], {'file_path': 'file', 'encoding': '"""utf8"""'}), "(file_path=file, encoding='utf8')\n", (13143, 13176), False, 'from langchain.document_loaders import UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader\n'), ((13258, 13286), 'langchain.document_loaders.UnstructuredFileLoader', 'UnstructuredFileLoader', (['file'], {}), '(file)\n', (13280, 13286), False, 'from langchain.document_loaders import UnstructuredPowerPointLoader, UnstructuredWordDocumentLoader, PyPDFLoader, UnstructuredFileLoader, CSVLoader, MWDumpLoader\n')]
import os import re import langchain import paperqa import paperscraper from langchain.base_language import BaseLanguageModel from langchain.tools import BaseTool from pypdf.errors import PdfReadError def paper_scraper(search: str, pdir: str = "query") -> dict: try: return paperscraper.search_papers(search, pdir=pdir) except KeyError: return {} def paper_search(llm, query): prompt = langchain.prompts.PromptTemplate( input_variables=["question"], template=""" I would like to find scholarly papers to answer this question: {question}. Your response must be at most 10 words long. 'A search query that would bring up papers that can answer this question would be: '""", ) query_chain = langchain.chains.llm.LLMChain(llm=llm, prompt=prompt) if not os.path.isdir("./query"): # todo: move to ckpt os.mkdir("query/") search = query_chain.run(query) print("\nSearch:", search) papers = paper_scraper(search, pdir=f"query/{re.sub(' ', '', search)}") return papers def scholar2result_llm(llm, query, k=5, max_sources=2): """Useful to answer questions that require technical knowledge. Ask a specific question.""" papers = paper_search(llm, query) if len(papers) == 0: return "Not enough papers found" docs = paperqa.Docs(llm=llm) not_loaded = 0 for path, data in papers.items(): try: docs.add(path, data["citation"]) except (ValueError, FileNotFoundError, PdfReadError): not_loaded += 1 print(f"\nFound {len(papers.items())} papers but couldn't load {not_loaded}") answer = docs.query(query, k=k, max_sources=max_sources).formatted_answer return answer class Scholar2ResultLLM(BaseTool): name = "LiteratureSearch" description = ( "Useful to answer questions that require technical " "knowledge. Ask a specific question." ) llm: BaseLanguageModel = None def __init__(self, llm): super().__init__() self.llm = llm def _run(self, query) -> str: return scholar2result_llm(self.llm, query) async def _arun(self, query) -> str: """Use the tool asynchronously.""" raise NotImplementedError("this tool does not support async")
[ "langchain.prompts.PromptTemplate", "langchain.chains.llm.LLMChain" ]
[((419, 753), 'langchain.prompts.PromptTemplate', 'langchain.prompts.PromptTemplate', ([], {'input_variables': "['question']", 'template': '"""\n I would like to find scholarly papers to answer\n this question: {question}. Your response must be at\n most 10 words long.\n \'A search query that would bring up papers that can answer\n this question would be: \'"""'}), '(input_variables=[\'question\'], template=\n """\n I would like to find scholarly papers to answer\n this question: {question}. Your response must be at\n most 10 words long.\n \'A search query that would bring up papers that can answer\n this question would be: \'"""\n )\n', (451, 753), False, 'import langchain\n'), ((786, 839), 'langchain.chains.llm.LLMChain', 'langchain.chains.llm.LLMChain', ([], {'llm': 'llm', 'prompt': 'prompt'}), '(llm=llm, prompt=prompt)\n', (815, 839), False, 'import langchain\n'), ((1360, 1381), 'paperqa.Docs', 'paperqa.Docs', ([], {'llm': 'llm'}), '(llm=llm)\n', (1372, 1381), False, 'import paperqa\n'), ((289, 334), 'paperscraper.search_papers', 'paperscraper.search_papers', (['search'], {'pdir': 'pdir'}), '(search, pdir=pdir)\n', (315, 334), False, 'import paperscraper\n'), ((851, 875), 'os.path.isdir', 'os.path.isdir', (['"""./query"""'], {}), "('./query')\n", (864, 875), False, 'import os\n'), ((907, 925), 'os.mkdir', 'os.mkdir', (['"""query/"""'], {}), "('query/')\n", (915, 925), False, 'import os\n'), ((1042, 1065), 're.sub', 're.sub', (['""" """', '""""""', 'search'], {}), "(' ', '', search)\n", (1048, 1065), False, 'import re\n')]
""" Class for Langchain chain, this chain makes a request to OpenAI to provide information in a given location and time period. """ import os import logging from pathlib import Path import langchain PROMPT_STRING = """ You just gave historical information for {location} around the time period of {time_period} and \n this is the answer you gave {previous_response}. \n Provide bullet points in relation to what you said previously with more information around this topic: {selected_text}. \n Provide no filler or conversation response, just detailed bullet points.\n If you are unsure of the history for this time period, explain that the for this area and time period is limited. \n\n """ class FollowUpChain(): def __init__(self, openai_api_key, model): super().__init__() self.openai_api_key = openai_api_key self.llm_chains = [] # Buffer to store LLMChain objects self.model = model # Add LLMChain objects to the buffer self.add_chain(self.get_history_chain()) def add_chain(self, llm_chain): self.llm_chains.append(llm_chain) def get_template(self, input): # Create the first PromptTemplate prompt = langchain.PromptTemplate( input_variables=input, template=PROMPT_STRING ) return prompt def get_history_chain(self): prompt_1 = self.get_template(input=["location", "time_period", "previous_response", "selected_text"]) return langchain.LLMChain(llm=langchain.OpenAI( openai_api_key=self.openai_api_key, model_name=self.model), prompt=prompt_1) def run(self, input_text): llm_1 = self.llm_chains[0].run(input_text) # Return the results return { "response": llm_1, }
[ "langchain.OpenAI", "langchain.PromptTemplate" ]
[((1201, 1272), 'langchain.PromptTemplate', 'langchain.PromptTemplate', ([], {'input_variables': 'input', 'template': 'PROMPT_STRING'}), '(input_variables=input, template=PROMPT_STRING)\n', (1225, 1272), False, 'import langchain\n'), ((1512, 1587), 'langchain.OpenAI', 'langchain.OpenAI', ([], {'openai_api_key': 'self.openai_api_key', 'model_name': 'self.model'}), '(openai_api_key=self.openai_api_key, model_name=self.model)\n', (1528, 1587), False, 'import langchain\n')]
from typing import TYPE_CHECKING if TYPE_CHECKING: from langchain.chains import LLMChain from langchain.prompts.few_shot import FewShotPromptTemplate def get_prompt(is_zh: bool = False, sydney: bool = False) -> 'FewShotPromptTemplate': from langchain.prompts.few_shot import FewShotPromptTemplate from langchain.prompts.prompt import PromptTemplate langwhat_example_prompt = PromptTemplate( input_variables=["question", "answer"], template="Q: {question}\n{answer}" ) langwhat_examples = [ { "question": "vJzDRsEKDa0", "answer": """ Might be: A YouTube video ID Description: The ID "vJzDRsEKDa0" is most likely a unique identifier for a specific video on YouTube. This alphanumeric code is automatically assigned to every video uploaded on the platform and is used to access the video directly or share it with other users. """, }, { "question": "https://langchain.readthedocs.io/", "answer": """ Might be: A website or online documentation Description: This website/document provides information about LangChain, a technology or platform that could be used for language processing or analysis. """, }, ] langwhat_examples_zh = [ { "question": "vJzDRsEKDa0", "answer": """ 可能是: YouTube 视频 ID 描述: “vJzDRsEKDa0”这个ID很可能是YouTube上特定视频的唯一标识符。 每上传一个视频,该平台都会自动分配这个字母数字代码,并用于直接访问或与其他用户共享该视频。 """, }, { "question": "https://langchain.readthedocs.io/", "answer": """ 可能是: 一个网站或在线文档 描述: 这个网站/文件提供有关LangChain的信息,它是一种可用于语言处理或分析的技术或平台。 """, }, ] prefix_sydney = ''' You are a helpful assistant that answer questions in the format below, you must answer the question only, do not ask what I want to know more about: """ '''.lstrip() suffix_sydney = '''"""\n\nBEGIN!\n\nQ: {input}''' suffix_openai = '''\nQ: {input}''' langwhat_prompt = FewShotPromptTemplate( example_prompt=langwhat_example_prompt, examples=langwhat_examples_zh if is_zh else langwhat_examples, suffix=suffix_sydney if sydney else suffix_openai, prefix=prefix_sydney if sydney else '', input_variables=["input"], ) return langwhat_prompt def get_llm_chain( is_zh: bool = False, sydney: bool = False, cookie_path: str | None = None ) -> 'LLMChain': from langchain.llms import OpenAIChat from langchain.chains import LLMChain from .llm import EdgeLLM if sydney: if cookie_path is None: raise ValueError("cookie_path is required for sydney.") llm = EdgeLLM(bing_cookie_path=cookie_path) else: llm = OpenAIChat() # type: ignore langwhat_prompt = get_prompt(is_zh=is_zh, sydney=sydney) chain = LLMChain(llm=llm, prompt=langwhat_prompt) return chain def parse_standard_answer_format(answer: str) -> tuple[str, str]: ans_lines = answer.strip().splitlines() might_be = ans_lines[1] description = '\n'.join(ans_lines[3:]) description = description.replace("\n", ",\r\n") # description = description.replace(",", ",\r\n") # description = description.replace(",", ",\r\n") # description = description.replace("。", "。\r\n") # description = description.replace(".", ".\r\n") return might_be, description def parse_chain_response(chain_response: dict[str, str]) -> tuple[str, str, str]: """Parse the response from the chain. Args: chain_response (dict[str, str]): The response from the chain. Returns: tuple[str, str]: The first element is the might be, the second element is the description. """ answer = chain_response['text'] return split_edgegpt_answer(answer) def split_edgegpt_answer(answer: str) -> tuple[str, str, str]: answer = answer.strip() if answer.startswith('['): references, answer_body = answer.split('\n\n', maxsplit=1) references = references.strip() answer_body = answer_body.strip() else: references = '' answer_body = answer return references, *parse_standard_answer_format(answer_body) def use_langchain_sqlite_cache(): import langchain from langchain.cache import SQLiteCache from .config import LANGCHAIN_SQLITE_CACHE_FILE langchain.llm_cache = SQLiteCache(database_path=str(LANGCHAIN_SQLITE_CACHE_FILE))
[ "langchain.chains.LLMChain", "langchain.prompts.prompt.PromptTemplate", "langchain.llms.OpenAIChat", "langchain.prompts.few_shot.FewShotPromptTemplate" ]
[((400, 498), 'langchain.prompts.prompt.PromptTemplate', 'PromptTemplate', ([], {'input_variables': "['question', 'answer']", 'template': '"""Q: {question}\n{answer}"""'}), '(input_variables=[\'question\', \'answer\'], template=\n """Q: {question}\n{answer}""")\n', (414, 498), False, 'from langchain.prompts.prompt import PromptTemplate\n'), ((1960, 2216), 'langchain.prompts.few_shot.FewShotPromptTemplate', 'FewShotPromptTemplate', ([], {'example_prompt': 'langwhat_example_prompt', 'examples': '(langwhat_examples_zh if is_zh else langwhat_examples)', 'suffix': '(suffix_sydney if sydney else suffix_openai)', 'prefix': "(prefix_sydney if sydney else '')", 'input_variables': "['input']"}), "(example_prompt=langwhat_example_prompt, examples=\n langwhat_examples_zh if is_zh else langwhat_examples, suffix=\n suffix_sydney if sydney else suffix_openai, prefix=prefix_sydney if\n sydney else '', input_variables=['input'])\n", (1981, 2216), False, 'from langchain.prompts.few_shot import FewShotPromptTemplate\n'), ((2800, 2841), 'langchain.chains.LLMChain', 'LLMChain', ([], {'llm': 'llm', 'prompt': 'langwhat_prompt'}), '(llm=llm, prompt=langwhat_prompt)\n', (2808, 2841), False, 'from langchain.chains import LLMChain\n'), ((2698, 2710), 'langchain.llms.OpenAIChat', 'OpenAIChat', ([], {}), '()\n', (2708, 2710), False, 'from langchain.llms import OpenAIChat\n')]
from langchain.text_splitter import CharacterTextSplitter from langchain.vectorstores.faiss import FAISS from langchain.embeddings import OpenAIEmbeddings from langchain.memory import ConversationBufferMemory from langchain.chains import ConversationalRetrievalChain from langchain.chat_models import ChatOpenAI from PyPDF2 import PdfReader import langchain import requests import io import docx import json langchain.verbose = False def get_text_from_files(file_list): text = "" for file in file_list: ext_name = file.split(".")[-1].lower() response = requests.get(file) if ext_name == "pdf": ##load pdf from url pdf_reader = PdfReader(io.BytesIO(response.content)) # pdf_reader = PdfReader(file) for page in pdf_reader.pages: text += page.extract_text() elif ext_name == "txt": #load txt from url text = response.content.decode("utf-8") elif ext_name == "docx": #load docx from url docx_file = io.BytesIO(response.content) doc = docx.Document(docx_file) for para in doc.paragraphs: text += para.text return text def get_text_chunks(text): text_splitter = CharacterTextSplitter( separator="\n", chunk_size=800, chunk_overlap=200, length_function=len, ) chunks = text_splitter.split_text(text) return chunks ##get keys from .env file import os # from dotenv import load_dotenv # load_dotenv() openai_api_key = os.getenv("OPENAI_API_KEY") def get_vectorstore(text_chunks): embeddings = OpenAIEmbeddings(openai_api_key=openai_api_key) vector_store = FAISS.from_texts(texts=text_chunks, embedding=embeddings) return vector_store def get_conversation_chain(vectorstore): llm = ChatOpenAI(openai_api_key=openai_api_key, temperature=0.25) memory = ConversationBufferMemory(memory_key='chat_history', return_messages=True) conversation_chain = ConversationalRetrievalChain.from_llm( llm=llm, retriever=vectorstore.as_retriever(), memory=memory, ) return conversation_chain def process(text_info,files_or_urls,questions): conversation=None if text_info: raw_text = str(text_info) else: raw_text = "" ## allow if only text is input if files_or_urls: raw_text = raw_text+ "\n" + str(get_text_from_files(files_or_urls)) text_chunks = get_text_chunks(raw_text) vectorstore = get_vectorstore(text_chunks) conversation = get_conversation_chain(vectorstore) for user_question in questions: response = conversation({'question': user_question}) chat_history = response['chat_history'] data={} for i, message in enumerate(chat_history): if i % 2 == 0: key=message.content else: data[key]=message.content return data def lambda_handler(event, context): # TODO implement result=process(text_info=event['text_info'],files_or_urls=event['files_or_urls'],questions=event['questions']) return { 'statusCode': 200, 'body': json.dumps(result) }
[ "langchain.text_splitter.CharacterTextSplitter", "langchain.memory.ConversationBufferMemory", "langchain.chat_models.ChatOpenAI", "langchain.vectorstores.faiss.FAISS.from_texts", "langchain.embeddings.OpenAIEmbeddings" ]
[((1580, 1607), 'os.getenv', 'os.getenv', (['"""OPENAI_API_KEY"""'], {}), "('OPENAI_API_KEY')\n", (1589, 1607), False, 'import os\n'), ((1282, 1379), 'langchain.text_splitter.CharacterTextSplitter', 'CharacterTextSplitter', ([], {'separator': '"""\n"""', 'chunk_size': '(800)', 'chunk_overlap': '(200)', 'length_function': 'len'}), "(separator='\\n', chunk_size=800, chunk_overlap=200,\n length_function=len)\n", (1303, 1379), False, 'from langchain.text_splitter import CharacterTextSplitter\n'), ((1660, 1707), 'langchain.embeddings.OpenAIEmbeddings', 'OpenAIEmbeddings', ([], {'openai_api_key': 'openai_api_key'}), '(openai_api_key=openai_api_key)\n', (1676, 1707), False, 'from langchain.embeddings import OpenAIEmbeddings\n'), ((1727, 1784), 'langchain.vectorstores.faiss.FAISS.from_texts', 'FAISS.from_texts', ([], {'texts': 'text_chunks', 'embedding': 'embeddings'}), '(texts=text_chunks, embedding=embeddings)\n', (1743, 1784), False, 'from langchain.vectorstores.faiss import FAISS\n'), ((1861, 1920), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': 'openai_api_key', 'temperature': '(0.25)'}), '(openai_api_key=openai_api_key, temperature=0.25)\n', (1871, 1920), False, 'from langchain.chat_models import ChatOpenAI\n'), ((1955, 2028), 'langchain.memory.ConversationBufferMemory', 'ConversationBufferMemory', ([], {'memory_key': '"""chat_history"""', 'return_messages': '(True)'}), "(memory_key='chat_history', return_messages=True)\n", (1979, 2028), False, 'from langchain.memory import ConversationBufferMemory\n'), ((578, 596), 'requests.get', 'requests.get', (['file'], {}), '(file)\n', (590, 596), False, 'import requests\n'), ((3216, 3234), 'json.dumps', 'json.dumps', (['result'], {}), '(result)\n', (3226, 3234), False, 'import json\n'), ((707, 735), 'io.BytesIO', 'io.BytesIO', (['response.content'], {}), '(response.content)\n', (717, 735), False, 'import io\n'), ((1072, 1100), 'io.BytesIO', 'io.BytesIO', (['response.content'], {}), '(response.content)\n', (1082, 1100), False, 'import io\n'), ((1119, 1143), 'docx.Document', 'docx.Document', (['docx_file'], {}), '(docx_file)\n', (1132, 1143), False, 'import docx\n')]
import zipfile from langchain.chat_models import ChatOpenAI from langchain.schema import ( HumanMessage, SystemMessage ) import langchain from langchain.cache import SQLiteCache langchain.llm_cache = SQLiteCache( database_path=".langchain.db" ) # caches queries that are the same. def generate_code(question, model_type, api_key): """ Generate Python code script. The script should only include code, no comments. What is returned is code that needs some light modification. """ task = "Generate Python Code Script. The script should only include code, no comments." chat_model = ChatOpenAI(temperature=0, model=model_type, openai_api_key = api_key) response = chat_model( messages=[ SystemMessage(content=task), HumanMessage(content=question) ], ) code = response.content return format_code(code) def format_code(code): """ Format the generated code to remove read_csv if its there. """ csv_line_index = code.find("read_csv") if csv_line_index > 0: before_csv_line_index = code[0:csv_line_index].rfind("\n") before_csv = ( "" if before_csv_line_index == -1 else code[0:before_csv_line_index] ) after_csv = code[csv_line_index:] after_csv_line_index = after_csv.find("\n") after_csv = ( "" if after_csv_line_index == -1 else after_csv[after_csv_line_index:] ) return before_csv + after_csv return code def format_primer(primer_desc, primer_code, question): """ Format the primer description and code. """ return f'"""\n{primer_desc}{question}\n"""\n{primer_code}' def generate_primer(df, df_name): """ Generate a primer for visualization. Include all the necessary imports. """ primer_desc = """ If a graph is needed, label the x and y axes appropriately. Add a title. Use st.plotly or plotly.express to visualize." Using Python version 3.9.12, only create a script using the dataframe "df" if the user asked for a graph. Do not give a y/n answer in the code or try to take any user input. Graph should be created using st.plotly or plotly.express. If there is only a single row, manipulate it to be able to chart the data. IF the given sql query is only one column, manipulate it to get it in the right format for a chart if it has multiple rows. If its just one column one row, don't chart, just display the table. Make sure that variable names are the same as the sql. """ primer_code = ( "import pandas as pd\nimport streamlit as st\nimport plotly.express as px\n" ) primer_code += f"df = {df_name}.copy()\n" return primer_desc, primer_code def unzip(file): with zipfile.ZipFile(file, 'r') as zip_ref: zip_ref.extractall()
[ "langchain.schema.HumanMessage", "langchain.schema.SystemMessage", "langchain.chat_models.ChatOpenAI", "langchain.cache.SQLiteCache" ]
[((211, 253), 'langchain.cache.SQLiteCache', 'SQLiteCache', ([], {'database_path': '""".langchain.db"""'}), "(database_path='.langchain.db')\n", (222, 253), False, 'from langchain.cache import SQLiteCache\n'), ((621, 688), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'temperature': '(0)', 'model': 'model_type', 'openai_api_key': 'api_key'}), '(temperature=0, model=model_type, openai_api_key=api_key)\n', (631, 688), False, 'from langchain.chat_models import ChatOpenAI\n'), ((2781, 2807), 'zipfile.ZipFile', 'zipfile.ZipFile', (['file', '"""r"""'], {}), "(file, 'r')\n", (2796, 2807), False, 'import zipfile\n'), ((749, 776), 'langchain.schema.SystemMessage', 'SystemMessage', ([], {'content': 'task'}), '(content=task)\n', (762, 776), False, 'from langchain.schema import HumanMessage, SystemMessage\n'), ((790, 820), 'langchain.schema.HumanMessage', 'HumanMessage', ([], {'content': 'question'}), '(content=question)\n', (802, 820), False, 'from langchain.schema import HumanMessage, SystemMessage\n')]
import langchain from langchain.chat_models import ChatOpenAI from langchain_core.tools import Tool langchain.verbose = True langchain.debug = True def get_chat(): return ChatOpenAI(model_name="gpt-3.5-turbo", temperature=0) def my_super_func(params): return 42 if __name__ == "__main__": tools = [ Tool.from_function( func=my_super_func, name="The_Answer", description="生命、宇宙、そして万物について究極の疑問の答え" ), ]
[ "langchain_core.tools.Tool.from_function", "langchain.chat_models.ChatOpenAI" ]
[((178, 231), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'model_name': '"""gpt-3.5-turbo"""', 'temperature': '(0)'}), "(model_name='gpt-3.5-turbo', temperature=0)\n", (188, 231), False, 'from langchain.chat_models import ChatOpenAI\n'), ((326, 427), 'langchain_core.tools.Tool.from_function', 'Tool.from_function', ([], {'func': 'my_super_func', 'name': '"""The_Answer"""', 'description': '"""生命、宇宙、そして万物について究極の疑問の答え"""'}), "(func=my_super_func, name='The_Answer', description=\n '生命、宇宙、そして万物について究極の疑問の答え')\n", (344, 427), False, 'from langchain_core.tools import Tool\n')]
import os import tkinter as tk from tkinter import Label, Entry, Button, Text, Scrollbar import langchain from langchain.text_splitter import RecursiveCharacterTextSplitter from langchain.chat_models import ChatOpenAI class ProjectEvaluatorApp: def __init__(self, root): self.root = root self.root.title('Project Evaluator') self.experience_label = Label(root, text='Rate your coding ability on a scale of 1 to 5:') self.experience_input = Entry(root) self.frameworks_label = Label(root, text='Which coding languages have you used?') self.frameworks_input = Entry(root) self.learn_label = Label(root, text='What would you like to learn?') self.learn_input = Entry(root) self.evaluate_button = Button(root, text='Evaluate Project', command=self.evaluate_project) self.result_text = Text(root, wrap='word', height=10, width=50) self.result_scrollbar = Scrollbar(root, command=self.result_text.yview) self.result_text['yscrollcommand'] = self.result_scrollbar.set self.experience_label.pack() self.experience_input.pack() self.frameworks_label.pack() self.frameworks_input.pack() self.learn_label.pack() self.learn_input.pack() self.evaluate_button.pack() self.result_text.pack(side='left', padx=5) self.result_scrollbar.pack(side='right', fill='y') # Corrected argument def evaluate_project(self): experience_level = self.experience_input.get() frameworks_used = self.frameworks_input.get() what_to_learn = self.learn_input.get() CREATE_CONTEXT_PROMPT = f"""Summarize this student's coding experience level and learning goals using the following information: The student's experience level on a scale of 0 to 5: {experience_level} The coding languages the student has used: {frameworks_used} The thing the student wants to learn: {what_to_learn}""" llm = ChatOpenAI( openai_api_key=os.environ["OPENAI_API_KEY"], model_name='gpt-4', temperature=0.0 ) context = llm.invoke(CREATE_CONTEXT_PROMPT) PROMPT = f"""You are evaluating potential project ideas. Each idea has a hierarchy based on difficulty and applicability in the industry, rated on a scale from 1 to 5. Your goal is to organize these ideas in ascending order of difficulty and descending order of applicability: {context} Given their coding experience, what simple example project could the student build to learn the topic they want to learn? """ project = llm.invoke(str(PROMPT)) self.result_text.config(state='normal') # Enable editing self.result_text.delete('1.0', 'end') # Clear preavious content self.result_text.insert('1.0', f"{project}\n") # Insert new content self.result_text.config(state='disabled') # Disable editing if __name__ == '__main__': root = tk.Tk() app = ProjectEvaluatorApp(root) root.mainloop()
[ "langchain.chat_models.ChatOpenAI" ]
[((3005, 3012), 'tkinter.Tk', 'tk.Tk', ([], {}), '()\n', (3010, 3012), True, 'import tkinter as tk\n'), ((380, 446), 'tkinter.Label', 'Label', (['root'], {'text': '"""Rate your coding ability on a scale of 1 to 5:"""'}), "(root, text='Rate your coding ability on a scale of 1 to 5:')\n", (385, 446), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((479, 490), 'tkinter.Entry', 'Entry', (['root'], {}), '(root)\n', (484, 490), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((524, 581), 'tkinter.Label', 'Label', (['root'], {'text': '"""Which coding languages have you used?"""'}), "(root, text='Which coding languages have you used?')\n", (529, 581), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((614, 625), 'tkinter.Entry', 'Entry', (['root'], {}), '(root)\n', (619, 625), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((654, 703), 'tkinter.Label', 'Label', (['root'], {'text': '"""What would you like to learn?"""'}), "(root, text='What would you like to learn?')\n", (659, 703), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((731, 742), 'tkinter.Entry', 'Entry', (['root'], {}), '(root)\n', (736, 742), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((775, 843), 'tkinter.Button', 'Button', (['root'], {'text': '"""Evaluate Project"""', 'command': 'self.evaluate_project'}), "(root, text='Evaluate Project', command=self.evaluate_project)\n", (781, 843), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((872, 916), 'tkinter.Text', 'Text', (['root'], {'wrap': '"""word"""', 'height': '(10)', 'width': '(50)'}), "(root, wrap='word', height=10, width=50)\n", (876, 916), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((949, 996), 'tkinter.Scrollbar', 'Scrollbar', (['root'], {'command': 'self.result_text.yview'}), '(root, command=self.result_text.yview)\n', (958, 996), False, 'from tkinter import Label, Entry, Button, Text, Scrollbar\n'), ((2024, 2120), 'langchain.chat_models.ChatOpenAI', 'ChatOpenAI', ([], {'openai_api_key': "os.environ['OPENAI_API_KEY']", 'model_name': '"""gpt-4"""', 'temperature': '(0.0)'}), "(openai_api_key=os.environ['OPENAI_API_KEY'], model_name='gpt-4',\n temperature=0.0)\n", (2034, 2120), False, 'from langchain.chat_models import ChatOpenAI\n')]
from __future__ import annotations from collections import OrderedDict from typing import Any, Dict, List, Optional, Tuple import langchain import numpy as np import orjson import pandas as pd from langchain.cache import InMemoryCache from peewee import ModelSelect, fn from .constants import * from .orm import Knowledge, Solution, Space, Task, database_proxy from .utils import format_config, get_llm SAVE_OPTIONS = orjson.OPT_SERIALIZE_NUMPY | orjson.OPT_SERIALIZE_DATACLASS langchain.llm_cache = InMemoryCache() def ingest_experience( history_df: pd.DataFrame, task_desc: Optional[Dict[str, str]], space_desc: str, space_id: str, ) -> Space: """Ingest experience from history dataframe. Parameters ---------- history_df: pandas.DataFrame The history of configurations. task_desc The task descriptions. space_desc The space description. space_id The space id. Returns ------- None Notes ----- The history_df should be a dataframe with the following columns: - CONFIG_0 - CONFIG_1 - ... - CONFIG_N - METRIC The task_desc should be a dict with the following format: { "task_id_0": "task_description_0", "task_id_1": "task_description_1", ... "task_id_N": "task_description_N" } """ history_df = history_df.drop_duplicates() history_df["TASK_ID"] = history_df["TASK_ID"].astype(str) if task_desc is None: task_desc = {k: "." for k in history_df["TASK_ID"].unique()} experience_df = history_df[ history_df[["TASK_ID", "SCORE"]] .groupby("TASK_ID") .rank(method="first", ascending=False)["SCORE"] <= TOP_K ] quantile_info = get_quantile_stat(experience_df) space = _ingest_space(space_id, space_desc, quantile_info) _ingest_task(history_df, task_desc) _ingest_solution(history_df, space) print("Ingested experience into database.") # save db database_proxy.commit() return space def _ingest_space( space_id: str, space_desc: str, quantile_info: Dict[str, List[float]] ) -> Space: with database_proxy.atomic(): try: space = Space.get(Space.space_id == space_id) print(f"Space {space_id} already exists, skip ingestion.") return space except: space = Space.create( space_id=space_id, desc=space_desc, quantile_info=orjson.dumps(quantile_info, option=SAVE_OPTIONS), ) print("Ingested space into database.") # save db database_proxy.commit() return space def _ingest_task(history_df: pd.DataFrame, row_task_desc: Dict[str, str]) -> None: embeddings = get_llm("embedding")() with database_proxy.atomic(): for task_id in history_df["TASK_ID"].unique(): try: Task.get(Task.task_id == task_id) print(f"Task {task_id} already exists, skip ingestion.") continue except Task.DoesNotExist: task_desc = canonicalize_task(row_task_desc[task_id]) embedding = np.asarray( embeddings.embed_query(task_desc), dtype=np.float32 ).tobytes() Task.create( task_id=task_id, embedding=embedding, desc=task_desc, row_desc=row_task_desc[task_id], ) print("Ingested task into database.") # save db database_proxy.commit() def _ingest_solution(history_df: pd.DataFrame, space: Space) -> None: with database_proxy.atomic(): solutions = [] for _, row in ( history_df.groupby("TASK_ID") .apply(lambda x: x.sort_values("SCORE", ascending=False)) .reset_index(drop=True) .iterrows() ): row_config_dict = { k[7:]: v for k, v in row.to_dict().items() if k.startswith("CONFIG_") } row_config = orjson.dumps((row_config_dict), option=SAVE_OPTIONS) try: Solution.get( Solution.task == row["TASK_ID"], Solution.space == space.space_id, Solution.row_config == row_config, ) print( f"Solution '{row_config}' for space {space.space_id} and task {row['TASK_ID']} already exists, skip ingestion." ) continue except Solution.DoesNotExist: solutions.append( { "task": row["TASK_ID"], "space": space, "metric": row["SCORE"], "row_config": row_config, "extra_metric": row.get("EXTRA_METRIC", ""), "demo": canonicalize_config( row_config_dict, orjson.loads(space.quantile_info) ), } ) Solution.insert_many(solutions).execute() print("Ingested solution into database.") # save db database_proxy.commit() def canonicalize_task(row_task_desc: str) -> str: """ Canonicalize task to a string. Parameters ---------- row_task_desc: str The task description. Returns ------- str The canonicalized task description. """ task_desc = f"Task: {row_task_desc}" return task_desc def canonicalize_config( config_term: Dict[str, Any], quantile_info: Dict[str, List[float]] ) -> str: """ Canonicalize configuration to a string. Parameters ---------- config_term The configuration term. quantile_info The meta train info for stat. Record quantile information for each hyperparameter. Returns ------- str The canonicalized configuration. """ demo = format_config(config_term, quantile_info, bin_map=bin_map) return demo def get_quantile_stat(experience_df: pd.DataFrame) -> Dict[str, List[float]]: """ Get quantile stat from experience dataframe. Parameters ---------- experience_df: pandas.DataFrame The experience dataframe. Returns ------- Dict[str, List[float]] The quantile stat. """ # map column name 'CONFIG_{}' to '{}', rest of columns are deleted meta_train_df_for_stat = experience_df.rename( columns=lambda x: x[7:] if x.startswith("CONFIG_") else None ) del meta_train_df_for_stat[None] try: quantile_info = { col: np.quantile( meta_train_df_for_stat[col][~meta_train_df_for_stat[col].isna()], list(bin_map.keys()), method="nearest", ) .astype(np.float32) .tolist() for col in meta_train_df_for_stat.select_dtypes(include=[np.number]) } except: quantile_info = { col: np.quantile( meta_train_df_for_stat[col][~meta_train_df_for_stat[col].isna()], list(bin_map.keys()), interpolation="nearest", ) .astype(np.float32) .tolist() for col in meta_train_df_for_stat.select_dtypes(include=[np.number]) } return quantile_info def gen_experience_per_task(space: Space, task: Task) -> str: """ Generate experience content from space and task. Parameters ---------- task_desc: str The task description. demos: str The demos. example: Configurations 1: ... Configurations 2: ... Configurations 3: ... Returns ------- str The experience content. """ demos = _gen_experience_demos(space, task) content = f"{task.desc}\n{demos}" return content def _gen_experience_demos(space: Space, task: Task) -> str: """ Generate experience demonstrations Parameters ---------- space: Space The space. task: Task The task. Returns ------- str The experience demonstrations per task. """ solutions = ( Solution.select() .where(Solution.task_id == task.task_id, Solution.space_id == space.space_id) .order_by(Solution.metric.desc()) .limit(TOP_K) ) demos = "\n".join( [ f"Configuration {i+1}: {solution.demo}" for i, solution in enumerate(solutions) ] ) return demos def _get_best_relevant_solutions(space: Space, task_desc: str) -> ModelSelect: """ Get the best relevant solution for a task. The relevance is measured by cosine similarity between task description embeddings, which affects the order of results. Parameters ---------- space: Space The space. task_desc: str The task description. Returns ------- ModelSelect The best relevant solution. """ SolutionAlias = Solution.alias() order_key = Task.embedding.cosine_distance(task_desc) subquery = ( SolutionAlias.select( SolutionAlias.demo, Task.task_id, Task.desc, Task.embedding, fn.ROW_NUMBER() .over( partition_by=[SolutionAlias.space, SolutionAlias.task], order_by=[SolutionAlias.metric.desc()], ) .alias("rnk"), ) .where(SolutionAlias.space == space) .join(Task, on=(SolutionAlias.task == Task.task_id)) .order_by(order_key) .alias("subq") ) query = ( Solution.select(subquery.c.task_id, subquery.c.demo, subquery.c.desc) .from_(subquery) .where(subquery.c.rnk <= TOP_K) ) return query def _get_best_solutions(space: Space) -> ModelSelect: """ Get the best solution for each task. Parameters ---------- space: Space The space. Returns ------- ModelSelect The best solution for each task. """ SolutionAlias = Solution.alias() subquery = ( SolutionAlias.select( SolutionAlias.demo, Task.task_id, Task.desc, Task.embedding, fn.ROW_NUMBER() .over( partition_by=[SolutionAlias.space, SolutionAlias.task], order_by=[SolutionAlias.metric.desc()], ) .alias("rnk"), ) .where(SolutionAlias.space == space) .join(Task, on=(SolutionAlias.task == Task.task_id)) .alias("subq") ) query = ( Solution.select(subquery.c.task_id, subquery.c.demo, subquery.c.desc) .from_(subquery) .where(subquery.c.rnk <= TOP_K) ) return query def gen_experience( space: Space, task_desc: Optional[str] = None ) -> Tuple[List[str], List[str]]: """ Generate experience content from space and optional task description. Parameters ---------- space: Space The space. task_desc The task description. Returns ------- List[str] The experience content. """ if task_desc is None: query = _get_best_solutions(space) else: query = _get_best_relevant_solutions(space, task_desc) examples = OrderedDict() for solution in query: if solution.task_id not in examples: examples[solution.task_id] = [solution.desc] examples[solution.task_id].append( f"Configuration {len(examples[solution.task_id])}: {solution.demo}" ) return list(examples.keys()), ["\n".join(e) for e in examples.values()]
[ "langchain.cache.InMemoryCache" ]
[((506, 521), 'langchain.cache.InMemoryCache', 'InMemoryCache', ([], {}), '()\n', (519, 521), False, 'from langchain.cache import InMemoryCache\n'), ((11527, 11540), 'collections.OrderedDict', 'OrderedDict', ([], {}), '()\n', (11538, 11540), False, 'from collections import OrderedDict\n'), ((4157, 4207), 'orjson.dumps', 'orjson.dumps', (['row_config_dict'], {'option': 'SAVE_OPTIONS'}), '(row_config_dict, option=SAVE_OPTIONS)\n', (4169, 4207), False, 'import orjson\n'), ((2552, 2600), 'orjson.dumps', 'orjson.dumps', (['quantile_info'], {'option': 'SAVE_OPTIONS'}), '(quantile_info, option=SAVE_OPTIONS)\n', (2564, 2600), False, 'import orjson\n'), ((5086, 5119), 'orjson.loads', 'orjson.loads', (['space.quantile_info'], {}), '(space.quantile_info)\n', (5098, 5119), False, 'import orjson\n'), ((10463, 10478), 'peewee.fn.ROW_NUMBER', 'fn.ROW_NUMBER', ([], {}), '()\n', (10476, 10478), False, 'from peewee import ModelSelect, fn\n'), ((9438, 9453), 'peewee.fn.ROW_NUMBER', 'fn.ROW_NUMBER', ([], {}), '()\n', (9451, 9453), False, 'from peewee import ModelSelect, fn\n')]
from typing import Any, Dict, List, Optional from langchain import PromptTemplate ,LLMChain import langchain from langchain.chat_models import ChatOpenAI ,AzureChatOpenAI from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler import sys import re import argparse import os from langchain.prompts.chat import ( ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate, ) from langchain.schema import AIMessage, HumanMessage, SystemMessage # 英文700单词,800+token, 3800+字符(含空格)。 翻译成中文 1200+字符(字),1400+token, 如果假设英文字符1byte, 中文字2bytes. 则电子存贮中文信息密度高30%-50%(考虑空格:否-是)。印刷信息密度高50-100% # max_words=700 & context=4K 对于普通文章是有余地的,但如包含大量代码或符号,则可能不够. GPT3.5 还是会偷懒,在这种场景下,把前面的润色了,后面的就copy.把 max_words 调小点会增加润色的比例 def split_text_into_chunks(text, max_words=430): # Split the text by newline lines = text.split('\n') chunks = [] current_chunk_words = 0 current_chunk = "" #按行处理分段 《max_word for line in lines: words = line.split(' ') if current_chunk_words + len(words) > max_words: # 分段截止到前一行 chunks.append(current_chunk) # 当前行新起一段 current_chunk = line current_chunk_words = len(words) if current_chunk_words > max_words : # 遇到超长的一行 # split the current_chunk into sentences 句子拆分 split_sentences = current_chunk.split('.') # 将句子合并成子段,直到子段的总词数达到max_words current_sub_chunk = "" current_word_count = 0 for sentence in split_sentences: sentence_words = sentence.split() if current_word_count + len(sentence_words) <= max_words: current_sub_chunk += sentence+". " current_word_count += len(sentence_words) else: chunks.append(current_sub_chunk.strip()) current_sub_chunk = sentence current_word_count = len(sentence_words) if current_sub_chunk: chunks.append(current_sub_chunk.strip()) current_chunk = "" current_chunk_words = 0 else: # Add a newline character to keep the original format current_chunk += '\n' + line current_chunk_words += len(words) # Add the last chunk if current_chunk: chunks.append(current_chunk) return chunks ############# main start ############################# prompt_messages_1stTrans = [ SystemMessage( content=( """你是一位精通简体中文的专业翻译,曾参与《纽约时报》和《经济学人》中文版的翻译工作,因此对于新闻和博客文章的翻译有深入的理解。我希望你能帮我将以下英文文章段落翻译成中文,风格与上述杂志的中文版相似。 规则: - 翻译时要准确传达原文事实和背景。 - 保留特定的英文术语、数字或名字,并在其前后加上空格,例如:"生成式 AI 产品","不超过 10 秒"。 - 保留复制原文的所有特殊符号,包括 空格,连续的换行'\n',制表缩进'\t'等 根据内容直译,不要遗漏任何信息""" ) ), HumanMessagePromptTemplate.from_template(""" ### 英文原文: {essay} ### 中文直译结果: """ ,input_variables=["essay"] ) ] prompt_messages_2ndtTrans = [ SystemMessage( content=( """你是一位专业中文翻译,擅长对翻译结果进行二次修改和润色成通俗易懂的中文,我希望你能帮我将以下英文文章的中文翻译结果重新意译和润色。 规则: - 这些博客文章包含机器学习或AI等专业知识相关,注意翻译时术语的准确性 - 保留特定的英文术语、数字或名字,并在其前后加上空格,例如:"生成式 AI 产品","不超过 10 秒"。 - 保留复制原文的所有特殊符号,包括 空格,连续的换行'\n',制表缩进'\t'等 - 基于直译结果重新意译和润色,意译和润色时务必对照原始英文,不要添加也不要遗漏内容,并以让翻译结果通俗易懂,符合中文表达习惯 """ ) ), HumanMessagePromptTemplate.from_template(""" ### 英文原文: {essay} ### 中文直译结果: {trans_1st} ### 请你基于直译结果重新意译和润色,意译和润色时务必对照英文原文,不要添加也不要遗漏内容,让翻译结果通俗易懂,符合中文表达习惯. ### 重新意译和润色后: """ ,input_variables=["essay","trans_1st"] ) ] langchain.verbose = True llmchat=AzureChatOpenAI(streaming=True,deployment_name="gpt35turbo-16k", max_tokens=1500, temperature=0, callbacks=[StreamingStdOutCallbackHandler()]) #llmchat=ChatOpenAI(streaming=True,model_name="gpt-4", max_tokens=1500, temperature=0, callbacks=[StreamingStdOutCallbackHandler()]) chat_prompt_1stTrans = ChatPromptTemplate(messages=prompt_messages_1stTrans) chat_prompt_2ndTrans = ChatPromptTemplate(messages=prompt_messages_2ndtTrans) chain1st = LLMChain(llm=llmchat, prompt=chat_prompt_1stTrans) chain2nd = LLMChain(llm=llmchat, prompt=chat_prompt_2ndTrans) # Parse arguments parser = argparse.ArgumentParser(description='直译+意译') parser.add_argument('fileName', type=str, help='English essay file') args = parser.parse_args() fileName=args.fileName output1stFileName = os.path.splitext(fileName)[0] + '_精译.md' print(f"\n\n######## output_file_name : {output1stFileName}") #fileName='SuperlinearReturns.md' output1stFileName=fileName.split('.')[0]+"_翻译.md" output1stText=f"\n\n###################### {output1stFileName} ##########\n\n" output2ndFileName=fileName.split('.')[0]+"_精译.md" output2ndText=f"\n\n###################### {output2ndFileName} ##########\n\n" output3rdFileName=fileName.split('.')[0]+"_翻译对照.md" output3rdText=f"\n\n###################### {output3rdFileName} ##########\n\n" with open(fileName, 'r', encoding='utf-8') as file: markdown_text = file.read() chunks = split_text_into_chunks(markdown_text) for txt in chunks: print(txt) for i, chunk in enumerate(chunks): #if i!=0 : # continue try : print(f"\n\n\n################################### chunk - {i} \n") inputs1= {"essay": chunk} response1 = chain1st.run(inputs1) output1stText = output1stText + response1 #print (f"\n 第一次翻译结果: " ,response1) inputs2= {"essay": chunk, "trans_1st":response1} response2=chain2nd.run(inputs2 ) output2ndText = output2ndText + response2 output3rdText = output3rdText + "\n\n--------------------------------------------------------------------------------------\n"+chunk+"\n\n------------------------------------------------------------------------------------\n"+response1+"\n\n-----------------------------------------------------------------------------------\n"+response2 except BaseException as e: print("$$$!!!! BaseException : ",e) continue with open(output1stFileName, 'w', encoding='utf-8') as file1: file1.write(output1stText) with open(output2ndFileName, 'w', encoding='utf-8') as file2: file2.write(output2ndText) with open(output3rdFileName, 'w', encoding='utf-8') as file3: file3.write(output3rdText)
[ "langchain.LLMChain", "langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler", "langchain.prompts.chat.ChatPromptTemplate", "langchain.schema.SystemMessage", "langchain.prompts.chat.HumanMessagePromptTemplate.from_template" ]
[((4074, 4127), 'langchain.prompts.chat.ChatPromptTemplate', 'ChatPromptTemplate', ([], {'messages': 'prompt_messages_1stTrans'}), '(messages=prompt_messages_1stTrans)\n', (4092, 4127), False, 'from langchain.prompts.chat import ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate\n'), ((4151, 4205), 'langchain.prompts.chat.ChatPromptTemplate', 'ChatPromptTemplate', ([], {'messages': 'prompt_messages_2ndtTrans'}), '(messages=prompt_messages_2ndtTrans)\n', (4169, 4205), False, 'from langchain.prompts.chat import ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate\n'), ((4218, 4268), 'langchain.LLMChain', 'LLMChain', ([], {'llm': 'llmchat', 'prompt': 'chat_prompt_1stTrans'}), '(llm=llmchat, prompt=chat_prompt_1stTrans)\n', (4226, 4268), False, 'from langchain import PromptTemplate, LLMChain\n'), ((4280, 4330), 'langchain.LLMChain', 'LLMChain', ([], {'llm': 'llmchat', 'prompt': 'chat_prompt_2ndTrans'}), '(llm=llmchat, prompt=chat_prompt_2ndTrans)\n', (4288, 4330), False, 'from langchain import PromptTemplate, LLMChain\n'), ((4361, 4405), 'argparse.ArgumentParser', 'argparse.ArgumentParser', ([], {'description': '"""直译+意译"""'}), "(description='直译+意译')\n", (4384, 4405), False, 'import argparse\n'), ((2690, 2962), 'langchain.schema.SystemMessage', 'SystemMessage', ([], {'content': '"""你是一位精通简体中文的专业翻译,曾参与《纽约时报》和《经济学人》中文版的翻译工作,因此对于新闻和博客文章的翻译有深入的理解。我希望你能帮我将以下英文文章段落翻译成中文,风格与上述杂志的中文版相似。\n规则:\n- 翻译时要准确传达原文事实和背景。\n- 保留特定的英文术语、数字或名字,并在其前后加上空格,例如:"生成式 AI 产品","不超过 10 秒"。\n- 保留复制原文的所有特殊符号,包括 空格,连续的换行\'\n\',制表缩进\'\t\'等\n 根据内容直译,不要遗漏任何信息"""'}), '(content=\n """你是一位精通简体中文的专业翻译,曾参与《纽约时报》和《经济学人》中文版的翻译工作,因此对于新闻和博客文章的翻译有深入的理解。我希望你能帮我将以下英文文章段落翻译成中文,风格与上述杂志的中文版相似。\n规则:\n- 翻译时要准确传达原文事实和背景。\n- 保留特定的英文术语、数字或名字,并在其前后加上空格,例如:"生成式 AI 产品","不超过 10 秒"。\n- 保留复制原文的所有特殊符号,包括 空格,连续的换行\'\n\',制表缩进\'\t\'等\n 根据内容直译,不要遗漏任何信息"""\n )\n', (2703, 2962), False, 'from langchain.schema import AIMessage, HumanMessage, SystemMessage\n'), ((3013, 3127), 'langchain.prompts.chat.HumanMessagePromptTemplate.from_template', 'HumanMessagePromptTemplate.from_template', (['"""\n### 英文原文: \n{essay}\n\n### 中文直译结果: \n"""'], {'input_variables': "['essay']"}), '(\n """\n### 英文原文: \n{essay}\n\n### 中文直译结果: \n""", input_variables=[\'essay\'])\n', (3053, 3127), False, 'from langchain.prompts.chat import ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate\n'), ((3161, 3466), 'langchain.schema.SystemMessage', 'SystemMessage', ([], {'content': '"""你是一位专业中文翻译,擅长对翻译结果进行二次修改和润色成通俗易懂的中文,我希望你能帮我将以下英文文章的中文翻译结果重新意译和润色。\n规则:\n- 这些博客文章包含机器学习或AI等专业知识相关,注意翻译时术语的准确性\n- 保留特定的英文术语、数字或名字,并在其前后加上空格,例如:"生成式 AI 产品","不超过 10 秒"。\n- 保留复制原文的所有特殊符号,包括 空格,连续的换行\'\n\',制表缩进\'\t\'等\n- 基于直译结果重新意译和润色,意译和润色时务必对照原始英文,不要添加也不要遗漏内容,并以让翻译结果通俗易懂,符合中文表达习惯\n"""'}), '(content=\n """你是一位专业中文翻译,擅长对翻译结果进行二次修改和润色成通俗易懂的中文,我希望你能帮我将以下英文文章的中文翻译结果重新意译和润色。\n规则:\n- 这些博客文章包含机器学习或AI等专业知识相关,注意翻译时术语的准确性\n- 保留特定的英文术语、数字或名字,并在其前后加上空格,例如:"生成式 AI 产品","不超过 10 秒"。\n- 保留复制原文的所有特殊符号,包括 空格,连续的换行\'\n\',制表缩进\'\t\'等\n- 基于直译结果重新意译和润色,意译和润色时务必对照原始英文,不要添加也不要遗漏内容,并以让翻译结果通俗易懂,符合中文表达习惯\n"""\n )\n', (3174, 3466), False, 'from langchain.schema import AIMessage, HumanMessage, SystemMessage\n'), ((3517, 3743), 'langchain.prompts.chat.HumanMessagePromptTemplate.from_template', 'HumanMessagePromptTemplate.from_template', (['"""\n### 英文原文: \n{essay}\n\n### 中文直译结果: \n{trans_1st}\n\n### 请你基于直译结果重新意译和润色,意译和润色时务必对照英文原文,不要添加也不要遗漏内容,让翻译结果通俗易懂,符合中文表达习惯.\n### 重新意译和润色后:\n"""'], {'input_variables': "['essay', 'trans_1st']"}), '(\n """\n### 英文原文: \n{essay}\n\n### 中文直译结果: \n{trans_1st}\n\n### 请你基于直译结果重新意译和润色,意译和润色时务必对照英文原文,不要添加也不要遗漏内容,让翻译结果通俗易懂,符合中文表达习惯.\n### 重新意译和润色后:\n"""\n , input_variables=[\'essay\', \'trans_1st\'])\n', (3557, 3743), False, 'from langchain.prompts.chat import ChatPromptTemplate, SystemMessagePromptTemplate, AIMessagePromptTemplate, HumanMessagePromptTemplate\n'), ((4545, 4571), 'os.path.splitext', 'os.path.splitext', (['fileName'], {}), '(fileName)\n', (4561, 4571), False, 'import os\n'), ((3878, 3910), 'langchain.callbacks.streaming_stdout.StreamingStdOutCallbackHandler', 'StreamingStdOutCallbackHandler', ([], {}), '()\n', (3908, 3910), False, 'from langchain.callbacks.streaming_stdout import StreamingStdOutCallbackHandler\n')]