Spaces:
Sleeping
Sleeping
import os | |
import cohere | |
from langchain.llms.base import LLM | |
co = cohere.Client(os.environ['COHERE_API_KEY']) # This is my trial API key | |
# 24222433-2a40-410c-af5c-65a5422282fb-ft | |
class CustomLLM(LLM): | |
model: str = os.environ['CODE_GEN_MODEL_ID'] # The custom model we used. | |
def _llm_type(self) -> str: | |
return "custom" | |
def _call(self,prompt: str,stop = None,run_manager = None,) -> str: | |
""" This is where the main logic of the """ | |
if stop is not None: | |
raise ValueError("stop kwargs are not permitted.") | |
response = co.chat(message=f'{prompt}', max_tokens=1700, temperature=0.9) | |
print(response.text) | |
return f' {response.text} ' | |
def _identifying_params(self): | |
"""Get the identifying parameters.""" | |
return {"model_type": f'COHERE_CUSTOM-<{self.model}>'} | |
""" Now, this thing can be used as a custom LLM. Use it in the LLM Chain thing. Done mi boy. """ |