Spaces:
Runtime error
Runtime error
import os | |
import torch | |
from langchain import HuggingFacePipeline | |
from transformers import AutoTokenizer | |
import transformers | |
from langchain import HuggingFaceHub | |
def get_openai_chat_model(API_key): | |
try: | |
from langchain.llms import OpenAI | |
except ImportError as err: | |
raise "{}, unable to load openAI. Please install openai and add OPENAIAPI_KEY" | |
os.environ["OPENAI_API_KEY"] = API_key | |
llm = OpenAI() | |
return llm | |
def get_llama_model(temperature=0,api_key=None,max_tokens=2048): | |
model_id = "meta-llama/Llama-2-7b-chat-hf" | |
llm = None | |
try: | |
tokenizer = AutoTokenizer.from_pretrained(model_id) | |
pipeline = transformers.pipeline("text-generation", | |
model=model_id, | |
tokenizer=tokenizer, | |
torch_dtype=torch.bfloat16, | |
trust_remote_code=True, | |
device_map="auto", | |
max_length=1000, | |
eos_token_id=tokenizer.eos_token_id | |
) | |
llm = HuggingFacePipeline(pipeline = pipeline, model_kwargs = {'temperature':temperature}) | |
except: | |
raise "User not autorized to access the Model" | |
return llm | |
def get_model_from_hub(api_key,temperature=0.1,max_tokens=2048,model_id="meta-llama/Llama-2-7b-chat-hf"): | |
llm = HuggingFaceHub(huggingfacehub_api_token=api_key, | |
repo_id=model_id, | |
model_kwargs={"temperature": temperature, "max_new_tokens": max_tokens}) | |
return llm |