import os #os.system('git clone https://github.com/tloen/alpaca-lora.git') import streamlit as st from streamlit_chat import message as st_message import bitsandbytes as bnb from transformers import GenerationConfig from transformers import AutoTokenizer, AutoConfig, LlamaForCausalLM, LlamaTokenizer tokenizer = LlamaTokenizer.from_pretrained("wxjiao/alpaca-7b") model2 = LlamaForCausalLM.from_pretrained( "wxjiao/alpaca-7b", load_in_8bit_fp32_cpu_offload=True, from_pretrained="auto", )