lunarflu HF staff commited on
Commit
b470e0c
·
verified ·
1 Parent(s): c4c5696
Files changed (1) hide show
  1. app.py +2 -0
app.py CHANGED
@@ -1,10 +1,12 @@
1
  import gradio as gr
 
2
  from transformers import AutoTokenizer, AutoModelForCausalLM
3
 
4
  # Load the model and tokenizer (you may need to adjust device_map or other settings depending on your hardware)
5
  tokenizer = AutoTokenizer.from_pretrained("TheBloke/Chronoboros-33B-GPTQ")
6
  model = AutoModelForCausalLM.from_pretrained("TheBloke/Chronoboros-33B-GPTQ", device_map="auto")
7
 
 
8
  def respond(message, history: list[tuple[str, str]], system_message, max_tokens, temperature, top_p):
9
  # Build the prompt using conversation history
10
  prompt = f"{system_message}\n"
 
1
  import gradio as gr
2
+ import spaces
3
  from transformers import AutoTokenizer, AutoModelForCausalLM
4
 
5
  # Load the model and tokenizer (you may need to adjust device_map or other settings depending on your hardware)
6
  tokenizer = AutoTokenizer.from_pretrained("TheBloke/Chronoboros-33B-GPTQ")
7
  model = AutoModelForCausalLM.from_pretrained("TheBloke/Chronoboros-33B-GPTQ", device_map="auto")
8
 
9
10
  def respond(message, history: list[tuple[str, str]], system_message, max_tokens, temperature, top_p):
11
  # Build the prompt using conversation history
12
  prompt = f"{system_message}\n"