reedmayhew commited on
Commit
c7c3eb0
·
verified ·
1 Parent(s): cb7408b

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -7
app.py CHANGED
@@ -11,7 +11,7 @@ HF_TOKEN = os.environ.get("HF_TOKEN", None)
11
 
12
  DESCRIPTION = '''
13
  <div>
14
- <h1 style="text-align: center;">deepseek-ai/DeepSeek-R1-Distill-Llama-8B</h1>
15
  </div>
16
  '''
17
 
@@ -44,8 +44,8 @@ h1 {
44
  """
45
 
46
  # Load the tokenizer and model
47
- tokenizer = AutoTokenizer.from_pretrained("deepseek-ai/DeepSeek-R1-Distill-Llama-8B")
48
- model = AutoModelForCausalLM.from_pretrained("deepseek-ai/DeepSeek-R1-Distill-Llama-8B", device_map="auto") # to("cuda:0")
49
  terminators = [
50
  tokenizer.eos_token_id,
51
  tokenizer.convert_tokens_to_ids("<|eot_id|>")
@@ -110,15 +110,15 @@ with gr.Blocks(fill_height=True, css=css) as demo:
110
  fill_height=True,
111
  additional_inputs_accordion=gr.Accordion(label="⚙️ Parameters", open=False, render=False),
112
  additional_inputs=[
113
- gr.Slider(minimum=0,
114
- maximum=1,
115
  step=0.1,
116
- value=0.5,
117
  label="Temperature",
118
  render=False),
119
  gr.Slider(minimum=128,
120
  maximum=4096,
121
- step=1,
122
  value=1024,
123
  label="Max new tokens",
124
  render=False ),
 
11
 
12
  DESCRIPTION = '''
13
  <div>
14
+ <h1 style="text-align: center;">DeepSeek-R1-Zero</h1>
15
  </div>
16
  '''
17
 
 
44
  """
45
 
46
  # Load the tokenizer and model
47
+ tokenizer = AutoTokenizer.from_pretrained("reedmayhew/DeepSeek-R1-Refined-Llama-3.1-8B-hf")
48
+ model = AutoModelForCausalLM.from_pretrained("reedmayhew/DeepSeek-R1-Refined-Llama-3.1-8B-hf", device_map="auto") # to("cuda:0")
49
  terminators = [
50
  tokenizer.eos_token_id,
51
  tokenizer.convert_tokens_to_ids("<|eot_id|>")
 
110
  fill_height=True,
111
  additional_inputs_accordion=gr.Accordion(label="⚙️ Parameters", open=False, render=False),
112
  additional_inputs=[
113
+ gr.Slider(minimum=0.6,
114
+ maximum=0.6,
115
  step=0.1,
116
+ value=0.6,
117
  label="Temperature",
118
  render=False),
119
  gr.Slider(minimum=128,
120
  maximum=4096,
121
+ step=64,
122
  value=1024,
123
  label="Max new tokens",
124
  render=False ),