amaltese commited on
Commit
55c0785
·
verified ·
1 Parent(s): 4cbdac3

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -1
app.py CHANGED
@@ -5,7 +5,12 @@ import torch
5
  # Load the Zephyr-7B-Alpha model (fully open and optimized for instruction-following)
6
  MODEL_NAME = "HuggingFaceH4/zephyr-7b-alpha"
7
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
8
- model = AutoModelForCausalLM.from_pretrained(MODEL_NAME, torch_dtype=torch.float16, device_map="auto")
 
 
 
 
 
9
 
10
  # Initialize conversation history if not present
11
  if "conversation" not in st.session_state:
 
5
  # Load the Zephyr-7B-Alpha model (fully open and optimized for instruction-following)
6
  MODEL_NAME = "HuggingFaceH4/zephyr-7b-alpha"
7
  tokenizer = AutoTokenizer.from_pretrained(MODEL_NAME)
8
+ model = AutoModelForCausalLM.from_pretrained(
9
+ MODEL_NAME,
10
+ torch_dtype=torch.float16,
11
+ device_map="cpu", # Forces CPU usage
12
+ low_cpu_mem_usage=True # Helps reduce memory spikes
13
+ )
14
 
15
  # Initialize conversation history if not present
16
  if "conversation" not in st.session_state: