Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -1,6 +1,9 @@
|
|
1 |
from fastapi import FastAPI
|
2 |
import time
|
3 |
import torch
|
|
|
|
|
|
|
4 |
|
5 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
6 |
device = "cpu" # the device to load the model onto
|
@@ -28,11 +31,12 @@ time5 = time.time()
|
|
28 |
print(time5-time4)
|
29 |
|
30 |
|
31 |
-
tokenizer2 = AutoTokenizer.from_pretrained("google/gemma-2-2b-it")
|
32 |
model2 = AutoModelForCausalLM.from_pretrained(
|
33 |
"google/gemma-2-2b-it",
|
34 |
device_map="auto",
|
35 |
torch_dtype=torch.bfloat16,
|
|
|
36 |
)
|
37 |
|
38 |
|
|
|
1 |
from fastapi import FastAPI
|
2 |
import time
|
3 |
import torch
|
4 |
+
import os
|
5 |
+
|
6 |
+
access_token = os.evn["read_access"]
|
7 |
|
8 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
9 |
device = "cpu" # the device to load the model onto
|
|
|
31 |
print(time5-time4)
|
32 |
|
33 |
|
34 |
+
tokenizer2 = AutoTokenizer.from_pretrained("google/gemma-2-2b-it", token=access_token)
|
35 |
model2 = AutoModelForCausalLM.from_pretrained(
|
36 |
"google/gemma-2-2b-it",
|
37 |
device_map="auto",
|
38 |
torch_dtype=torch.bfloat16,
|
39 |
+
token=access_token
|
40 |
)
|
41 |
|
42 |
|