Spaces:
Sleeping
Sleeping
Commit
·
70f5edf
1
Parent(s):
bc7bf8f
model change FinetunedByYalcin5
Browse files
app.py
CHANGED
@@ -18,10 +18,12 @@ tokenizer = None
|
|
18 |
def get_model_and_tokenizer(model_id):
|
19 |
global model, tokenizer
|
20 |
try:
|
|
|
21 |
# Load the tokenizer
|
22 |
tokenizer = AutoTokenizer.from_pretrained(model_id, use_fast=False)
|
23 |
tokenizer.pad_token = tokenizer.eos_token
|
24 |
|
|
|
25 |
# Load the model
|
26 |
model = AutoModelForCausalLM.from_pretrained(model_id) #, device_map="auto")
|
27 |
model.config.use_cache = False
|
@@ -73,7 +75,7 @@ def handle_post_request():
|
|
73 |
return jsonify({"error": "No JSON data provided"}), 400
|
74 |
|
75 |
message = data.get("inputs", "No message provided.")
|
76 |
-
model_id = data.get("model_id", "YALCINKAYA/
|
77 |
|
78 |
try:
|
79 |
# Generate a response from the model
|
|
|
18 |
def get_model_and_tokenizer(model_id):
|
19 |
global model, tokenizer
|
20 |
try:
|
21 |
+
print(f"Loading tokenizer for model_id: {model_id}")
|
22 |
# Load the tokenizer
|
23 |
tokenizer = AutoTokenizer.from_pretrained(model_id, use_fast=False)
|
24 |
tokenizer.pad_token = tokenizer.eos_token
|
25 |
|
26 |
+
print(f"Loading model and for model_id: {model_id}")
|
27 |
# Load the model
|
28 |
model = AutoModelForCausalLM.from_pretrained(model_id) #, device_map="auto")
|
29 |
model.config.use_cache = False
|
|
|
75 |
return jsonify({"error": "No JSON data provided"}), 400
|
76 |
|
77 |
message = data.get("inputs", "No message provided.")
|
78 |
+
model_id = data.get("model_id", "YALCINKAYA/FinetunedByYalcin5") # Default model if not provided
|
79 |
|
80 |
try:
|
81 |
# Generate a response from the model
|