1024m commited on
Commit
aa82a83
·
verified ·
1 Parent(s): 7da50f2

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -6
app.py CHANGED
@@ -3,14 +3,13 @@ import torch
3
  import time
4
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
5
  from threading import Thread
 
 
 
6
  print("Loading model and tokenizer...")
7
  model_name = "large-traversaal/Phi-4-Hindi"
8
  tokenizer = AutoTokenizer.from_pretrained(model_name)
9
- model = AutoModelForCausalLM.from_pretrained(
10
- model_name,
11
- torch_dtype=torch.bfloat16,
12
- device_map="auto"
13
- )
14
  print("Model and tokenizer loaded successfully!")
15
  def generate_response(message, temperature, max_new_tokens, top_p):
16
  print(f"Input: {message}")
@@ -30,12 +29,21 @@ def generate_response(message, temperature, max_new_tokens, top_p):
30
  result = []
31
  for text in streamer:
32
  result.append(text)
33
- yield "".join(result)
 
 
 
 
34
  end_time = time.time()
35
  time_taken = end_time - start_time
36
  output_text = "".join(result)
 
 
37
  print(f"Output: {output_text}")
38
  print(f"Time taken: {time_taken:.2f} seconds")
 
 
 
39
  with gr.Blocks() as demo:
40
  gr.Markdown("# Phi-4-Hindi Demo")
41
  with gr.Row():
 
3
  import time
4
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
5
  from threading import Thread
6
+ import time
7
+ import pytz
8
+ from datetime import datetime
9
  print("Loading model and tokenizer...")
10
  model_name = "large-traversaal/Phi-4-Hindi"
11
  tokenizer = AutoTokenizer.from_pretrained(model_name)
12
+ model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=torch.bfloat16, device_map="auto")
 
 
 
 
13
  print("Model and tokenizer loaded successfully!")
14
  def generate_response(message, temperature, max_new_tokens, top_p):
15
  print(f"Input: {message}")
 
29
  result = []
30
  for text in streamer:
31
  result.append(text)
32
+ current_output = "".join(result)
33
+ if current_output.startswith(message):
34
+ yield current_output[len(message):]
35
+ else:
36
+ yield current_output
37
  end_time = time.time()
38
  time_taken = end_time - start_time
39
  output_text = "".join(result)
40
+ if output_text.startswith(message):
41
+ output_text = output_text[len(message):]
42
  print(f"Output: {output_text}")
43
  print(f"Time taken: {time_taken:.2f} seconds")
44
+ pst_timezone = pytz.timezone('America/Los_Angeles')
45
+ current_time_pst = datetime.now(pst_timezone).strftime("%Y-%m-%d %H:%M:%S %Z%z")
46
+ print(f"Current timestamp (PST): {current_time_pst}")
47
  with gr.Blocks() as demo:
48
  gr.Markdown("# Phi-4-Hindi Demo")
49
  with gr.Row():