MaxLSB commited on
Commit
f6b834f
·
verified ·
1 Parent(s): 00105bc

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -1
app.py CHANGED
@@ -1,4 +1,5 @@
1
  import os
 
2
  import threading
3
  import gradio as gr
4
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
@@ -6,6 +7,8 @@ from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStream
6
  # Hugging Face token
7
  hf_token = os.environ["HUGGINGFACEHUB_API_TOKEN"]
8
 
 
 
9
  # Global model & tokenizer
10
  tokenizer = None
11
  model = None
@@ -36,7 +39,11 @@ def respond(message, max_tokens, temperature, top_p):
36
  eos_token_id=tokenizer.eos_token_id,
37
  )
38
 
39
- thread = threading.Thread(target=model.generate, kwargs=generate_kwargs)
 
 
 
 
40
  thread.start()
41
 
42
  response = ""
 
1
  import os
2
+ import torch
3
  import threading
4
  import gradio as gr
5
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
 
7
  # Hugging Face token
8
  hf_token = os.environ["HUGGINGFACEHUB_API_TOKEN"]
9
 
10
+ torch.set_num_threads(4)
11
+
12
  # Global model & tokenizer
13
  tokenizer = None
14
  model = None
 
39
  eos_token_id=tokenizer.eos_token_id,
40
  )
41
 
42
+ def run():
43
+ with torch.no_grad():
44
+ model.generate(**generate_kwargs)
45
+
46
+ thread = threading.Thread(target=run)
47
  thread.start()
48
 
49
  response = ""