TURX commited on
Commit
827f4e7
·
1 Parent(s): 3d3aed0
Files changed (1) hide show
  1. main.py +7 -4
main.py CHANGED
@@ -8,12 +8,15 @@ from funix.hint import HTML
8
 
9
  low_memory = True # Set to True to run on mobile devices
10
 
 
 
 
11
  ku_gpt_tokenizer = AutoTokenizer.from_pretrained("ku-nlp/gpt2-medium-japanese-char")
12
- chj_gpt_tokenizer = AutoTokenizer.from_pretrained("TURX/chj-gpt2")
13
- wakagpt_tokenizer = AutoTokenizer.from_pretrained("TURX/wakagpt")
14
  ku_gpt_model = AutoModelForCausalLM.from_pretrained("ku-nlp/gpt2-medium-japanese-char")
15
- chj_gpt_model = AutoModelForCausalLM.from_pretrained("TURX/chj-gpt2")
16
- wakagpt_model = AutoModelForCausalLM.from_pretrained("TURX/wakagpt")
17
 
18
  print("Models loaded successfully.")
19
 
 
8
 
9
  low_memory = True # Set to True to run on mobile devices
10
 
11
+ import os
12
+ hf_token = os.environ.get("HF_TOKEN")
13
+
14
  ku_gpt_tokenizer = AutoTokenizer.from_pretrained("ku-nlp/gpt2-medium-japanese-char")
15
+ chj_gpt_tokenizer = AutoTokenizer.from_pretrained("TURX/chj-gpt2", token=hf_token)
16
+ wakagpt_tokenizer = AutoTokenizer.from_pretrained("TURX/wakagpt", token=hf_token)
17
  ku_gpt_model = AutoModelForCausalLM.from_pretrained("ku-nlp/gpt2-medium-japanese-char")
18
+ chj_gpt_model = AutoModelForCausalLM.from_pretrained("TURX/chj-gpt2", token=hf_token)
19
+ wakagpt_model = AutoModelForCausalLM.from_pretrained("TURX/wakagpt", token=hf_token)
20
 
21
  print("Models loaded successfully.")
22