CineAI commited on
Commit
84d7a8d
·
verified ·
1 Parent(s): 5697cd9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -21
README.md CHANGED
@@ -163,27 +163,25 @@ datasets:
163
 
164
  You can use it with a script
165
 
166
- if do_run_lora_model:
167
- if is_own:
168
- model, tokenizer = FastLanguageModel.from_pretrained(
169
- model_name="CineAI/Llama32-3B-CoT",
170
- max_seq_length=max_length,
171
- dtype=dtype,
172
- load_in_4bit=load_in_4bit
173
- )
174
-
175
- FastLanguageModel.for_inference(model)
176
-
177
- inputs = tokenizer.apply_chat_template(
178
- message,
179
- tokenize = True,
180
- add_generation_prompt = True, # Must add for generation
181
- return_tensors = "pt",
182
- ).to(device)
183
-
184
- text_streamer = TextStreamer(tokenizer, skip_prompt = True)
185
- _ = model.generate(input_ids = inputs, streamer = text_streamer, max_new_tokens = max_new_tokens,
186
- use_cache = True, temperature = temperature, min_p = min_p)
187
 
188
  # Uploaded model
189
 
 
163
 
164
  You can use it with a script
165
 
166
+ model, tokenizer = FastLanguageModel.from_pretrained(
167
+ model_name="CineAI/Llama32-3B-CoT",
168
+ max_seq_length=max_length,
169
+ dtype=dtype,
170
+ load_in_4bit=load_in_4bit
171
+ )
172
+
173
+ FastLanguageModel.for_inference(model)
174
+
175
+ inputs = tokenizer.apply_chat_template(
176
+ message,
177
+ tokenize = True,
178
+ add_generation_prompt = True, # Must add for generation
179
+ return_tensors = "pt",
180
+ ).to(device)
181
+
182
+ text_streamer = TextStreamer(tokenizer, skip_prompt = True)
183
+ _ = model.generate(input_ids = inputs, streamer = text_streamer, max_new_tokens = max_new_tokens,
184
+ use_cache = True, temperature = temperature, min_p = min_p)
 
 
185
 
186
  # Uploaded model
187