Felladrin commited on
Commit
faf04b5
·
verified ·
1 Parent(s): 84b801b

Update usage code example

Browse files
Files changed (1) hide show
  1. README.md +11 -7
README.md CHANGED
@@ -23,7 +23,7 @@ pip install transformers==4.49.0 torch==2.6.0
23
  ```
24
 
25
  ```python
26
- from transformers import pipeline, TextStreamer, AutoModelForCausalLM, AutoTokenizer
27
  import torch
28
 
29
  model_path = "Felladrin/Qwen2-96M"
@@ -32,15 +32,19 @@ device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
32
  tokenizer = AutoTokenizer.from_pretrained(model_path)
33
  model = AutoModelForCausalLM.from_pretrained(model_path).to(device)
34
  streamer = TextStreamer(tokenizer)
35
- generate = pipeline("text-generation", model=model, tokenizer=tokenizer, device=device)
36
- inputs = tokenizer(prompt, return_tensors="pt").to(device)
37
- model.generate(
38
- inputs.input_ids,
39
- attention_mask=inputs.attention_mask,
40
  streamer=streamer,
41
- max_length=tokenizer.model_max_length,
 
 
42
  eos_token_id=tokenizer.eos_token_id,
43
  pad_token_id=tokenizer.pad_token_id,
 
 
44
  do_sample=True,
45
  repetition_penalty=1.05,
46
  )
 
23
  ```
24
 
25
  ```python
26
+ from transformers import pipeline, AutoModelForCausalLM, AutoTokenizer, TextStreamer
27
  import torch
28
 
29
  model_path = "Felladrin/Qwen2-96M"
 
32
  tokenizer = AutoTokenizer.from_pretrained(model_path)
33
  model = AutoModelForCausalLM.from_pretrained(model_path).to(device)
34
  streamer = TextStreamer(tokenizer)
35
+ generate = pipeline(
36
+ "text-generation",
37
+ model=model,
38
+ tokenizer=tokenizer,
39
+ device=device,
40
  streamer=streamer,
41
+ )
42
+ generate(
43
+ prompt,
44
  eos_token_id=tokenizer.eos_token_id,
45
  pad_token_id=tokenizer.pad_token_id,
46
+ max_length=tokenizer.model_max_length,
47
+ truncation=True,
48
  do_sample=True,
49
  repetition_penalty=1.05,
50
  )