akahana commited on
Commit
cf26196
·
verified ·
1 Parent(s): 86b39d6

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +48 -0
README.md ADDED
@@ -0,0 +1,48 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - wikimedia/wikipedia
5
+ language:
6
+ - id
7
+ base_model:
8
+ - openai-community/gpt2
9
+ ---
10
+
11
+ ```python
12
+ from transformers import AutoModelForCausalLM, AutoTokenizer
13
+ from transformers import TextStreamer
14
+
15
+ model_name = "akahana/wikipedia-gpt2"
16
+
17
+ model = AutoModelForCausalLM.from_pretrained(
18
+ model_name,
19
+ torch_dtype="auto",
20
+ device_map="auto"
21
+ )
22
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
23
+
24
+ wikipedia_prompt = """Artikel Wikipedia
25
+ [[Judul]]
26
+ {}
27
+
28
+
29
+ [[Artikel]]
30
+ {}"""
31
+
32
+
33
+ title = "Hal Holbrook"
34
+ prompt = wikipedia_prompt.format(title, "")
35
+
36
+ model_inputs = tokenizer([prompt], return_tensors="pt").to(model.device)
37
+
38
+ generated_ids = model.generate(
39
+ **model_inputs,
40
+ max_new_tokens=512,
41
+ )
42
+ generated_ids = [
43
+ output_ids[len(input_ids):] for input_ids, output_ids in zip(model_inputs.input_ids, generated_ids)
44
+ ]
45
+
46
+ response = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
47
+ response
48
+ ```