Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -6,11 +6,11 @@ import torch
|
|
6 |
import os
|
7 |
|
8 |
# Load the pre-trained language model and tokenizer from Hugging Face
|
9 |
-
model_name = "gpt2-
|
10 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
11 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
12 |
|
13 |
-
# Set up a text generation pipeline
|
14 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer, device=0 if torch.cuda.is_available() else -1)
|
15 |
|
16 |
# Function to generate Trump-like response
|
|
|
6 |
import os
|
7 |
|
8 |
# Load the pre-trained language model and tokenizer from Hugging Face
|
9 |
+
model_name = "gpt2-large" # Use "EleutherAI/gpt-j-6B" for GPT-J (requires more resources)
|
10 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
11 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
12 |
|
13 |
+
# Set up a text generation pipeline with the larger model
|
14 |
generator = pipeline("text-generation", model=model, tokenizer=tokenizer, device=0 if torch.cuda.is_available() else -1)
|
15 |
|
16 |
# Function to generate Trump-like response
|