Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -5,7 +5,7 @@ import torch
|
|
5 |
import os
|
6 |
|
7 |
# Load the pre-trained language model and tokenizer from Hugging Face
|
8 |
-
model_name = "gpt2-
|
9 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
10 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
11 |
|
@@ -18,7 +18,7 @@ def generate_response(prompt):
|
|
18 |
trump_prompt = f"Donald Trump says: {prompt}"
|
19 |
|
20 |
# Generate the response
|
21 |
-
response = generator(trump_prompt, max_length=
|
22 |
return response[0]['generated_text']
|
23 |
|
24 |
# Function to convert text to audio using gTTS
|
|
|
5 |
import os
|
6 |
|
7 |
# Load the pre-trained language model and tokenizer from Hugging Face
|
8 |
+
model_name = "gpt2-medium" # Use "EleutherAI/gpt-j-6B" for GPT-J (requires more resources)
|
9 |
tokenizer = AutoTokenizer.from_pretrained(model_name)
|
10 |
model = AutoModelForCausalLM.from_pretrained(model_name)
|
11 |
|
|
|
18 |
trump_prompt = f"Donald Trump says: {prompt}"
|
19 |
|
20 |
# Generate the response
|
21 |
+
response = generator(trump_prompt, max_length=50, num_return_sequences=1, temperature=0.8)
|
22 |
return response[0]['generated_text']
|
23 |
|
24 |
# Function to convert text to audio using gTTS
|