Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -9,26 +9,26 @@ def load_model():
|
|
9 |
# Initialize the model
|
10 |
generator = load_model()
|
11 |
|
12 |
-
st.title("
|
13 |
-
st.write("Enter a prompt
|
14 |
|
15 |
# User input
|
16 |
user_input = st.text_input("Enter your prompt:")
|
17 |
|
18 |
if user_input:
|
19 |
-
#
|
20 |
trump_prompt = f"In the style of Donald Trump: {user_input}"
|
21 |
|
22 |
-
# Generate the response
|
23 |
response = generator(
|
24 |
trump_prompt,
|
25 |
-
max_length=
|
26 |
num_return_sequences=1,
|
27 |
do_sample=True,
|
28 |
-
temperature=0.
|
29 |
-
top_p=0.
|
30 |
)
|
31 |
|
32 |
-
# Extract the generated text
|
33 |
generated_text = response[0]["generated_text"]
|
34 |
st.write("Response:", generated_text)
|
|
|
9 |
# Initialize the model
|
10 |
generator = load_model()
|
11 |
|
12 |
+
st.title("Trump-Inspired Text Generator")
|
13 |
+
st.write("Enter a prompt to receive a response.")
|
14 |
|
15 |
# User input
|
16 |
user_input = st.text_input("Enter your prompt:")
|
17 |
|
18 |
if user_input:
|
19 |
+
# Hidden style prompt for model guidance, but not shown to user
|
20 |
trump_prompt = f"In the style of Donald Trump: {user_input}"
|
21 |
|
22 |
+
# Generate the response with adjusted parameters
|
23 |
response = generator(
|
24 |
trump_prompt,
|
25 |
+
max_length=30,
|
26 |
num_return_sequences=1,
|
27 |
do_sample=True,
|
28 |
+
temperature=0.9,
|
29 |
+
top_p=0.85
|
30 |
)
|
31 |
|
32 |
+
# Extract and display the generated text
|
33 |
generated_text = response[0]["generated_text"]
|
34 |
st.write("Response:", generated_text)
|