Spaces:
Running
Running
Akshatha Arodi
commited on
Commit
·
4a143b3
1
Parent(s):
a24f5ec
gradio_app.py
Browse files- hackaprompt/gradio_app.py +10 -4
hackaprompt/gradio_app.py
CHANGED
@@ -118,10 +118,16 @@ with gr.Blocks() as demo:
|
|
118 |
evaluator_0 = gr.State(get_evaluator(level=0, completer=None))
|
119 |
|
120 |
gr.Markdown(
|
121 |
-
|
122 |
-
|
123 |
-
|
124 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
125 |
|
126 |
### Your Task
|
127 |
Your goal is to manipulate a language model using carefully crafted prompts to bypass safeguards and generate a specific response. This exercise will help you:
|
|
|
118 |
evaluator_0 = gr.State(get_evaluator(level=0, completer=None))
|
119 |
|
120 |
gr.Markdown(
|
121 |
+
"""
|
122 |
+
<div style="display: flex; align-items: center;">
|
123 |
+
<img src="https://mila.quebec/sites/default/files/media-library/image/4032/milalogowebcoulrgb.png" width="120" style="margin-right: 0px;">
|
124 |
+
<h1 style="margin: 0;">Hands-on Exercise: Exploring Jailbreaking & Gen AI Risks</h1>
|
125 |
+
</div>
|
126 |
+
|
127 |
+
This exercise is part of the TRAIL Responsible Gen AI Risk Module, designed to provide hands-on experience with the challenges of Gen AI.
|
128 |
+
|
129 |
+
In the presentation, we explored key risks associated with generative AI, including **jailbreaking, prompt engineering exploits, and data leakage**. Now, this hands-on exercise will give you practical experience in understanding how these vulnerabilities work in real-time.
|
130 |
+
While real-world risks can be severe, this exercise presents a controlled, simplified version.
|
131 |
|
132 |
### Your Task
|
133 |
Your goal is to manipulate a language model using carefully crafted prompts to bypass safeguards and generate a specific response. This exercise will help you:
|