Akshatha Arodi commited on
Commit
4a143b3
·
1 Parent(s): a24f5ec

gradio_app.py

Browse files
Files changed (1) hide show
  1. hackaprompt/gradio_app.py +10 -4
hackaprompt/gradio_app.py CHANGED
@@ -118,10 +118,16 @@ with gr.Blocks() as demo:
118
  evaluator_0 = gr.State(get_evaluator(level=0, completer=None))
119
 
120
  gr.Markdown(
121
- """
122
- # TRAIL Hands-on Exercise: Exploring Jailbreaking & Gen AI Risks
123
-
124
- In the presentation, we explored key risks associated with generative AI, including **jailbreaking, prompt engineering exploits, and data leakage**. Now, this hands-on exercise will give you practical experience in understanding how these vulnerabilities work in real-time.
 
 
 
 
 
 
125
 
126
  ### Your Task
127
  Your goal is to manipulate a language model using carefully crafted prompts to bypass safeguards and generate a specific response. This exercise will help you:
 
118
  evaluator_0 = gr.State(get_evaluator(level=0, completer=None))
119
 
120
  gr.Markdown(
121
+ """
122
+ <div style="display: flex; align-items: center;">
123
+ <img src="https://mila.quebec/sites/default/files/media-library/image/4032/milalogowebcoulrgb.png" width="120" style="margin-right: 0px;">
124
+ <h1 style="margin: 0;">Hands-on Exercise: Exploring Jailbreaking & Gen AI Risks</h1>
125
+ </div>
126
+
127
+ This exercise is part of the TRAIL Responsible Gen AI Risk Module, designed to provide hands-on experience with the challenges of Gen AI.
128
+
129
+ In the presentation, we explored key risks associated with generative AI, including **jailbreaking, prompt engineering exploits, and data leakage**. Now, this hands-on exercise will give you practical experience in understanding how these vulnerabilities work in real-time.
130
+ While real-world risks can be severe, this exercise presents a controlled, simplified version.
131
 
132
  ### Your Task
133
  Your goal is to manipulate a language model using carefully crafted prompts to bypass safeguards and generate a specific response. This exercise will help you: